repo
stringlengths
7
55
path
stringlengths
4
223
func_name
stringlengths
1
134
original_string
stringlengths
75
104k
language
stringclasses
1 value
code
stringlengths
75
104k
code_tokens
listlengths
19
28.4k
docstring
stringlengths
1
46.9k
docstring_tokens
listlengths
1
1.97k
sha
stringlengths
40
40
url
stringlengths
87
315
partition
stringclasses
1 value
mretegan/crispy
crispy/modules/orca/parser.py
OutputData._parse_tensor
def _parse_tensor(self, indices=False): '''Parse a tensor.''' if indices: self.line = self._skip_lines(1) tensor = np.zeros((3, 3)) for i in range(3): tokens = self.line.split() if indices: tensor[i][0] = float(tokens[1]) tensor[i][1] = float(tokens[2]) tensor[i][2] = float(tokens[3]) else: tensor[i][0] = float(tokens[0]) tensor[i][1] = float(tokens[1]) tensor[i][2] = float(tokens[2]) self.line = self._skip_lines(1) return tensor
python
def _parse_tensor(self, indices=False): '''Parse a tensor.''' if indices: self.line = self._skip_lines(1) tensor = np.zeros((3, 3)) for i in range(3): tokens = self.line.split() if indices: tensor[i][0] = float(tokens[1]) tensor[i][1] = float(tokens[2]) tensor[i][2] = float(tokens[3]) else: tensor[i][0] = float(tokens[0]) tensor[i][1] = float(tokens[1]) tensor[i][2] = float(tokens[2]) self.line = self._skip_lines(1) return tensor
[ "def", "_parse_tensor", "(", "self", ",", "indices", "=", "False", ")", ":", "if", "indices", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "tensor", "=", "np", ".", "zeros", "(", "(", "3", ",", "3", ")", ")", "for", "i", "in", "range", "(", "3", ")", ":", "tokens", "=", "self", ".", "line", ".", "split", "(", ")", "if", "indices", ":", "tensor", "[", "i", "]", "[", "0", "]", "=", "float", "(", "tokens", "[", "1", "]", ")", "tensor", "[", "i", "]", "[", "1", "]", "=", "float", "(", "tokens", "[", "2", "]", ")", "tensor", "[", "i", "]", "[", "2", "]", "=", "float", "(", "tokens", "[", "3", "]", ")", "else", ":", "tensor", "[", "i", "]", "[", "0", "]", "=", "float", "(", "tokens", "[", "0", "]", ")", "tensor", "[", "i", "]", "[", "1", "]", "=", "float", "(", "tokens", "[", "1", "]", ")", "tensor", "[", "i", "]", "[", "2", "]", "=", "float", "(", "tokens", "[", "2", "]", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "return", "tensor" ]
Parse a tensor.
[ "Parse", "a", "tensor", "." ]
7e241ac1a48d34ca769f3a6183c430360b5f6725
https://github.com/mretegan/crispy/blob/7e241ac1a48d34ca769f3a6183c430360b5f6725/crispy/modules/orca/parser.py#L157-L174
train
mretegan/crispy
crispy/modules/orca/parser.py
OutputData.parse
def parse(self): '''Iterate over the lines and extract the required data.''' for self.line in self.output: # Parse general data: charge, multiplicity, coordinates, etc. self.index = 0 if self.line[1:13] == 'Total Charge': tokens = self.line.split() self.charge = int(tokens[-1]) if (self.line[1:13] or self.line[0:12]) == 'Multiplicity': tokens = self.line.split() self.multiplicity = int(tokens[-1]) if self.line[0:33] == 'CARTESIAN COORDINATES (ANGSTROEM)': if not hasattr(self, 'names'): self.names = dict() if not hasattr(self, 'coords'): self.coords = dict() self.line = self._skip_lines(2) names = list() coords = list() while self.line.strip(): tokens = self.line.split() names.append(tokens[0]) x = float(tokens[1]) y = float(tokens[2]) z = float(tokens[3]) coords.append((x, y, z)) self.line = next(self.output) self.names = np.array(names) self.coords[self.index] = np.array(coords) if self.line[22:50] == 'MULLIKEN POPULATION ANALYSIS': if not hasattr(self, 'populations'): self.populations = dict() self.line = self._skip_lines(6) populations = list() while self.line.strip() and 'Sum' not in self.line: tokens = self.line.split() populations.append((float(tokens[-2]), float(tokens[-1]))) self.line = next(self.output) self.populations['mulliken'][self.index] = np.array(populations) # noqa # Parse data from the EPR/NMR module if self.line[37:44] == 'EPR/NMR': self.eprnmr = dict() if self.line[0:19] == 'ELECTRONIC G-MATRIX': self.line = self._skip_lines(4) self.eprnmr['g']['tensor'] = self._parse_tensor() if self.line[0:27] == 'ZERO-FIELD-SPLITTING TENSOR': self.line = self._skip_lines(4) self.eprnmr['zfs']['tensor'] = self._parse_tensor() if self.line[1:8] == 'Nucleus': tokens = self.line.split() nucleus = int(re.findall(r'\d+', tokens[1])[0]) while 'Raw HFC' not in self.line: self.line = self._skip_lines(1) self.line = self._skip_lines(2) self.eprnmr['hfc'][nucleus]['tensor'] = self._parse_tensor() self.line = self._skip_lines(1) self.eprnmr['hfc'][nucleus]['fc'] = self._parse_components() self.eprnmr['hfc'][nucleus]['sd'] = self._parse_components() self.line = self._skip_lines(1) self.eprnmr['hfc'][nucleus]['orb'] = self._parse_components() self.eprnmr['hfc'][nucleus]['dia'] = self._parse_components() # Parse data from the MRCI module if self.line[36:43] == 'M R C I': self.mrci = dict() if self.line[1:19] == 'SPIN-SPIN COUPLING': self.line = self._skip_lines(4) self.mrci['zfs']['ssc']['tensor'] = self._parse_tensor() if self.line[1:30] == '2ND ORDER SPIN-ORBIT COUPLING': while 'Second' not in self.line: self.line = self._skip_lines(1) self.line = self._skip_lines(1) self.mrci['zfs']['soc']['second_order']['0']['tensor'] = self._parse_tensor() # noqa self.line = self._skip_lines(2) self.mrci['zfs']['soc']['second_order']['m']['tensor'] = self._parse_tensor() # noqa self.line = self._skip_lines(2) self.mrci['zfs']['soc']['second_order']['p']['tensor'] = self._parse_tensor() # noqa if self.line[1:42] == 'EFFECTIVE HAMILTONIAN SPIN-ORBIT COUPLING': self.line = self._skip_lines(4) self.mrci['zfs']['soc']['heff']['tensor'] = self._parse_tensor()
python
def parse(self): '''Iterate over the lines and extract the required data.''' for self.line in self.output: # Parse general data: charge, multiplicity, coordinates, etc. self.index = 0 if self.line[1:13] == 'Total Charge': tokens = self.line.split() self.charge = int(tokens[-1]) if (self.line[1:13] or self.line[0:12]) == 'Multiplicity': tokens = self.line.split() self.multiplicity = int(tokens[-1]) if self.line[0:33] == 'CARTESIAN COORDINATES (ANGSTROEM)': if not hasattr(self, 'names'): self.names = dict() if not hasattr(self, 'coords'): self.coords = dict() self.line = self._skip_lines(2) names = list() coords = list() while self.line.strip(): tokens = self.line.split() names.append(tokens[0]) x = float(tokens[1]) y = float(tokens[2]) z = float(tokens[3]) coords.append((x, y, z)) self.line = next(self.output) self.names = np.array(names) self.coords[self.index] = np.array(coords) if self.line[22:50] == 'MULLIKEN POPULATION ANALYSIS': if not hasattr(self, 'populations'): self.populations = dict() self.line = self._skip_lines(6) populations = list() while self.line.strip() and 'Sum' not in self.line: tokens = self.line.split() populations.append((float(tokens[-2]), float(tokens[-1]))) self.line = next(self.output) self.populations['mulliken'][self.index] = np.array(populations) # noqa # Parse data from the EPR/NMR module if self.line[37:44] == 'EPR/NMR': self.eprnmr = dict() if self.line[0:19] == 'ELECTRONIC G-MATRIX': self.line = self._skip_lines(4) self.eprnmr['g']['tensor'] = self._parse_tensor() if self.line[0:27] == 'ZERO-FIELD-SPLITTING TENSOR': self.line = self._skip_lines(4) self.eprnmr['zfs']['tensor'] = self._parse_tensor() if self.line[1:8] == 'Nucleus': tokens = self.line.split() nucleus = int(re.findall(r'\d+', tokens[1])[0]) while 'Raw HFC' not in self.line: self.line = self._skip_lines(1) self.line = self._skip_lines(2) self.eprnmr['hfc'][nucleus]['tensor'] = self._parse_tensor() self.line = self._skip_lines(1) self.eprnmr['hfc'][nucleus]['fc'] = self._parse_components() self.eprnmr['hfc'][nucleus]['sd'] = self._parse_components() self.line = self._skip_lines(1) self.eprnmr['hfc'][nucleus]['orb'] = self._parse_components() self.eprnmr['hfc'][nucleus]['dia'] = self._parse_components() # Parse data from the MRCI module if self.line[36:43] == 'M R C I': self.mrci = dict() if self.line[1:19] == 'SPIN-SPIN COUPLING': self.line = self._skip_lines(4) self.mrci['zfs']['ssc']['tensor'] = self._parse_tensor() if self.line[1:30] == '2ND ORDER SPIN-ORBIT COUPLING': while 'Second' not in self.line: self.line = self._skip_lines(1) self.line = self._skip_lines(1) self.mrci['zfs']['soc']['second_order']['0']['tensor'] = self._parse_tensor() # noqa self.line = self._skip_lines(2) self.mrci['zfs']['soc']['second_order']['m']['tensor'] = self._parse_tensor() # noqa self.line = self._skip_lines(2) self.mrci['zfs']['soc']['second_order']['p']['tensor'] = self._parse_tensor() # noqa if self.line[1:42] == 'EFFECTIVE HAMILTONIAN SPIN-ORBIT COUPLING': self.line = self._skip_lines(4) self.mrci['zfs']['soc']['heff']['tensor'] = self._parse_tensor()
[ "def", "parse", "(", "self", ")", ":", "for", "self", ".", "line", "in", "self", ".", "output", ":", "# Parse general data: charge, multiplicity, coordinates, etc.", "self", ".", "index", "=", "0", "if", "self", ".", "line", "[", "1", ":", "13", "]", "==", "'Total Charge'", ":", "tokens", "=", "self", ".", "line", ".", "split", "(", ")", "self", ".", "charge", "=", "int", "(", "tokens", "[", "-", "1", "]", ")", "if", "(", "self", ".", "line", "[", "1", ":", "13", "]", "or", "self", ".", "line", "[", "0", ":", "12", "]", ")", "==", "'Multiplicity'", ":", "tokens", "=", "self", ".", "line", ".", "split", "(", ")", "self", ".", "multiplicity", "=", "int", "(", "tokens", "[", "-", "1", "]", ")", "if", "self", ".", "line", "[", "0", ":", "33", "]", "==", "'CARTESIAN COORDINATES (ANGSTROEM)'", ":", "if", "not", "hasattr", "(", "self", ",", "'names'", ")", ":", "self", ".", "names", "=", "dict", "(", ")", "if", "not", "hasattr", "(", "self", ",", "'coords'", ")", ":", "self", ".", "coords", "=", "dict", "(", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "2", ")", "names", "=", "list", "(", ")", "coords", "=", "list", "(", ")", "while", "self", ".", "line", ".", "strip", "(", ")", ":", "tokens", "=", "self", ".", "line", ".", "split", "(", ")", "names", ".", "append", "(", "tokens", "[", "0", "]", ")", "x", "=", "float", "(", "tokens", "[", "1", "]", ")", "y", "=", "float", "(", "tokens", "[", "2", "]", ")", "z", "=", "float", "(", "tokens", "[", "3", "]", ")", "coords", ".", "append", "(", "(", "x", ",", "y", ",", "z", ")", ")", "self", ".", "line", "=", "next", "(", "self", ".", "output", ")", "self", ".", "names", "=", "np", ".", "array", "(", "names", ")", "self", ".", "coords", "[", "self", ".", "index", "]", "=", "np", ".", "array", "(", "coords", ")", "if", "self", ".", "line", "[", "22", ":", "50", "]", "==", "'MULLIKEN POPULATION ANALYSIS'", ":", "if", "not", "hasattr", "(", "self", ",", "'populations'", ")", ":", "self", ".", "populations", "=", "dict", "(", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "6", ")", "populations", "=", "list", "(", ")", "while", "self", ".", "line", ".", "strip", "(", ")", "and", "'Sum'", "not", "in", "self", ".", "line", ":", "tokens", "=", "self", ".", "line", ".", "split", "(", ")", "populations", ".", "append", "(", "(", "float", "(", "tokens", "[", "-", "2", "]", ")", ",", "float", "(", "tokens", "[", "-", "1", "]", ")", ")", ")", "self", ".", "line", "=", "next", "(", "self", ".", "output", ")", "self", ".", "populations", "[", "'mulliken'", "]", "[", "self", ".", "index", "]", "=", "np", ".", "array", "(", "populations", ")", "# noqa", "# Parse data from the EPR/NMR module", "if", "self", ".", "line", "[", "37", ":", "44", "]", "==", "'EPR/NMR'", ":", "self", ".", "eprnmr", "=", "dict", "(", ")", "if", "self", ".", "line", "[", "0", ":", "19", "]", "==", "'ELECTRONIC G-MATRIX'", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "4", ")", "self", ".", "eprnmr", "[", "'g'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "if", "self", ".", "line", "[", "0", ":", "27", "]", "==", "'ZERO-FIELD-SPLITTING TENSOR'", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "4", ")", "self", ".", "eprnmr", "[", "'zfs'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "if", "self", ".", "line", "[", "1", ":", "8", "]", "==", "'Nucleus'", ":", "tokens", "=", "self", ".", "line", ".", "split", "(", ")", "nucleus", "=", "int", "(", "re", ".", "findall", "(", "r'\\d+'", ",", "tokens", "[", "1", "]", ")", "[", "0", "]", ")", "while", "'Raw HFC'", "not", "in", "self", ".", "line", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "2", ")", "self", ".", "eprnmr", "[", "'hfc'", "]", "[", "nucleus", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "self", ".", "eprnmr", "[", "'hfc'", "]", "[", "nucleus", "]", "[", "'fc'", "]", "=", "self", ".", "_parse_components", "(", ")", "self", ".", "eprnmr", "[", "'hfc'", "]", "[", "nucleus", "]", "[", "'sd'", "]", "=", "self", ".", "_parse_components", "(", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "self", ".", "eprnmr", "[", "'hfc'", "]", "[", "nucleus", "]", "[", "'orb'", "]", "=", "self", ".", "_parse_components", "(", ")", "self", ".", "eprnmr", "[", "'hfc'", "]", "[", "nucleus", "]", "[", "'dia'", "]", "=", "self", ".", "_parse_components", "(", ")", "# Parse data from the MRCI module", "if", "self", ".", "line", "[", "36", ":", "43", "]", "==", "'M R C I'", ":", "self", ".", "mrci", "=", "dict", "(", ")", "if", "self", ".", "line", "[", "1", ":", "19", "]", "==", "'SPIN-SPIN COUPLING'", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "4", ")", "self", ".", "mrci", "[", "'zfs'", "]", "[", "'ssc'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "if", "self", ".", "line", "[", "1", ":", "30", "]", "==", "'2ND ORDER SPIN-ORBIT COUPLING'", ":", "while", "'Second'", "not", "in", "self", ".", "line", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "1", ")", "self", ".", "mrci", "[", "'zfs'", "]", "[", "'soc'", "]", "[", "'second_order'", "]", "[", "'0'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "# noqa", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "2", ")", "self", ".", "mrci", "[", "'zfs'", "]", "[", "'soc'", "]", "[", "'second_order'", "]", "[", "'m'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "# noqa", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "2", ")", "self", ".", "mrci", "[", "'zfs'", "]", "[", "'soc'", "]", "[", "'second_order'", "]", "[", "'p'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")", "# noqa", "if", "self", ".", "line", "[", "1", ":", "42", "]", "==", "'EFFECTIVE HAMILTONIAN SPIN-ORBIT COUPLING'", ":", "self", ".", "line", "=", "self", ".", "_skip_lines", "(", "4", ")", "self", ".", "mrci", "[", "'zfs'", "]", "[", "'soc'", "]", "[", "'heff'", "]", "[", "'tensor'", "]", "=", "self", ".", "_parse_tensor", "(", ")" ]
Iterate over the lines and extract the required data.
[ "Iterate", "over", "the", "lines", "and", "extract", "the", "required", "data", "." ]
7e241ac1a48d34ca769f3a6183c430360b5f6725
https://github.com/mretegan/crispy/blob/7e241ac1a48d34ca769f3a6183c430360b5f6725/crispy/modules/orca/parser.py#L185-L275
train
xeBuz/Flask-Validator
flask_validator/validator.py
FlaskValidator.__validate
def __validate(self, target, value, oldvalue, initiator): """ Method executed when the event 'set' is triggered. :param target: Object triggered :param value: New value :param oldvalue: Previous value :param initiator: Column modified :return: :raise ValidateError: """ if value == oldvalue: return value if self.allow_null and value is None: return value if self.check_value(value): return value else: if self.throw_exception: if self.message: self.message = self.message.format( field=self.field, new_value=value, old_value=oldvalue, key=initiator.key) raise ValidateError(self.message) else: raise ValidateError('Value %s from column %s is not valid' % (value, initiator.key)) return oldvalue
python
def __validate(self, target, value, oldvalue, initiator): """ Method executed when the event 'set' is triggered. :param target: Object triggered :param value: New value :param oldvalue: Previous value :param initiator: Column modified :return: :raise ValidateError: """ if value == oldvalue: return value if self.allow_null and value is None: return value if self.check_value(value): return value else: if self.throw_exception: if self.message: self.message = self.message.format( field=self.field, new_value=value, old_value=oldvalue, key=initiator.key) raise ValidateError(self.message) else: raise ValidateError('Value %s from column %s is not valid' % (value, initiator.key)) return oldvalue
[ "def", "__validate", "(", "self", ",", "target", ",", "value", ",", "oldvalue", ",", "initiator", ")", ":", "if", "value", "==", "oldvalue", ":", "return", "value", "if", "self", ".", "allow_null", "and", "value", "is", "None", ":", "return", "value", "if", "self", ".", "check_value", "(", "value", ")", ":", "return", "value", "else", ":", "if", "self", ".", "throw_exception", ":", "if", "self", ".", "message", ":", "self", ".", "message", "=", "self", ".", "message", ".", "format", "(", "field", "=", "self", ".", "field", ",", "new_value", "=", "value", ",", "old_value", "=", "oldvalue", ",", "key", "=", "initiator", ".", "key", ")", "raise", "ValidateError", "(", "self", ".", "message", ")", "else", ":", "raise", "ValidateError", "(", "'Value %s from column %s is not valid'", "%", "(", "value", ",", "initiator", ".", "key", ")", ")", "return", "oldvalue" ]
Method executed when the event 'set' is triggered. :param target: Object triggered :param value: New value :param oldvalue: Previous value :param initiator: Column modified :return: :raise ValidateError:
[ "Method", "executed", "when", "the", "event", "set", "is", "triggered", "." ]
ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9
https://github.com/xeBuz/Flask-Validator/blob/ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9/flask_validator/validator.py#L35-L62
train
xeBuz/Flask-Validator
flask_validator/validator.py
FlaskValidator.__create_event
def __create_event(self): """ Create an SQLAlchemy event listening the 'set' in a particular column. :rtype : object """ if not event.contains(self.field, 'set', self.__validate): event.listen(self.field, 'set', self.__validate, retval=True)
python
def __create_event(self): """ Create an SQLAlchemy event listening the 'set' in a particular column. :rtype : object """ if not event.contains(self.field, 'set', self.__validate): event.listen(self.field, 'set', self.__validate, retval=True)
[ "def", "__create_event", "(", "self", ")", ":", "if", "not", "event", ".", "contains", "(", "self", ".", "field", ",", "'set'", ",", "self", ".", "__validate", ")", ":", "event", ".", "listen", "(", "self", ".", "field", ",", "'set'", ",", "self", ".", "__validate", ",", "retval", "=", "True", ")" ]
Create an SQLAlchemy event listening the 'set' in a particular column. :rtype : object
[ "Create", "an", "SQLAlchemy", "event", "listening", "the", "set", "in", "a", "particular", "column", "." ]
ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9
https://github.com/xeBuz/Flask-Validator/blob/ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9/flask_validator/validator.py#L64-L70
train
xeBuz/Flask-Validator
flask_validator/validator.py
FlaskValidator.stop
def stop(self): """ Remove the listener to stop the validation """ if event.contains(self.field, 'set', self.__validate): event.remove(self.field, 'set', self.__validate)
python
def stop(self): """ Remove the listener to stop the validation """ if event.contains(self.field, 'set', self.__validate): event.remove(self.field, 'set', self.__validate)
[ "def", "stop", "(", "self", ")", ":", "if", "event", ".", "contains", "(", "self", ".", "field", ",", "'set'", ",", "self", ".", "__validate", ")", ":", "event", ".", "remove", "(", "self", ".", "field", ",", "'set'", ",", "self", ".", "__validate", ")" ]
Remove the listener to stop the validation
[ "Remove", "the", "listener", "to", "stop", "the", "validation" ]
ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9
https://github.com/xeBuz/Flask-Validator/blob/ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9/flask_validator/validator.py#L80-L84
train
xeBuz/Flask-Validator
flask_validator/validator.py
FlaskValidator.start
def start(self): """ Restart the listener """ if not event.contains(self.field, 'set', self.__validate): self.__create_event()
python
def start(self): """ Restart the listener """ if not event.contains(self.field, 'set', self.__validate): self.__create_event()
[ "def", "start", "(", "self", ")", ":", "if", "not", "event", ".", "contains", "(", "self", ".", "field", ",", "'set'", ",", "self", ".", "__validate", ")", ":", "self", ".", "__create_event", "(", ")" ]
Restart the listener
[ "Restart", "the", "listener" ]
ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9
https://github.com/xeBuz/Flask-Validator/blob/ef3dd0a24300c88cb728e6dc1a221e7e7127e1f9/flask_validator/validator.py#L86-L90
train
doanguyen/lasotuvi
lasotuvi/DiaBan.py
diaBan.nhapDaiHan
def nhapDaiHan(self, cucSo, gioiTinh): """Nhap dai han Args: cucSo (TYPE): Description gioiTinh (TYPE): Description Returns: TYPE: Description """ for cung in self.thapNhiCung: khoangCach = khoangCachCung(cung.cungSo, self.cungMenh, gioiTinh) cung.daiHan(cucSo + khoangCach * 10) return self
python
def nhapDaiHan(self, cucSo, gioiTinh): """Nhap dai han Args: cucSo (TYPE): Description gioiTinh (TYPE): Description Returns: TYPE: Description """ for cung in self.thapNhiCung: khoangCach = khoangCachCung(cung.cungSo, self.cungMenh, gioiTinh) cung.daiHan(cucSo + khoangCach * 10) return self
[ "def", "nhapDaiHan", "(", "self", ",", "cucSo", ",", "gioiTinh", ")", ":", "for", "cung", "in", "self", ".", "thapNhiCung", ":", "khoangCach", "=", "khoangCachCung", "(", "cung", ".", "cungSo", ",", "self", ".", "cungMenh", ",", "gioiTinh", ")", "cung", ".", "daiHan", "(", "cucSo", "+", "khoangCach", "*", "10", ")", "return", "self" ]
Nhap dai han Args: cucSo (TYPE): Description gioiTinh (TYPE): Description Returns: TYPE: Description
[ "Nhap", "dai", "han" ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/DiaBan.py#L153-L166
train
tansey/gfl
pygfl/easy.py
solve_gfl
def solve_gfl(data, edges=None, weights=None, minlam=0.2, maxlam=1000.0, numlam=30, alpha=0.2, inflate=2., converge=1e-6, maxsteps=1000000, lam=None, verbose=0, missing_val=None, full_path=False, loss='normal'): '''A very easy-to-use version of GFL solver that just requires the data and the edges.''' #Fix no edge cases if edges.shape[0] < 1: return data #Keep initial edges init_edges = edges if verbose: print('Decomposing graph into trails') if loss == 'binomial': flat_data = data[0].flatten() nonmissing_flat_data = flat_data, data[1].flatten() else: flat_data = data.flatten() nonmissing_flat_data = flat_data if edges is None: if loss == 'binomial': if verbose: print('Using default edge set of a grid of same shape as the data: {0}'.format(data[0].shape)) edges = hypercube_edges(data[0].shape) else: if verbose: print('Using default edge set of a grid of same shape as the data: {0}'.format(data.shape)) edges = hypercube_edges(data.shape) if missing_val is not None: if verbose: print('Removing all data points whose data value is {0}'.format(missing_val)) edges = [(e1,e2) for (e1,e2) in edges if flat_data[e1] != missing_val and flat_data[e2] != missing_val] if loss == 'binomial': nonmissing_flat_data = flat_data[flat_data != missing_val], nonmissing_flat_data[1][flat_data != missing_val] else: nonmissing_flat_data = flat_data[flat_data != missing_val] ########### Setup the graph g = Graph() g.add_edges_from(edges) chains = decompose_graph(g, heuristic='greedy') ntrails, trails, breakpoints, edges = chains_to_trails(chains) if verbose: print('Setting up trail solver') ########### Setup the solver if loss == 'normal': solver = TrailSolver(alpha, inflate, maxsteps, converge) elif loss == 'logistic': solver = LogisticTrailSolver(alpha, inflate, maxsteps, converge) elif loss == 'binomial': solver = BinomialTrailSolver(alpha, inflate, maxsteps, converge) else: raise NotImplementedError('Loss must be normal, logistic, or binomial') # Set the data and pre-cache any necessary structures solver.set_data(nonmissing_flat_data, edges, ntrails, trails, breakpoints, weights=weights) if verbose: print('Solving') ########### Run the solver if lam: # Fixed lambda beta = solver.solve(lam) else: # Grid search to find the best lambda beta = solver.solution_path(minlam, maxlam, numlam, verbose=max(0, verbose-1)) if not full_path: beta = beta['best'] ########### Fix disconnected nodes mask = np.ones_like(beta) mask[init_edges[:,0]] = 0 mask[init_edges[:,1]] = 0 beta[mask>0] = data[mask>0] return beta
python
def solve_gfl(data, edges=None, weights=None, minlam=0.2, maxlam=1000.0, numlam=30, alpha=0.2, inflate=2., converge=1e-6, maxsteps=1000000, lam=None, verbose=0, missing_val=None, full_path=False, loss='normal'): '''A very easy-to-use version of GFL solver that just requires the data and the edges.''' #Fix no edge cases if edges.shape[0] < 1: return data #Keep initial edges init_edges = edges if verbose: print('Decomposing graph into trails') if loss == 'binomial': flat_data = data[0].flatten() nonmissing_flat_data = flat_data, data[1].flatten() else: flat_data = data.flatten() nonmissing_flat_data = flat_data if edges is None: if loss == 'binomial': if verbose: print('Using default edge set of a grid of same shape as the data: {0}'.format(data[0].shape)) edges = hypercube_edges(data[0].shape) else: if verbose: print('Using default edge set of a grid of same shape as the data: {0}'.format(data.shape)) edges = hypercube_edges(data.shape) if missing_val is not None: if verbose: print('Removing all data points whose data value is {0}'.format(missing_val)) edges = [(e1,e2) for (e1,e2) in edges if flat_data[e1] != missing_val and flat_data[e2] != missing_val] if loss == 'binomial': nonmissing_flat_data = flat_data[flat_data != missing_val], nonmissing_flat_data[1][flat_data != missing_val] else: nonmissing_flat_data = flat_data[flat_data != missing_val] ########### Setup the graph g = Graph() g.add_edges_from(edges) chains = decompose_graph(g, heuristic='greedy') ntrails, trails, breakpoints, edges = chains_to_trails(chains) if verbose: print('Setting up trail solver') ########### Setup the solver if loss == 'normal': solver = TrailSolver(alpha, inflate, maxsteps, converge) elif loss == 'logistic': solver = LogisticTrailSolver(alpha, inflate, maxsteps, converge) elif loss == 'binomial': solver = BinomialTrailSolver(alpha, inflate, maxsteps, converge) else: raise NotImplementedError('Loss must be normal, logistic, or binomial') # Set the data and pre-cache any necessary structures solver.set_data(nonmissing_flat_data, edges, ntrails, trails, breakpoints, weights=weights) if verbose: print('Solving') ########### Run the solver if lam: # Fixed lambda beta = solver.solve(lam) else: # Grid search to find the best lambda beta = solver.solution_path(minlam, maxlam, numlam, verbose=max(0, verbose-1)) if not full_path: beta = beta['best'] ########### Fix disconnected nodes mask = np.ones_like(beta) mask[init_edges[:,0]] = 0 mask[init_edges[:,1]] = 0 beta[mask>0] = data[mask>0] return beta
[ "def", "solve_gfl", "(", "data", ",", "edges", "=", "None", ",", "weights", "=", "None", ",", "minlam", "=", "0.2", ",", "maxlam", "=", "1000.0", ",", "numlam", "=", "30", ",", "alpha", "=", "0.2", ",", "inflate", "=", "2.", ",", "converge", "=", "1e-6", ",", "maxsteps", "=", "1000000", ",", "lam", "=", "None", ",", "verbose", "=", "0", ",", "missing_val", "=", "None", ",", "full_path", "=", "False", ",", "loss", "=", "'normal'", ")", ":", "#Fix no edge cases", "if", "edges", ".", "shape", "[", "0", "]", "<", "1", ":", "return", "data", "#Keep initial edges", "init_edges", "=", "edges", "if", "verbose", ":", "print", "(", "'Decomposing graph into trails'", ")", "if", "loss", "==", "'binomial'", ":", "flat_data", "=", "data", "[", "0", "]", ".", "flatten", "(", ")", "nonmissing_flat_data", "=", "flat_data", ",", "data", "[", "1", "]", ".", "flatten", "(", ")", "else", ":", "flat_data", "=", "data", ".", "flatten", "(", ")", "nonmissing_flat_data", "=", "flat_data", "if", "edges", "is", "None", ":", "if", "loss", "==", "'binomial'", ":", "if", "verbose", ":", "print", "(", "'Using default edge set of a grid of same shape as the data: {0}'", ".", "format", "(", "data", "[", "0", "]", ".", "shape", ")", ")", "edges", "=", "hypercube_edges", "(", "data", "[", "0", "]", ".", "shape", ")", "else", ":", "if", "verbose", ":", "print", "(", "'Using default edge set of a grid of same shape as the data: {0}'", ".", "format", "(", "data", ".", "shape", ")", ")", "edges", "=", "hypercube_edges", "(", "data", ".", "shape", ")", "if", "missing_val", "is", "not", "None", ":", "if", "verbose", ":", "print", "(", "'Removing all data points whose data value is {0}'", ".", "format", "(", "missing_val", ")", ")", "edges", "=", "[", "(", "e1", ",", "e2", ")", "for", "(", "e1", ",", "e2", ")", "in", "edges", "if", "flat_data", "[", "e1", "]", "!=", "missing_val", "and", "flat_data", "[", "e2", "]", "!=", "missing_val", "]", "if", "loss", "==", "'binomial'", ":", "nonmissing_flat_data", "=", "flat_data", "[", "flat_data", "!=", "missing_val", "]", ",", "nonmissing_flat_data", "[", "1", "]", "[", "flat_data", "!=", "missing_val", "]", "else", ":", "nonmissing_flat_data", "=", "flat_data", "[", "flat_data", "!=", "missing_val", "]", "########### Setup the graph", "g", "=", "Graph", "(", ")", "g", ".", "add_edges_from", "(", "edges", ")", "chains", "=", "decompose_graph", "(", "g", ",", "heuristic", "=", "'greedy'", ")", "ntrails", ",", "trails", ",", "breakpoints", ",", "edges", "=", "chains_to_trails", "(", "chains", ")", "if", "verbose", ":", "print", "(", "'Setting up trail solver'", ")", "########### Setup the solver", "if", "loss", "==", "'normal'", ":", "solver", "=", "TrailSolver", "(", "alpha", ",", "inflate", ",", "maxsteps", ",", "converge", ")", "elif", "loss", "==", "'logistic'", ":", "solver", "=", "LogisticTrailSolver", "(", "alpha", ",", "inflate", ",", "maxsteps", ",", "converge", ")", "elif", "loss", "==", "'binomial'", ":", "solver", "=", "BinomialTrailSolver", "(", "alpha", ",", "inflate", ",", "maxsteps", ",", "converge", ")", "else", ":", "raise", "NotImplementedError", "(", "'Loss must be normal, logistic, or binomial'", ")", "# Set the data and pre-cache any necessary structures", "solver", ".", "set_data", "(", "nonmissing_flat_data", ",", "edges", ",", "ntrails", ",", "trails", ",", "breakpoints", ",", "weights", "=", "weights", ")", "if", "verbose", ":", "print", "(", "'Solving'", ")", "########### Run the solver", "if", "lam", ":", "# Fixed lambda", "beta", "=", "solver", ".", "solve", "(", "lam", ")", "else", ":", "# Grid search to find the best lambda", "beta", "=", "solver", ".", "solution_path", "(", "minlam", ",", "maxlam", ",", "numlam", ",", "verbose", "=", "max", "(", "0", ",", "verbose", "-", "1", ")", ")", "if", "not", "full_path", ":", "beta", "=", "beta", "[", "'best'", "]", "########### Fix disconnected nodes", "mask", "=", "np", ".", "ones_like", "(", "beta", ")", "mask", "[", "init_edges", "[", ":", ",", "0", "]", "]", "=", "0", "mask", "[", "init_edges", "[", ":", ",", "1", "]", "]", "=", "0", "beta", "[", "mask", ">", "0", "]", "=", "data", "[", "mask", ">", "0", "]", "return", "beta" ]
A very easy-to-use version of GFL solver that just requires the data and the edges.
[ "A", "very", "easy", "-", "to", "-", "use", "version", "of", "GFL", "solver", "that", "just", "requires", "the", "data", "and", "the", "edges", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/easy.py#L27-L113
train
doanguyen/lasotuvi
lasotuvi/AmDuong.py
ngayThangNam
def ngayThangNam(nn, tt, nnnn, duongLich=True, timeZone=7): """Summary Args: nn (TYPE): ngay tt (TYPE): thang nnnn (TYPE): nam duongLich (bool, optional): bool timeZone (int, optional): +7 Vietnam Returns: TYPE: Description Raises: Exception: Description """ thangNhuan = 0 # if nnnn > 1000 and nnnn < 3000 and nn > 0 and \ if nn > 0 and \ nn < 32 and tt < 13 and tt > 0: if duongLich is True: [nn, tt, nnnn, thangNhuan] = S2L(nn, tt, nnnn, timeZone=timeZone) return [nn, tt, nnnn, thangNhuan] else: raise Exception("Ngày, tháng, năm không chính xác.")
python
def ngayThangNam(nn, tt, nnnn, duongLich=True, timeZone=7): """Summary Args: nn (TYPE): ngay tt (TYPE): thang nnnn (TYPE): nam duongLich (bool, optional): bool timeZone (int, optional): +7 Vietnam Returns: TYPE: Description Raises: Exception: Description """ thangNhuan = 0 # if nnnn > 1000 and nnnn < 3000 and nn > 0 and \ if nn > 0 and \ nn < 32 and tt < 13 and tt > 0: if duongLich is True: [nn, tt, nnnn, thangNhuan] = S2L(nn, tt, nnnn, timeZone=timeZone) return [nn, tt, nnnn, thangNhuan] else: raise Exception("Ngày, tháng, năm không chính xác.")
[ "def", "ngayThangNam", "(", "nn", ",", "tt", ",", "nnnn", ",", "duongLich", "=", "True", ",", "timeZone", "=", "7", ")", ":", "thangNhuan", "=", "0", "# if nnnn > 1000 and nnnn < 3000 and nn > 0 and \\", "if", "nn", ">", "0", "and", "nn", "<", "32", "and", "tt", "<", "13", "and", "tt", ">", "0", ":", "if", "duongLich", "is", "True", ":", "[", "nn", ",", "tt", ",", "nnnn", ",", "thangNhuan", "]", "=", "S2L", "(", "nn", ",", "tt", ",", "nnnn", ",", "timeZone", "=", "timeZone", ")", "return", "[", "nn", ",", "tt", ",", "nnnn", ",", "thangNhuan", "]", "else", ":", "raise", "Exception", "(", "\"Ngày, tháng, năm không chính xác.\")", "" ]
Summary Args: nn (TYPE): ngay tt (TYPE): thang nnnn (TYPE): nam duongLich (bool, optional): bool timeZone (int, optional): +7 Vietnam Returns: TYPE: Description Raises: Exception: Description
[ "Summary" ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/AmDuong.py#L218-L242
train
doanguyen/lasotuvi
lasotuvi/AmDuong.py
canChiNgay
def canChiNgay(nn, tt, nnnn, duongLich=True, timeZone=7, thangNhuan=False): """Summary Args: nn (int): ngày tt (int): tháng nnnn (int): năm duongLich (bool, optional): True nếu là dương lịch, False âm lịch timeZone (int, optional): Múi giờ thangNhuan (bool, optional): Có phải là tháng nhuận không? Returns: TYPE: Description """ if duongLich is False: [nn, tt, nnnn] = L2S(nn, tt, nnnn, thangNhuan, timeZone) jd = jdFromDate(nn, tt, nnnn) # print jd canNgay = (jd + 9) % 10 + 1 chiNgay = (jd + 1) % 12 + 1 return [canNgay, chiNgay]
python
def canChiNgay(nn, tt, nnnn, duongLich=True, timeZone=7, thangNhuan=False): """Summary Args: nn (int): ngày tt (int): tháng nnnn (int): năm duongLich (bool, optional): True nếu là dương lịch, False âm lịch timeZone (int, optional): Múi giờ thangNhuan (bool, optional): Có phải là tháng nhuận không? Returns: TYPE: Description """ if duongLich is False: [nn, tt, nnnn] = L2S(nn, tt, nnnn, thangNhuan, timeZone) jd = jdFromDate(nn, tt, nnnn) # print jd canNgay = (jd + 9) % 10 + 1 chiNgay = (jd + 1) % 12 + 1 return [canNgay, chiNgay]
[ "def", "canChiNgay", "(", "nn", ",", "tt", ",", "nnnn", ",", "duongLich", "=", "True", ",", "timeZone", "=", "7", ",", "thangNhuan", "=", "False", ")", ":", "if", "duongLich", "is", "False", ":", "[", "nn", ",", "tt", ",", "nnnn", "]", "=", "L2S", "(", "nn", ",", "tt", ",", "nnnn", ",", "thangNhuan", ",", "timeZone", ")", "jd", "=", "jdFromDate", "(", "nn", ",", "tt", ",", "nnnn", ")", "# print jd", "canNgay", "=", "(", "jd", "+", "9", ")", "%", "10", "+", "1", "chiNgay", "=", "(", "jd", "+", "1", ")", "%", "12", "+", "1", "return", "[", "canNgay", ",", "chiNgay", "]" ]
Summary Args: nn (int): ngày tt (int): tháng nnnn (int): năm duongLich (bool, optional): True nếu là dương lịch, False âm lịch timeZone (int, optional): Múi giờ thangNhuan (bool, optional): Có phải là tháng nhuận không? Returns: TYPE: Description
[ "Summary" ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/AmDuong.py#L245-L265
train
doanguyen/lasotuvi
lasotuvi/AmDuong.py
ngayThangNamCanChi
def ngayThangNamCanChi(nn, tt, nnnn, duongLich=True, timeZone=7): """chuyển đổi năm, tháng âm/dương lịch sang Can, Chi trong tiếng Việt. Không tính đến can ngày vì phải chuyển đổi qua lịch Julius. Hàm tìm can ngày là hàm canChiNgay(nn, tt, nnnn, duongLich=True,\ timeZone=7, thangNhuan=False) Args: nn (int): Ngày tt (int): Tháng nnnn (int): Năm Returns: TYPE: Description """ if duongLich is True: [nn, tt, nnnn, thangNhuan] = \ ngayThangNam(nn, tt, nnnn, timeZone=timeZone) # Can của tháng canThang = (nnnn * 12 + tt + 3) % 10 + 1 # Can chi của năm canNamSinh = (nnnn + 6) % 10 + 1 chiNam = (nnnn + 8) % 12 + 1 return [canThang, canNamSinh, chiNam]
python
def ngayThangNamCanChi(nn, tt, nnnn, duongLich=True, timeZone=7): """chuyển đổi năm, tháng âm/dương lịch sang Can, Chi trong tiếng Việt. Không tính đến can ngày vì phải chuyển đổi qua lịch Julius. Hàm tìm can ngày là hàm canChiNgay(nn, tt, nnnn, duongLich=True,\ timeZone=7, thangNhuan=False) Args: nn (int): Ngày tt (int): Tháng nnnn (int): Năm Returns: TYPE: Description """ if duongLich is True: [nn, tt, nnnn, thangNhuan] = \ ngayThangNam(nn, tt, nnnn, timeZone=timeZone) # Can của tháng canThang = (nnnn * 12 + tt + 3) % 10 + 1 # Can chi của năm canNamSinh = (nnnn + 6) % 10 + 1 chiNam = (nnnn + 8) % 12 + 1 return [canThang, canNamSinh, chiNam]
[ "def", "ngayThangNamCanChi", "(", "nn", ",", "tt", ",", "nnnn", ",", "duongLich", "=", "True", ",", "timeZone", "=", "7", ")", ":", "if", "duongLich", "is", "True", ":", "[", "nn", ",", "tt", ",", "nnnn", ",", "thangNhuan", "]", "=", "ngayThangNam", "(", "nn", ",", "tt", ",", "nnnn", ",", "timeZone", "=", "timeZone", ")", "# Can của tháng", "canThang", "=", "(", "nnnn", "*", "12", "+", "tt", "+", "3", ")", "%", "10", "+", "1", "# Can chi của năm", "canNamSinh", "=", "(", "nnnn", "+", "6", ")", "%", "10", "+", "1", "chiNam", "=", "(", "nnnn", "+", "8", ")", "%", "12", "+", "1", "return", "[", "canThang", ",", "canNamSinh", ",", "chiNam", "]" ]
chuyển đổi năm, tháng âm/dương lịch sang Can, Chi trong tiếng Việt. Không tính đến can ngày vì phải chuyển đổi qua lịch Julius. Hàm tìm can ngày là hàm canChiNgay(nn, tt, nnnn, duongLich=True,\ timeZone=7, thangNhuan=False) Args: nn (int): Ngày tt (int): Tháng nnnn (int): Năm Returns: TYPE: Description
[ "chuyển", "đổi", "năm", "tháng", "âm", "/", "dương", "lịch", "sang", "Can", "Chi", "trong", "tiếng", "Việt", ".", "Không", "tính", "đến", "can", "ngày", "vì", "phải", "chuyển", "đổi", "qua", "lịch", "Julius", "." ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/AmDuong.py#L281-L305
train
doanguyen/lasotuvi
lasotuvi/AmDuong.py
nguHanh
def nguHanh(tenHanh): """ Args: tenHanh (string): Tên Hành trong ngũ hành, Kim hoặc K, Moc hoặc M, Thuy hoặc T, Hoa hoặc H, Tho hoặc O Returns: Dictionary: ID của Hành, tên đầy đủ của Hành, số Cục của Hành Raises: Exception: Description """ if tenHanh in ["Kim", "K"]: return {"id": 1, "tenHanh": "Kim", "cuc": 4, "tenCuc": "Kim tứ Cục", "css": "hanhKim"} elif tenHanh == "Moc" or tenHanh == "M": return {"id": 2, "tenHanh": "Mộc", "cuc": 3, "tenCuc": "Mộc tam Cục", "css": "hanhMoc"} elif tenHanh == "Thuy" or tenHanh == "T": return {"id": 3, "tenHanh": "Thủy", "cuc": 2, "tenCuc": "Thủy nhị Cục", "css": "hanhThuy"} elif tenHanh == "Hoa" or tenHanh == "H": return {"id": 4, "tenHanh": "Hỏa", "cuc": 6, "tenCuc": "Hỏa lục Cục", "css": "hanhHoa"} elif tenHanh == "Tho" or tenHanh == "O": return {"id": 5, "tenHanh": "Thổ", "cuc": 5, "tenCuc": "Thổ ngũ Cục", "css": "hanhTho"} else: raise Exception( "Tên Hành phải thuộc Kim (K), Mộc (M), Thủy (T), \ Hỏa (H) hoặc Thổ (O)")
python
def nguHanh(tenHanh): """ Args: tenHanh (string): Tên Hành trong ngũ hành, Kim hoặc K, Moc hoặc M, Thuy hoặc T, Hoa hoặc H, Tho hoặc O Returns: Dictionary: ID của Hành, tên đầy đủ của Hành, số Cục của Hành Raises: Exception: Description """ if tenHanh in ["Kim", "K"]: return {"id": 1, "tenHanh": "Kim", "cuc": 4, "tenCuc": "Kim tứ Cục", "css": "hanhKim"} elif tenHanh == "Moc" or tenHanh == "M": return {"id": 2, "tenHanh": "Mộc", "cuc": 3, "tenCuc": "Mộc tam Cục", "css": "hanhMoc"} elif tenHanh == "Thuy" or tenHanh == "T": return {"id": 3, "tenHanh": "Thủy", "cuc": 2, "tenCuc": "Thủy nhị Cục", "css": "hanhThuy"} elif tenHanh == "Hoa" or tenHanh == "H": return {"id": 4, "tenHanh": "Hỏa", "cuc": 6, "tenCuc": "Hỏa lục Cục", "css": "hanhHoa"} elif tenHanh == "Tho" or tenHanh == "O": return {"id": 5, "tenHanh": "Thổ", "cuc": 5, "tenCuc": "Thổ ngũ Cục", "css": "hanhTho"} else: raise Exception( "Tên Hành phải thuộc Kim (K), Mộc (M), Thủy (T), \ Hỏa (H) hoặc Thổ (O)")
[ "def", "nguHanh", "(", "tenHanh", ")", ":", "if", "tenHanh", "in", "[", "\"Kim\"", ",", "\"K\"", "]", ":", "return", "{", "\"id\"", ":", "1", ",", "\"tenHanh\"", ":", "\"Kim\"", ",", "\"cuc\"", ":", "4", ",", "\"tenCuc\"", ":", "\"Kim tứ Cục\",", "", "\"css\"", ":", "\"hanhKim\"", "}", "elif", "tenHanh", "==", "\"Moc\"", "or", "tenHanh", "==", "\"M\"", ":", "return", "{", "\"id\"", ":", "2", ",", "\"tenHanh\"", ":", "\"Mộc\", ", "\"", "uc\": ", "3", " ", "\"", "enCuc\": ", "\"", "ộc tam Cục\",", "", "\"css\"", ":", "\"hanhMoc\"", "}", "elif", "tenHanh", "==", "\"Thuy\"", "or", "tenHanh", "==", "\"T\"", ":", "return", "{", "\"id\"", ":", "3", ",", "\"tenHanh\"", ":", "\"Thủy\", ", "\"", "uc\": ", "2", " ", "\"", "enCuc\": ", "\"", "hủy nhị Cục\",", "", "\"css\"", ":", "\"hanhThuy\"", "}", "elif", "tenHanh", "==", "\"Hoa\"", "or", "tenHanh", "==", "\"H\"", ":", "return", "{", "\"id\"", ":", "4", ",", "\"tenHanh\"", ":", "\"Hỏa\", ", "\"", "uc\": ", "6", " ", "\"", "enCuc\": ", "\"", "ỏa lục Cục\",", "", "\"css\"", ":", "\"hanhHoa\"", "}", "elif", "tenHanh", "==", "\"Tho\"", "or", "tenHanh", "==", "\"O\"", ":", "return", "{", "\"id\"", ":", "5", ",", "\"tenHanh\"", ":", "\"Thổ\", ", "\"", "uc\": ", "5", " ", "\"", "enCuc\": ", "\"", "hổ ngũ Cục\",", "", "\"css\"", ":", "\"hanhTho\"", "}", "else", ":", "raise", "Exception", "(", "\"Tên Hành phải thuộc Kim (K), Mộc (M), Thủy (T), \\\n Hỏa (H) hoặc Thổ (O)\")", "" ]
Args: tenHanh (string): Tên Hành trong ngũ hành, Kim hoặc K, Moc hoặc M, Thuy hoặc T, Hoa hoặc H, Tho hoặc O Returns: Dictionary: ID của Hành, tên đầy đủ của Hành, số Cục của Hành Raises: Exception: Description
[ "Args", ":", "tenHanh", "(", "string", ")", ":", "Tên", "Hành", "trong", "ngũ", "hành", "Kim", "hoặc", "K", "Moc", "hoặc", "M", "Thuy", "hoặc", "T", "Hoa", "hoặc", "H", "Tho", "hoặc", "O" ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/AmDuong.py#L308-L338
train
doanguyen/lasotuvi
lasotuvi/AmDuong.py
nguHanhNapAm
def nguHanhNapAm(diaChi, thienCan, xuatBanMenh=False): """Sử dụng Ngũ Hành nạp âm để tính Hành của năm. Args: diaChi (integer): Số thứ tự của địa chi (Tý=1, Sửu=2,...) thienCan (integer): Số thứ tự của thiên can (Giáp=1, Ất=2,...) Returns: Trả về chữ viết tắt Hành của năm (K, T, H, O, M) """ banMenh = { "K1": "HẢI TRUNG KIM", "T1": "GIÁNG HẠ THỦY", "H1": "TÍCH LỊCH HỎA", "O1": "BÍCH THƯỢNG THỔ", "M1": "TANG ÐỐ MỘC", "T2": "ÐẠI KHÊ THỦY", "H2": "LƯ TRUNG HỎA", "O2": "THÀNH ÐẦU THỔ", "M2": "TÒNG BÁ MỘC", "K2": "KIM BẠCH KIM", "H3": "PHÚ ÐĂNG HỎA", "O3": "SA TRUNG THỔ", "M3": "ÐẠI LÂM MỘC", "K3": "BẠCH LẠP KIM", "T3": "TRƯỜNG LƯU THỦY", "K4": "SA TRUNG KIM", "T4": "THIÊN HÀ THỦY", "H4": "THIÊN THƯỢNG HỎA", "O4": "LỘ BÀN THỔ", "M4": "DƯƠNG LIỄU MỘC", "T5": "TRUYỀN TRUNG THỦY", "H5": "SƠN HẠ HỎA", "O5": "ÐẠI TRẠCH THỔ", "M5": "THẠCH LỰU MỘC", "K5": "KIẾM PHONG KIM", "H6": "SƠN ÐẦU HỎA", "O6": "ỐC THƯỢNG THỔ", "M6": "BÌNH ÐỊA MỘC", "K6": "XOA XUYẾN KIM", "T6": "ÐẠI HẢI THỦY"} matranNapAm = [ [0, "G", "Ất", "Bính", "Đinh", "Mậu", "Kỷ", "Canh", "Tân", "N", "Q"], [1, "K1", False, "T1", False, "H1", False, "O1", False, "M1", False], [2, False, "K1", False, "T1", False, "H1", False, "O1", False, "M1"], [3, "T2", False, "H2", False, "O2", False, "M2", False, "K2", False], [4, False, "T2", False, "H2", False, "O2", False, "M2", False, "K2"], [5, "H3", False, "O3", False, "M3", False, "K3", False, "T3", False], [6, False, "H3", False, "O3", False, "M3", False, "K3", False, "T3"], [7, "K4", False, "T4", False, "H4", False, "O4", False, "M4", False], [8, False, "K4", False, "T4", False, "H4", False, "O4", False, "M4"], [9, "T5", False, "H5", False, "O5", False, "M5", False, "K5", False], [10, False, "T5", False, "H5", False, "O5", False, "M5", False, "K5"], [11, "H6", False, "O6", False, "M6", False, "K6", False, "T6", False], [12, False, "H6", False, "O6", False, "M6", False, "K6", False, "T6"] ] try: nh = matranNapAm[diaChi][thienCan] if nh[0] in ["K", "M", "T", "H", "O"]: if xuatBanMenh is True: return banMenh[nh] else: return nh[0] except: raise Exception(nguHanhNapAm.__doc__)
python
def nguHanhNapAm(diaChi, thienCan, xuatBanMenh=False): """Sử dụng Ngũ Hành nạp âm để tính Hành của năm. Args: diaChi (integer): Số thứ tự của địa chi (Tý=1, Sửu=2,...) thienCan (integer): Số thứ tự của thiên can (Giáp=1, Ất=2,...) Returns: Trả về chữ viết tắt Hành của năm (K, T, H, O, M) """ banMenh = { "K1": "HẢI TRUNG KIM", "T1": "GIÁNG HẠ THỦY", "H1": "TÍCH LỊCH HỎA", "O1": "BÍCH THƯỢNG THỔ", "M1": "TANG ÐỐ MỘC", "T2": "ÐẠI KHÊ THỦY", "H2": "LƯ TRUNG HỎA", "O2": "THÀNH ÐẦU THỔ", "M2": "TÒNG BÁ MỘC", "K2": "KIM BẠCH KIM", "H3": "PHÚ ÐĂNG HỎA", "O3": "SA TRUNG THỔ", "M3": "ÐẠI LÂM MỘC", "K3": "BẠCH LẠP KIM", "T3": "TRƯỜNG LƯU THỦY", "K4": "SA TRUNG KIM", "T4": "THIÊN HÀ THỦY", "H4": "THIÊN THƯỢNG HỎA", "O4": "LỘ BÀN THỔ", "M4": "DƯƠNG LIỄU MỘC", "T5": "TRUYỀN TRUNG THỦY", "H5": "SƠN HẠ HỎA", "O5": "ÐẠI TRẠCH THỔ", "M5": "THẠCH LỰU MỘC", "K5": "KIẾM PHONG KIM", "H6": "SƠN ÐẦU HỎA", "O6": "ỐC THƯỢNG THỔ", "M6": "BÌNH ÐỊA MỘC", "K6": "XOA XUYẾN KIM", "T6": "ÐẠI HẢI THỦY"} matranNapAm = [ [0, "G", "Ất", "Bính", "Đinh", "Mậu", "Kỷ", "Canh", "Tân", "N", "Q"], [1, "K1", False, "T1", False, "H1", False, "O1", False, "M1", False], [2, False, "K1", False, "T1", False, "H1", False, "O1", False, "M1"], [3, "T2", False, "H2", False, "O2", False, "M2", False, "K2", False], [4, False, "T2", False, "H2", False, "O2", False, "M2", False, "K2"], [5, "H3", False, "O3", False, "M3", False, "K3", False, "T3", False], [6, False, "H3", False, "O3", False, "M3", False, "K3", False, "T3"], [7, "K4", False, "T4", False, "H4", False, "O4", False, "M4", False], [8, False, "K4", False, "T4", False, "H4", False, "O4", False, "M4"], [9, "T5", False, "H5", False, "O5", False, "M5", False, "K5", False], [10, False, "T5", False, "H5", False, "O5", False, "M5", False, "K5"], [11, "H6", False, "O6", False, "M6", False, "K6", False, "T6", False], [12, False, "H6", False, "O6", False, "M6", False, "K6", False, "T6"] ] try: nh = matranNapAm[diaChi][thienCan] if nh[0] in ["K", "M", "T", "H", "O"]: if xuatBanMenh is True: return banMenh[nh] else: return nh[0] except: raise Exception(nguHanhNapAm.__doc__)
[ "def", "nguHanhNapAm", "(", "diaChi", ",", "thienCan", ",", "xuatBanMenh", "=", "False", ")", ":", "banMenh", "=", "{", "\"K1\"", ":", "\"HẢI TRUNG KIM\",", "", "\"T1\"", ":", "\"GIÁNG HẠ THỦY\",", "", "\"H1\"", ":", "\"TÍCH LỊCH HỎA\",", "", "\"O1\"", ":", "\"BÍCH THƯỢNG THỔ\",", "", "\"M1\"", ":", "\"TANG ÐỐ MỘC\",", "", "\"T2\"", ":", "\"ÐẠI KHÊ THỦY\",", "", "\"H2\"", ":", "\"LƯ TRUNG HỎA\",", "", "\"O2\"", ":", "\"THÀNH ÐẦU THỔ\",", "", "\"M2\"", ":", "\"TÒNG BÁ MỘC\",", "", "\"K2\"", ":", "\"KIM BẠCH KIM\",", "", "\"H3\"", ":", "\"PHÚ ÐĂNG HỎA\",", "", "\"O3\"", ":", "\"SA TRUNG THỔ\",", "", "\"M3\"", ":", "\"ÐẠI LÂM MỘC\",", "", "\"K3\"", ":", "\"BẠCH LẠP KIM\",", "", "\"T3\"", ":", "\"TRƯỜNG LƯU THỦY\",", "", "\"K4\"", ":", "\"SA TRUNG KIM\"", ",", "\"T4\"", ":", "\"THIÊN HÀ THỦY\",", "", "\"H4\"", ":", "\"THIÊN THƯỢNG HỎA\",", "", "\"O4\"", ":", "\"LỘ BÀN THỔ\",", "", "\"M4\"", ":", "\"DƯƠNG LIỄU MỘC\",", "", "\"T5\"", ":", "\"TRUYỀN TRUNG THỦY\",", "", "\"H5\"", ":", "\"SƠN HẠ HỎA\",", "", "\"O5\"", ":", "\"ÐẠI TRẠCH THỔ\",", "", "\"M5\"", ":", "\"THẠCH LỰU MỘC\",", "", "\"K5\"", ":", "\"KIẾM PHONG KIM\",", "", "\"H6\"", ":", "\"SƠN ÐẦU HỎA\",", "", "\"O6\"", ":", "\"ỐC THƯỢNG THỔ\",", "", "\"M6\"", ":", "\"BÌNH ÐỊA MỘC\",", "", "\"K6\"", ":", "\"XOA XUYẾN KIM\",", "", "\"T6\"", ":", "\"ÐẠI HẢI THỦY\"}", "", "matranNapAm", "=", "[", "[", "0", ",", "\"G\"", ",", "\"Ất\", ", "\"", "ính\", \"", "Đ", "nh\", \"M", "ậ", "\", \"Kỷ\"", ",", "\"Canh\"", ",", "\"Tân\",", " ", "N\", \"Q", "\"", ",", "", "", "", "", "[", "1", ",", "\"K1\"", ",", "False", ",", "\"T1\"", ",", "False", ",", "\"H1\"", ",", "False", ",", "\"O1\"", ",", "False", ",", "\"M1\"", ",", "False", "]", ",", "[", "2", ",", "False", ",", "\"K1\"", ",", "False", ",", "\"T1\"", ",", "False", ",", "\"H1\"", ",", "False", ",", "\"O1\"", ",", "False", ",", "\"M1\"", "]", ",", "[", "3", ",", "\"T2\"", ",", "False", ",", "\"H2\"", ",", "False", ",", "\"O2\"", ",", "False", ",", "\"M2\"", ",", "False", ",", "\"K2\"", ",", "False", "]", ",", "[", "4", ",", "False", ",", "\"T2\"", ",", "False", ",", "\"H2\"", ",", "False", ",", "\"O2\"", ",", "False", ",", "\"M2\"", ",", "False", ",", "\"K2\"", "]", ",", "[", "5", ",", "\"H3\"", ",", "False", ",", "\"O3\"", ",", "False", ",", "\"M3\"", ",", "False", ",", "\"K3\"", ",", "False", ",", "\"T3\"", ",", "False", "]", ",", "[", "6", ",", "False", ",", "\"H3\"", ",", "False", ",", "\"O3\"", ",", "False", ",", "\"M3\"", ",", "False", ",", "\"K3\"", ",", "False", ",", "\"T3\"", "]", ",", "[", "7", ",", "\"K4\"", ",", "False", ",", "\"T4\"", ",", "False", ",", "\"H4\"", ",", "False", ",", "\"O4\"", ",", "False", ",", "\"M4\"", ",", "False", "]", ",", "[", "8", ",", "False", ",", "\"K4\"", ",", "False", ",", "\"T4\"", ",", "False", ",", "\"H4\"", ",", "False", ",", "\"O4\"", ",", "False", ",", "\"M4\"", "]", ",", "[", "9", ",", "\"T5\"", ",", "False", ",", "\"H5\"", ",", "False", ",", "\"O5\"", ",", "False", ",", "\"M5\"", ",", "False", ",", "\"K5\"", ",", "False", "]", ",", "[", "10", ",", "False", ",", "\"T5\"", ",", "False", ",", "\"H5\"", ",", "False", ",", "\"O5\"", ",", "False", ",", "\"M5\"", ",", "False", ",", "\"K5\"", "]", ",", "[", "11", ",", "\"H6\"", ",", "False", ",", "\"O6\"", ",", "False", ",", "\"M6\"", ",", "False", ",", "\"K6\"", ",", "False", ",", "\"T6\"", ",", "False", "]", ",", "[", "12", ",", "False", ",", "\"H6\"", ",", "False", ",", "\"O6\"", ",", "False", ",", "\"M6\"", ",", "False", ",", "\"K6\"", ",", "False", ",", "\"T6\"", "]", "]", "try", ":", "nh", "=", "matranNapAm", "[", "diaChi", "]", "[", "thienCan", "]", "if", "nh", "[", "0", "]", "in", "[", "\"K\"", ",", "\"M\"", ",", "\"T\"", ",", "\"H\"", ",", "\"O\"", "]", ":", "if", "xuatBanMenh", "is", "True", ":", "return", "banMenh", "[", "nh", "]", "else", ":", "return", "nh", "[", "0", "]", "except", ":", "raise", "Exception", "(", "nguHanhNapAm", ".", "__doc__", ")" ]
Sử dụng Ngũ Hành nạp âm để tính Hành của năm. Args: diaChi (integer): Số thứ tự của địa chi (Tý=1, Sửu=2,...) thienCan (integer): Số thứ tự của thiên can (Giáp=1, Ất=2,...) Returns: Trả về chữ viết tắt Hành của năm (K, T, H, O, M)
[ "Sử", "dụng", "Ngũ", "Hành", "nạp", "âm", "để", "tính", "Hành", "của", "năm", "." ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/AmDuong.py#L361-L425
train
doanguyen/lasotuvi
lasotuvi/AmDuong.py
timTuVi
def timTuVi(cuc, ngaySinhAmLich): """Tìm vị trí của sao Tử vi Args: cuc (TYPE): Description ngaySinhAmLich (TYPE): Description Returns: TYPE: Description Raises: Exception: Description """ cungDan = 3 # Vị trí cung Dần ban đầu là 3 cucBanDau = cuc if cuc not in [2, 3, 4, 5, 6]: # Tránh trường hợp infinite loop raise Exception("Số cục phải là 2, 3, 4, 5, 6") while cuc < ngaySinhAmLich: cuc += cucBanDau cungDan += 1 # Dịch vị trí cung Dần saiLech = cuc - ngaySinhAmLich if saiLech % 2 is 1: saiLech = -saiLech # Nếu sai lệch là chẵn thì tiến, lẻ thì lùi return dichCung(cungDan, saiLech)
python
def timTuVi(cuc, ngaySinhAmLich): """Tìm vị trí của sao Tử vi Args: cuc (TYPE): Description ngaySinhAmLich (TYPE): Description Returns: TYPE: Description Raises: Exception: Description """ cungDan = 3 # Vị trí cung Dần ban đầu là 3 cucBanDau = cuc if cuc not in [2, 3, 4, 5, 6]: # Tránh trường hợp infinite loop raise Exception("Số cục phải là 2, 3, 4, 5, 6") while cuc < ngaySinhAmLich: cuc += cucBanDau cungDan += 1 # Dịch vị trí cung Dần saiLech = cuc - ngaySinhAmLich if saiLech % 2 is 1: saiLech = -saiLech # Nếu sai lệch là chẵn thì tiến, lẻ thì lùi return dichCung(cungDan, saiLech)
[ "def", "timTuVi", "(", "cuc", ",", "ngaySinhAmLich", ")", ":", "cungDan", "=", "3", "# Vị trí cung Dần ban đầu là 3", "cucBanDau", "=", "cuc", "if", "cuc", "not", "in", "[", "2", ",", "3", ",", "4", ",", "5", ",", "6", "]", ":", "# Tránh trường hợp infinite loop", "raise", "Exception", "(", "\"Số cục phải là 2, 3, 4, 5, 6\")", "", "while", "cuc", "<", "ngaySinhAmLich", ":", "cuc", "+=", "cucBanDau", "cungDan", "+=", "1", "# Dịch vị trí cung Dần", "saiLech", "=", "cuc", "-", "ngaySinhAmLich", "if", "saiLech", "%", "2", "is", "1", ":", "saiLech", "=", "-", "saiLech", "# Nếu sai lệch là chẵn thì tiến, lẻ thì lùi", "return", "dichCung", "(", "cungDan", ",", "saiLech", ")" ]
Tìm vị trí của sao Tử vi Args: cuc (TYPE): Description ngaySinhAmLich (TYPE): Description Returns: TYPE: Description Raises: Exception: Description
[ "Tìm", "vị", "trí", "của", "sao", "Tử", "vi" ]
98383a3056f0a0633d6937d364c37eb788661c0d
https://github.com/doanguyen/lasotuvi/blob/98383a3056f0a0633d6937d364c37eb788661c0d/lasotuvi/AmDuong.py#L452-L475
train
zsethna/OLGA
olga/utils.py
nt2aa
def nt2aa(ntseq): """Translate a nucleotide sequence into an amino acid sequence. Parameters ---------- ntseq : str Nucleotide sequence composed of A, C, G, or T (uppercase or lowercase) Returns ------- aaseq : str Amino acid sequence Example -------- >>> nt2aa('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') 'CAWSVAPDRGGYTF' """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3, 'a': 0, 'c': 1, 'g': 2, 't': 3} aa_dict ='KQE*TPASRRG*ILVLNHDYTPASSRGCILVFKQE*TPASRRGWMLVLNHDYTPASSRGCILVF' return ''.join([aa_dict[nt2num[ntseq[i]] + 4*nt2num[ntseq[i+1]] + 16*nt2num[ntseq[i+2]]] for i in range(0, len(ntseq), 3) if i+2 < len(ntseq)])
python
def nt2aa(ntseq): """Translate a nucleotide sequence into an amino acid sequence. Parameters ---------- ntseq : str Nucleotide sequence composed of A, C, G, or T (uppercase or lowercase) Returns ------- aaseq : str Amino acid sequence Example -------- >>> nt2aa('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') 'CAWSVAPDRGGYTF' """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3, 'a': 0, 'c': 1, 'g': 2, 't': 3} aa_dict ='KQE*TPASRRG*ILVLNHDYTPASSRGCILVFKQE*TPASRRGWMLVLNHDYTPASSRGCILVF' return ''.join([aa_dict[nt2num[ntseq[i]] + 4*nt2num[ntseq[i+1]] + 16*nt2num[ntseq[i+2]]] for i in range(0, len(ntseq), 3) if i+2 < len(ntseq)])
[ "def", "nt2aa", "(", "ntseq", ")", ":", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", ",", "'a'", ":", "0", ",", "'c'", ":", "1", ",", "'g'", ":", "2", ",", "'t'", ":", "3", "}", "aa_dict", "=", "'KQE*TPASRRG*ILVLNHDYTPASSRGCILVFKQE*TPASRRGWMLVLNHDYTPASSRGCILVF'", "return", "''", ".", "join", "(", "[", "aa_dict", "[", "nt2num", "[", "ntseq", "[", "i", "]", "]", "+", "4", "*", "nt2num", "[", "ntseq", "[", "i", "+", "1", "]", "]", "+", "16", "*", "nt2num", "[", "ntseq", "[", "i", "+", "2", "]", "]", "]", "for", "i", "in", "range", "(", "0", ",", "len", "(", "ntseq", ")", ",", "3", ")", "if", "i", "+", "2", "<", "len", "(", "ntseq", ")", "]", ")" ]
Translate a nucleotide sequence into an amino acid sequence. Parameters ---------- ntseq : str Nucleotide sequence composed of A, C, G, or T (uppercase or lowercase) Returns ------- aaseq : str Amino acid sequence Example -------- >>> nt2aa('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') 'CAWSVAPDRGGYTF'
[ "Translate", "a", "nucleotide", "sequence", "into", "an", "amino", "acid", "sequence", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L31-L53
train
zsethna/OLGA
olga/utils.py
nt2codon_rep
def nt2codon_rep(ntseq): """Represent nucleotide sequence by sequence of codon symbols. 'Translates' the nucleotide sequence into a symbolic representation of 'amino acids' where each codon gets its own unique character symbol. These characters should be reserved only for representing the 64 individual codons --- note that this means it is important that this function matches the corresponding function in the preprocess script and that any custom alphabet does not use these symbols. Defining symbols for each individual codon allows for Pgen computation of inframe nucleotide sequences. Parameters ---------- ntseq : str A Nucleotide sequence (normally a CDR3 nucleotide sequence) to be 'translated' into the codon - symbol representation. Can be either uppercase or lowercase, but only composed of A, C, G, or T. Returns ------- codon_rep : str The codon - symbolic representation of ntseq. Note that if len(ntseq) == 3L --> len(codon_rep) == L Example -------- >>> nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') '\xbb\x96\xab\xb8\x8e\xb6\xa5\x92\xa8\xba\x9a\x93\x94\x9f' """ # nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3, 'a': 0, 'c': 1, 'g': 2, 't': 3} #Use single characters not in use to represent each individual codon --- this function is called in constructing the codon dictionary codon_rep ='\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf' return ''.join([codon_rep[nt2num[ntseq[i]] + 4*nt2num[ntseq[i+1]] + 16*nt2num[ntseq[i+2]]] for i in range(0, len(ntseq), 3) if i+2 < len(ntseq)])
python
def nt2codon_rep(ntseq): """Represent nucleotide sequence by sequence of codon symbols. 'Translates' the nucleotide sequence into a symbolic representation of 'amino acids' where each codon gets its own unique character symbol. These characters should be reserved only for representing the 64 individual codons --- note that this means it is important that this function matches the corresponding function in the preprocess script and that any custom alphabet does not use these symbols. Defining symbols for each individual codon allows for Pgen computation of inframe nucleotide sequences. Parameters ---------- ntseq : str A Nucleotide sequence (normally a CDR3 nucleotide sequence) to be 'translated' into the codon - symbol representation. Can be either uppercase or lowercase, but only composed of A, C, G, or T. Returns ------- codon_rep : str The codon - symbolic representation of ntseq. Note that if len(ntseq) == 3L --> len(codon_rep) == L Example -------- >>> nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') '\xbb\x96\xab\xb8\x8e\xb6\xa5\x92\xa8\xba\x9a\x93\x94\x9f' """ # nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3, 'a': 0, 'c': 1, 'g': 2, 't': 3} #Use single characters not in use to represent each individual codon --- this function is called in constructing the codon dictionary codon_rep ='\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf' return ''.join([codon_rep[nt2num[ntseq[i]] + 4*nt2num[ntseq[i+1]] + 16*nt2num[ntseq[i+2]]] for i in range(0, len(ntseq), 3) if i+2 < len(ntseq)])
[ "def", "nt2codon_rep", "(", "ntseq", ")", ":", "#", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", ",", "'a'", ":", "0", ",", "'c'", ":", "1", ",", "'g'", ":", "2", ",", "'t'", ":", "3", "}", "#Use single characters not in use to represent each individual codon --- this function is called in constructing the codon dictionary", "codon_rep", "=", "'\\x80\\x81\\x82\\x83\\x84\\x85\\x86\\x87\\x88\\x89\\x8a\\x8b\\x8c\\x8d\\x8e\\x8f\\x90\\x91\\x92\\x93\\x94\\x95\\x96\\x97\\x98\\x99\\x9a\\x9b\\x9c\\x9d\\x9e\\x9f\\xa0\\xa1\\xa2\\xa3\\xa4\\xa5\\xa6\\xa7\\xa8\\xa9\\xaa\\xab\\xac\\xad\\xae\\xaf\\xb0\\xb1\\xb2\\xb3\\xb4\\xb5\\xb6\\xb7\\xb8\\xb9\\xba\\xbb\\xbc\\xbd\\xbe\\xbf'", "return", "''", ".", "join", "(", "[", "codon_rep", "[", "nt2num", "[", "ntseq", "[", "i", "]", "]", "+", "4", "*", "nt2num", "[", "ntseq", "[", "i", "+", "1", "]", "]", "+", "16", "*", "nt2num", "[", "ntseq", "[", "i", "+", "2", "]", "]", "]", "for", "i", "in", "range", "(", "0", ",", "len", "(", "ntseq", ")", ",", "3", ")", "if", "i", "+", "2", "<", "len", "(", "ntseq", ")", "]", ")" ]
Represent nucleotide sequence by sequence of codon symbols. 'Translates' the nucleotide sequence into a symbolic representation of 'amino acids' where each codon gets its own unique character symbol. These characters should be reserved only for representing the 64 individual codons --- note that this means it is important that this function matches the corresponding function in the preprocess script and that any custom alphabet does not use these symbols. Defining symbols for each individual codon allows for Pgen computation of inframe nucleotide sequences. Parameters ---------- ntseq : str A Nucleotide sequence (normally a CDR3 nucleotide sequence) to be 'translated' into the codon - symbol representation. Can be either uppercase or lowercase, but only composed of A, C, G, or T. Returns ------- codon_rep : str The codon - symbolic representation of ntseq. Note that if len(ntseq) == 3L --> len(codon_rep) == L Example -------- >>> nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') '\xbb\x96\xab\xb8\x8e\xb6\xa5\x92\xa8\xba\x9a\x93\x94\x9f'
[ "Represent", "nucleotide", "sequence", "by", "sequence", "of", "codon", "symbols", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L55-L92
train
zsethna/OLGA
olga/utils.py
cutR_seq
def cutR_seq(seq, cutR, max_palindrome): """Cut genomic sequence from the right. Parameters ---------- seq : str Nucleotide sequence to be cut from the right cutR : int cutR - max_palindrome = how many nucleotides to cut from the right. Negative cutR implies complementary palindromic insertions. max_palindrome : int Length of the maximum palindromic insertion. Returns ------- seq : str Nucleotide sequence after being cut from the right Examples -------- >>> cutR_seq('TGCGCCAGCAGTGAGTC', 0, 4) 'TGCGCCAGCAGTGAGTCGACT' >>> cutR_seq('TGCGCCAGCAGTGAGTC', 8, 4) 'TGCGCCAGCAGTG' """ complement_dict = {'A': 'T', 'C': 'G', 'G': 'C', 'T': 'A'} #can include lower case if wanted if cutR < max_palindrome: seq = seq + ''.join([complement_dict[nt] for nt in seq[cutR - max_palindrome:]][::-1]) #reverse complement palindrome insertions else: seq = seq[:len(seq) - cutR + max_palindrome] #deletions return seq
python
def cutR_seq(seq, cutR, max_palindrome): """Cut genomic sequence from the right. Parameters ---------- seq : str Nucleotide sequence to be cut from the right cutR : int cutR - max_palindrome = how many nucleotides to cut from the right. Negative cutR implies complementary palindromic insertions. max_palindrome : int Length of the maximum palindromic insertion. Returns ------- seq : str Nucleotide sequence after being cut from the right Examples -------- >>> cutR_seq('TGCGCCAGCAGTGAGTC', 0, 4) 'TGCGCCAGCAGTGAGTCGACT' >>> cutR_seq('TGCGCCAGCAGTGAGTC', 8, 4) 'TGCGCCAGCAGTG' """ complement_dict = {'A': 'T', 'C': 'G', 'G': 'C', 'T': 'A'} #can include lower case if wanted if cutR < max_palindrome: seq = seq + ''.join([complement_dict[nt] for nt in seq[cutR - max_palindrome:]][::-1]) #reverse complement palindrome insertions else: seq = seq[:len(seq) - cutR + max_palindrome] #deletions return seq
[ "def", "cutR_seq", "(", "seq", ",", "cutR", ",", "max_palindrome", ")", ":", "complement_dict", "=", "{", "'A'", ":", "'T'", ",", "'C'", ":", "'G'", ",", "'G'", ":", "'C'", ",", "'T'", ":", "'A'", "}", "#can include lower case if wanted", "if", "cutR", "<", "max_palindrome", ":", "seq", "=", "seq", "+", "''", ".", "join", "(", "[", "complement_dict", "[", "nt", "]", "for", "nt", "in", "seq", "[", "cutR", "-", "max_palindrome", ":", "]", "]", "[", ":", ":", "-", "1", "]", ")", "#reverse complement palindrome insertions", "else", ":", "seq", "=", "seq", "[", ":", "len", "(", "seq", ")", "-", "cutR", "+", "max_palindrome", "]", "#deletions", "return", "seq" ]
Cut genomic sequence from the right. Parameters ---------- seq : str Nucleotide sequence to be cut from the right cutR : int cutR - max_palindrome = how many nucleotides to cut from the right. Negative cutR implies complementary palindromic insertions. max_palindrome : int Length of the maximum palindromic insertion. Returns ------- seq : str Nucleotide sequence after being cut from the right Examples -------- >>> cutR_seq('TGCGCCAGCAGTGAGTC', 0, 4) 'TGCGCCAGCAGTGAGTCGACT' >>> cutR_seq('TGCGCCAGCAGTGAGTC', 8, 4) 'TGCGCCAGCAGTG'
[ "Cut", "genomic", "sequence", "from", "the", "right", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L95-L127
train
zsethna/OLGA
olga/utils.py
cutL_seq
def cutL_seq(seq, cutL, max_palindrome): """Cut genomic sequence from the left. Parameters ---------- seq : str Nucleotide sequence to be cut from the right cutL : int cutL - max_palindrome = how many nucleotides to cut from the left. Negative cutL implies complementary palindromic insertions. max_palindrome : int Length of the maximum palindromic insertion. Returns ------- seq : str Nucleotide sequence after being cut from the left Examples -------- >>> cutL_seq('TGAACACTGAAGCTTTCTTT', 8, 4) 'CACTGAAGCTTTCTTT' >>> cutL_seq('TGAACACTGAAGCTTTCTTT', 0, 4) 'TTCATGAACACTGAAGCTTTCTTT' """ complement_dict = {'A': 'T', 'C': 'G', 'G': 'C', 'T': 'A'} #can include lower case if wanted if cutL < max_palindrome: seq = ''.join([complement_dict[nt] for nt in seq[:max_palindrome - cutL]][::-1]) + seq #reverse complement palindrome insertions else: seq = seq[cutL-max_palindrome:] #deletions return seq
python
def cutL_seq(seq, cutL, max_palindrome): """Cut genomic sequence from the left. Parameters ---------- seq : str Nucleotide sequence to be cut from the right cutL : int cutL - max_palindrome = how many nucleotides to cut from the left. Negative cutL implies complementary palindromic insertions. max_palindrome : int Length of the maximum palindromic insertion. Returns ------- seq : str Nucleotide sequence after being cut from the left Examples -------- >>> cutL_seq('TGAACACTGAAGCTTTCTTT', 8, 4) 'CACTGAAGCTTTCTTT' >>> cutL_seq('TGAACACTGAAGCTTTCTTT', 0, 4) 'TTCATGAACACTGAAGCTTTCTTT' """ complement_dict = {'A': 'T', 'C': 'G', 'G': 'C', 'T': 'A'} #can include lower case if wanted if cutL < max_palindrome: seq = ''.join([complement_dict[nt] for nt in seq[:max_palindrome - cutL]][::-1]) + seq #reverse complement palindrome insertions else: seq = seq[cutL-max_palindrome:] #deletions return seq
[ "def", "cutL_seq", "(", "seq", ",", "cutL", ",", "max_palindrome", ")", ":", "complement_dict", "=", "{", "'A'", ":", "'T'", ",", "'C'", ":", "'G'", ",", "'G'", ":", "'C'", ",", "'T'", ":", "'A'", "}", "#can include lower case if wanted", "if", "cutL", "<", "max_palindrome", ":", "seq", "=", "''", ".", "join", "(", "[", "complement_dict", "[", "nt", "]", "for", "nt", "in", "seq", "[", ":", "max_palindrome", "-", "cutL", "]", "]", "[", ":", ":", "-", "1", "]", ")", "+", "seq", "#reverse complement palindrome insertions", "else", ":", "seq", "=", "seq", "[", "cutL", "-", "max_palindrome", ":", "]", "#deletions", "return", "seq" ]
Cut genomic sequence from the left. Parameters ---------- seq : str Nucleotide sequence to be cut from the right cutL : int cutL - max_palindrome = how many nucleotides to cut from the left. Negative cutL implies complementary palindromic insertions. max_palindrome : int Length of the maximum palindromic insertion. Returns ------- seq : str Nucleotide sequence after being cut from the left Examples -------- >>> cutL_seq('TGAACACTGAAGCTTTCTTT', 8, 4) 'CACTGAAGCTTTCTTT' >>> cutL_seq('TGAACACTGAAGCTTTCTTT', 0, 4) 'TTCATGAACACTGAAGCTTTCTTT'
[ "Cut", "genomic", "sequence", "from", "the", "left", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L129-L162
train
zsethna/OLGA
olga/utils.py
construct_codons_dict
def construct_codons_dict(alphabet_file = None): """Generate the sub_codons_right dictionary of codon suffixes. syntax of custom alphabet_files: char: list,of,amino,acids,or,codons,separated,by,commas Parameters ---------- alphabet_file : str File name for a custom alphabet definition. If no file is provided, the default alphabet is used, i.e. standard amino acids, undetermined amino acids (B, J, X, and Z), and single codon symbols. Returns ------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. """ #Some symbols can't be used in the CDR3 sequences in order to allow for #regular expression parsing and general manipulation. protected_symbols = [' ', '\t', '\n', '\x0b', '\x0c', '\r', ':', ',', ';', '[', ']', '{', '}', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9'] #construct list of all 64 codons codons = [i + j + k for i in 'ACGT' for j in 'ACGT' for k in 'ACGT'] codons_dict = {} #add standard amino acids symbols to the dict (i.e. 'ACDEFGHIKLMNPQRSTVWY*'). #these symbols CANNOT be overwritten by custom alphabet files for codon in codons: codons_dict[nt2aa(codon)] = codons_dict.get(nt2aa(codon), []) + [codon] #add single codon symbols to allow for inframe ntseq pgen computation #'\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf' #these symbols CANNOT be overwritten by custom alphabet files for codon in codons: codons_dict[nt2codon_rep(codon)] = [codon] #Check to see if custom alphabet file is supplied, else use default alphabet #Include standard ambigious amino acids. #these symbols CAN be overwritten by custom alphabet files expanded_alphabet = {} expanded_alphabet['B'] = ['D','N'] expanded_alphabet['J'] = ['I', 'L'] expanded_alphabet['X'] = ['A','C','D','E','F','G','H','I','K','L','M','N','P','Q','R','S','T','V','W','Y'] expanded_alphabet['Z'] = ['E', 'Q'] if alphabet_file is not None: #Use custom alphabet file definitions alphabet_f = open(alphabet_file, 'r') for line in alphabet_f: #assumed syntax is of a line is: #s: a1, a2, a3, a4, a5, ..., aN #where s is a single character symbol that isn't reserved, and all #of the a's are either amino acid symbols or codons. Whitespaces #will be stripped as will brackets if the a's are presented as a #list. c_symbol = line.split(':', 1)[0].strip(''.join(protected_symbols)) #Note there shouldn't be any additional colons -- this is a protected symbol. c_aa_codon_list_str = line.split(':', 1)[1] expanded_alphabet[c_symbol] = [x.strip(''.join(protected_symbols)) for x in c_aa_codon_list_str.split(',')] alphabet_f.close() for symbol in expanded_alphabet.keys(): #Double check that the symbol isn't already used (important particularly for the single codon representation) if symbol in codons_dict.keys(): print symbol + " is already used as an 'amino acid' symbol for codons: " print codons_dict[symbol] continue elif not len(symbol) == 1: #Check that the custom symbol is a single character print "Can't use " + symbol + " as a custom 'amino acid' definitions as such symbols must be single characters." continue elif symbol in protected_symbols: #This elif shouldn't trigger due to the stripping of protected symbols. print symbol + " is a protected character" current_codon_collection = set() for x in expanded_alphabet[symbol]: if x in codons_dict.keys(): #Check if reference to an amino acid or other amino acid symbol current_codon_collection = current_codon_collection.union(codons_dict[x]) #If so, add those codons to the new collection elif x.upper() in codons: #Check if specifying a single codon current_codon_collection.add(x.upper()) #If so, add the codon to the new collection elif len(x) == 0: #fully stripped away continue else: #If not, don't recognize the addition and continue. print 'Unfamiliar amino acid symbol or codon: ' + x continue codons_dict[symbol] = list(current_codon_collection) return codons_dict
python
def construct_codons_dict(alphabet_file = None): """Generate the sub_codons_right dictionary of codon suffixes. syntax of custom alphabet_files: char: list,of,amino,acids,or,codons,separated,by,commas Parameters ---------- alphabet_file : str File name for a custom alphabet definition. If no file is provided, the default alphabet is used, i.e. standard amino acids, undetermined amino acids (B, J, X, and Z), and single codon symbols. Returns ------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. """ #Some symbols can't be used in the CDR3 sequences in order to allow for #regular expression parsing and general manipulation. protected_symbols = [' ', '\t', '\n', '\x0b', '\x0c', '\r', ':', ',', ';', '[', ']', '{', '}', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9'] #construct list of all 64 codons codons = [i + j + k for i in 'ACGT' for j in 'ACGT' for k in 'ACGT'] codons_dict = {} #add standard amino acids symbols to the dict (i.e. 'ACDEFGHIKLMNPQRSTVWY*'). #these symbols CANNOT be overwritten by custom alphabet files for codon in codons: codons_dict[nt2aa(codon)] = codons_dict.get(nt2aa(codon), []) + [codon] #add single codon symbols to allow for inframe ntseq pgen computation #'\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf' #these symbols CANNOT be overwritten by custom alphabet files for codon in codons: codons_dict[nt2codon_rep(codon)] = [codon] #Check to see if custom alphabet file is supplied, else use default alphabet #Include standard ambigious amino acids. #these symbols CAN be overwritten by custom alphabet files expanded_alphabet = {} expanded_alphabet['B'] = ['D','N'] expanded_alphabet['J'] = ['I', 'L'] expanded_alphabet['X'] = ['A','C','D','E','F','G','H','I','K','L','M','N','P','Q','R','S','T','V','W','Y'] expanded_alphabet['Z'] = ['E', 'Q'] if alphabet_file is not None: #Use custom alphabet file definitions alphabet_f = open(alphabet_file, 'r') for line in alphabet_f: #assumed syntax is of a line is: #s: a1, a2, a3, a4, a5, ..., aN #where s is a single character symbol that isn't reserved, and all #of the a's are either amino acid symbols or codons. Whitespaces #will be stripped as will brackets if the a's are presented as a #list. c_symbol = line.split(':', 1)[0].strip(''.join(protected_symbols)) #Note there shouldn't be any additional colons -- this is a protected symbol. c_aa_codon_list_str = line.split(':', 1)[1] expanded_alphabet[c_symbol] = [x.strip(''.join(protected_symbols)) for x in c_aa_codon_list_str.split(',')] alphabet_f.close() for symbol in expanded_alphabet.keys(): #Double check that the symbol isn't already used (important particularly for the single codon representation) if symbol in codons_dict.keys(): print symbol + " is already used as an 'amino acid' symbol for codons: " print codons_dict[symbol] continue elif not len(symbol) == 1: #Check that the custom symbol is a single character print "Can't use " + symbol + " as a custom 'amino acid' definitions as such symbols must be single characters." continue elif symbol in protected_symbols: #This elif shouldn't trigger due to the stripping of protected symbols. print symbol + " is a protected character" current_codon_collection = set() for x in expanded_alphabet[symbol]: if x in codons_dict.keys(): #Check if reference to an amino acid or other amino acid symbol current_codon_collection = current_codon_collection.union(codons_dict[x]) #If so, add those codons to the new collection elif x.upper() in codons: #Check if specifying a single codon current_codon_collection.add(x.upper()) #If so, add the codon to the new collection elif len(x) == 0: #fully stripped away continue else: #If not, don't recognize the addition and continue. print 'Unfamiliar amino acid symbol or codon: ' + x continue codons_dict[symbol] = list(current_codon_collection) return codons_dict
[ "def", "construct_codons_dict", "(", "alphabet_file", "=", "None", ")", ":", "#Some symbols can't be used in the CDR3 sequences in order to allow for", "#regular expression parsing and general manipulation.", "protected_symbols", "=", "[", "' '", ",", "'\\t'", ",", "'\\n'", ",", "'\\x0b'", ",", "'\\x0c'", ",", "'\\r'", ",", "':'", ",", "','", ",", "';'", ",", "'['", ",", "']'", ",", "'{'", ",", "'}'", ",", "'0'", ",", "'1'", ",", "'2'", ",", "'3'", ",", "'4'", ",", "'5'", ",", "'6'", ",", "'7'", ",", "'8'", ",", "'9'", "]", "#construct list of all 64 codons", "codons", "=", "[", "i", "+", "j", "+", "k", "for", "i", "in", "'ACGT'", "for", "j", "in", "'ACGT'", "for", "k", "in", "'ACGT'", "]", "codons_dict", "=", "{", "}", "#add standard amino acids symbols to the dict (i.e. 'ACDEFGHIKLMNPQRSTVWY*').", "#these symbols CANNOT be overwritten by custom alphabet files", "for", "codon", "in", "codons", ":", "codons_dict", "[", "nt2aa", "(", "codon", ")", "]", "=", "codons_dict", ".", "get", "(", "nt2aa", "(", "codon", ")", ",", "[", "]", ")", "+", "[", "codon", "]", "#add single codon symbols to allow for inframe ntseq pgen computation", "#'\\x80\\x81\\x82\\x83\\x84\\x85\\x86\\x87\\x88\\x89\\x8a\\x8b\\x8c\\x8d\\x8e\\x8f\\x90\\x91\\x92\\x93\\x94\\x95\\x96\\x97\\x98\\x99\\x9a\\x9b\\x9c\\x9d\\x9e\\x9f\\xa0\\xa1\\xa2\\xa3\\xa4\\xa5\\xa6\\xa7\\xa8\\xa9\\xaa\\xab\\xac\\xad\\xae\\xaf\\xb0\\xb1\\xb2\\xb3\\xb4\\xb5\\xb6\\xb7\\xb8\\xb9\\xba\\xbb\\xbc\\xbd\\xbe\\xbf'", "#these symbols CANNOT be overwritten by custom alphabet files", "for", "codon", "in", "codons", ":", "codons_dict", "[", "nt2codon_rep", "(", "codon", ")", "]", "=", "[", "codon", "]", "#Check to see if custom alphabet file is supplied, else use default alphabet", "#Include standard ambigious amino acids.", "#these symbols CAN be overwritten by custom alphabet files", "expanded_alphabet", "=", "{", "}", "expanded_alphabet", "[", "'B'", "]", "=", "[", "'D'", ",", "'N'", "]", "expanded_alphabet", "[", "'J'", "]", "=", "[", "'I'", ",", "'L'", "]", "expanded_alphabet", "[", "'X'", "]", "=", "[", "'A'", ",", "'C'", ",", "'D'", ",", "'E'", ",", "'F'", ",", "'G'", ",", "'H'", ",", "'I'", ",", "'K'", ",", "'L'", ",", "'M'", ",", "'N'", ",", "'P'", ",", "'Q'", ",", "'R'", ",", "'S'", ",", "'T'", ",", "'V'", ",", "'W'", ",", "'Y'", "]", "expanded_alphabet", "[", "'Z'", "]", "=", "[", "'E'", ",", "'Q'", "]", "if", "alphabet_file", "is", "not", "None", ":", "#Use custom alphabet file definitions", "alphabet_f", "=", "open", "(", "alphabet_file", ",", "'r'", ")", "for", "line", "in", "alphabet_f", ":", "#assumed syntax is of a line is:", "#s: a1, a2, a3, a4, a5, ..., aN", "#where s is a single character symbol that isn't reserved, and all ", "#of the a's are either amino acid symbols or codons. Whitespaces", "#will be stripped as will brackets if the a's are presented as a", "#list.", "c_symbol", "=", "line", ".", "split", "(", "':'", ",", "1", ")", "[", "0", "]", ".", "strip", "(", "''", ".", "join", "(", "protected_symbols", ")", ")", "#Note there shouldn't be any additional colons -- this is a protected symbol.", "c_aa_codon_list_str", "=", "line", ".", "split", "(", "':'", ",", "1", ")", "[", "1", "]", "expanded_alphabet", "[", "c_symbol", "]", "=", "[", "x", ".", "strip", "(", "''", ".", "join", "(", "protected_symbols", ")", ")", "for", "x", "in", "c_aa_codon_list_str", ".", "split", "(", "','", ")", "]", "alphabet_f", ".", "close", "(", ")", "for", "symbol", "in", "expanded_alphabet", ".", "keys", "(", ")", ":", "#Double check that the symbol isn't already used (important particularly for the single codon representation)", "if", "symbol", "in", "codons_dict", ".", "keys", "(", ")", ":", "print", "symbol", "+", "\" is already used as an 'amino acid' symbol for codons: \"", "print", "codons_dict", "[", "symbol", "]", "continue", "elif", "not", "len", "(", "symbol", ")", "==", "1", ":", "#Check that the custom symbol is a single character", "print", "\"Can't use \"", "+", "symbol", "+", "\" as a custom 'amino acid' definitions as such symbols must be single characters.\"", "continue", "elif", "symbol", "in", "protected_symbols", ":", "#This elif shouldn't trigger due to the stripping of protected symbols.", "print", "symbol", "+", "\" is a protected character\"", "current_codon_collection", "=", "set", "(", ")", "for", "x", "in", "expanded_alphabet", "[", "symbol", "]", ":", "if", "x", "in", "codons_dict", ".", "keys", "(", ")", ":", "#Check if reference to an amino acid or other amino acid symbol", "current_codon_collection", "=", "current_codon_collection", ".", "union", "(", "codons_dict", "[", "x", "]", ")", "#If so, add those codons to the new collection", "elif", "x", ".", "upper", "(", ")", "in", "codons", ":", "#Check if specifying a single codon", "current_codon_collection", ".", "add", "(", "x", ".", "upper", "(", ")", ")", "#If so, add the codon to the new collection", "elif", "len", "(", "x", ")", "==", "0", ":", "#fully stripped away", "continue", "else", ":", "#If not, don't recognize the addition and continue.", "print", "'Unfamiliar amino acid symbol or codon: '", "+", "x", "continue", "codons_dict", "[", "symbol", "]", "=", "list", "(", "current_codon_collection", ")", "return", "codons_dict" ]
Generate the sub_codons_right dictionary of codon suffixes. syntax of custom alphabet_files: char: list,of,amino,acids,or,codons,separated,by,commas Parameters ---------- alphabet_file : str File name for a custom alphabet definition. If no file is provided, the default alphabet is used, i.e. standard amino acids, undetermined amino acids (B, J, X, and Z), and single codon symbols. Returns ------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol.
[ "Generate", "the", "sub_codons_right", "dictionary", "of", "codon", "suffixes", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L165-L259
train
zsethna/OLGA
olga/utils.py
generate_sub_codons_left
def generate_sub_codons_left(codons_dict): """Generate the sub_codons_left dictionary of codon prefixes. Parameters ---------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. Returns ------- sub_codons_left : dict Dictionary of the 1 and 2 nucleotide prefixes (read from 5') for each codon in an 'amino acid' grouping """ sub_codons_left = {} for aa in codons_dict.keys(): sub_codons_left[aa] = list(set([x[0] for x in codons_dict[aa]] + [x[:2] for x in codons_dict[aa]])) return sub_codons_left
python
def generate_sub_codons_left(codons_dict): """Generate the sub_codons_left dictionary of codon prefixes. Parameters ---------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. Returns ------- sub_codons_left : dict Dictionary of the 1 and 2 nucleotide prefixes (read from 5') for each codon in an 'amino acid' grouping """ sub_codons_left = {} for aa in codons_dict.keys(): sub_codons_left[aa] = list(set([x[0] for x in codons_dict[aa]] + [x[:2] for x in codons_dict[aa]])) return sub_codons_left
[ "def", "generate_sub_codons_left", "(", "codons_dict", ")", ":", "sub_codons_left", "=", "{", "}", "for", "aa", "in", "codons_dict", ".", "keys", "(", ")", ":", "sub_codons_left", "[", "aa", "]", "=", "list", "(", "set", "(", "[", "x", "[", "0", "]", "for", "x", "in", "codons_dict", "[", "aa", "]", "]", "+", "[", "x", "[", ":", "2", "]", "for", "x", "in", "codons_dict", "[", "aa", "]", "]", ")", ")", "return", "sub_codons_left" ]
Generate the sub_codons_left dictionary of codon prefixes. Parameters ---------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. Returns ------- sub_codons_left : dict Dictionary of the 1 and 2 nucleotide prefixes (read from 5') for each codon in an 'amino acid' grouping
[ "Generate", "the", "sub_codons_left", "dictionary", "of", "codon", "prefixes", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L261-L281
train
zsethna/OLGA
olga/utils.py
generate_sub_codons_right
def generate_sub_codons_right(codons_dict): """Generate the sub_codons_right dictionary of codon suffixes. Parameters ---------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. Returns ------- sub_codons_right : dict Dictionary of the 1 and 2 nucleotide suffixes (read from 5') for each codon in an 'amino acid' grouping. """ sub_codons_right = {} for aa in codons_dict.keys(): sub_codons_right[aa] = list(set([x[-1] for x in codons_dict[aa]] + [x[-2:] for x in codons_dict[aa]])) return sub_codons_right
python
def generate_sub_codons_right(codons_dict): """Generate the sub_codons_right dictionary of codon suffixes. Parameters ---------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. Returns ------- sub_codons_right : dict Dictionary of the 1 and 2 nucleotide suffixes (read from 5') for each codon in an 'amino acid' grouping. """ sub_codons_right = {} for aa in codons_dict.keys(): sub_codons_right[aa] = list(set([x[-1] for x in codons_dict[aa]] + [x[-2:] for x in codons_dict[aa]])) return sub_codons_right
[ "def", "generate_sub_codons_right", "(", "codons_dict", ")", ":", "sub_codons_right", "=", "{", "}", "for", "aa", "in", "codons_dict", ".", "keys", "(", ")", ":", "sub_codons_right", "[", "aa", "]", "=", "list", "(", "set", "(", "[", "x", "[", "-", "1", "]", "for", "x", "in", "codons_dict", "[", "aa", "]", "]", "+", "[", "x", "[", "-", "2", ":", "]", "for", "x", "in", "codons_dict", "[", "aa", "]", "]", ")", ")", "return", "sub_codons_right" ]
Generate the sub_codons_right dictionary of codon suffixes. Parameters ---------- codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. Returns ------- sub_codons_right : dict Dictionary of the 1 and 2 nucleotide suffixes (read from 5') for each codon in an 'amino acid' grouping.
[ "Generate", "the", "sub_codons_right", "dictionary", "of", "codon", "suffixes", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L283-L303
train
zsethna/OLGA
olga/utils.py
determine_seq_type
def determine_seq_type(seq, aa_alphabet): """Determine the type of a sequence. Parameters ---------- seq : str Sequence to be typed. aa_alphabet : str String of all characters recoginized as 'amino acids'. (i.e. the keys of codons_dict: aa_alphabet = ''.join(codons_dict.keys()) ) Returns ------- seq_type : str The type of sequence (ntseq, aaseq, regex, None) seq is. Example -------- >>> determine_seq_type('TGTGCCAGCAGTTCCGAAGGGGCGGGAGGGCCCTCCCTGAGAGGTCATGAGCAGTTCTTC', aa_alphabet) 'ntseq' >>> determine_seq_type('CSARDX[TV]GNX{0,}', aa_alphabet) 'regex """ if all([x in 'ACGTacgt' for x in seq]): return 'ntseq' elif all([x in aa_alphabet for x in seq]): return 'aaseq' elif all([x in aa_alphabet + '[]{}0123456789,']): return 'regex'
python
def determine_seq_type(seq, aa_alphabet): """Determine the type of a sequence. Parameters ---------- seq : str Sequence to be typed. aa_alphabet : str String of all characters recoginized as 'amino acids'. (i.e. the keys of codons_dict: aa_alphabet = ''.join(codons_dict.keys()) ) Returns ------- seq_type : str The type of sequence (ntseq, aaseq, regex, None) seq is. Example -------- >>> determine_seq_type('TGTGCCAGCAGTTCCGAAGGGGCGGGAGGGCCCTCCCTGAGAGGTCATGAGCAGTTCTTC', aa_alphabet) 'ntseq' >>> determine_seq_type('CSARDX[TV]GNX{0,}', aa_alphabet) 'regex """ if all([x in 'ACGTacgt' for x in seq]): return 'ntseq' elif all([x in aa_alphabet for x in seq]): return 'aaseq' elif all([x in aa_alphabet + '[]{}0123456789,']): return 'regex'
[ "def", "determine_seq_type", "(", "seq", ",", "aa_alphabet", ")", ":", "if", "all", "(", "[", "x", "in", "'ACGTacgt'", "for", "x", "in", "seq", "]", ")", ":", "return", "'ntseq'", "elif", "all", "(", "[", "x", "in", "aa_alphabet", "for", "x", "in", "seq", "]", ")", ":", "return", "'aaseq'", "elif", "all", "(", "[", "x", "in", "aa_alphabet", "+", "'[]{}0123456789,'", "]", ")", ":", "return", "'regex'" ]
Determine the type of a sequence. Parameters ---------- seq : str Sequence to be typed. aa_alphabet : str String of all characters recoginized as 'amino acids'. (i.e. the keys of codons_dict: aa_alphabet = ''.join(codons_dict.keys()) ) Returns ------- seq_type : str The type of sequence (ntseq, aaseq, regex, None) seq is. Example -------- >>> determine_seq_type('TGTGCCAGCAGTTCCGAAGGGGCGGGAGGGCCCTCCCTGAGAGGTCATGAGCAGTTCTTC', aa_alphabet) 'ntseq' >>> determine_seq_type('CSARDX[TV]GNX{0,}', aa_alphabet) 'regex
[ "Determine", "the", "type", "of", "a", "sequence", ".", "Parameters", "----------" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L305-L336
train
zsethna/OLGA
olga/utils.py
calc_steady_state_dist
def calc_steady_state_dist(R): """Calculate the steady state dist of a 4 state markov transition matrix. Parameters ---------- R : ndarray Markov transition matrix Returns ------- p_ss : ndarray Steady state probability distribution """ #Calc steady state distribution for a dinucleotide bias matrix w, v = np.linalg.eig(R) for i in range(4): if np.abs(w[i] - 1) < 1e-8: return np.real(v[:, i] / np.sum(v[:, i])) return -1
python
def calc_steady_state_dist(R): """Calculate the steady state dist of a 4 state markov transition matrix. Parameters ---------- R : ndarray Markov transition matrix Returns ------- p_ss : ndarray Steady state probability distribution """ #Calc steady state distribution for a dinucleotide bias matrix w, v = np.linalg.eig(R) for i in range(4): if np.abs(w[i] - 1) < 1e-8: return np.real(v[:, i] / np.sum(v[:, i])) return -1
[ "def", "calc_steady_state_dist", "(", "R", ")", ":", "#Calc steady state distribution for a dinucleotide bias matrix", "w", ",", "v", "=", "np", ".", "linalg", ".", "eig", "(", "R", ")", "for", "i", "in", "range", "(", "4", ")", ":", "if", "np", ".", "abs", "(", "w", "[", "i", "]", "-", "1", ")", "<", "1e-8", ":", "return", "np", ".", "real", "(", "v", "[", ":", ",", "i", "]", "/", "np", ".", "sum", "(", "v", "[", ":", ",", "i", "]", ")", ")", "return", "-", "1" ]
Calculate the steady state dist of a 4 state markov transition matrix. Parameters ---------- R : ndarray Markov transition matrix Returns ------- p_ss : ndarray Steady state probability distribution
[ "Calculate", "the", "steady", "state", "dist", "of", "a", "4", "state", "markov", "transition", "matrix", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/utils.py#L340-L361
train
zsethna/OLGA
olga/sequence_generation.py
rnd_ins_seq
def rnd_ins_seq(ins_len, C_R, CP_first_nt): """Generate a random insertion nucleotide sequence of length ins_len. Draws the sequence identity (for a set length) from the distribution defined by the dinucleotide markov model of transition matrix R. Parameters ---------- ins_len : int Length of nucleotide sequence to be inserted. C_R : ndarray (4, 4) array of the cumulative transition probabilities defined by the Markov transition matrix R CP_first_nt : ndarray (4,) array of the cumulative probabilities for the first inserted nucleotide Returns ------- seq : str Randomly generated insertion sequence of length ins_len. Examples -------- >>> rnd_ins_seq(7, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'GATGGAC' >>> rnd_ins_seq(7, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'ACCCCCG' >>> rnd_ins_seq(3, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'GCC' """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num2nt = 'ACGT' if ins_len == 0: return '' seq = num2nt[CP_first_nt.searchsorted(np.random.random())] ins_len += -1 while ins_len > 0: seq += num2nt[C_R[nt2num[seq[-1]], :].searchsorted(np.random.random())] ins_len += -1 return seq
python
def rnd_ins_seq(ins_len, C_R, CP_first_nt): """Generate a random insertion nucleotide sequence of length ins_len. Draws the sequence identity (for a set length) from the distribution defined by the dinucleotide markov model of transition matrix R. Parameters ---------- ins_len : int Length of nucleotide sequence to be inserted. C_R : ndarray (4, 4) array of the cumulative transition probabilities defined by the Markov transition matrix R CP_first_nt : ndarray (4,) array of the cumulative probabilities for the first inserted nucleotide Returns ------- seq : str Randomly generated insertion sequence of length ins_len. Examples -------- >>> rnd_ins_seq(7, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'GATGGAC' >>> rnd_ins_seq(7, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'ACCCCCG' >>> rnd_ins_seq(3, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'GCC' """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num2nt = 'ACGT' if ins_len == 0: return '' seq = num2nt[CP_first_nt.searchsorted(np.random.random())] ins_len += -1 while ins_len > 0: seq += num2nt[C_R[nt2num[seq[-1]], :].searchsorted(np.random.random())] ins_len += -1 return seq
[ "def", "rnd_ins_seq", "(", "ins_len", ",", "C_R", ",", "CP_first_nt", ")", ":", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "num2nt", "=", "'ACGT'", "if", "ins_len", "==", "0", ":", "return", "''", "seq", "=", "num2nt", "[", "CP_first_nt", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "]", "ins_len", "+=", "-", "1", "while", "ins_len", ">", "0", ":", "seq", "+=", "num2nt", "[", "C_R", "[", "nt2num", "[", "seq", "[", "-", "1", "]", "]", ",", ":", "]", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "]", "ins_len", "+=", "-", "1", "return", "seq" ]
Generate a random insertion nucleotide sequence of length ins_len. Draws the sequence identity (for a set length) from the distribution defined by the dinucleotide markov model of transition matrix R. Parameters ---------- ins_len : int Length of nucleotide sequence to be inserted. C_R : ndarray (4, 4) array of the cumulative transition probabilities defined by the Markov transition matrix R CP_first_nt : ndarray (4,) array of the cumulative probabilities for the first inserted nucleotide Returns ------- seq : str Randomly generated insertion sequence of length ins_len. Examples -------- >>> rnd_ins_seq(7, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'GATGGAC' >>> rnd_ins_seq(7, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'ACCCCCG' >>> rnd_ins_seq(3, CP_generative_model['C_Rvd'], CP_generative_model['C_first_nt_bias_insVD']) 'GCC'
[ "Generate", "a", "random", "insertion", "nucleotide", "sequence", "of", "length", "ins_len", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/sequence_generation.py#L457-L502
train
zsethna/OLGA
olga/sequence_generation.py
SequenceGenerationVDJ.gen_rnd_prod_CDR3
def gen_rnd_prod_CDR3(self, conserved_J_residues = 'FVW'): """Generate a productive CDR3 seq from a Monte Carlo draw of the model. Parameters ---------- conserved_J_residues : str, optional Conserved amino acid residues defining the CDR3 on the J side (normally F, V, and/or W) Returns ------- ntseq : str Productive CDR3 nucleotide sequence aaseq : str CDR3 amino acid sequence (aaseq = nt2aa(ntseq)) V_choice : int Index of V allele chosen to generate the CDR3 seq J_choice : int Index of J allele chosen to generate the CDR3 seq """ coding_pass = False while ~coding_pass: recomb_events = self.choose_random_recomb_events() V_seq = self.cutV_genomic_CDR3_segs[recomb_events['V']] #This both checks that the position of the conserved C is #identified and that the V isn't fully deleted out of the CDR3 #region if len(V_seq) <= max(recomb_events['delV'], 0): continue D_seq = self.cutD_genomic_CDR3_segs[recomb_events['D']] J_seq = self.cutJ_genomic_CDR3_segs[recomb_events['J']] #We check that the D and J aren't deleted more than allowed. Note #the generative model really should reflect this structure already if len(D_seq) < (recomb_events['delDl'] + recomb_events['delDr']) or len(J_seq) < recomb_events['delJ']: continue V_seq = V_seq[:len(V_seq) - recomb_events['delV']] D_seq = D_seq[recomb_events['delDl']:len(D_seq)-recomb_events['delDr']] J_seq = J_seq[recomb_events['delJ']:] if (len(V_seq)+ len(D_seq) + len(J_seq) + recomb_events['insVD'] + recomb_events['insDJ']) % 3 != 0: continue insVD_seq = rnd_ins_seq(recomb_events['insVD'], self.C_Rvd, self.C_first_nt_bias_insVD) insDJ_seq = rnd_ins_seq(recomb_events['insDJ'], self.C_Rdj, self.C_first_nt_bias_insDJ)[::-1] #have to reverse the DJ seq #Translate to amino acid sequence, see if productive ntseq = V_seq + insVD_seq + D_seq + insDJ_seq + J_seq aaseq = nt2aa(ntseq) if '*' not in aaseq and aaseq[0]=='C' and aaseq[-1] in conserved_J_residues: return ntseq, aaseq, recomb_events['V'], recomb_events['J']
python
def gen_rnd_prod_CDR3(self, conserved_J_residues = 'FVW'): """Generate a productive CDR3 seq from a Monte Carlo draw of the model. Parameters ---------- conserved_J_residues : str, optional Conserved amino acid residues defining the CDR3 on the J side (normally F, V, and/or W) Returns ------- ntseq : str Productive CDR3 nucleotide sequence aaseq : str CDR3 amino acid sequence (aaseq = nt2aa(ntseq)) V_choice : int Index of V allele chosen to generate the CDR3 seq J_choice : int Index of J allele chosen to generate the CDR3 seq """ coding_pass = False while ~coding_pass: recomb_events = self.choose_random_recomb_events() V_seq = self.cutV_genomic_CDR3_segs[recomb_events['V']] #This both checks that the position of the conserved C is #identified and that the V isn't fully deleted out of the CDR3 #region if len(V_seq) <= max(recomb_events['delV'], 0): continue D_seq = self.cutD_genomic_CDR3_segs[recomb_events['D']] J_seq = self.cutJ_genomic_CDR3_segs[recomb_events['J']] #We check that the D and J aren't deleted more than allowed. Note #the generative model really should reflect this structure already if len(D_seq) < (recomb_events['delDl'] + recomb_events['delDr']) or len(J_seq) < recomb_events['delJ']: continue V_seq = V_seq[:len(V_seq) - recomb_events['delV']] D_seq = D_seq[recomb_events['delDl']:len(D_seq)-recomb_events['delDr']] J_seq = J_seq[recomb_events['delJ']:] if (len(V_seq)+ len(D_seq) + len(J_seq) + recomb_events['insVD'] + recomb_events['insDJ']) % 3 != 0: continue insVD_seq = rnd_ins_seq(recomb_events['insVD'], self.C_Rvd, self.C_first_nt_bias_insVD) insDJ_seq = rnd_ins_seq(recomb_events['insDJ'], self.C_Rdj, self.C_first_nt_bias_insDJ)[::-1] #have to reverse the DJ seq #Translate to amino acid sequence, see if productive ntseq = V_seq + insVD_seq + D_seq + insDJ_seq + J_seq aaseq = nt2aa(ntseq) if '*' not in aaseq and aaseq[0]=='C' and aaseq[-1] in conserved_J_residues: return ntseq, aaseq, recomb_events['V'], recomb_events['J']
[ "def", "gen_rnd_prod_CDR3", "(", "self", ",", "conserved_J_residues", "=", "'FVW'", ")", ":", "coding_pass", "=", "False", "while", "~", "coding_pass", ":", "recomb_events", "=", "self", ".", "choose_random_recomb_events", "(", ")", "V_seq", "=", "self", ".", "cutV_genomic_CDR3_segs", "[", "recomb_events", "[", "'V'", "]", "]", "#This both checks that the position of the conserved C is", "#identified and that the V isn't fully deleted out of the CDR3", "#region", "if", "len", "(", "V_seq", ")", "<=", "max", "(", "recomb_events", "[", "'delV'", "]", ",", "0", ")", ":", "continue", "D_seq", "=", "self", ".", "cutD_genomic_CDR3_segs", "[", "recomb_events", "[", "'D'", "]", "]", "J_seq", "=", "self", ".", "cutJ_genomic_CDR3_segs", "[", "recomb_events", "[", "'J'", "]", "]", "#We check that the D and J aren't deleted more than allowed. Note", "#the generative model really should reflect this structure already", "if", "len", "(", "D_seq", ")", "<", "(", "recomb_events", "[", "'delDl'", "]", "+", "recomb_events", "[", "'delDr'", "]", ")", "or", "len", "(", "J_seq", ")", "<", "recomb_events", "[", "'delJ'", "]", ":", "continue", "V_seq", "=", "V_seq", "[", ":", "len", "(", "V_seq", ")", "-", "recomb_events", "[", "'delV'", "]", "]", "D_seq", "=", "D_seq", "[", "recomb_events", "[", "'delDl'", "]", ":", "len", "(", "D_seq", ")", "-", "recomb_events", "[", "'delDr'", "]", "]", "J_seq", "=", "J_seq", "[", "recomb_events", "[", "'delJ'", "]", ":", "]", "if", "(", "len", "(", "V_seq", ")", "+", "len", "(", "D_seq", ")", "+", "len", "(", "J_seq", ")", "+", "recomb_events", "[", "'insVD'", "]", "+", "recomb_events", "[", "'insDJ'", "]", ")", "%", "3", "!=", "0", ":", "continue", "insVD_seq", "=", "rnd_ins_seq", "(", "recomb_events", "[", "'insVD'", "]", ",", "self", ".", "C_Rvd", ",", "self", ".", "C_first_nt_bias_insVD", ")", "insDJ_seq", "=", "rnd_ins_seq", "(", "recomb_events", "[", "'insDJ'", "]", ",", "self", ".", "C_Rdj", ",", "self", ".", "C_first_nt_bias_insDJ", ")", "[", ":", ":", "-", "1", "]", "#have to reverse the DJ seq", "#Translate to amino acid sequence, see if productive", "ntseq", "=", "V_seq", "+", "insVD_seq", "+", "D_seq", "+", "insDJ_seq", "+", "J_seq", "aaseq", "=", "nt2aa", "(", "ntseq", ")", "if", "'*'", "not", "in", "aaseq", "and", "aaseq", "[", "0", "]", "==", "'C'", "and", "aaseq", "[", "-", "1", "]", "in", "conserved_J_residues", ":", "return", "ntseq", ",", "aaseq", ",", "recomb_events", "[", "'V'", "]", ",", "recomb_events", "[", "'J'", "]" ]
Generate a productive CDR3 seq from a Monte Carlo draw of the model. Parameters ---------- conserved_J_residues : str, optional Conserved amino acid residues defining the CDR3 on the J side (normally F, V, and/or W) Returns ------- ntseq : str Productive CDR3 nucleotide sequence aaseq : str CDR3 amino acid sequence (aaseq = nt2aa(ntseq)) V_choice : int Index of V allele chosen to generate the CDR3 seq J_choice : int Index of J allele chosen to generate the CDR3 seq
[ "Generate", "a", "productive", "CDR3", "seq", "from", "a", "Monte", "Carlo", "draw", "of", "the", "model", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/sequence_generation.py#L187-L243
train
zsethna/OLGA
olga/sequence_generation.py
SequenceGenerationVDJ.choose_random_recomb_events
def choose_random_recomb_events(self): """Sample the genomic model for VDJ recombination events. Returns ------- recomb_events : dict Dictionary of the VDJ recombination events. These are integers determining gene choice, deletions, and number of insertions. Example -------- >>> sequence_generation.choose_random_recomb_events() {'D': 0, 'J': 13, 'V': 36, 'delDl': 2, 'delDr': 13, 'delJ': 10, 'delV': 5, 'insDJ': 6, 'insVD': 9} """ recomb_events = {} recomb_events['V'] = self.CPV.searchsorted(np.random.random()) #For 2D arrays make sure to take advantage of a mod expansion to find indicies DJ_choice = self.CPDJ.searchsorted(np.random.random()) recomb_events['D'] = DJ_choice/self.num_J_genes recomb_events['J'] = DJ_choice % self.num_J_genes #Refer to the correct slices for the dependent distributions recomb_events['delV'] = self.given_V_CPdelV[recomb_events['V'], :].searchsorted(np.random.random()) recomb_events['delJ'] = self.given_J_CPdelJ[recomb_events['J'], :].searchsorted(np.random.random()) delDldelDr_choice = self.given_D_CPdelDldelDr[recomb_events['D'], :].searchsorted(np.random.random()) recomb_events['delDl'] = delDldelDr_choice/self.num_delDr_poss recomb_events['delDr'] = delDldelDr_choice % self.num_delDr_poss recomb_events['insVD'] = self.CinsVD.searchsorted(np.random.random()) recomb_events['insDJ'] = self.CinsDJ.searchsorted(np.random.random()) return recomb_events
python
def choose_random_recomb_events(self): """Sample the genomic model for VDJ recombination events. Returns ------- recomb_events : dict Dictionary of the VDJ recombination events. These are integers determining gene choice, deletions, and number of insertions. Example -------- >>> sequence_generation.choose_random_recomb_events() {'D': 0, 'J': 13, 'V': 36, 'delDl': 2, 'delDr': 13, 'delJ': 10, 'delV': 5, 'insDJ': 6, 'insVD': 9} """ recomb_events = {} recomb_events['V'] = self.CPV.searchsorted(np.random.random()) #For 2D arrays make sure to take advantage of a mod expansion to find indicies DJ_choice = self.CPDJ.searchsorted(np.random.random()) recomb_events['D'] = DJ_choice/self.num_J_genes recomb_events['J'] = DJ_choice % self.num_J_genes #Refer to the correct slices for the dependent distributions recomb_events['delV'] = self.given_V_CPdelV[recomb_events['V'], :].searchsorted(np.random.random()) recomb_events['delJ'] = self.given_J_CPdelJ[recomb_events['J'], :].searchsorted(np.random.random()) delDldelDr_choice = self.given_D_CPdelDldelDr[recomb_events['D'], :].searchsorted(np.random.random()) recomb_events['delDl'] = delDldelDr_choice/self.num_delDr_poss recomb_events['delDr'] = delDldelDr_choice % self.num_delDr_poss recomb_events['insVD'] = self.CinsVD.searchsorted(np.random.random()) recomb_events['insDJ'] = self.CinsDJ.searchsorted(np.random.random()) return recomb_events
[ "def", "choose_random_recomb_events", "(", "self", ")", ":", "recomb_events", "=", "{", "}", "recomb_events", "[", "'V'", "]", "=", "self", ".", "CPV", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "#For 2D arrays make sure to take advantage of a mod expansion to find indicies", "DJ_choice", "=", "self", ".", "CPDJ", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'D'", "]", "=", "DJ_choice", "/", "self", ".", "num_J_genes", "recomb_events", "[", "'J'", "]", "=", "DJ_choice", "%", "self", ".", "num_J_genes", "#Refer to the correct slices for the dependent distributions", "recomb_events", "[", "'delV'", "]", "=", "self", ".", "given_V_CPdelV", "[", "recomb_events", "[", "'V'", "]", ",", ":", "]", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'delJ'", "]", "=", "self", ".", "given_J_CPdelJ", "[", "recomb_events", "[", "'J'", "]", ",", ":", "]", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "delDldelDr_choice", "=", "self", ".", "given_D_CPdelDldelDr", "[", "recomb_events", "[", "'D'", "]", ",", ":", "]", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'delDl'", "]", "=", "delDldelDr_choice", "/", "self", ".", "num_delDr_poss", "recomb_events", "[", "'delDr'", "]", "=", "delDldelDr_choice", "%", "self", ".", "num_delDr_poss", "recomb_events", "[", "'insVD'", "]", "=", "self", ".", "CinsVD", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'insDJ'", "]", "=", "self", ".", "CinsDJ", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "return", "recomb_events" ]
Sample the genomic model for VDJ recombination events. Returns ------- recomb_events : dict Dictionary of the VDJ recombination events. These are integers determining gene choice, deletions, and number of insertions. Example -------- >>> sequence_generation.choose_random_recomb_events() {'D': 0, 'J': 13, 'V': 36, 'delDl': 2, 'delDr': 13, 'delJ': 10, 'delV': 5, 'insDJ': 6, 'insVD': 9}
[ "Sample", "the", "genomic", "model", "for", "VDJ", "recombination", "events", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/sequence_generation.py#L245-L283
train
zsethna/OLGA
olga/sequence_generation.py
SequenceGenerationVJ.gen_rnd_prod_CDR3
def gen_rnd_prod_CDR3(self, conserved_J_residues = 'FVW'): """Generate a productive CDR3 seq from a Monte Carlo draw of the model. Parameters ---------- conserved_J_residues : str, optional Conserved amino acid residues defining the CDR3 on the J side (normally F, V, and/or W) Returns ------- ntseq : str Productive CDR3 nucleotide sequence aaseq : str CDR3 amino acid sequence (aaseq = nt2aa(ntseq)) V_choice : int Index of V allele chosen to generate the CDR3 seq J_choice : int Index of J allele chosen to generate the CDR3 seq """ coding_pass = False while ~coding_pass: recomb_events = self.choose_random_recomb_events() V_seq = self.cutV_genomic_CDR3_segs[recomb_events['V']] #This both checks that the position of the conserved C is #identified and that the V isn't fully deleted out of the CDR3 #region if len(V_seq) <= max(recomb_events['delV'], 0): continue J_seq = self.cutJ_genomic_CDR3_segs[recomb_events['J']] #We check that J isn't deleted more than allowed. Note the #generative model really should reflect this structure already if len(J_seq) < recomb_events['delJ']: continue V_seq = V_seq[:len(V_seq) - recomb_events['delV']] J_seq = J_seq[recomb_events['delJ']:] if (len(V_seq)+len(J_seq) + recomb_events['insVJ']) % 3 != 0: continue insVJ_seq = rnd_ins_seq(recomb_events['insVJ'], self.C_Rvj, self.C_first_nt_bias_insVJ) #Translate to amino acid sequence, see if productive ntseq = V_seq + insVJ_seq + J_seq aaseq = nt2aa(ntseq) if '*' not in aaseq and aaseq[0]=='C' and aaseq[-1] in conserved_J_residues: return ntseq, aaseq, recomb_events['V'], recomb_events['J']
python
def gen_rnd_prod_CDR3(self, conserved_J_residues = 'FVW'): """Generate a productive CDR3 seq from a Monte Carlo draw of the model. Parameters ---------- conserved_J_residues : str, optional Conserved amino acid residues defining the CDR3 on the J side (normally F, V, and/or W) Returns ------- ntseq : str Productive CDR3 nucleotide sequence aaseq : str CDR3 amino acid sequence (aaseq = nt2aa(ntseq)) V_choice : int Index of V allele chosen to generate the CDR3 seq J_choice : int Index of J allele chosen to generate the CDR3 seq """ coding_pass = False while ~coding_pass: recomb_events = self.choose_random_recomb_events() V_seq = self.cutV_genomic_CDR3_segs[recomb_events['V']] #This both checks that the position of the conserved C is #identified and that the V isn't fully deleted out of the CDR3 #region if len(V_seq) <= max(recomb_events['delV'], 0): continue J_seq = self.cutJ_genomic_CDR3_segs[recomb_events['J']] #We check that J isn't deleted more than allowed. Note the #generative model really should reflect this structure already if len(J_seq) < recomb_events['delJ']: continue V_seq = V_seq[:len(V_seq) - recomb_events['delV']] J_seq = J_seq[recomb_events['delJ']:] if (len(V_seq)+len(J_seq) + recomb_events['insVJ']) % 3 != 0: continue insVJ_seq = rnd_ins_seq(recomb_events['insVJ'], self.C_Rvj, self.C_first_nt_bias_insVJ) #Translate to amino acid sequence, see if productive ntseq = V_seq + insVJ_seq + J_seq aaseq = nt2aa(ntseq) if '*' not in aaseq and aaseq[0]=='C' and aaseq[-1] in conserved_J_residues: return ntseq, aaseq, recomb_events['V'], recomb_events['J']
[ "def", "gen_rnd_prod_CDR3", "(", "self", ",", "conserved_J_residues", "=", "'FVW'", ")", ":", "coding_pass", "=", "False", "while", "~", "coding_pass", ":", "recomb_events", "=", "self", ".", "choose_random_recomb_events", "(", ")", "V_seq", "=", "self", ".", "cutV_genomic_CDR3_segs", "[", "recomb_events", "[", "'V'", "]", "]", "#This both checks that the position of the conserved C is", "#identified and that the V isn't fully deleted out of the CDR3", "#region", "if", "len", "(", "V_seq", ")", "<=", "max", "(", "recomb_events", "[", "'delV'", "]", ",", "0", ")", ":", "continue", "J_seq", "=", "self", ".", "cutJ_genomic_CDR3_segs", "[", "recomb_events", "[", "'J'", "]", "]", "#We check that J isn't deleted more than allowed. Note the", "#generative model really should reflect this structure already", "if", "len", "(", "J_seq", ")", "<", "recomb_events", "[", "'delJ'", "]", ":", "continue", "V_seq", "=", "V_seq", "[", ":", "len", "(", "V_seq", ")", "-", "recomb_events", "[", "'delV'", "]", "]", "J_seq", "=", "J_seq", "[", "recomb_events", "[", "'delJ'", "]", ":", "]", "if", "(", "len", "(", "V_seq", ")", "+", "len", "(", "J_seq", ")", "+", "recomb_events", "[", "'insVJ'", "]", ")", "%", "3", "!=", "0", ":", "continue", "insVJ_seq", "=", "rnd_ins_seq", "(", "recomb_events", "[", "'insVJ'", "]", ",", "self", ".", "C_Rvj", ",", "self", ".", "C_first_nt_bias_insVJ", ")", "#Translate to amino acid sequence, see if productive", "ntseq", "=", "V_seq", "+", "insVJ_seq", "+", "J_seq", "aaseq", "=", "nt2aa", "(", "ntseq", ")", "if", "'*'", "not", "in", "aaseq", "and", "aaseq", "[", "0", "]", "==", "'C'", "and", "aaseq", "[", "-", "1", "]", "in", "conserved_J_residues", ":", "return", "ntseq", ",", "aaseq", ",", "recomb_events", "[", "'V'", "]", ",", "recomb_events", "[", "'J'", "]" ]
Generate a productive CDR3 seq from a Monte Carlo draw of the model. Parameters ---------- conserved_J_residues : str, optional Conserved amino acid residues defining the CDR3 on the J side (normally F, V, and/or W) Returns ------- ntseq : str Productive CDR3 nucleotide sequence aaseq : str CDR3 amino acid sequence (aaseq = nt2aa(ntseq)) V_choice : int Index of V allele chosen to generate the CDR3 seq J_choice : int Index of J allele chosen to generate the CDR3 seq
[ "Generate", "a", "productive", "CDR3", "seq", "from", "a", "Monte", "Carlo", "draw", "of", "the", "model", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/sequence_generation.py#L368-L422
train
zsethna/OLGA
olga/sequence_generation.py
SequenceGenerationVJ.choose_random_recomb_events
def choose_random_recomb_events(self): """Sample the genomic model for VDJ recombination events. Returns ------- recomb_events : dict Dictionary of the VDJ recombination events. These are integers determining gene choice, deletions, and number of insertions. Example -------- >>> sequence_generation.choose_random_recomb_events() {'J': 13, 'V': 36, 'delJ': 10, 'delV': 5, 'insVJ': 3} """ recomb_events = {} #For 2D arrays make sure to take advantage of a mod expansion to find indicies VJ_choice = self.CPVJ.searchsorted(np.random.random()) recomb_events['V'] = VJ_choice/self.num_J_genes recomb_events['J'] = VJ_choice % self.num_J_genes #Refer to the correct slices for the dependent distributions recomb_events['delV'] = self.given_V_CPdelV[recomb_events['V'], :].searchsorted(np.random.random()) recomb_events['delJ'] = self.given_J_CPdelJ[recomb_events['J'], :].searchsorted(np.random.random()) recomb_events['insVJ'] = self.CPinsVJ.searchsorted(np.random.random()) return recomb_events
python
def choose_random_recomb_events(self): """Sample the genomic model for VDJ recombination events. Returns ------- recomb_events : dict Dictionary of the VDJ recombination events. These are integers determining gene choice, deletions, and number of insertions. Example -------- >>> sequence_generation.choose_random_recomb_events() {'J': 13, 'V': 36, 'delJ': 10, 'delV': 5, 'insVJ': 3} """ recomb_events = {} #For 2D arrays make sure to take advantage of a mod expansion to find indicies VJ_choice = self.CPVJ.searchsorted(np.random.random()) recomb_events['V'] = VJ_choice/self.num_J_genes recomb_events['J'] = VJ_choice % self.num_J_genes #Refer to the correct slices for the dependent distributions recomb_events['delV'] = self.given_V_CPdelV[recomb_events['V'], :].searchsorted(np.random.random()) recomb_events['delJ'] = self.given_J_CPdelJ[recomb_events['J'], :].searchsorted(np.random.random()) recomb_events['insVJ'] = self.CPinsVJ.searchsorted(np.random.random()) return recomb_events
[ "def", "choose_random_recomb_events", "(", "self", ")", ":", "recomb_events", "=", "{", "}", "#For 2D arrays make sure to take advantage of a mod expansion to find indicies", "VJ_choice", "=", "self", ".", "CPVJ", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'V'", "]", "=", "VJ_choice", "/", "self", ".", "num_J_genes", "recomb_events", "[", "'J'", "]", "=", "VJ_choice", "%", "self", ".", "num_J_genes", "#Refer to the correct slices for the dependent distributions", "recomb_events", "[", "'delV'", "]", "=", "self", ".", "given_V_CPdelV", "[", "recomb_events", "[", "'V'", "]", ",", ":", "]", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'delJ'", "]", "=", "self", ".", "given_J_CPdelJ", "[", "recomb_events", "[", "'J'", "]", ",", ":", "]", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "recomb_events", "[", "'insVJ'", "]", "=", "self", ".", "CPinsVJ", ".", "searchsorted", "(", "np", ".", "random", ".", "random", "(", ")", ")", "return", "recomb_events" ]
Sample the genomic model for VDJ recombination events. Returns ------- recomb_events : dict Dictionary of the VDJ recombination events. These are integers determining gene choice, deletions, and number of insertions. Example -------- >>> sequence_generation.choose_random_recomb_events() {'J': 13, 'V': 36, 'delJ': 10, 'delV': 5, 'insVJ': 3}
[ "Sample", "the", "genomic", "model", "for", "VDJ", "recombination", "events", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/sequence_generation.py#L424-L454
train
evonove/django-money-rates
djmoney_rates/backends.py
BaseRateBackend.update_rates
def update_rates(self): """ Creates or updates rates for a source """ source, created = RateSource.objects.get_or_create(name=self.get_source_name()) source.base_currency = self.get_base_currency() source.save() for currency, value in six.iteritems(self.get_rates()): try: rate = Rate.objects.get(source=source, currency=currency) except Rate.DoesNotExist: rate = Rate(source=source, currency=currency) rate.value = value rate.save()
python
def update_rates(self): """ Creates or updates rates for a source """ source, created = RateSource.objects.get_or_create(name=self.get_source_name()) source.base_currency = self.get_base_currency() source.save() for currency, value in six.iteritems(self.get_rates()): try: rate = Rate.objects.get(source=source, currency=currency) except Rate.DoesNotExist: rate = Rate(source=source, currency=currency) rate.value = value rate.save()
[ "def", "update_rates", "(", "self", ")", ":", "source", ",", "created", "=", "RateSource", ".", "objects", ".", "get_or_create", "(", "name", "=", "self", ".", "get_source_name", "(", ")", ")", "source", ".", "base_currency", "=", "self", ".", "get_base_currency", "(", ")", "source", ".", "save", "(", ")", "for", "currency", ",", "value", "in", "six", ".", "iteritems", "(", "self", ".", "get_rates", "(", ")", ")", ":", "try", ":", "rate", "=", "Rate", ".", "objects", ".", "get", "(", "source", "=", "source", ",", "currency", "=", "currency", ")", "except", "Rate", ".", "DoesNotExist", ":", "rate", "=", "Rate", "(", "source", "=", "source", ",", "currency", "=", "currency", ")", "rate", ".", "value", "=", "value", "rate", ".", "save", "(", ")" ]
Creates or updates rates for a source
[ "Creates", "or", "updates", "rates", "for", "a", "source" ]
ac1f7636b9a38d3e153eb833019342c4d88634c2
https://github.com/evonove/django-money-rates/blob/ac1f7636b9a38d3e153eb833019342c4d88634c2/djmoney_rates/backends.py#L52-L67
train
tansey/gfl
pygfl/bayes.py
sample_gtf
def sample_gtf(data, D, k, likelihood='gaussian', prior='laplace', lambda_hyperparams=None, lam_walk_stdev=0.01, lam0=1., dp_hyperparameter=None, w_hyperparameters=None, iterations=7000, burn=2000, thin=10, robust=False, empirical=False, verbose=False): '''Generate samples from the generalized graph trend filtering distribution via a modified Swendsen-Wang slice sampling algorithm. Options for likelihood: gaussian, binomial, poisson. Options for prior: laplace, doublepareto.''' Dk = get_delta(D, k) dk_rows, dk_rowbreaks, dk_cols, dk_vals = decompose_delta(Dk) if likelihood == 'gaussian': y, w = data elif likelihood == 'binomial': trials, successes = data elif likelihood == 'poisson': obs = data else: raise Exception('Unknown likelihood type: {0}'.format(likelihood)) if prior == 'laplace': if lambda_hyperparams == None: lambda_hyperparams = (1., 1.) elif prior == 'laplacegamma': if lambda_hyperparams == None: lambda_hyperparams = (1., 1.) if dp_hyperparameter == None: dp_hyperparameter = 1. elif prior == 'doublepareto' or prior == 'doublepareto2': if lambda_hyperparams == None: lambda_hyperparams = (1.0, 1.0) if dp_hyperparameter == None: dp_hyperparameter = 0.1 elif prior == 'cauchy': if lambda_hyperparams == None: lambda_hyperparams = (1.0, 1.0) else: raise Exception('Unknown prior type: {0}.'.format(prior)) if robust and w_hyperparameters is None: w_hyperparameters = (1., 1.) # Run the Gibbs sampler sample_size = (iterations - burn) / thin beta_samples = np.zeros((sample_size, D.shape[1]), dtype='double') lam_samples = np.zeros(sample_size, dtype='double') if likelihood == 'gaussian': if prior == 'laplace': gflbayes_gaussian_laplace(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'laplacegamma': if robust: gflbayes_gaussian_laplace_gamma_robust(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, w_hyperparameters[0], w_hyperparameters[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) else: gflbayes_gaussian_laplace_gamma(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto': gflbayes_gaussian_doublepareto(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto2': gflbayes_gaussian_doublepareto2(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'cauchy': gflbayes_gaussian_cauchy(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif likelihood == 'binomial': if prior == 'laplace': gflbayes_binomial_laplace(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto': gflbayes_binomial_doublepareto(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'laplacegamma': if empirical: gflbayes_empirical_binomial_laplace_gamma(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lam0, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) else: gflbayes_binomial_laplace_gamma(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif likelihood == 'poisson': if prior == 'laplace': gflbayes_poisson_laplace(len(obs), obs, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto': gflbayes_poisson_doublepareto(len(obs), obs, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) else: raise Exception('Unknown likelihood type: {0}'.format(likelihood)) return (beta_samples,lam_samples)
python
def sample_gtf(data, D, k, likelihood='gaussian', prior='laplace', lambda_hyperparams=None, lam_walk_stdev=0.01, lam0=1., dp_hyperparameter=None, w_hyperparameters=None, iterations=7000, burn=2000, thin=10, robust=False, empirical=False, verbose=False): '''Generate samples from the generalized graph trend filtering distribution via a modified Swendsen-Wang slice sampling algorithm. Options for likelihood: gaussian, binomial, poisson. Options for prior: laplace, doublepareto.''' Dk = get_delta(D, k) dk_rows, dk_rowbreaks, dk_cols, dk_vals = decompose_delta(Dk) if likelihood == 'gaussian': y, w = data elif likelihood == 'binomial': trials, successes = data elif likelihood == 'poisson': obs = data else: raise Exception('Unknown likelihood type: {0}'.format(likelihood)) if prior == 'laplace': if lambda_hyperparams == None: lambda_hyperparams = (1., 1.) elif prior == 'laplacegamma': if lambda_hyperparams == None: lambda_hyperparams = (1., 1.) if dp_hyperparameter == None: dp_hyperparameter = 1. elif prior == 'doublepareto' or prior == 'doublepareto2': if lambda_hyperparams == None: lambda_hyperparams = (1.0, 1.0) if dp_hyperparameter == None: dp_hyperparameter = 0.1 elif prior == 'cauchy': if lambda_hyperparams == None: lambda_hyperparams = (1.0, 1.0) else: raise Exception('Unknown prior type: {0}.'.format(prior)) if robust and w_hyperparameters is None: w_hyperparameters = (1., 1.) # Run the Gibbs sampler sample_size = (iterations - burn) / thin beta_samples = np.zeros((sample_size, D.shape[1]), dtype='double') lam_samples = np.zeros(sample_size, dtype='double') if likelihood == 'gaussian': if prior == 'laplace': gflbayes_gaussian_laplace(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'laplacegamma': if robust: gflbayes_gaussian_laplace_gamma_robust(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, w_hyperparameters[0], w_hyperparameters[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) else: gflbayes_gaussian_laplace_gamma(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto': gflbayes_gaussian_doublepareto(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto2': gflbayes_gaussian_doublepareto2(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'cauchy': gflbayes_gaussian_cauchy(len(y), y, w, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif likelihood == 'binomial': if prior == 'laplace': gflbayes_binomial_laplace(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto': gflbayes_binomial_doublepareto(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'laplacegamma': if empirical: gflbayes_empirical_binomial_laplace_gamma(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lam0, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) else: gflbayes_binomial_laplace_gamma(len(trials), trials, successes, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif likelihood == 'poisson': if prior == 'laplace': gflbayes_poisson_laplace(len(obs), obs, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) elif prior == 'doublepareto': gflbayes_poisson_doublepareto(len(obs), obs, dk_rows, dk_rowbreaks, dk_cols, dk_vals, lambda_hyperparams[0], lambda_hyperparams[1], lam_walk_stdev, lam0, dp_hyperparameter, iterations, burn, thin, double_matrix_to_c_pointer(beta_samples), lam_samples) else: raise Exception('Unknown likelihood type: {0}'.format(likelihood)) return (beta_samples,lam_samples)
[ "def", "sample_gtf", "(", "data", ",", "D", ",", "k", ",", "likelihood", "=", "'gaussian'", ",", "prior", "=", "'laplace'", ",", "lambda_hyperparams", "=", "None", ",", "lam_walk_stdev", "=", "0.01", ",", "lam0", "=", "1.", ",", "dp_hyperparameter", "=", "None", ",", "w_hyperparameters", "=", "None", ",", "iterations", "=", "7000", ",", "burn", "=", "2000", ",", "thin", "=", "10", ",", "robust", "=", "False", ",", "empirical", "=", "False", ",", "verbose", "=", "False", ")", ":", "Dk", "=", "get_delta", "(", "D", ",", "k", ")", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", "=", "decompose_delta", "(", "Dk", ")", "if", "likelihood", "==", "'gaussian'", ":", "y", ",", "w", "=", "data", "elif", "likelihood", "==", "'binomial'", ":", "trials", ",", "successes", "=", "data", "elif", "likelihood", "==", "'poisson'", ":", "obs", "=", "data", "else", ":", "raise", "Exception", "(", "'Unknown likelihood type: {0}'", ".", "format", "(", "likelihood", ")", ")", "if", "prior", "==", "'laplace'", ":", "if", "lambda_hyperparams", "==", "None", ":", "lambda_hyperparams", "=", "(", "1.", ",", "1.", ")", "elif", "prior", "==", "'laplacegamma'", ":", "if", "lambda_hyperparams", "==", "None", ":", "lambda_hyperparams", "=", "(", "1.", ",", "1.", ")", "if", "dp_hyperparameter", "==", "None", ":", "dp_hyperparameter", "=", "1.", "elif", "prior", "==", "'doublepareto'", "or", "prior", "==", "'doublepareto2'", ":", "if", "lambda_hyperparams", "==", "None", ":", "lambda_hyperparams", "=", "(", "1.0", ",", "1.0", ")", "if", "dp_hyperparameter", "==", "None", ":", "dp_hyperparameter", "=", "0.1", "elif", "prior", "==", "'cauchy'", ":", "if", "lambda_hyperparams", "==", "None", ":", "lambda_hyperparams", "=", "(", "1.0", ",", "1.0", ")", "else", ":", "raise", "Exception", "(", "'Unknown prior type: {0}.'", ".", "format", "(", "prior", ")", ")", "if", "robust", "and", "w_hyperparameters", "is", "None", ":", "w_hyperparameters", "=", "(", "1.", ",", "1.", ")", "# Run the Gibbs sampler", "sample_size", "=", "(", "iterations", "-", "burn", ")", "/", "thin", "beta_samples", "=", "np", ".", "zeros", "(", "(", "sample_size", ",", "D", ".", "shape", "[", "1", "]", ")", ",", "dtype", "=", "'double'", ")", "lam_samples", "=", "np", ".", "zeros", "(", "sample_size", ",", "dtype", "=", "'double'", ")", "if", "likelihood", "==", "'gaussian'", ":", "if", "prior", "==", "'laplace'", ":", "gflbayes_gaussian_laplace", "(", "len", "(", "y", ")", ",", "y", ",", "w", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'laplacegamma'", ":", "if", "robust", ":", "gflbayes_gaussian_laplace_gamma_robust", "(", "len", "(", "y", ")", ",", "y", ",", "w", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "dp_hyperparameter", ",", "w_hyperparameters", "[", "0", "]", ",", "w_hyperparameters", "[", "1", "]", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "else", ":", "gflbayes_gaussian_laplace_gamma", "(", "len", "(", "y", ")", ",", "y", ",", "w", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "dp_hyperparameter", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'doublepareto'", ":", "gflbayes_gaussian_doublepareto", "(", "len", "(", "y", ")", ",", "y", ",", "w", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "lam_walk_stdev", ",", "lam0", ",", "dp_hyperparameter", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'doublepareto2'", ":", "gflbayes_gaussian_doublepareto2", "(", "len", "(", "y", ")", ",", "y", ",", "w", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "dp_hyperparameter", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'cauchy'", ":", "gflbayes_gaussian_cauchy", "(", "len", "(", "y", ")", ",", "y", ",", "w", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "lam_walk_stdev", ",", "lam0", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "likelihood", "==", "'binomial'", ":", "if", "prior", "==", "'laplace'", ":", "gflbayes_binomial_laplace", "(", "len", "(", "trials", ")", ",", "trials", ",", "successes", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'doublepareto'", ":", "gflbayes_binomial_doublepareto", "(", "len", "(", "trials", ")", ",", "trials", ",", "successes", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "lam_walk_stdev", ",", "lam0", ",", "dp_hyperparameter", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'laplacegamma'", ":", "if", "empirical", ":", "gflbayes_empirical_binomial_laplace_gamma", "(", "len", "(", "trials", ")", ",", "trials", ",", "successes", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lam0", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "else", ":", "gflbayes_binomial_laplace_gamma", "(", "len", "(", "trials", ")", ",", "trials", ",", "successes", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "dp_hyperparameter", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "likelihood", "==", "'poisson'", ":", "if", "prior", "==", "'laplace'", ":", "gflbayes_poisson_laplace", "(", "len", "(", "obs", ")", ",", "obs", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "elif", "prior", "==", "'doublepareto'", ":", "gflbayes_poisson_doublepareto", "(", "len", "(", "obs", ")", ",", "obs", ",", "dk_rows", ",", "dk_rowbreaks", ",", "dk_cols", ",", "dk_vals", ",", "lambda_hyperparams", "[", "0", "]", ",", "lambda_hyperparams", "[", "1", "]", ",", "lam_walk_stdev", ",", "lam0", ",", "dp_hyperparameter", ",", "iterations", ",", "burn", ",", "thin", ",", "double_matrix_to_c_pointer", "(", "beta_samples", ")", ",", "lam_samples", ")", "else", ":", "raise", "Exception", "(", "'Unknown likelihood type: {0}'", ".", "format", "(", "likelihood", ")", ")", "return", "(", "beta_samples", ",", "lam_samples", ")" ]
Generate samples from the generalized graph trend filtering distribution via a modified Swendsen-Wang slice sampling algorithm. Options for likelihood: gaussian, binomial, poisson. Options for prior: laplace, doublepareto.
[ "Generate", "samples", "from", "the", "generalized", "graph", "trend", "filtering", "distribution", "via", "a", "modified", "Swendsen", "-", "Wang", "slice", "sampling", "algorithm", ".", "Options", "for", "likelihood", ":", "gaussian", "binomial", "poisson", ".", "Options", "for", "prior", ":", "laplace", "doublepareto", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/bayes.py#L138-L274
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.compute_regex_CDR3_template_pgen
def compute_regex_CDR3_template_pgen(self, regex_seq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True, raise_overload_warning = True): """Compute Pgen for all seqs consistent with regular expression regex_seq. Computes Pgen for a (limited vocabulary) regular expression of CDR3 amino acid sequences, conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. Please note that this function will list out all the sequences that correspond to the regular expression and then calculate the Pgen of each sequence in succession. THIS CAN BE SLOW. Consider defining a custom alphabet to represent any undetermined amino acids as this will greatly speed up the computations. For example, if the symbol ^ is defined as [AGR] in a custom alphabet, then instead of running compute_regex_CDR3_template_pgen('CASS[AGR]SARPEQFF', ppp), which will compute Pgen for 3 sequences, the single sequence 'CASS^SARPEQFF' can be considered. (Examples are TCRB sequences/model) Parameters ---------- regex_seq : str The regular expression string that represents the CDR3 sequences to be listed then their Pgens computed and summed. V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. raise_overload_warning : bool A flag to warn of more than 10000 seqs corresponding to the regex_seq Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_regex_CDR3_template_pgen('CASS[AGR]SARPEQFF') 8.1090898050318022e-10 >>> generation_probability.compute_regex_CDR3_template_pgen('CASSAX{0,5}SARPEQFF') 6.8468778040965569e-10 """ V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) CDR3_seqs = self.list_seqs_from_regex(regex_seq, print_warnings, raise_overload_warning) pgen = 0 for CDR3_seq in CDR3_seqs: if len(CDR3_seq) == 0: continue pgen += self.compute_CDR3_pgen(CDR3_seq, V_usage_mask, J_usage_mask) return pgen
python
def compute_regex_CDR3_template_pgen(self, regex_seq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True, raise_overload_warning = True): """Compute Pgen for all seqs consistent with regular expression regex_seq. Computes Pgen for a (limited vocabulary) regular expression of CDR3 amino acid sequences, conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. Please note that this function will list out all the sequences that correspond to the regular expression and then calculate the Pgen of each sequence in succession. THIS CAN BE SLOW. Consider defining a custom alphabet to represent any undetermined amino acids as this will greatly speed up the computations. For example, if the symbol ^ is defined as [AGR] in a custom alphabet, then instead of running compute_regex_CDR3_template_pgen('CASS[AGR]SARPEQFF', ppp), which will compute Pgen for 3 sequences, the single sequence 'CASS^SARPEQFF' can be considered. (Examples are TCRB sequences/model) Parameters ---------- regex_seq : str The regular expression string that represents the CDR3 sequences to be listed then their Pgens computed and summed. V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. raise_overload_warning : bool A flag to warn of more than 10000 seqs corresponding to the regex_seq Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_regex_CDR3_template_pgen('CASS[AGR]SARPEQFF') 8.1090898050318022e-10 >>> generation_probability.compute_regex_CDR3_template_pgen('CASSAX{0,5}SARPEQFF') 6.8468778040965569e-10 """ V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) CDR3_seqs = self.list_seqs_from_regex(regex_seq, print_warnings, raise_overload_warning) pgen = 0 for CDR3_seq in CDR3_seqs: if len(CDR3_seq) == 0: continue pgen += self.compute_CDR3_pgen(CDR3_seq, V_usage_mask, J_usage_mask) return pgen
[ "def", "compute_regex_CDR3_template_pgen", "(", "self", ",", "regex_seq", ",", "V_usage_mask_in", "=", "None", ",", "J_usage_mask_in", "=", "None", ",", "print_warnings", "=", "True", ",", "raise_overload_warning", "=", "True", ")", ":", "V_usage_mask", ",", "J_usage_mask", "=", "self", ".", "format_usage_masks", "(", "V_usage_mask_in", ",", "J_usage_mask_in", ",", "print_warnings", ")", "CDR3_seqs", "=", "self", ".", "list_seqs_from_regex", "(", "regex_seq", ",", "print_warnings", ",", "raise_overload_warning", ")", "pgen", "=", "0", "for", "CDR3_seq", "in", "CDR3_seqs", ":", "if", "len", "(", "CDR3_seq", ")", "==", "0", ":", "continue", "pgen", "+=", "self", ".", "compute_CDR3_pgen", "(", "CDR3_seq", ",", "V_usage_mask", ",", "J_usage_mask", ")", "return", "pgen" ]
Compute Pgen for all seqs consistent with regular expression regex_seq. Computes Pgen for a (limited vocabulary) regular expression of CDR3 amino acid sequences, conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. Please note that this function will list out all the sequences that correspond to the regular expression and then calculate the Pgen of each sequence in succession. THIS CAN BE SLOW. Consider defining a custom alphabet to represent any undetermined amino acids as this will greatly speed up the computations. For example, if the symbol ^ is defined as [AGR] in a custom alphabet, then instead of running compute_regex_CDR3_template_pgen('CASS[AGR]SARPEQFF', ppp), which will compute Pgen for 3 sequences, the single sequence 'CASS^SARPEQFF' can be considered. (Examples are TCRB sequences/model) Parameters ---------- regex_seq : str The regular expression string that represents the CDR3 sequences to be listed then their Pgens computed and summed. V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. raise_overload_warning : bool A flag to warn of more than 10000 seqs corresponding to the regex_seq Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_regex_CDR3_template_pgen('CASS[AGR]SARPEQFF') 8.1090898050318022e-10 >>> generation_probability.compute_regex_CDR3_template_pgen('CASSAX{0,5}SARPEQFF') 6.8468778040965569e-10
[ "Compute", "Pgen", "for", "all", "seqs", "consistent", "with", "regular", "expression", "regex_seq", ".", "Computes", "Pgen", "for", "a", "(", "limited", "vocabulary", ")", "regular", "expression", "of", "CDR3", "amino", "acid", "sequences", "conditioned", "on", "the", "V", "genes", "/", "alleles", "indicated", "in", "V_usage_mask_in", "and", "the", "J", "genes", "/", "alleles", "in", "J_usage_mask_in", ".", "Please", "note", "that", "this", "function", "will", "list", "out", "all", "the", "sequences", "that", "correspond", "to", "the", "regular", "expression", "and", "then", "calculate", "the", "Pgen", "of", "each", "sequence", "in", "succession", ".", "THIS", "CAN", "BE", "SLOW", ".", "Consider", "defining", "a", "custom", "alphabet", "to", "represent", "any", "undetermined", "amino", "acids", "as", "this", "will", "greatly", "speed", "up", "the", "computations", ".", "For", "example", "if", "the", "symbol", "^", "is", "defined", "as", "[", "AGR", "]", "in", "a", "custom", "alphabet", "then", "instead", "of", "running", "compute_regex_CDR3_template_pgen", "(", "CASS", "[", "AGR", "]", "SARPEQFF", "ppp", ")", "which", "will", "compute", "Pgen", "for", "3", "sequences", "the", "single", "sequence", "CASS^SARPEQFF", "can", "be", "considered", ".", "(", "Examples", "are", "TCRB", "sequences", "/", "model", ")", "Parameters", "----------", "regex_seq", ":", "str", "The", "regular", "expression", "string", "that", "represents", "the", "CDR3", "sequences", "to", "be", "listed", "then", "their", "Pgens", "computed", "and", "summed", ".", "V_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "V", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "V", "alleles", ".", "J_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "J", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "J", "alleles", ".", "print_warnings", ":", "bool", "Determines", "whether", "warnings", "are", "printed", "or", "not", ".", "Default", "ON", ".", "raise_overload_warning", ":", "bool", "A", "flag", "to", "warn", "of", "more", "than", "10000", "seqs", "corresponding", "to", "the", "regex_seq", "Returns", "-------", "pgen", ":", "float", "The", "generation", "probability", "(", "Pgen", ")", "of", "the", "sequence", "Examples", "--------", ">>>", "generation_probability", ".", "compute_regex_CDR3_template_pgen", "(", "CASS", "[", "AGR", "]", "SARPEQFF", ")", "8", ".", "1090898050318022e", "-", "10", ">>>", "generation_probability", ".", "compute_regex_CDR3_template_pgen", "(", "CASSAX", "{", "0", "5", "}", "SARPEQFF", ")", "6", ".", "8468778040965569e", "-", "10" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L156-L213
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.compute_aa_CDR3_pgen
def compute_aa_CDR3_pgen(self, CDR3_seq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True): """Compute Pgen for the amino acid sequence CDR3_seq. Conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' -- the standard amino acids, plus any custom symbols for an expanded codon alphabet (note the standard ambiguous amino acids -- B, J, X, and Z -- are included by default). V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_aa_CDR3_pgen('CAWSVAPDRGGYTF') 1.5756106696284584e-10 >>> generation_probability.compute_aa_CDR3_pgen('CAWSVAPDRGGYTF', 'TRBV30*01', 'TRBJ1-2*01') 1.203646865765782e-10 >>> generation_probability.compute_aa_CDR3_pgen('CAWXXXXXXXGYTF') 7.8102586432014974e-05 """ if len(CDR3_seq) == 0: return 0 for aa in CDR3_seq: if aa not in self.codons_dict.keys(): #Check to make sure all symbols are accounted for if print_warnings: print 'Invalid amino acid CDR3 sequence --- unfamiliar symbol: ' + aa return 0 V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) return self.compute_CDR3_pgen(CDR3_seq, V_usage_mask, J_usage_mask)
python
def compute_aa_CDR3_pgen(self, CDR3_seq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True): """Compute Pgen for the amino acid sequence CDR3_seq. Conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' -- the standard amino acids, plus any custom symbols for an expanded codon alphabet (note the standard ambiguous amino acids -- B, J, X, and Z -- are included by default). V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_aa_CDR3_pgen('CAWSVAPDRGGYTF') 1.5756106696284584e-10 >>> generation_probability.compute_aa_CDR3_pgen('CAWSVAPDRGGYTF', 'TRBV30*01', 'TRBJ1-2*01') 1.203646865765782e-10 >>> generation_probability.compute_aa_CDR3_pgen('CAWXXXXXXXGYTF') 7.8102586432014974e-05 """ if len(CDR3_seq) == 0: return 0 for aa in CDR3_seq: if aa not in self.codons_dict.keys(): #Check to make sure all symbols are accounted for if print_warnings: print 'Invalid amino acid CDR3 sequence --- unfamiliar symbol: ' + aa return 0 V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) return self.compute_CDR3_pgen(CDR3_seq, V_usage_mask, J_usage_mask)
[ "def", "compute_aa_CDR3_pgen", "(", "self", ",", "CDR3_seq", ",", "V_usage_mask_in", "=", "None", ",", "J_usage_mask_in", "=", "None", ",", "print_warnings", "=", "True", ")", ":", "if", "len", "(", "CDR3_seq", ")", "==", "0", ":", "return", "0", "for", "aa", "in", "CDR3_seq", ":", "if", "aa", "not", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "#Check to make sure all symbols are accounted for", "if", "print_warnings", ":", "print", "'Invalid amino acid CDR3 sequence --- unfamiliar symbol: '", "+", "aa", "return", "0", "V_usage_mask", ",", "J_usage_mask", "=", "self", ".", "format_usage_masks", "(", "V_usage_mask_in", ",", "J_usage_mask_in", ",", "print_warnings", ")", "return", "self", ".", "compute_CDR3_pgen", "(", "CDR3_seq", ",", "V_usage_mask", ",", "J_usage_mask", ")" ]
Compute Pgen for the amino acid sequence CDR3_seq. Conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' -- the standard amino acids, plus any custom symbols for an expanded codon alphabet (note the standard ambiguous amino acids -- B, J, X, and Z -- are included by default). V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_aa_CDR3_pgen('CAWSVAPDRGGYTF') 1.5756106696284584e-10 >>> generation_probability.compute_aa_CDR3_pgen('CAWSVAPDRGGYTF', 'TRBV30*01', 'TRBJ1-2*01') 1.203646865765782e-10 >>> generation_probability.compute_aa_CDR3_pgen('CAWXXXXXXXGYTF') 7.8102586432014974e-05
[ "Compute", "Pgen", "for", "the", "amino", "acid", "sequence", "CDR3_seq", ".", "Conditioned", "on", "the", "V", "genes", "/", "alleles", "indicated", "in", "V_usage_mask_in", "and", "the", "J", "genes", "/", "alleles", "in", "J_usage_mask_in", ".", "(", "Examples", "are", "TCRB", "sequences", "/", "model", ")", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "--", "the", "standard", "amino", "acids", "plus", "any", "custom", "symbols", "for", "an", "expanded", "codon", "alphabet", "(", "note", "the", "standard", "ambiguous", "amino", "acids", "--", "B", "J", "X", "and", "Z", "--", "are", "included", "by", "default", ")", ".", "V_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "V", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "V", "alleles", ".", "J_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "J", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "J", "alleles", ".", "print_warnings", ":", "bool", "Determines", "whether", "warnings", "are", "printed", "or", "not", ".", "Default", "ON", ".", "Returns", "-------", "pgen", ":", "float", "The", "generation", "probability", "(", "Pgen", ")", "of", "the", "sequence", "Examples", "--------", ">>>", "generation_probability", ".", "compute_aa_CDR3_pgen", "(", "CAWSVAPDRGGYTF", ")", "1", ".", "5756106696284584e", "-", "10", ">>>", "generation_probability", ".", "compute_aa_CDR3_pgen", "(", "CAWSVAPDRGGYTF", "TRBV30", "*", "01", "TRBJ1", "-", "2", "*", "01", ")", "1", ".", "203646865765782e", "-", "10", ">>>", "generation_probability", ".", "compute_aa_CDR3_pgen", "(", "CAWXXXXXXXGYTF", ")", "7", ".", "8102586432014974e", "-", "05" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L216-L264
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.compute_hamming_dist_1_pgen
def compute_hamming_dist_1_pgen(self, CDR3_seq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True): """Compute Pgen of all seqs hamming dist 1 (in amino acids) from CDR3_seq. Please note that this function will list out all the sequences that are hamming distance 1 from the base sequence and then calculate the Pgen of each sequence in succession. THIS CAN BE SLOW as it computes Pgen for L+1 sequences where L = len(CDR3_seq). (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of amino acids (ONLY the standard amino acids). Pgens for all sequences of hamming distance 1 (in amino acid sequence) are summed. V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float The sum of generation probabilities (Pgens) of the sequences at most hamming distance 1 (in amino acids) from CDR3_seq. """ #make sure that the symbol X is defined as the fully undetermined amino acid: #X ~ ACDEFGHIKLMNPQRSTVWY V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) if len(CDR3_seq) == 0: return 0 for aa in CDR3_seq: if aa not in 'ACDEFGHIKLMNPQRSTVWY': #Check to make sure all symbols are accounted for if print_warnings: print 'Invalid amino acid CDR3 sequence --- unfamiliar symbol: ' + aa return 0 tot_pgen = 0 for i in range(len(CDR3_seq)): tot_pgen += self.compute_CDR3_pgen(CDR3_seq[:i] + 'X' + CDR3_seq[i+1:], V_usage_mask, J_usage_mask) tot_pgen += -(len(CDR3_seq) - 1)*self.compute_CDR3_pgen(CDR3_seq, V_usage_mask, J_usage_mask) return tot_pgen
python
def compute_hamming_dist_1_pgen(self, CDR3_seq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True): """Compute Pgen of all seqs hamming dist 1 (in amino acids) from CDR3_seq. Please note that this function will list out all the sequences that are hamming distance 1 from the base sequence and then calculate the Pgen of each sequence in succession. THIS CAN BE SLOW as it computes Pgen for L+1 sequences where L = len(CDR3_seq). (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of amino acids (ONLY the standard amino acids). Pgens for all sequences of hamming distance 1 (in amino acid sequence) are summed. V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float The sum of generation probabilities (Pgens) of the sequences at most hamming distance 1 (in amino acids) from CDR3_seq. """ #make sure that the symbol X is defined as the fully undetermined amino acid: #X ~ ACDEFGHIKLMNPQRSTVWY V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) if len(CDR3_seq) == 0: return 0 for aa in CDR3_seq: if aa not in 'ACDEFGHIKLMNPQRSTVWY': #Check to make sure all symbols are accounted for if print_warnings: print 'Invalid amino acid CDR3 sequence --- unfamiliar symbol: ' + aa return 0 tot_pgen = 0 for i in range(len(CDR3_seq)): tot_pgen += self.compute_CDR3_pgen(CDR3_seq[:i] + 'X' + CDR3_seq[i+1:], V_usage_mask, J_usage_mask) tot_pgen += -(len(CDR3_seq) - 1)*self.compute_CDR3_pgen(CDR3_seq, V_usage_mask, J_usage_mask) return tot_pgen
[ "def", "compute_hamming_dist_1_pgen", "(", "self", ",", "CDR3_seq", ",", "V_usage_mask_in", "=", "None", ",", "J_usage_mask_in", "=", "None", ",", "print_warnings", "=", "True", ")", ":", "#make sure that the symbol X is defined as the fully undetermined amino acid:", "#X ~ ACDEFGHIKLMNPQRSTVWY", "V_usage_mask", ",", "J_usage_mask", "=", "self", ".", "format_usage_masks", "(", "V_usage_mask_in", ",", "J_usage_mask_in", ",", "print_warnings", ")", "if", "len", "(", "CDR3_seq", ")", "==", "0", ":", "return", "0", "for", "aa", "in", "CDR3_seq", ":", "if", "aa", "not", "in", "'ACDEFGHIKLMNPQRSTVWY'", ":", "#Check to make sure all symbols are accounted for", "if", "print_warnings", ":", "print", "'Invalid amino acid CDR3 sequence --- unfamiliar symbol: '", "+", "aa", "return", "0", "tot_pgen", "=", "0", "for", "i", "in", "range", "(", "len", "(", "CDR3_seq", ")", ")", ":", "tot_pgen", "+=", "self", ".", "compute_CDR3_pgen", "(", "CDR3_seq", "[", ":", "i", "]", "+", "'X'", "+", "CDR3_seq", "[", "i", "+", "1", ":", "]", ",", "V_usage_mask", ",", "J_usage_mask", ")", "tot_pgen", "+=", "-", "(", "len", "(", "CDR3_seq", ")", "-", "1", ")", "*", "self", ".", "compute_CDR3_pgen", "(", "CDR3_seq", ",", "V_usage_mask", ",", "J_usage_mask", ")", "return", "tot_pgen" ]
Compute Pgen of all seqs hamming dist 1 (in amino acids) from CDR3_seq. Please note that this function will list out all the sequences that are hamming distance 1 from the base sequence and then calculate the Pgen of each sequence in succession. THIS CAN BE SLOW as it computes Pgen for L+1 sequences where L = len(CDR3_seq). (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of amino acids (ONLY the standard amino acids). Pgens for all sequences of hamming distance 1 (in amino acid sequence) are summed. V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float The sum of generation probabilities (Pgens) of the sequences at most hamming distance 1 (in amino acids) from CDR3_seq.
[ "Compute", "Pgen", "of", "all", "seqs", "hamming", "dist", "1", "(", "in", "amino", "acids", ")", "from", "CDR3_seq", ".", "Please", "note", "that", "this", "function", "will", "list", "out", "all", "the", "sequences", "that", "are", "hamming", "distance", "1", "from", "the", "base", "sequence", "and", "then", "calculate", "the", "Pgen", "of", "each", "sequence", "in", "succession", ".", "THIS", "CAN", "BE", "SLOW", "as", "it", "computes", "Pgen", "for", "L", "+", "1", "sequences", "where", "L", "=", "len", "(", "CDR3_seq", ")", ".", "(", "Examples", "are", "TCRB", "sequences", "/", "model", ")", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "ONLY", "the", "standard", "amino", "acids", ")", ".", "Pgens", "for", "all", "sequences", "of", "hamming", "distance", "1", "(", "in", "amino", "acid", "sequence", ")", "are", "summed", ".", "V_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "V", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "V", "alleles", ".", "J_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "J", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "J", "alleles", ".", "print_warnings", ":", "bool", "Determines", "whether", "warnings", "are", "printed", "or", "not", ".", "Default", "ON", ".", "Returns", "-------", "pgen", ":", "float", "The", "sum", "of", "generation", "probabilities", "(", "Pgens", ")", "of", "the", "sequences", "at", "most", "hamming", "distance", "1", "(", "in", "amino", "acids", ")", "from", "CDR3_seq", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L266-L316
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.compute_nt_CDR3_pgen
def compute_nt_CDR3_pgen(self, CDR3_ntseq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True): """Compute Pgen for the inframe nucleotide sequence CDR3_ntseq. Conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. (Examples are TCRB sequences/model) Parameters ---------- CDR3_ntseq : str Inframe nucleotide sequence composed of ONLY A, C, G, or T (either uppercase or lowercase). V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float64 The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_nt_CDR3_pgen('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') 3.2674893012379071e-12 >>> generation_probability.compute_nt_CDR3_pgen('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC', 'TRBV30*01', 'TRBJ1-2*01') 2.3986503758867323e-12 """ if not len(CDR3_ntseq)%3 == 0: #Make sure sequence is inframe if print_warnings: print 'Invalid nucleotide CDR3 sequence --- out of frame sequence' return 0 elif len(CDR3_ntseq) == 0: return 0 else: for nt in CDR3_ntseq: if nt not in 'ACGTacgt': if print_warnings: print 'Invalid nucleotide CDR3 sequence --- unfamiliar nucleotide: ' + nt return 0 V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) return self.compute_CDR3_pgen(nt2codon_rep(CDR3_ntseq), V_usage_mask, J_usage_mask)
python
def compute_nt_CDR3_pgen(self, CDR3_ntseq, V_usage_mask_in = None, J_usage_mask_in = None, print_warnings = True): """Compute Pgen for the inframe nucleotide sequence CDR3_ntseq. Conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. (Examples are TCRB sequences/model) Parameters ---------- CDR3_ntseq : str Inframe nucleotide sequence composed of ONLY A, C, G, or T (either uppercase or lowercase). V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float64 The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_nt_CDR3_pgen('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') 3.2674893012379071e-12 >>> generation_probability.compute_nt_CDR3_pgen('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC', 'TRBV30*01', 'TRBJ1-2*01') 2.3986503758867323e-12 """ if not len(CDR3_ntseq)%3 == 0: #Make sure sequence is inframe if print_warnings: print 'Invalid nucleotide CDR3 sequence --- out of frame sequence' return 0 elif len(CDR3_ntseq) == 0: return 0 else: for nt in CDR3_ntseq: if nt not in 'ACGTacgt': if print_warnings: print 'Invalid nucleotide CDR3 sequence --- unfamiliar nucleotide: ' + nt return 0 V_usage_mask, J_usage_mask = self.format_usage_masks(V_usage_mask_in, J_usage_mask_in, print_warnings) return self.compute_CDR3_pgen(nt2codon_rep(CDR3_ntseq), V_usage_mask, J_usage_mask)
[ "def", "compute_nt_CDR3_pgen", "(", "self", ",", "CDR3_ntseq", ",", "V_usage_mask_in", "=", "None", ",", "J_usage_mask_in", "=", "None", ",", "print_warnings", "=", "True", ")", ":", "if", "not", "len", "(", "CDR3_ntseq", ")", "%", "3", "==", "0", ":", "#Make sure sequence is inframe", "if", "print_warnings", ":", "print", "'Invalid nucleotide CDR3 sequence --- out of frame sequence'", "return", "0", "elif", "len", "(", "CDR3_ntseq", ")", "==", "0", ":", "return", "0", "else", ":", "for", "nt", "in", "CDR3_ntseq", ":", "if", "nt", "not", "in", "'ACGTacgt'", ":", "if", "print_warnings", ":", "print", "'Invalid nucleotide CDR3 sequence --- unfamiliar nucleotide: '", "+", "nt", "return", "0", "V_usage_mask", ",", "J_usage_mask", "=", "self", ".", "format_usage_masks", "(", "V_usage_mask_in", ",", "J_usage_mask_in", ",", "print_warnings", ")", "return", "self", ".", "compute_CDR3_pgen", "(", "nt2codon_rep", "(", "CDR3_ntseq", ")", ",", "V_usage_mask", ",", "J_usage_mask", ")" ]
Compute Pgen for the inframe nucleotide sequence CDR3_ntseq. Conditioned on the V genes/alleles indicated in V_usage_mask_in and the J genes/alleles in J_usage_mask_in. (Examples are TCRB sequences/model) Parameters ---------- CDR3_ntseq : str Inframe nucleotide sequence composed of ONLY A, C, G, or T (either uppercase or lowercase). V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- pgen : float64 The generation probability (Pgen) of the sequence Examples -------- >>> generation_probability.compute_nt_CDR3_pgen('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC') 3.2674893012379071e-12 >>> generation_probability.compute_nt_CDR3_pgen('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC', 'TRBV30*01', 'TRBJ1-2*01') 2.3986503758867323e-12
[ "Compute", "Pgen", "for", "the", "inframe", "nucleotide", "sequence", "CDR3_ntseq", ".", "Conditioned", "on", "the", "V", "genes", "/", "alleles", "indicated", "in", "V_usage_mask_in", "and", "the", "J", "genes", "/", "alleles", "in", "J_usage_mask_in", ".", "(", "Examples", "are", "TCRB", "sequences", "/", "model", ")", "Parameters", "----------", "CDR3_ntseq", ":", "str", "Inframe", "nucleotide", "sequence", "composed", "of", "ONLY", "A", "C", "G", "or", "T", "(", "either", "uppercase", "or", "lowercase", ")", ".", "V_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "V", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "V", "alleles", ".", "J_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "J", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "J", "alleles", ".", "print_warnings", ":", "bool", "Determines", "whether", "warnings", "are", "printed", "or", "not", ".", "Default", "ON", ".", "Returns", "-------", "pgen", ":", "float64", "The", "generation", "probability", "(", "Pgen", ")", "of", "the", "sequence", "Examples", "--------", ">>>", "generation_probability", ".", "compute_nt_CDR3_pgen", "(", "TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC", ")", "3", ".", "2674893012379071e", "-", "12", ">>>", "generation_probability", ".", "compute_nt_CDR3_pgen", "(", "TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC", "TRBV30", "*", "01", "TRBJ1", "-", "2", "*", "01", ")", "2", ".", "3986503758867323e", "-", "12" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L318-L368
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.format_usage_masks
def format_usage_masks(self, V_usage_mask_in, J_usage_mask_in, print_warnings = True): """Format raw usage masks into lists of indices. Usage masks allows the Pgen computation to be conditioned on the V and J gene/allele identities. The inputted masks are lists of strings, or a single string, of the names of the genes or alleles to be conditioned on. The default mask includes all productive V or J genes. Parameters ---------- V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- V_usage_mask : list of integers Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list of integers Indices of the J alleles to be considered in the Pgen computation Examples -------- >>> generation_probability.format_usage_masks('TRBV27*01','TRBJ1-1*01') ([34], [0]) >>> generation_probability.format_usage_masks('TRBV27*01', '') ([34], [0, 1, 2, 3, 4, 7, 8, 9, 10, 11, 12, 13]) >>> generation_probability.format_usage_masks(['TRBV27*01', 'TRBV13*01'], 'TRBJ1-1*01') ([34, 18], [0]) """ #Format the V usage mask if V_usage_mask_in is None: #Default case, use all productive V genes with non-zero probability #V_usage_mask = [v for v, V in enumerate(ppp['cutV_genomic_CDR3_segs']) if len(V) > 0] V_usage_mask = self.d_V_usage_mask elif isinstance(V_usage_mask_in, list): e_V_usage_mask = set() for v in V_usage_mask_in: try: e_V_usage_mask = e_V_usage_mask.union(self.V_mask_mapping[v]) except KeyError: if print_warnings: print 'Unfamiliar V gene/allele: ' + v pass if len(e_V_usage_mask) == 0: if print_warnings: print 'No recognized V genes/alleles. Using default V_usage_mask' V_usage_mask = self.d_V_usage_mask else: V_usage_mask = list(e_V_usage_mask) else: try: V_usage_mask = self.V_mask_mapping[V_usage_mask_in] except KeyError: #Do raise error here as the mask will be empty if print_warnings: print 'Unfamiliar V usage mask: ' + str(V_usage_mask_in) + ', please check the allowed V alleles. Using default V_usage_mask' V_usage_mask = self.d_V_usage_mask #Format the J usage mask if J_usage_mask_in is None: #Default case, use all productive J genes with non-zero probability #J_usage_mask = [j for j, J in enumerate(ppp['cutJ_genomic_CDR3_segs']) if len(J) > 0] J_usage_mask = self.d_J_usage_mask elif isinstance(J_usage_mask_in, list): e_J_usage_mask = set() for j in J_usage_mask_in: try: e_J_usage_mask = e_J_usage_mask.union(self.J_mask_mapping[j]) except KeyError: if print_warnings: print 'Unfamiliar J gene/allele: ' + j pass if len(e_J_usage_mask) == 0: if print_warnings: print 'No recognized J genes/alleles. Using default J_usage_mask' J_usage_mask = self.d_J_usage_mask else: J_usage_mask = list(e_J_usage_mask) else: try: J_usage_mask = self.J_mask_mapping[J_usage_mask_in] except KeyError: #Do raise error here as the mask will be empty if print_warnings: print 'Unfamiliar J usage mask: ' + str(J_usage_mask_in) + ', please check the allowed J alleles. Using default J_usage_mask' J_usage_mask = self.d_J_usage_mask return V_usage_mask, J_usage_mask
python
def format_usage_masks(self, V_usage_mask_in, J_usage_mask_in, print_warnings = True): """Format raw usage masks into lists of indices. Usage masks allows the Pgen computation to be conditioned on the V and J gene/allele identities. The inputted masks are lists of strings, or a single string, of the names of the genes or alleles to be conditioned on. The default mask includes all productive V or J genes. Parameters ---------- V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- V_usage_mask : list of integers Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list of integers Indices of the J alleles to be considered in the Pgen computation Examples -------- >>> generation_probability.format_usage_masks('TRBV27*01','TRBJ1-1*01') ([34], [0]) >>> generation_probability.format_usage_masks('TRBV27*01', '') ([34], [0, 1, 2, 3, 4, 7, 8, 9, 10, 11, 12, 13]) >>> generation_probability.format_usage_masks(['TRBV27*01', 'TRBV13*01'], 'TRBJ1-1*01') ([34, 18], [0]) """ #Format the V usage mask if V_usage_mask_in is None: #Default case, use all productive V genes with non-zero probability #V_usage_mask = [v for v, V in enumerate(ppp['cutV_genomic_CDR3_segs']) if len(V) > 0] V_usage_mask = self.d_V_usage_mask elif isinstance(V_usage_mask_in, list): e_V_usage_mask = set() for v in V_usage_mask_in: try: e_V_usage_mask = e_V_usage_mask.union(self.V_mask_mapping[v]) except KeyError: if print_warnings: print 'Unfamiliar V gene/allele: ' + v pass if len(e_V_usage_mask) == 0: if print_warnings: print 'No recognized V genes/alleles. Using default V_usage_mask' V_usage_mask = self.d_V_usage_mask else: V_usage_mask = list(e_V_usage_mask) else: try: V_usage_mask = self.V_mask_mapping[V_usage_mask_in] except KeyError: #Do raise error here as the mask will be empty if print_warnings: print 'Unfamiliar V usage mask: ' + str(V_usage_mask_in) + ', please check the allowed V alleles. Using default V_usage_mask' V_usage_mask = self.d_V_usage_mask #Format the J usage mask if J_usage_mask_in is None: #Default case, use all productive J genes with non-zero probability #J_usage_mask = [j for j, J in enumerate(ppp['cutJ_genomic_CDR3_segs']) if len(J) > 0] J_usage_mask = self.d_J_usage_mask elif isinstance(J_usage_mask_in, list): e_J_usage_mask = set() for j in J_usage_mask_in: try: e_J_usage_mask = e_J_usage_mask.union(self.J_mask_mapping[j]) except KeyError: if print_warnings: print 'Unfamiliar J gene/allele: ' + j pass if len(e_J_usage_mask) == 0: if print_warnings: print 'No recognized J genes/alleles. Using default J_usage_mask' J_usage_mask = self.d_J_usage_mask else: J_usage_mask = list(e_J_usage_mask) else: try: J_usage_mask = self.J_mask_mapping[J_usage_mask_in] except KeyError: #Do raise error here as the mask will be empty if print_warnings: print 'Unfamiliar J usage mask: ' + str(J_usage_mask_in) + ', please check the allowed J alleles. Using default J_usage_mask' J_usage_mask = self.d_J_usage_mask return V_usage_mask, J_usage_mask
[ "def", "format_usage_masks", "(", "self", ",", "V_usage_mask_in", ",", "J_usage_mask_in", ",", "print_warnings", "=", "True", ")", ":", "#Format the V usage mask", "if", "V_usage_mask_in", "is", "None", ":", "#Default case, use all productive V genes with non-zero probability", "#V_usage_mask = [v for v, V in enumerate(ppp['cutV_genomic_CDR3_segs']) if len(V) > 0]", "V_usage_mask", "=", "self", ".", "d_V_usage_mask", "elif", "isinstance", "(", "V_usage_mask_in", ",", "list", ")", ":", "e_V_usage_mask", "=", "set", "(", ")", "for", "v", "in", "V_usage_mask_in", ":", "try", ":", "e_V_usage_mask", "=", "e_V_usage_mask", ".", "union", "(", "self", ".", "V_mask_mapping", "[", "v", "]", ")", "except", "KeyError", ":", "if", "print_warnings", ":", "print", "'Unfamiliar V gene/allele: '", "+", "v", "pass", "if", "len", "(", "e_V_usage_mask", ")", "==", "0", ":", "if", "print_warnings", ":", "print", "'No recognized V genes/alleles. Using default V_usage_mask'", "V_usage_mask", "=", "self", ".", "d_V_usage_mask", "else", ":", "V_usage_mask", "=", "list", "(", "e_V_usage_mask", ")", "else", ":", "try", ":", "V_usage_mask", "=", "self", ".", "V_mask_mapping", "[", "V_usage_mask_in", "]", "except", "KeyError", ":", "#Do raise error here as the mask will be empty", "if", "print_warnings", ":", "print", "'Unfamiliar V usage mask: '", "+", "str", "(", "V_usage_mask_in", ")", "+", "', please check the allowed V alleles. Using default V_usage_mask'", "V_usage_mask", "=", "self", ".", "d_V_usage_mask", "#Format the J usage mask", "if", "J_usage_mask_in", "is", "None", ":", "#Default case, use all productive J genes with non-zero probability", "#J_usage_mask = [j for j, J in enumerate(ppp['cutJ_genomic_CDR3_segs']) if len(J) > 0]", "J_usage_mask", "=", "self", ".", "d_J_usage_mask", "elif", "isinstance", "(", "J_usage_mask_in", ",", "list", ")", ":", "e_J_usage_mask", "=", "set", "(", ")", "for", "j", "in", "J_usage_mask_in", ":", "try", ":", "e_J_usage_mask", "=", "e_J_usage_mask", ".", "union", "(", "self", ".", "J_mask_mapping", "[", "j", "]", ")", "except", "KeyError", ":", "if", "print_warnings", ":", "print", "'Unfamiliar J gene/allele: '", "+", "j", "pass", "if", "len", "(", "e_J_usage_mask", ")", "==", "0", ":", "if", "print_warnings", ":", "print", "'No recognized J genes/alleles. Using default J_usage_mask'", "J_usage_mask", "=", "self", ".", "d_J_usage_mask", "else", ":", "J_usage_mask", "=", "list", "(", "e_J_usage_mask", ")", "else", ":", "try", ":", "J_usage_mask", "=", "self", ".", "J_mask_mapping", "[", "J_usage_mask_in", "]", "except", "KeyError", ":", "#Do raise error here as the mask will be empty", "if", "print_warnings", ":", "print", "'Unfamiliar J usage mask: '", "+", "str", "(", "J_usage_mask_in", ")", "+", "', please check the allowed J alleles. Using default J_usage_mask'", "J_usage_mask", "=", "self", ".", "d_J_usage_mask", "return", "V_usage_mask", ",", "J_usage_mask" ]
Format raw usage masks into lists of indices. Usage masks allows the Pgen computation to be conditioned on the V and J gene/allele identities. The inputted masks are lists of strings, or a single string, of the names of the genes or alleles to be conditioned on. The default mask includes all productive V or J genes. Parameters ---------- V_usage_mask_in : str or list An object to indicate which V alleles should be considered. The default input is None which returns the list of all productive V alleles. J_usage_mask_in : str or list An object to indicate which J alleles should be considered. The default input is None which returns the list of all productive J alleles. print_warnings : bool Determines whether warnings are printed or not. Default ON. Returns ------- V_usage_mask : list of integers Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list of integers Indices of the J alleles to be considered in the Pgen computation Examples -------- >>> generation_probability.format_usage_masks('TRBV27*01','TRBJ1-1*01') ([34], [0]) >>> generation_probability.format_usage_masks('TRBV27*01', '') ([34], [0, 1, 2, 3, 4, 7, 8, 9, 10, 11, 12, 13]) >>> generation_probability.format_usage_masks(['TRBV27*01', 'TRBV13*01'], 'TRBJ1-1*01') ([34, 18], [0])
[ "Format", "raw", "usage", "masks", "into", "lists", "of", "indices", ".", "Usage", "masks", "allows", "the", "Pgen", "computation", "to", "be", "conditioned", "on", "the", "V", "and", "J", "gene", "/", "allele", "identities", ".", "The", "inputted", "masks", "are", "lists", "of", "strings", "or", "a", "single", "string", "of", "the", "names", "of", "the", "genes", "or", "alleles", "to", "be", "conditioned", "on", ".", "The", "default", "mask", "includes", "all", "productive", "V", "or", "J", "genes", ".", "Parameters", "----------", "V_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "V", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "V", "alleles", ".", "J_usage_mask_in", ":", "str", "or", "list", "An", "object", "to", "indicate", "which", "J", "alleles", "should", "be", "considered", ".", "The", "default", "input", "is", "None", "which", "returns", "the", "list", "of", "all", "productive", "J", "alleles", ".", "print_warnings", ":", "bool", "Determines", "whether", "warnings", "are", "printed", "or", "not", ".", "Default", "ON", ".", "Returns", "-------", "V_usage_mask", ":", "list", "of", "integers", "Indices", "of", "the", "V", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "J_usage_mask", ":", "list", "of", "integers", "Indices", "of", "the", "J", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "Examples", "--------", ">>>", "generation_probability", ".", "format_usage_masks", "(", "TRBV27", "*", "01", "TRBJ1", "-", "1", "*", "01", ")", "(", "[", "34", "]", "[", "0", "]", ")", ">>>", "generation_probability", ".", "format_usage_masks", "(", "TRBV27", "*", "01", ")", "(", "[", "34", "]", "[", "0", "1", "2", "3", "4", "7", "8", "9", "10", "11", "12", "13", "]", ")", ">>>", "generation_probability", ".", "format_usage_masks", "(", "[", "TRBV27", "*", "01", "TRBV13", "*", "01", "]", "TRBJ1", "-", "1", "*", "01", ")", "(", "[", "34", "18", "]", "[", "0", "]", ")" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L377-L470
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.list_seqs_from_regex
def list_seqs_from_regex(self, regex_seq, print_warnings = True, raise_overload_warning = True): """List sequences that match regular expression template. This function parses a limited regular expression vocabulary, and lists all the sequences consistent with the regular expression. Supported regex syntax: [] and {}. Cannot have two {} in a row. Note we can't use Kline star (*) as this is the symbol for a stop codon --- use {}. Parameters ---------- regex_seq : str The regular expression string that represents the sequences to be listed. print_warnings : bool Determines whether warnings are printed or not. Default ON. raise_overload_warning : bool A flag to warn of more than 10000 seqs corresponding to the regex_seq Returns ------- CDR3_seqs : list A list of CDR3 sequences that correspond to the regex_seq Examples -------- >>> generation_probability.list_seqs_from_regex('CASS[AGR]SARPEQFF') ['CASSGSARPEQFF', 'CASSRSARPEQFF', 'CASSASARPEQFF'] >>> generation_probability.list_seqs_from_regex('CASSAX{0,5}SARPEQFF') ['CASSASARPEQFF', 'CASSAXXXXSARPEQFF', 'CASSAXXSARPEQFF', 'CASSAXXXXXSARPEQFF', 'CASSAXXXSARPEQFF', 'CASSAXSARPEQFF'] """ aa_symbols = ''.join(self.codons_dict) default_max_reps = 40 #Check to make sure that expression is of the right form/symbols #Identify bracket expressions bracket_ex = [x for x in re.findall('\[[' + aa_symbols + ']*?\]|\{\d+,{0,1}\d*\}', regex_seq)] split_seq = re.split('\[[' + aa_symbols + ']*?\]|\{\d+,{0,1}\d*\}', regex_seq) #Check that all remaining characters are in the codon dict for aa in ''.join(split_seq): if aa not in aa_symbols: if print_warnings: print 'Unfamiliar symbol representing a codon:' + aa + ' --- check codon dictionary or the regex syntax' return [] regex_list = [split_seq[i/2] if i%2 == 0 else bracket_ex[i/2] for i in range(len(bracket_ex) + len(split_seq)) if not (i%2 == 0 and len(split_seq[i/2]) == 0)] max_num_seqs = 1 for l, ex in enumerate(regex_list[::-1]): i = len(regex_list) - l - 1 if ex[0] == '[': #bracket expression #check characters for aa in ex.strip('[]'): if aa not in aa_symbols: if print_warnings: print 'Unfamiliar symbol representing a codon:' + aa + ' --- check codon dictionary' return [] max_num_seqs *= len(ex) - 2 elif ex[0] == '{': #curly bracket if i == 0: if print_warnings: print "Can't have {} expression at start of sequence" return [] elif isinstance(regex_list[i-1], list): if print_warnings: print "Two {} expressions in a row is not supported" return [] elif regex_list[i-1][0] == '[': syms = regex_list[i-1].strip('[]') regex_list[i-1] = '' else: syms = regex_list[i-1][-1] regex_list[i-1] = regex_list[i-1][:-1] if ',' not in ex: new_expression = [int(ex.strip('{}')), int(ex.strip('{}')), syms] max_num_seqs *= len(syms)**new_expression[0] else: try: new_expression = [int(ex.strip('{}').split(',')[0]), int(ex.strip('{}').split(',')[1]), syms] except ValueError: #No max limit --- use default new_expression = [int(ex.strip('{}').split(',')[0]), default_max_reps, syms] if new_expression[0] > new_expression[1]: if print_warnings: print 'Check regex syntax --- should be {min,max}' return [] max_num_seqs *= sum([len(syms)**n for n in range(new_expression[0], new_expression[1]+1)])/len(syms) #print new_expression regex_list[i] = new_expression if max_num_seqs > 10000 and raise_overload_warning: if print_warnings: answer = raw_input('Warning large number of sequences (estimated ' + str(max_num_seqs) + ' seqs) match the regular expression. Possible memory and time issues. Continue? (y/n)') if not answer == 'y': print 'Canceling...' return [] else: return [] #print regex_list CDR3_seqs = [''] for l, ex in enumerate(regex_list[::-1]): i = len(regex_list) - l - 1 if isinstance(ex, list): #curly bracket case c_seqs = [''] f_seqs = [] for j in range(ex[1] + 1): if j in range(ex[0], ex[1]+1): f_seqs += c_seqs c_seqs = [aa + c_seq for aa in ex[2] for c_seq in c_seqs] CDR3_seqs = [f_seq + CDR3_seq for f_seq in f_seqs for CDR3_seq in CDR3_seqs] elif len(ex) == 0: pass elif ex[0] == '[': #square bracket case CDR3_seqs = [aa + CDR3_seq for aa in ex.strip('[]') for CDR3_seq in CDR3_seqs] else: CDR3_seqs = [ex + CDR3_seq for CDR3_seq in CDR3_seqs] return list(set(CDR3_seqs))
python
def list_seqs_from_regex(self, regex_seq, print_warnings = True, raise_overload_warning = True): """List sequences that match regular expression template. This function parses a limited regular expression vocabulary, and lists all the sequences consistent with the regular expression. Supported regex syntax: [] and {}. Cannot have two {} in a row. Note we can't use Kline star (*) as this is the symbol for a stop codon --- use {}. Parameters ---------- regex_seq : str The regular expression string that represents the sequences to be listed. print_warnings : bool Determines whether warnings are printed or not. Default ON. raise_overload_warning : bool A flag to warn of more than 10000 seqs corresponding to the regex_seq Returns ------- CDR3_seqs : list A list of CDR3 sequences that correspond to the regex_seq Examples -------- >>> generation_probability.list_seqs_from_regex('CASS[AGR]SARPEQFF') ['CASSGSARPEQFF', 'CASSRSARPEQFF', 'CASSASARPEQFF'] >>> generation_probability.list_seqs_from_regex('CASSAX{0,5}SARPEQFF') ['CASSASARPEQFF', 'CASSAXXXXSARPEQFF', 'CASSAXXSARPEQFF', 'CASSAXXXXXSARPEQFF', 'CASSAXXXSARPEQFF', 'CASSAXSARPEQFF'] """ aa_symbols = ''.join(self.codons_dict) default_max_reps = 40 #Check to make sure that expression is of the right form/symbols #Identify bracket expressions bracket_ex = [x for x in re.findall('\[[' + aa_symbols + ']*?\]|\{\d+,{0,1}\d*\}', regex_seq)] split_seq = re.split('\[[' + aa_symbols + ']*?\]|\{\d+,{0,1}\d*\}', regex_seq) #Check that all remaining characters are in the codon dict for aa in ''.join(split_seq): if aa not in aa_symbols: if print_warnings: print 'Unfamiliar symbol representing a codon:' + aa + ' --- check codon dictionary or the regex syntax' return [] regex_list = [split_seq[i/2] if i%2 == 0 else bracket_ex[i/2] for i in range(len(bracket_ex) + len(split_seq)) if not (i%2 == 0 and len(split_seq[i/2]) == 0)] max_num_seqs = 1 for l, ex in enumerate(regex_list[::-1]): i = len(regex_list) - l - 1 if ex[0] == '[': #bracket expression #check characters for aa in ex.strip('[]'): if aa not in aa_symbols: if print_warnings: print 'Unfamiliar symbol representing a codon:' + aa + ' --- check codon dictionary' return [] max_num_seqs *= len(ex) - 2 elif ex[0] == '{': #curly bracket if i == 0: if print_warnings: print "Can't have {} expression at start of sequence" return [] elif isinstance(regex_list[i-1], list): if print_warnings: print "Two {} expressions in a row is not supported" return [] elif regex_list[i-1][0] == '[': syms = regex_list[i-1].strip('[]') regex_list[i-1] = '' else: syms = regex_list[i-1][-1] regex_list[i-1] = regex_list[i-1][:-1] if ',' not in ex: new_expression = [int(ex.strip('{}')), int(ex.strip('{}')), syms] max_num_seqs *= len(syms)**new_expression[0] else: try: new_expression = [int(ex.strip('{}').split(',')[0]), int(ex.strip('{}').split(',')[1]), syms] except ValueError: #No max limit --- use default new_expression = [int(ex.strip('{}').split(',')[0]), default_max_reps, syms] if new_expression[0] > new_expression[1]: if print_warnings: print 'Check regex syntax --- should be {min,max}' return [] max_num_seqs *= sum([len(syms)**n for n in range(new_expression[0], new_expression[1]+1)])/len(syms) #print new_expression regex_list[i] = new_expression if max_num_seqs > 10000 and raise_overload_warning: if print_warnings: answer = raw_input('Warning large number of sequences (estimated ' + str(max_num_seqs) + ' seqs) match the regular expression. Possible memory and time issues. Continue? (y/n)') if not answer == 'y': print 'Canceling...' return [] else: return [] #print regex_list CDR3_seqs = [''] for l, ex in enumerate(regex_list[::-1]): i = len(regex_list) - l - 1 if isinstance(ex, list): #curly bracket case c_seqs = [''] f_seqs = [] for j in range(ex[1] + 1): if j in range(ex[0], ex[1]+1): f_seqs += c_seqs c_seqs = [aa + c_seq for aa in ex[2] for c_seq in c_seqs] CDR3_seqs = [f_seq + CDR3_seq for f_seq in f_seqs for CDR3_seq in CDR3_seqs] elif len(ex) == 0: pass elif ex[0] == '[': #square bracket case CDR3_seqs = [aa + CDR3_seq for aa in ex.strip('[]') for CDR3_seq in CDR3_seqs] else: CDR3_seqs = [ex + CDR3_seq for CDR3_seq in CDR3_seqs] return list(set(CDR3_seqs))
[ "def", "list_seqs_from_regex", "(", "self", ",", "regex_seq", ",", "print_warnings", "=", "True", ",", "raise_overload_warning", "=", "True", ")", ":", "aa_symbols", "=", "''", ".", "join", "(", "self", ".", "codons_dict", ")", "default_max_reps", "=", "40", "#Check to make sure that expression is of the right form/symbols", "#Identify bracket expressions", "bracket_ex", "=", "[", "x", "for", "x", "in", "re", ".", "findall", "(", "'\\[['", "+", "aa_symbols", "+", "']*?\\]|\\{\\d+,{0,1}\\d*\\}'", ",", "regex_seq", ")", "]", "split_seq", "=", "re", ".", "split", "(", "'\\[['", "+", "aa_symbols", "+", "']*?\\]|\\{\\d+,{0,1}\\d*\\}'", ",", "regex_seq", ")", "#Check that all remaining characters are in the codon dict", "for", "aa", "in", "''", ".", "join", "(", "split_seq", ")", ":", "if", "aa", "not", "in", "aa_symbols", ":", "if", "print_warnings", ":", "print", "'Unfamiliar symbol representing a codon:'", "+", "aa", "+", "' --- check codon dictionary or the regex syntax'", "return", "[", "]", "regex_list", "=", "[", "split_seq", "[", "i", "/", "2", "]", "if", "i", "%", "2", "==", "0", "else", "bracket_ex", "[", "i", "/", "2", "]", "for", "i", "in", "range", "(", "len", "(", "bracket_ex", ")", "+", "len", "(", "split_seq", ")", ")", "if", "not", "(", "i", "%", "2", "==", "0", "and", "len", "(", "split_seq", "[", "i", "/", "2", "]", ")", "==", "0", ")", "]", "max_num_seqs", "=", "1", "for", "l", ",", "ex", "in", "enumerate", "(", "regex_list", "[", ":", ":", "-", "1", "]", ")", ":", "i", "=", "len", "(", "regex_list", ")", "-", "l", "-", "1", "if", "ex", "[", "0", "]", "==", "'['", ":", "#bracket expression", "#check characters", "for", "aa", "in", "ex", ".", "strip", "(", "'[]'", ")", ":", "if", "aa", "not", "in", "aa_symbols", ":", "if", "print_warnings", ":", "print", "'Unfamiliar symbol representing a codon:'", "+", "aa", "+", "' --- check codon dictionary'", "return", "[", "]", "max_num_seqs", "*=", "len", "(", "ex", ")", "-", "2", "elif", "ex", "[", "0", "]", "==", "'{'", ":", "#curly bracket", "if", "i", "==", "0", ":", "if", "print_warnings", ":", "print", "\"Can't have {} expression at start of sequence\"", "return", "[", "]", "elif", "isinstance", "(", "regex_list", "[", "i", "-", "1", "]", ",", "list", ")", ":", "if", "print_warnings", ":", "print", "\"Two {} expressions in a row is not supported\"", "return", "[", "]", "elif", "regex_list", "[", "i", "-", "1", "]", "[", "0", "]", "==", "'['", ":", "syms", "=", "regex_list", "[", "i", "-", "1", "]", ".", "strip", "(", "'[]'", ")", "regex_list", "[", "i", "-", "1", "]", "=", "''", "else", ":", "syms", "=", "regex_list", "[", "i", "-", "1", "]", "[", "-", "1", "]", "regex_list", "[", "i", "-", "1", "]", "=", "regex_list", "[", "i", "-", "1", "]", "[", ":", "-", "1", "]", "if", "','", "not", "in", "ex", ":", "new_expression", "=", "[", "int", "(", "ex", ".", "strip", "(", "'{}'", ")", ")", ",", "int", "(", "ex", ".", "strip", "(", "'{}'", ")", ")", ",", "syms", "]", "max_num_seqs", "*=", "len", "(", "syms", ")", "**", "new_expression", "[", "0", "]", "else", ":", "try", ":", "new_expression", "=", "[", "int", "(", "ex", ".", "strip", "(", "'{}'", ")", ".", "split", "(", "','", ")", "[", "0", "]", ")", ",", "int", "(", "ex", ".", "strip", "(", "'{}'", ")", ".", "split", "(", "','", ")", "[", "1", "]", ")", ",", "syms", "]", "except", "ValueError", ":", "#No max limit --- use default", "new_expression", "=", "[", "int", "(", "ex", ".", "strip", "(", "'{}'", ")", ".", "split", "(", "','", ")", "[", "0", "]", ")", ",", "default_max_reps", ",", "syms", "]", "if", "new_expression", "[", "0", "]", ">", "new_expression", "[", "1", "]", ":", "if", "print_warnings", ":", "print", "'Check regex syntax --- should be {min,max}'", "return", "[", "]", "max_num_seqs", "*=", "sum", "(", "[", "len", "(", "syms", ")", "**", "n", "for", "n", "in", "range", "(", "new_expression", "[", "0", "]", ",", "new_expression", "[", "1", "]", "+", "1", ")", "]", ")", "/", "len", "(", "syms", ")", "#print new_expression", "regex_list", "[", "i", "]", "=", "new_expression", "if", "max_num_seqs", ">", "10000", "and", "raise_overload_warning", ":", "if", "print_warnings", ":", "answer", "=", "raw_input", "(", "'Warning large number of sequences (estimated '", "+", "str", "(", "max_num_seqs", ")", "+", "' seqs) match the regular expression. Possible memory and time issues. Continue? (y/n)'", ")", "if", "not", "answer", "==", "'y'", ":", "print", "'Canceling...'", "return", "[", "]", "else", ":", "return", "[", "]", "#print regex_list", "CDR3_seqs", "=", "[", "''", "]", "for", "l", ",", "ex", "in", "enumerate", "(", "regex_list", "[", ":", ":", "-", "1", "]", ")", ":", "i", "=", "len", "(", "regex_list", ")", "-", "l", "-", "1", "if", "isinstance", "(", "ex", ",", "list", ")", ":", "#curly bracket case", "c_seqs", "=", "[", "''", "]", "f_seqs", "=", "[", "]", "for", "j", "in", "range", "(", "ex", "[", "1", "]", "+", "1", ")", ":", "if", "j", "in", "range", "(", "ex", "[", "0", "]", ",", "ex", "[", "1", "]", "+", "1", ")", ":", "f_seqs", "+=", "c_seqs", "c_seqs", "=", "[", "aa", "+", "c_seq", "for", "aa", "in", "ex", "[", "2", "]", "for", "c_seq", "in", "c_seqs", "]", "CDR3_seqs", "=", "[", "f_seq", "+", "CDR3_seq", "for", "f_seq", "in", "f_seqs", "for", "CDR3_seq", "in", "CDR3_seqs", "]", "elif", "len", "(", "ex", ")", "==", "0", ":", "pass", "elif", "ex", "[", "0", "]", "==", "'['", ":", "#square bracket case", "CDR3_seqs", "=", "[", "aa", "+", "CDR3_seq", "for", "aa", "in", "ex", ".", "strip", "(", "'[]'", ")", "for", "CDR3_seq", "in", "CDR3_seqs", "]", "else", ":", "CDR3_seqs", "=", "[", "ex", "+", "CDR3_seq", "for", "CDR3_seq", "in", "CDR3_seqs", "]", "return", "list", "(", "set", "(", "CDR3_seqs", ")", ")" ]
List sequences that match regular expression template. This function parses a limited regular expression vocabulary, and lists all the sequences consistent with the regular expression. Supported regex syntax: [] and {}. Cannot have two {} in a row. Note we can't use Kline star (*) as this is the symbol for a stop codon --- use {}. Parameters ---------- regex_seq : str The regular expression string that represents the sequences to be listed. print_warnings : bool Determines whether warnings are printed or not. Default ON. raise_overload_warning : bool A flag to warn of more than 10000 seqs corresponding to the regex_seq Returns ------- CDR3_seqs : list A list of CDR3 sequences that correspond to the regex_seq Examples -------- >>> generation_probability.list_seqs_from_regex('CASS[AGR]SARPEQFF') ['CASSGSARPEQFF', 'CASSRSARPEQFF', 'CASSASARPEQFF'] >>> generation_probability.list_seqs_from_regex('CASSAX{0,5}SARPEQFF') ['CASSASARPEQFF', 'CASSAXXXXSARPEQFF', 'CASSAXXSARPEQFF', 'CASSAXXXXXSARPEQFF', 'CASSAXXXSARPEQFF', 'CASSAXSARPEQFF']
[ "List", "sequences", "that", "match", "regular", "expression", "template", ".", "This", "function", "parses", "a", "limited", "regular", "expression", "vocabulary", "and", "lists", "all", "the", "sequences", "consistent", "with", "the", "regular", "expression", ".", "Supported", "regex", "syntax", ":", "[]", "and", "{}", ".", "Cannot", "have", "two", "{}", "in", "a", "row", ".", "Note", "we", "can", "t", "use", "Kline", "star", "(", "*", ")", "as", "this", "is", "the", "symbol", "for", "a", "stop", "codon", "---", "use", "{}", ".", "Parameters", "----------", "regex_seq", ":", "str", "The", "regular", "expression", "string", "that", "represents", "the", "sequences", "to", "be", "listed", ".", "print_warnings", ":", "bool", "Determines", "whether", "warnings", "are", "printed", "or", "not", ".", "Default", "ON", ".", "raise_overload_warning", ":", "bool", "A", "flag", "to", "warn", "of", "more", "than", "10000", "seqs", "corresponding", "to", "the", "regex_seq", "Returns", "-------", "CDR3_seqs", ":", "list", "A", "list", "of", "CDR3", "sequences", "that", "correspond", "to", "the", "regex_seq", "Examples", "--------", ">>>", "generation_probability", ".", "list_seqs_from_regex", "(", "CASS", "[", "AGR", "]", "SARPEQFF", ")", "[", "CASSGSARPEQFF", "CASSRSARPEQFF", "CASSASARPEQFF", "]", ">>>", "generation_probability", ".", "list_seqs_from_regex", "(", "CASSAX", "{", "0", "5", "}", "SARPEQFF", ")", "[", "CASSASARPEQFF", "CASSAXXXXSARPEQFF", "CASSAXXSARPEQFF", "CASSAXXXXXSARPEQFF", "CASSAXXXSARPEQFF", "CASSAXSARPEQFF", "]" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L472-L601
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.max_nt_to_aa_alignment_left
def max_nt_to_aa_alignment_left(self, CDR3_seq, ntseq): """Find maximum match between CDR3_seq and ntseq from the left. This function returns the length of the maximum length nucleotide subsequence of ntseq contiguous from the left (or 5' end) that is consistent with the 'amino acid' sequence CDR3_seq. Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). ntseq : str Genomic (V locus) nucleotide sequence to match. Returns ------- max_alignment : int Maximum length (in nucleotides) nucleotide sequence that matches the CDR3 'amino acid' sequence. Example -------- >>> generation_probability.max_nt_to_aa_alignment_left('CASSSEGAGGPSLRGHEQFF', 'TGTGCCAGCAGTTTATCGATA') 13 """ max_alignment = 0 if len(ntseq) == 0: return 0 aa_aligned = True while aa_aligned: if ntseq[max_alignment:max_alignment+3] in self.codons_dict[CDR3_seq[max_alignment/3]]: max_alignment += 3 if max_alignment/3 == len(CDR3_seq): return max_alignment else: break aa_aligned = False last_codon = ntseq[max_alignment:max_alignment+3] codon_frag = '' for nt in last_codon: codon_frag += nt if codon_frag in self.sub_codons_left[CDR3_seq[max_alignment/3]]: max_alignment += 1 else: break return max_alignment
python
def max_nt_to_aa_alignment_left(self, CDR3_seq, ntseq): """Find maximum match between CDR3_seq and ntseq from the left. This function returns the length of the maximum length nucleotide subsequence of ntseq contiguous from the left (or 5' end) that is consistent with the 'amino acid' sequence CDR3_seq. Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). ntseq : str Genomic (V locus) nucleotide sequence to match. Returns ------- max_alignment : int Maximum length (in nucleotides) nucleotide sequence that matches the CDR3 'amino acid' sequence. Example -------- >>> generation_probability.max_nt_to_aa_alignment_left('CASSSEGAGGPSLRGHEQFF', 'TGTGCCAGCAGTTTATCGATA') 13 """ max_alignment = 0 if len(ntseq) == 0: return 0 aa_aligned = True while aa_aligned: if ntseq[max_alignment:max_alignment+3] in self.codons_dict[CDR3_seq[max_alignment/3]]: max_alignment += 3 if max_alignment/3 == len(CDR3_seq): return max_alignment else: break aa_aligned = False last_codon = ntseq[max_alignment:max_alignment+3] codon_frag = '' for nt in last_codon: codon_frag += nt if codon_frag in self.sub_codons_left[CDR3_seq[max_alignment/3]]: max_alignment += 1 else: break return max_alignment
[ "def", "max_nt_to_aa_alignment_left", "(", "self", ",", "CDR3_seq", ",", "ntseq", ")", ":", "max_alignment", "=", "0", "if", "len", "(", "ntseq", ")", "==", "0", ":", "return", "0", "aa_aligned", "=", "True", "while", "aa_aligned", ":", "if", "ntseq", "[", "max_alignment", ":", "max_alignment", "+", "3", "]", "in", "self", ".", "codons_dict", "[", "CDR3_seq", "[", "max_alignment", "/", "3", "]", "]", ":", "max_alignment", "+=", "3", "if", "max_alignment", "/", "3", "==", "len", "(", "CDR3_seq", ")", ":", "return", "max_alignment", "else", ":", "break", "aa_aligned", "=", "False", "last_codon", "=", "ntseq", "[", "max_alignment", ":", "max_alignment", "+", "3", "]", "codon_frag", "=", "''", "for", "nt", "in", "last_codon", ":", "codon_frag", "+=", "nt", "if", "codon_frag", "in", "self", ".", "sub_codons_left", "[", "CDR3_seq", "[", "max_alignment", "/", "3", "]", "]", ":", "max_alignment", "+=", "1", "else", ":", "break", "return", "max_alignment" ]
Find maximum match between CDR3_seq and ntseq from the left. This function returns the length of the maximum length nucleotide subsequence of ntseq contiguous from the left (or 5' end) that is consistent with the 'amino acid' sequence CDR3_seq. Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). ntseq : str Genomic (V locus) nucleotide sequence to match. Returns ------- max_alignment : int Maximum length (in nucleotides) nucleotide sequence that matches the CDR3 'amino acid' sequence. Example -------- >>> generation_probability.max_nt_to_aa_alignment_left('CASSSEGAGGPSLRGHEQFF', 'TGTGCCAGCAGTTTATCGATA') 13
[ "Find", "maximum", "match", "between", "CDR3_seq", "and", "ntseq", "from", "the", "left", ".", "This", "function", "returns", "the", "length", "of", "the", "maximum", "length", "nucleotide", "subsequence", "of", "ntseq", "contiguous", "from", "the", "left", "(", "or", "5", "end", ")", "that", "is", "consistent", "with", "the", "amino", "acid", "sequence", "CDR3_seq", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "ntseq", ":", "str", "Genomic", "(", "V", "locus", ")", "nucleotide", "sequence", "to", "match", ".", "Returns", "-------", "max_alignment", ":", "int", "Maximum", "length", "(", "in", "nucleotides", ")", "nucleotide", "sequence", "that", "matches", "the", "CDR3", "amino", "acid", "sequence", ".", "Example", "--------", ">>>", "generation_probability", ".", "max_nt_to_aa_alignment_left", "(", "CASSSEGAGGPSLRGHEQFF", "TGTGCCAGCAGTTTATCGATA", ")", "13" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L604-L652
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbability.max_nt_to_aa_alignment_right
def max_nt_to_aa_alignment_right(self, CDR3_seq, ntseq): """Find maximum match between CDR3_seq and ntseq from the right. This function returns the length of the maximum length nucleotide subsequence of ntseq contiguous from the right (or 3' end) that is consistent with the 'amino acid' sequence CDR3_seq Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). ntseq : str Genomic (J locus) nucleotide sequence to match. Returns ------- max_alignment : int Maximum length (in nucleotides) nucleotide sequence that matches the CDR3 'amino acid' sequence. Example -------- >>> generation_probability.max_nt_to_aa_alignment_right('CASSSEGAGGPSLRGHEQFF', 'TTCATGAACACTGAAGCTTTCTTT') 6 """ r_CDR3_seq = CDR3_seq[::-1] #reverse CDR3_seq r_ntseq = ntseq[::-1] #reverse ntseq max_alignment = 0 if len(ntseq) == 0: return 0 aa_aligned = True while aa_aligned: if r_ntseq[max_alignment:max_alignment+3][::-1] in self.codons_dict[r_CDR3_seq[max_alignment/3]]: max_alignment += 3 if max_alignment/3 == len(CDR3_seq): return max_alignment else: break aa_aligned = False r_last_codon = r_ntseq[max_alignment:max_alignment+3] codon_frag = '' for nt in r_last_codon: codon_frag = nt + codon_frag if codon_frag in self.sub_codons_right[r_CDR3_seq[max_alignment/3]]: max_alignment += 1 else: break return max_alignment
python
def max_nt_to_aa_alignment_right(self, CDR3_seq, ntseq): """Find maximum match between CDR3_seq and ntseq from the right. This function returns the length of the maximum length nucleotide subsequence of ntseq contiguous from the right (or 3' end) that is consistent with the 'amino acid' sequence CDR3_seq Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). ntseq : str Genomic (J locus) nucleotide sequence to match. Returns ------- max_alignment : int Maximum length (in nucleotides) nucleotide sequence that matches the CDR3 'amino acid' sequence. Example -------- >>> generation_probability.max_nt_to_aa_alignment_right('CASSSEGAGGPSLRGHEQFF', 'TTCATGAACACTGAAGCTTTCTTT') 6 """ r_CDR3_seq = CDR3_seq[::-1] #reverse CDR3_seq r_ntseq = ntseq[::-1] #reverse ntseq max_alignment = 0 if len(ntseq) == 0: return 0 aa_aligned = True while aa_aligned: if r_ntseq[max_alignment:max_alignment+3][::-1] in self.codons_dict[r_CDR3_seq[max_alignment/3]]: max_alignment += 3 if max_alignment/3 == len(CDR3_seq): return max_alignment else: break aa_aligned = False r_last_codon = r_ntseq[max_alignment:max_alignment+3] codon_frag = '' for nt in r_last_codon: codon_frag = nt + codon_frag if codon_frag in self.sub_codons_right[r_CDR3_seq[max_alignment/3]]: max_alignment += 1 else: break return max_alignment
[ "def", "max_nt_to_aa_alignment_right", "(", "self", ",", "CDR3_seq", ",", "ntseq", ")", ":", "r_CDR3_seq", "=", "CDR3_seq", "[", ":", ":", "-", "1", "]", "#reverse CDR3_seq", "r_ntseq", "=", "ntseq", "[", ":", ":", "-", "1", "]", "#reverse ntseq", "max_alignment", "=", "0", "if", "len", "(", "ntseq", ")", "==", "0", ":", "return", "0", "aa_aligned", "=", "True", "while", "aa_aligned", ":", "if", "r_ntseq", "[", "max_alignment", ":", "max_alignment", "+", "3", "]", "[", ":", ":", "-", "1", "]", "in", "self", ".", "codons_dict", "[", "r_CDR3_seq", "[", "max_alignment", "/", "3", "]", "]", ":", "max_alignment", "+=", "3", "if", "max_alignment", "/", "3", "==", "len", "(", "CDR3_seq", ")", ":", "return", "max_alignment", "else", ":", "break", "aa_aligned", "=", "False", "r_last_codon", "=", "r_ntseq", "[", "max_alignment", ":", "max_alignment", "+", "3", "]", "codon_frag", "=", "''", "for", "nt", "in", "r_last_codon", ":", "codon_frag", "=", "nt", "+", "codon_frag", "if", "codon_frag", "in", "self", ".", "sub_codons_right", "[", "r_CDR3_seq", "[", "max_alignment", "/", "3", "]", "]", ":", "max_alignment", "+=", "1", "else", ":", "break", "return", "max_alignment" ]
Find maximum match between CDR3_seq and ntseq from the right. This function returns the length of the maximum length nucleotide subsequence of ntseq contiguous from the right (or 3' end) that is consistent with the 'amino acid' sequence CDR3_seq Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). ntseq : str Genomic (J locus) nucleotide sequence to match. Returns ------- max_alignment : int Maximum length (in nucleotides) nucleotide sequence that matches the CDR3 'amino acid' sequence. Example -------- >>> generation_probability.max_nt_to_aa_alignment_right('CASSSEGAGGPSLRGHEQFF', 'TTCATGAACACTGAAGCTTTCTTT') 6
[ "Find", "maximum", "match", "between", "CDR3_seq", "and", "ntseq", "from", "the", "right", ".", "This", "function", "returns", "the", "length", "of", "the", "maximum", "length", "nucleotide", "subsequence", "of", "ntseq", "contiguous", "from", "the", "right", "(", "or", "3", "end", ")", "that", "is", "consistent", "with", "the", "amino", "acid", "sequence", "CDR3_seq", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "ntseq", ":", "str", "Genomic", "(", "J", "locus", ")", "nucleotide", "sequence", "to", "match", ".", "Returns", "-------", "max_alignment", ":", "int", "Maximum", "length", "(", "in", "nucleotides", ")", "nucleotide", "sequence", "that", "matches", "the", "CDR3", "amino", "acid", "sequence", ".", "Example", "--------", ">>>", "generation_probability", ".", "max_nt_to_aa_alignment_right", "(", "CASSSEGAGGPSLRGHEQFF", "TTCATGAACACTGAAGCTTTCTTT", ")", "6" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L654-L703
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVDJ.compute_CDR3_pgen
def compute_CDR3_pgen(self, CDR3_seq, V_usage_mask, J_usage_mask): """Compute Pgen for CDR3 'amino acid' sequence CDR3_seq from VDJ model. Conditioned on the already formatted V genes/alleles indicated in V_usage_mask and the J genes/alleles in J_usage_mask. (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> compute_CDR3_pgen('CAWSVAPDRGGYTF', ppp, [42], [1]) 1.203646865765782e-10 >>> compute_CDR3_pgen(nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC'), ppp, [42], [1]) 2.3986503758867323e-12 >>> compute_CDR3_pgen('\xbb\x96\xab\xb8\x8e\xb6\xa5\x92\xa8\xba\x9a\x93\x94\x9f', ppp, [42], [1]) 2.3986503758867323e-12 """ #Genomic V alignment/matching (contribution from P(V, delV)), return Pi_V Pi_V, max_V_align = self.compute_Pi_V(CDR3_seq, V_usage_mask) #Include VD insertions (Rvd and PinsVD) to get the total contribution from the left (3') side. Return Pi_L Pi_L = self.compute_Pi_L(CDR3_seq, Pi_V, max_V_align) #Genomic J alignment/matching (contribution from P(D, J, delJ)), return Pi_J_given_D Pi_J_given_D, max_J_align = self.compute_Pi_J_given_D(CDR3_seq, J_usage_mask) #Include DJ insertions (Rdj and PinsDJ), return Pi_JinsDJ_given_D Pi_JinsDJ_given_D = self.compute_Pi_JinsDJ_given_D(CDR3_seq, Pi_J_given_D, max_J_align) #Include D genomic contribution (P(delDl, delDr | D)) to complete the contribution from the right (5') side. Return Pi_R Pi_R = self.compute_Pi_R(CDR3_seq, Pi_JinsDJ_given_D) pgen = 0 #zip Pi_L and Pi_R together to get total pgen for pos in range(len(CDR3_seq)*3 - 1): pgen += np.dot(Pi_L[:, pos], Pi_R[:, pos+1]) return pgen
python
def compute_CDR3_pgen(self, CDR3_seq, V_usage_mask, J_usage_mask): """Compute Pgen for CDR3 'amino acid' sequence CDR3_seq from VDJ model. Conditioned on the already formatted V genes/alleles indicated in V_usage_mask and the J genes/alleles in J_usage_mask. (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> compute_CDR3_pgen('CAWSVAPDRGGYTF', ppp, [42], [1]) 1.203646865765782e-10 >>> compute_CDR3_pgen(nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC'), ppp, [42], [1]) 2.3986503758867323e-12 >>> compute_CDR3_pgen('\xbb\x96\xab\xb8\x8e\xb6\xa5\x92\xa8\xba\x9a\x93\x94\x9f', ppp, [42], [1]) 2.3986503758867323e-12 """ #Genomic V alignment/matching (contribution from P(V, delV)), return Pi_V Pi_V, max_V_align = self.compute_Pi_V(CDR3_seq, V_usage_mask) #Include VD insertions (Rvd and PinsVD) to get the total contribution from the left (3') side. Return Pi_L Pi_L = self.compute_Pi_L(CDR3_seq, Pi_V, max_V_align) #Genomic J alignment/matching (contribution from P(D, J, delJ)), return Pi_J_given_D Pi_J_given_D, max_J_align = self.compute_Pi_J_given_D(CDR3_seq, J_usage_mask) #Include DJ insertions (Rdj and PinsDJ), return Pi_JinsDJ_given_D Pi_JinsDJ_given_D = self.compute_Pi_JinsDJ_given_D(CDR3_seq, Pi_J_given_D, max_J_align) #Include D genomic contribution (P(delDl, delDr | D)) to complete the contribution from the right (5') side. Return Pi_R Pi_R = self.compute_Pi_R(CDR3_seq, Pi_JinsDJ_given_D) pgen = 0 #zip Pi_L and Pi_R together to get total pgen for pos in range(len(CDR3_seq)*3 - 1): pgen += np.dot(Pi_L[:, pos], Pi_R[:, pos+1]) return pgen
[ "def", "compute_CDR3_pgen", "(", "self", ",", "CDR3_seq", ",", "V_usage_mask", ",", "J_usage_mask", ")", ":", "#Genomic V alignment/matching (contribution from P(V, delV)), return Pi_V", "Pi_V", ",", "max_V_align", "=", "self", ".", "compute_Pi_V", "(", "CDR3_seq", ",", "V_usage_mask", ")", "#Include VD insertions (Rvd and PinsVD) to get the total contribution from the left (3') side. Return Pi_L", "Pi_L", "=", "self", ".", "compute_Pi_L", "(", "CDR3_seq", ",", "Pi_V", ",", "max_V_align", ")", "#Genomic J alignment/matching (contribution from P(D, J, delJ)), return Pi_J_given_D", "Pi_J_given_D", ",", "max_J_align", "=", "self", ".", "compute_Pi_J_given_D", "(", "CDR3_seq", ",", "J_usage_mask", ")", "#Include DJ insertions (Rdj and PinsDJ), return Pi_JinsDJ_given_D", "Pi_JinsDJ_given_D", "=", "self", ".", "compute_Pi_JinsDJ_given_D", "(", "CDR3_seq", ",", "Pi_J_given_D", ",", "max_J_align", ")", "#Include D genomic contribution (P(delDl, delDr | D)) to complete the contribution from the right (5') side. Return Pi_R", "Pi_R", "=", "self", ".", "compute_Pi_R", "(", "CDR3_seq", ",", "Pi_JinsDJ_given_D", ")", "pgen", "=", "0", "#zip Pi_L and Pi_R together to get total pgen", "for", "pos", "in", "range", "(", "len", "(", "CDR3_seq", ")", "*", "3", "-", "1", ")", ":", "pgen", "+=", "np", ".", "dot", "(", "Pi_L", "[", ":", ",", "pos", "]", ",", "Pi_R", "[", ":", ",", "pos", "+", "1", "]", ")", "return", "pgen" ]
Compute Pgen for CDR3 'amino acid' sequence CDR3_seq from VDJ model. Conditioned on the already formatted V genes/alleles indicated in V_usage_mask and the J genes/alleles in J_usage_mask. (Examples are TCRB sequences/model) Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> compute_CDR3_pgen('CAWSVAPDRGGYTF', ppp, [42], [1]) 1.203646865765782e-10 >>> compute_CDR3_pgen(nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC'), ppp, [42], [1]) 2.3986503758867323e-12 >>> compute_CDR3_pgen('\xbb\x96\xab\xb8\x8e\xb6\xa5\x92\xa8\xba\x9a\x93\x94\x9f', ppp, [42], [1]) 2.3986503758867323e-12
[ "Compute", "Pgen", "for", "CDR3", "amino", "acid", "sequence", "CDR3_seq", "from", "VDJ", "model", ".", "Conditioned", "on", "the", "already", "formatted", "V", "genes", "/", "alleles", "indicated", "in", "V_usage_mask", "and", "the", "J", "genes", "/", "alleles", "in", "J_usage_mask", ".", "(", "Examples", "are", "TCRB", "sequences", "/", "model", ")", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "V_usage_mask", ":", "list", "Indices", "of", "the", "V", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "J_usage_mask", ":", "list", "Indices", "of", "the", "J", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "Returns", "-------", "pgen", ":", "float", "The", "generation", "probability", "(", "Pgen", ")", "of", "the", "sequence", "Examples", "--------", ">>>", "compute_CDR3_pgen", "(", "CAWSVAPDRGGYTF", "ppp", "[", "42", "]", "[", "1", "]", ")", "1", ".", "203646865765782e", "-", "10", ">>>", "compute_CDR3_pgen", "(", "nt2codon_rep", "(", "TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC", ")", "ppp", "[", "42", "]", "[", "1", "]", ")", "2", ".", "3986503758867323e", "-", "12", ">>>", "compute_CDR3_pgen", "(", "\\", "xbb", "\\", "x96", "\\", "xab", "\\", "xb8", "\\", "x8e", "\\", "xb6", "\\", "xa5", "\\", "x92", "\\", "xa8", "\\", "xba", "\\", "x9a", "\\", "x93", "\\", "x94", "\\", "x9f", "ppp", "[", "42", "]", "[", "1", "]", ")", "2", ".", "3986503758867323e", "-", "12" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L849-L904
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVDJ.compute_Pi_V
def compute_Pi_V(self, CDR3_seq, V_usage_mask): """Compute Pi_V. This function returns the Pi array from the model factors of the V genomic contributions, P(V)*P(delV|V). This corresponds to V_{x_1}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation self.cutV_genomic_CDR3_segs : list of strings List of all the V genomic nucleotide sequences trimmed to begin at the conserved C residue and with the maximum number of palindromic insertions appended. self.PVdelV_nt_pos_vec : list of ndarrays For each V allele, format P(V)*P(delV|V) into the correct form for a Pi array or V_{x_1}. This is only done for the first and last position in each codon. self.PVdelV_2nd_nt_pos_per_aa_vec : list of dicts For each V allele, and each 'amino acid', format P(V)*P(delV|V) for positions in the middle of a codon into the correct form for a Pi array or V_{x_1} given the 'amino acid'. Returns ------- Pi_V : ndarray (4, 3L) array corresponding to V_{x_1}. max_V_align: int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. """ #Note, the cutV_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at the conserved C Pi_V = np.zeros((4, len(CDR3_seq)*3)) #Holds the aggregate weight for each nt possiblity and position alignment_lengths = [] for V_in in V_usage_mask: try: cutV_gen_seg = self.cutV_genomic_CDR3_segs[V_in] except IndexError: print 'Check provided V usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_left(CDR3_seq, cutV_gen_seg) alignment_lengths += [current_alignment_length] current_Pi_V = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PVdelV_nt_pos_vec current_Pi_V[:, :current_alignment_length] = self.PVdelV_nt_pos_vec[V_in][:, :current_alignment_length] for pos in range(1, current_alignment_length, 3): #for middle nt use PVdelV_2nd_nt_pos_per_aa_vec current_Pi_V[:, pos] = self.PVdelV_2nd_nt_pos_per_aa_vec[V_in][CDR3_seq[pos/3]][:, pos] Pi_V[:, :current_alignment_length] += current_Pi_V[:, :current_alignment_length] return Pi_V, max(alignment_lengths)
python
def compute_Pi_V(self, CDR3_seq, V_usage_mask): """Compute Pi_V. This function returns the Pi array from the model factors of the V genomic contributions, P(V)*P(delV|V). This corresponds to V_{x_1}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation self.cutV_genomic_CDR3_segs : list of strings List of all the V genomic nucleotide sequences trimmed to begin at the conserved C residue and with the maximum number of palindromic insertions appended. self.PVdelV_nt_pos_vec : list of ndarrays For each V allele, format P(V)*P(delV|V) into the correct form for a Pi array or V_{x_1}. This is only done for the first and last position in each codon. self.PVdelV_2nd_nt_pos_per_aa_vec : list of dicts For each V allele, and each 'amino acid', format P(V)*P(delV|V) for positions in the middle of a codon into the correct form for a Pi array or V_{x_1} given the 'amino acid'. Returns ------- Pi_V : ndarray (4, 3L) array corresponding to V_{x_1}. max_V_align: int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. """ #Note, the cutV_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at the conserved C Pi_V = np.zeros((4, len(CDR3_seq)*3)) #Holds the aggregate weight for each nt possiblity and position alignment_lengths = [] for V_in in V_usage_mask: try: cutV_gen_seg = self.cutV_genomic_CDR3_segs[V_in] except IndexError: print 'Check provided V usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_left(CDR3_seq, cutV_gen_seg) alignment_lengths += [current_alignment_length] current_Pi_V = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PVdelV_nt_pos_vec current_Pi_V[:, :current_alignment_length] = self.PVdelV_nt_pos_vec[V_in][:, :current_alignment_length] for pos in range(1, current_alignment_length, 3): #for middle nt use PVdelV_2nd_nt_pos_per_aa_vec current_Pi_V[:, pos] = self.PVdelV_2nd_nt_pos_per_aa_vec[V_in][CDR3_seq[pos/3]][:, pos] Pi_V[:, :current_alignment_length] += current_Pi_V[:, :current_alignment_length] return Pi_V, max(alignment_lengths)
[ "def", "compute_Pi_V", "(", "self", ",", "CDR3_seq", ",", "V_usage_mask", ")", ":", "#Note, the cutV_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template.", "#furthermore, the genomic sequence should be pruned to start at the conserved C", "Pi_V", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "#Holds the aggregate weight for each nt possiblity and position", "alignment_lengths", "=", "[", "]", "for", "V_in", "in", "V_usage_mask", ":", "try", ":", "cutV_gen_seg", "=", "self", ".", "cutV_genomic_CDR3_segs", "[", "V_in", "]", "except", "IndexError", ":", "print", "'Check provided V usage mask. Contains indicies out of allowed range.'", "continue", "current_alignment_length", "=", "self", ".", "max_nt_to_aa_alignment_left", "(", "CDR3_seq", ",", "cutV_gen_seg", ")", "alignment_lengths", "+=", "[", "current_alignment_length", "]", "current_Pi_V", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "if", "current_alignment_length", ">", "0", ":", "#For first and last nt in a codon use PVdelV_nt_pos_vec", "current_Pi_V", "[", ":", ",", ":", "current_alignment_length", "]", "=", "self", ".", "PVdelV_nt_pos_vec", "[", "V_in", "]", "[", ":", ",", ":", "current_alignment_length", "]", "for", "pos", "in", "range", "(", "1", ",", "current_alignment_length", ",", "3", ")", ":", "#for middle nt use PVdelV_2nd_nt_pos_per_aa_vec", "current_Pi_V", "[", ":", ",", "pos", "]", "=", "self", ".", "PVdelV_2nd_nt_pos_per_aa_vec", "[", "V_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "pos", "]", "Pi_V", "[", ":", ",", ":", "current_alignment_length", "]", "+=", "current_Pi_V", "[", ":", ",", ":", "current_alignment_length", "]", "return", "Pi_V", ",", "max", "(", "alignment_lengths", ")" ]
Compute Pi_V. This function returns the Pi array from the model factors of the V genomic contributions, P(V)*P(delV|V). This corresponds to V_{x_1}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation self.cutV_genomic_CDR3_segs : list of strings List of all the V genomic nucleotide sequences trimmed to begin at the conserved C residue and with the maximum number of palindromic insertions appended. self.PVdelV_nt_pos_vec : list of ndarrays For each V allele, format P(V)*P(delV|V) into the correct form for a Pi array or V_{x_1}. This is only done for the first and last position in each codon. self.PVdelV_2nd_nt_pos_per_aa_vec : list of dicts For each V allele, and each 'amino acid', format P(V)*P(delV|V) for positions in the middle of a codon into the correct form for a Pi array or V_{x_1} given the 'amino acid'. Returns ------- Pi_V : ndarray (4, 3L) array corresponding to V_{x_1}. max_V_align: int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask.
[ "Compute", "Pi_V", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "V", "genomic", "contributions", "P", "(", "V", ")", "*", "P", "(", "delV|V", ")", ".", "This", "corresponds", "to", "V_", "{", "x_1", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "V_usage_mask", ":", "list", "Indices", "of", "the", "V", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "self", ".", "cutV_genomic_CDR3_segs", ":", "list", "of", "strings", "List", "of", "all", "the", "V", "genomic", "nucleotide", "sequences", "trimmed", "to", "begin", "at", "the", "conserved", "C", "residue", "and", "with", "the", "maximum", "number", "of", "palindromic", "insertions", "appended", ".", "self", ".", "PVdelV_nt_pos_vec", ":", "list", "of", "ndarrays", "For", "each", "V", "allele", "format", "P", "(", "V", ")", "*", "P", "(", "delV|V", ")", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "V_", "{", "x_1", "}", ".", "This", "is", "only", "done", "for", "the", "first", "and", "last", "position", "in", "each", "codon", ".", "self", ".", "PVdelV_2nd_nt_pos_per_aa_vec", ":", "list", "of", "dicts", "For", "each", "V", "allele", "and", "each", "amino", "acid", "format", "P", "(", "V", ")", "*", "P", "(", "delV|V", ")", "for", "positions", "in", "the", "middle", "of", "a", "codon", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "V_", "{", "x_1", "}", "given", "the", "amino", "acid", ".", "Returns", "-------", "Pi_V", ":", "ndarray", "(", "4", "3L", ")", "array", "corresponding", "to", "V_", "{", "x_1", "}", ".", "max_V_align", ":", "int", "Maximum", "alignment", "of", "the", "CDR3_seq", "to", "any", "genomic", "V", "allele", "allowed", "by", "V_usage_mask", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L907-L968
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVDJ.compute_Pi_L
def compute_Pi_L(self, CDR3_seq, Pi_V, max_V_align): """Compute Pi_L. This function returns the Pi array from the model factors of the V genomic contributions, P(V)*P(delV|V), and the VD (N1) insertions, first_nt_bias_insVD(m_1)PinsVD(\ell_{VD})\prod_{i=2}^{\ell_{VD}}Rvd(m_i|m_{i-1}). This corresponds to V_{x_1}{M^{x_1}}_{x_2}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_V : ndarray (4, 3L) array corresponding to V_{x_1}. max_V_align : int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. self.PinsVD : ndarray Probability distribution of the VD (N1) insertion sequence length self.first_nt_bias_insVD : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the VD junction. self.zero_nt_bias_insVD : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the VD insertion. zero_nt_bias_insVD = Rvd^{-1}first_nt_bias_insVD self.Tvd : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Svd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the first position. self.Dvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.lTvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion starting in the first position. self.lDvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for VD insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_L : ndarray (4, 3L) array corresponding to V_{x_1}{M^{x_1}}_{x_2}. """ #max_insertions = 30 #len(PinsVD) - 1 should zeropad the last few spots max_insertions = len(self.PinsVD) - 1 Pi_L = np.zeros((4, len(CDR3_seq)*3)) #start position is first nt in a codon for init_pos in range(0, max_V_align, 3): #Zero insertions Pi_L[:, init_pos] += self.PinsVD[0]*Pi_V[:, init_pos] #One insertion Pi_L[:, init_pos+1] += self.PinsVD[1]*np.dot(self.lDvd[CDR3_seq[init_pos/3]], Pi_V[:, init_pos]) #Two insertions and compute the base nt vec for the standard loop current_base_nt_vec = np.dot(self.lTvd[CDR3_seq[init_pos/3]], Pi_V[:, init_pos]) Pi_L[0, init_pos+2] += self.PinsVD[2]*np.sum(current_base_nt_vec) base_ins = 2 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_L[:, init_pos+base_ins+1] += self.PinsVD[base_ins + 1]*np.dot(self.Svd[aa], current_base_nt_vec) Pi_L[:, init_pos+base_ins+2] += self.PinsVD[base_ins + 2]*np.dot(self.Dvd[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvd[aa], current_base_nt_vec) Pi_L[0, init_pos+base_ins+3] += self.PinsVD[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is second nt in a codon for init_pos in range(1, max_V_align, 3): #Zero insertions Pi_L[:, init_pos] += self.PinsVD[0]*Pi_V[:, init_pos] #One insertion --- we first compute our p vec by pairwise mult with the ss distr current_base_nt_vec = np.multiply(Pi_V[:, init_pos], self.first_nt_bias_insVD) Pi_L[0, init_pos+1] += self.PinsVD[1]*np.sum(current_base_nt_vec) base_ins = 1 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_L[:, init_pos+base_ins+1] += self.PinsVD[base_ins + 1]*np.dot(self.Svd[aa], current_base_nt_vec) Pi_L[:, init_pos+base_ins+2] += self.PinsVD[base_ins + 2]*np.dot(self.Dvd[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvd[aa], current_base_nt_vec) Pi_L[0, init_pos+base_ins+3] += self.PinsVD[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is last nt in a codon for init_pos in range(2, max_V_align, 3): #Zero insertions Pi_L[0, init_pos] += self.PinsVD[0]*Pi_V[0, init_pos] #current_base_nt_vec = first_nt_bias_insVD*Pi_V[0, init_pos] #Okay for steady state current_base_nt_vec = self.zero_nt_bias_insVD*Pi_V[0, init_pos] base_ins = 0 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_L[:, init_pos+base_ins+1] += self.PinsVD[base_ins + 1]*np.dot(self.Svd[aa], current_base_nt_vec) Pi_L[:, init_pos+base_ins+2] += self.PinsVD[base_ins + 2]*np.dot(self.Dvd[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvd[aa], current_base_nt_vec) Pi_L[0, init_pos+base_ins+3] += self.PinsVD[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 return Pi_L
python
def compute_Pi_L(self, CDR3_seq, Pi_V, max_V_align): """Compute Pi_L. This function returns the Pi array from the model factors of the V genomic contributions, P(V)*P(delV|V), and the VD (N1) insertions, first_nt_bias_insVD(m_1)PinsVD(\ell_{VD})\prod_{i=2}^{\ell_{VD}}Rvd(m_i|m_{i-1}). This corresponds to V_{x_1}{M^{x_1}}_{x_2}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_V : ndarray (4, 3L) array corresponding to V_{x_1}. max_V_align : int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. self.PinsVD : ndarray Probability distribution of the VD (N1) insertion sequence length self.first_nt_bias_insVD : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the VD junction. self.zero_nt_bias_insVD : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the VD insertion. zero_nt_bias_insVD = Rvd^{-1}first_nt_bias_insVD self.Tvd : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Svd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the first position. self.Dvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.lTvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion starting in the first position. self.lDvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for VD insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_L : ndarray (4, 3L) array corresponding to V_{x_1}{M^{x_1}}_{x_2}. """ #max_insertions = 30 #len(PinsVD) - 1 should zeropad the last few spots max_insertions = len(self.PinsVD) - 1 Pi_L = np.zeros((4, len(CDR3_seq)*3)) #start position is first nt in a codon for init_pos in range(0, max_V_align, 3): #Zero insertions Pi_L[:, init_pos] += self.PinsVD[0]*Pi_V[:, init_pos] #One insertion Pi_L[:, init_pos+1] += self.PinsVD[1]*np.dot(self.lDvd[CDR3_seq[init_pos/3]], Pi_V[:, init_pos]) #Two insertions and compute the base nt vec for the standard loop current_base_nt_vec = np.dot(self.lTvd[CDR3_seq[init_pos/3]], Pi_V[:, init_pos]) Pi_L[0, init_pos+2] += self.PinsVD[2]*np.sum(current_base_nt_vec) base_ins = 2 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_L[:, init_pos+base_ins+1] += self.PinsVD[base_ins + 1]*np.dot(self.Svd[aa], current_base_nt_vec) Pi_L[:, init_pos+base_ins+2] += self.PinsVD[base_ins + 2]*np.dot(self.Dvd[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvd[aa], current_base_nt_vec) Pi_L[0, init_pos+base_ins+3] += self.PinsVD[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is second nt in a codon for init_pos in range(1, max_V_align, 3): #Zero insertions Pi_L[:, init_pos] += self.PinsVD[0]*Pi_V[:, init_pos] #One insertion --- we first compute our p vec by pairwise mult with the ss distr current_base_nt_vec = np.multiply(Pi_V[:, init_pos], self.first_nt_bias_insVD) Pi_L[0, init_pos+1] += self.PinsVD[1]*np.sum(current_base_nt_vec) base_ins = 1 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_L[:, init_pos+base_ins+1] += self.PinsVD[base_ins + 1]*np.dot(self.Svd[aa], current_base_nt_vec) Pi_L[:, init_pos+base_ins+2] += self.PinsVD[base_ins + 2]*np.dot(self.Dvd[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvd[aa], current_base_nt_vec) Pi_L[0, init_pos+base_ins+3] += self.PinsVD[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is last nt in a codon for init_pos in range(2, max_V_align, 3): #Zero insertions Pi_L[0, init_pos] += self.PinsVD[0]*Pi_V[0, init_pos] #current_base_nt_vec = first_nt_bias_insVD*Pi_V[0, init_pos] #Okay for steady state current_base_nt_vec = self.zero_nt_bias_insVD*Pi_V[0, init_pos] base_ins = 0 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_L[:, init_pos+base_ins+1] += self.PinsVD[base_ins + 1]*np.dot(self.Svd[aa], current_base_nt_vec) Pi_L[:, init_pos+base_ins+2] += self.PinsVD[base_ins + 2]*np.dot(self.Dvd[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvd[aa], current_base_nt_vec) Pi_L[0, init_pos+base_ins+3] += self.PinsVD[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 return Pi_L
[ "def", "compute_Pi_L", "(", "self", ",", "CDR3_seq", ",", "Pi_V", ",", "max_V_align", ")", ":", "#max_insertions = 30 #len(PinsVD) - 1 should zeropad the last few spots", "max_insertions", "=", "len", "(", "self", ".", "PinsVD", ")", "-", "1", "Pi_L", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "#start position is first nt in a codon", "for", "init_pos", "in", "range", "(", "0", ",", "max_V_align", ",", "3", ")", ":", "#Zero insertions", "Pi_L", "[", ":", ",", "init_pos", "]", "+=", "self", ".", "PinsVD", "[", "0", "]", "*", "Pi_V", "[", ":", ",", "init_pos", "]", "#One insertion", "Pi_L", "[", ":", ",", "init_pos", "+", "1", "]", "+=", "self", ".", "PinsVD", "[", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "lDvd", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ",", "Pi_V", "[", ":", ",", "init_pos", "]", ")", "#Two insertions and compute the base nt vec for the standard loop ", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "lTvd", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ",", "Pi_V", "[", ":", ",", "init_pos", "]", ")", "Pi_L", "[", "0", ",", "init_pos", "+", "2", "]", "+=", "self", ".", "PinsVD", "[", "2", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "=", "2", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "+", "1", ":", "init_pos", "/", "3", "+", "max_insertions", "/", "3", "]", ":", "Pi_L", "[", ":", ",", "init_pos", "+", "base_ins", "+", "1", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Svd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_L", "[", ":", ",", "init_pos", "+", "base_ins", "+", "2", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Dvd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tvd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_L", "[", "0", ",", "init_pos", "+", "base_ins", "+", "3", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "#start position is second nt in a codon", "for", "init_pos", "in", "range", "(", "1", ",", "max_V_align", ",", "3", ")", ":", "#Zero insertions", "Pi_L", "[", ":", ",", "init_pos", "]", "+=", "self", ".", "PinsVD", "[", "0", "]", "*", "Pi_V", "[", ":", ",", "init_pos", "]", "#One insertion --- we first compute our p vec by pairwise mult with the ss distr", "current_base_nt_vec", "=", "np", ".", "multiply", "(", "Pi_V", "[", ":", ",", "init_pos", "]", ",", "self", ".", "first_nt_bias_insVD", ")", "Pi_L", "[", "0", ",", "init_pos", "+", "1", "]", "+=", "self", ".", "PinsVD", "[", "1", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "=", "1", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "+", "1", ":", "init_pos", "/", "3", "+", "max_insertions", "/", "3", "]", ":", "Pi_L", "[", ":", ",", "init_pos", "+", "base_ins", "+", "1", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Svd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_L", "[", ":", ",", "init_pos", "+", "base_ins", "+", "2", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Dvd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tvd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_L", "[", "0", ",", "init_pos", "+", "base_ins", "+", "3", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "#start position is last nt in a codon ", "for", "init_pos", "in", "range", "(", "2", ",", "max_V_align", ",", "3", ")", ":", "#Zero insertions", "Pi_L", "[", "0", ",", "init_pos", "]", "+=", "self", ".", "PinsVD", "[", "0", "]", "*", "Pi_V", "[", "0", ",", "init_pos", "]", "#current_base_nt_vec = first_nt_bias_insVD*Pi_V[0, init_pos] #Okay for steady state", "current_base_nt_vec", "=", "self", ".", "zero_nt_bias_insVD", "*", "Pi_V", "[", "0", ",", "init_pos", "]", "base_ins", "=", "0", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "+", "1", ":", "init_pos", "/", "3", "+", "max_insertions", "/", "3", "]", ":", "Pi_L", "[", ":", ",", "init_pos", "+", "base_ins", "+", "1", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Svd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_L", "[", ":", ",", "init_pos", "+", "base_ins", "+", "2", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Dvd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tvd", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_L", "[", "0", ",", "init_pos", "+", "base_ins", "+", "3", "]", "+=", "self", ".", "PinsVD", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "return", "Pi_L" ]
Compute Pi_L. This function returns the Pi array from the model factors of the V genomic contributions, P(V)*P(delV|V), and the VD (N1) insertions, first_nt_bias_insVD(m_1)PinsVD(\ell_{VD})\prod_{i=2}^{\ell_{VD}}Rvd(m_i|m_{i-1}). This corresponds to V_{x_1}{M^{x_1}}_{x_2}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_V : ndarray (4, 3L) array corresponding to V_{x_1}. max_V_align : int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. self.PinsVD : ndarray Probability distribution of the VD (N1) insertion sequence length self.first_nt_bias_insVD : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the VD junction. self.zero_nt_bias_insVD : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the VD insertion. zero_nt_bias_insVD = Rvd^{-1}first_nt_bias_insVD self.Tvd : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Svd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the first position. self.Dvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.lTvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion starting in the first position. self.lDvd : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for VD insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_L : ndarray (4, 3L) array corresponding to V_{x_1}{M^{x_1}}_{x_2}.
[ "Compute", "Pi_L", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "V", "genomic", "contributions", "P", "(", "V", ")", "*", "P", "(", "delV|V", ")", "and", "the", "VD", "(", "N1", ")", "insertions", "first_nt_bias_insVD", "(", "m_1", ")", "PinsVD", "(", "\\", "ell_", "{", "VD", "}", ")", "\\", "prod_", "{", "i", "=", "2", "}", "^", "{", "\\", "ell_", "{", "VD", "}}", "Rvd", "(", "m_i|m_", "{", "i", "-", "1", "}", ")", ".", "This", "corresponds", "to", "V_", "{", "x_1", "}", "{", "M^", "{", "x_1", "}}", "_", "{", "x_2", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "Pi_V", ":", "ndarray", "(", "4", "3L", ")", "array", "corresponding", "to", "V_", "{", "x_1", "}", ".", "max_V_align", ":", "int", "Maximum", "alignment", "of", "the", "CDR3_seq", "to", "any", "genomic", "V", "allele", "allowed", "by", "V_usage_mask", ".", "self", ".", "PinsVD", ":", "ndarray", "Probability", "distribution", "of", "the", "VD", "(", "N1", ")", "insertion", "sequence", "length", "self", ".", "first_nt_bias_insVD", ":", "ndarray", "(", "4", ")", "array", "of", "the", "probability", "distribution", "of", "the", "indentity", "of", "the", "first", "nucleotide", "insertion", "for", "the", "VD", "junction", ".", "self", ".", "zero_nt_bias_insVD", ":", "ndarray", "(", "4", ")", "array", "of", "the", "probability", "distribution", "of", "the", "indentity", "of", "the", "the", "nucleotide", "BEFORE", "the", "VD", "insertion", ".", "zero_nt_bias_insVD", "=", "Rvd^", "{", "-", "1", "}", "first_nt_bias_insVD", "self", ".", "Tvd", ":", "dict", "Dictionary", "of", "full", "codon", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", ".", "self", ".", "Svd", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "ending", "in", "the", "first", "position", ".", "self", ".", "Dvd", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "ending", "in", "the", "second", "position", ".", "self", ".", "lTvd", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "starting", "in", "the", "first", "position", ".", "self", ".", "lDvd", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "VD", "insertion", "starting", "in", "the", "first", "position", "and", "ending", "in", "the", "second", "position", "of", "the", "same", "codon", ".", "Returns", "-------", "Pi_L", ":", "ndarray", "(", "4", "3L", ")", "array", "corresponding", "to", "V_", "{", "x_1", "}", "{", "M^", "{", "x_1", "}}", "_", "{", "x_2", "}", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L971-L1088
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVDJ.compute_Pi_J_given_D
def compute_Pi_J_given_D(self, CDR3_seq, J_usage_mask): """Compute Pi_J conditioned on D. This function returns the Pi array from the model factors of the D and J genomic contributions, P(D, J)*P(delJ|J) = P(D|J)P(J)P(delJ|J). This corresponds to J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation. self.cutJ_genomic_CDR3_segs : list List of all the J genomic nucleotide sequences trimmed to begin at the conserved 3' residue (F/W) and with the maximum number of palindromic insertions appended. self.PD_given_J : ndarray Probability distribution of D conditioned on J, i.e. P(D|J). self.PJdelJ_nt_pos_vec : list of ndarrays For each J allele, format P(J)*P(delJ|J) into the correct form for a Pi array or J(D)^{x_4}. This is only done for the first and last position in each codon. self.PJdelJ_2nd_nt_pos_per_aa_vec : list of dicts For each J allele, and each 'amino acid', format P(J)*P(delJ|J) for positions in the middle of a codon into the correct form for a Pi array or J(D)^{x_4} given the 'amino acid'. Returns ------- Pi_J_given_D : list List of (4, 3L) ndarrays corresponding to J(D)^{x_4}. max_J_align: int Maximum alignment of the CDR3_seq to any genomic J allele allowed by J_usage_mask. """ #Note, the cutJ_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at a conserved region on the J side num_D_genes = self.PD_given_J.shape[0] Pi_J_given_D = [np.zeros((4, len(CDR3_seq)*3)) for i in range(num_D_genes)] #Holds the aggregate weight for each nt possiblity and position alignment_lengths = [] for J_in in J_usage_mask: try: cutJ_gen_seg = self.cutJ_genomic_CDR3_segs[J_in] except IndexError: print 'Check provided V usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_right(CDR3_seq, cutJ_gen_seg) alignment_lengths += [current_alignment_length] current_Pi_J = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PJdelJ_nt_pos_vec current_Pi_J[:, -current_alignment_length:] = self.PJdelJ_nt_pos_vec[J_in][:, -current_alignment_length:] for pos in range(-2, -current_alignment_length-1, -3): #for middle nt use PJdelJ_2nd_nt_pos_per_aa_vec current_Pi_J[:, pos] = self.PJdelJ_2nd_nt_pos_per_aa_vec[J_in][CDR3_seq[pos/3]][:, pos] for D_in, pd_given_j in enumerate(self.PD_given_J[:, J_in]): Pi_J_given_D[D_in][:, -current_alignment_length:] += pd_given_j*current_Pi_J[:, -current_alignment_length:] return Pi_J_given_D, max(alignment_lengths)
python
def compute_Pi_J_given_D(self, CDR3_seq, J_usage_mask): """Compute Pi_J conditioned on D. This function returns the Pi array from the model factors of the D and J genomic contributions, P(D, J)*P(delJ|J) = P(D|J)P(J)P(delJ|J). This corresponds to J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation. self.cutJ_genomic_CDR3_segs : list List of all the J genomic nucleotide sequences trimmed to begin at the conserved 3' residue (F/W) and with the maximum number of palindromic insertions appended. self.PD_given_J : ndarray Probability distribution of D conditioned on J, i.e. P(D|J). self.PJdelJ_nt_pos_vec : list of ndarrays For each J allele, format P(J)*P(delJ|J) into the correct form for a Pi array or J(D)^{x_4}. This is only done for the first and last position in each codon. self.PJdelJ_2nd_nt_pos_per_aa_vec : list of dicts For each J allele, and each 'amino acid', format P(J)*P(delJ|J) for positions in the middle of a codon into the correct form for a Pi array or J(D)^{x_4} given the 'amino acid'. Returns ------- Pi_J_given_D : list List of (4, 3L) ndarrays corresponding to J(D)^{x_4}. max_J_align: int Maximum alignment of the CDR3_seq to any genomic J allele allowed by J_usage_mask. """ #Note, the cutJ_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at a conserved region on the J side num_D_genes = self.PD_given_J.shape[0] Pi_J_given_D = [np.zeros((4, len(CDR3_seq)*3)) for i in range(num_D_genes)] #Holds the aggregate weight for each nt possiblity and position alignment_lengths = [] for J_in in J_usage_mask: try: cutJ_gen_seg = self.cutJ_genomic_CDR3_segs[J_in] except IndexError: print 'Check provided V usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_right(CDR3_seq, cutJ_gen_seg) alignment_lengths += [current_alignment_length] current_Pi_J = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PJdelJ_nt_pos_vec current_Pi_J[:, -current_alignment_length:] = self.PJdelJ_nt_pos_vec[J_in][:, -current_alignment_length:] for pos in range(-2, -current_alignment_length-1, -3): #for middle nt use PJdelJ_2nd_nt_pos_per_aa_vec current_Pi_J[:, pos] = self.PJdelJ_2nd_nt_pos_per_aa_vec[J_in][CDR3_seq[pos/3]][:, pos] for D_in, pd_given_j in enumerate(self.PD_given_J[:, J_in]): Pi_J_given_D[D_in][:, -current_alignment_length:] += pd_given_j*current_Pi_J[:, -current_alignment_length:] return Pi_J_given_D, max(alignment_lengths)
[ "def", "compute_Pi_J_given_D", "(", "self", ",", "CDR3_seq", ",", "J_usage_mask", ")", ":", "#Note, the cutJ_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template.", "#furthermore, the genomic sequence should be pruned to start at a conserved region on the J side", "num_D_genes", "=", "self", ".", "PD_given_J", ".", "shape", "[", "0", "]", "Pi_J_given_D", "=", "[", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "for", "i", "in", "range", "(", "num_D_genes", ")", "]", "#Holds the aggregate weight for each nt possiblity and position", "alignment_lengths", "=", "[", "]", "for", "J_in", "in", "J_usage_mask", ":", "try", ":", "cutJ_gen_seg", "=", "self", ".", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", "except", "IndexError", ":", "print", "'Check provided V usage mask. Contains indicies out of allowed range.'", "continue", "current_alignment_length", "=", "self", ".", "max_nt_to_aa_alignment_right", "(", "CDR3_seq", ",", "cutJ_gen_seg", ")", "alignment_lengths", "+=", "[", "current_alignment_length", "]", "current_Pi_J", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "if", "current_alignment_length", ">", "0", ":", "#For first and last nt in a codon use PJdelJ_nt_pos_vec", "current_Pi_J", "[", ":", ",", "-", "current_alignment_length", ":", "]", "=", "self", ".", "PJdelJ_nt_pos_vec", "[", "J_in", "]", "[", ":", ",", "-", "current_alignment_length", ":", "]", "for", "pos", "in", "range", "(", "-", "2", ",", "-", "current_alignment_length", "-", "1", ",", "-", "3", ")", ":", "#for middle nt use PJdelJ_2nd_nt_pos_per_aa_vec", "current_Pi_J", "[", ":", ",", "pos", "]", "=", "self", ".", "PJdelJ_2nd_nt_pos_per_aa_vec", "[", "J_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "pos", "]", "for", "D_in", ",", "pd_given_j", "in", "enumerate", "(", "self", ".", "PD_given_J", "[", ":", ",", "J_in", "]", ")", ":", "Pi_J_given_D", "[", "D_in", "]", "[", ":", ",", "-", "current_alignment_length", ":", "]", "+=", "pd_given_j", "*", "current_Pi_J", "[", ":", ",", "-", "current_alignment_length", ":", "]", "return", "Pi_J_given_D", ",", "max", "(", "alignment_lengths", ")" ]
Compute Pi_J conditioned on D. This function returns the Pi array from the model factors of the D and J genomic contributions, P(D, J)*P(delJ|J) = P(D|J)P(J)P(delJ|J). This corresponds to J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation. self.cutJ_genomic_CDR3_segs : list List of all the J genomic nucleotide sequences trimmed to begin at the conserved 3' residue (F/W) and with the maximum number of palindromic insertions appended. self.PD_given_J : ndarray Probability distribution of D conditioned on J, i.e. P(D|J). self.PJdelJ_nt_pos_vec : list of ndarrays For each J allele, format P(J)*P(delJ|J) into the correct form for a Pi array or J(D)^{x_4}. This is only done for the first and last position in each codon. self.PJdelJ_2nd_nt_pos_per_aa_vec : list of dicts For each J allele, and each 'amino acid', format P(J)*P(delJ|J) for positions in the middle of a codon into the correct form for a Pi array or J(D)^{x_4} given the 'amino acid'. Returns ------- Pi_J_given_D : list List of (4, 3L) ndarrays corresponding to J(D)^{x_4}. max_J_align: int Maximum alignment of the CDR3_seq to any genomic J allele allowed by J_usage_mask.
[ "Compute", "Pi_J", "conditioned", "on", "D", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "D", "and", "J", "genomic", "contributions", "P", "(", "D", "J", ")", "*", "P", "(", "delJ|J", ")", "=", "P", "(", "D|J", ")", "P", "(", "J", ")", "P", "(", "delJ|J", ")", ".", "This", "corresponds", "to", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "J_usage_mask", ":", "list", "Indices", "of", "the", "J", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", ".", "self", ".", "cutJ_genomic_CDR3_segs", ":", "list", "List", "of", "all", "the", "J", "genomic", "nucleotide", "sequences", "trimmed", "to", "begin", "at", "the", "conserved", "3", "residue", "(", "F", "/", "W", ")", "and", "with", "the", "maximum", "number", "of", "palindromic", "insertions", "appended", ".", "self", ".", "PD_given_J", ":", "ndarray", "Probability", "distribution", "of", "D", "conditioned", "on", "J", "i", ".", "e", ".", "P", "(", "D|J", ")", ".", "self", ".", "PJdelJ_nt_pos_vec", ":", "list", "of", "ndarrays", "For", "each", "J", "allele", "format", "P", "(", "J", ")", "*", "P", "(", "delJ|J", ")", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "This", "is", "only", "done", "for", "the", "first", "and", "last", "position", "in", "each", "codon", ".", "self", ".", "PJdelJ_2nd_nt_pos_per_aa_vec", ":", "list", "of", "dicts", "For", "each", "J", "allele", "and", "each", "amino", "acid", "format", "P", "(", "J", ")", "*", "P", "(", "delJ|J", ")", "for", "positions", "in", "the", "middle", "of", "a", "codon", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "J", "(", "D", ")", "^", "{", "x_4", "}", "given", "the", "amino", "acid", ".", "Returns", "-------", "Pi_J_given_D", ":", "list", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "max_J_align", ":", "int", "Maximum", "alignment", "of", "the", "CDR3_seq", "to", "any", "genomic", "J", "allele", "allowed", "by", "J_usage_mask", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1091-L1159
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVDJ.compute_Pi_JinsDJ_given_D
def compute_Pi_JinsDJ_given_D(self, CDR3_seq, Pi_J_given_D, max_J_align): """Compute Pi_JinsDJ conditioned on D. This function returns the Pi array from the model factors of the J genomic contributions, P(D,J)*P(delJ|J), and the DJ (N2) insertions, first_nt_bias_insDJ(n_1)PinsDJ(\ell_{DJ})\prod_{i=2}^{\ell_{DJ}}Rdj(n_i|n_{i-1}) conditioned on D identity. This corresponds to {N^{x_3}}_{x_4}J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_J_given_D : ndarray List of (4, 3L) ndarrays corresponding to J(D)^{x_4}. max_J_align : int Maximum alignment of the CDR3_seq to any genomic J allele allowed by J_usage_mask. self.PinsDJ : ndarray Probability distribution of the DJ (N2) insertion sequence length self.first_nt_bias_insDJ : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the DJ junction. self.zero_nt_bias_insDJ : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the DJ insertion. Note, as the Markov model at the DJ junction goes 3' to 5' this is the position AFTER the insertions reading left to right. self.Tdj : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Sdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the DJ insertion ending in the first position. self.Ddj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.rTdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the DJ insertion starting in the first position. self.rDdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for DJ insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_JinsDJ_given_D : list List of (4, 3L) ndarrays corresponding to {N^{x_3}}_{x_4}J(D)^{x_4}. """ #max_insertions = 30 #len(PinsVD) - 1 should zeropad the last few spots max_insertions = len(self.PinsDJ) - 1 Pi_JinsDJ_given_D = [np.zeros((4, len(CDR3_seq)*3)) for i in range(len(Pi_J_given_D))] for D_in in range(len(Pi_J_given_D)): #start position is first nt in a codon for init_pos in range(-1, -(max_J_align+1), -3): #Zero insertions Pi_JinsDJ_given_D[D_in][:, init_pos] += self.PinsDJ[0]*Pi_J_given_D[D_in][:, init_pos] #One insertion Pi_JinsDJ_given_D[D_in][:, init_pos-1] += self.PinsDJ[1]*np.dot(self.rDdj[CDR3_seq[init_pos/3]], Pi_J_given_D[D_in][:, init_pos]) #Two insertions and compute the base nt vec for the standard loop current_base_nt_vec = np.dot(self.rTdj[CDR3_seq[init_pos/3]], Pi_J_given_D[D_in][:, init_pos]) Pi_JinsDJ_given_D[D_in][0, init_pos-2] += self.PinsDJ[2]*np.sum(current_base_nt_vec) base_ins = 2 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 - 1: init_pos/3 - max_insertions/3:-1]: Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-1] += self.PinsDJ[base_ins + 1]*np.dot(self.Sdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-2] += self.PinsDJ[base_ins + 2]*np.dot(self.Ddj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][0, init_pos-base_ins-3] += self.PinsDJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is second nt in a codon for init_pos in range(-2, -(max_J_align+1), -3): #Zero insertions Pi_JinsDJ_given_D[D_in][:, init_pos] += self.PinsDJ[0]*Pi_J_given_D[D_in][:, init_pos] #One insertion --- we first compute our p vec by pairwise mult with the ss distr current_base_nt_vec = np.multiply(Pi_J_given_D[D_in][:, init_pos], self.first_nt_bias_insDJ) Pi_JinsDJ_given_D[D_in][0, init_pos-1] += self.PinsDJ[1]*np.sum(current_base_nt_vec) base_ins = 1 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 - 1: init_pos/3 - max_insertions/3:-1]: Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-1] += self.PinsDJ[base_ins + 1]*np.dot(self.Sdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-2] += self.PinsDJ[base_ins + 2]*np.dot(self.Ddj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][0, init_pos-base_ins-3] += self.PinsDJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is last nt in a codon for init_pos in range(-3, -(max_J_align+1), -3): #Zero insertions Pi_JinsDJ_given_D[D_in][0, init_pos] += self.PinsDJ[0]*Pi_J_given_D[D_in][0, init_pos] #current_base_nt_vec = first_nt_bias_insDJ*Pi_J_given_D[D_in][0, init_pos] #Okay for steady state current_base_nt_vec = self.zero_nt_bias_insDJ*Pi_J_given_D[D_in][0, init_pos] base_ins = 0 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 - 1: init_pos/3 - max_insertions/3:-1]: Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-1] += self.PinsDJ[base_ins + 1]*np.dot(self.Sdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-2] += self.PinsDJ[base_ins + 2]*np.dot(self.Ddj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][0, init_pos-base_ins-3] += self.PinsDJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 return Pi_JinsDJ_given_D
python
def compute_Pi_JinsDJ_given_D(self, CDR3_seq, Pi_J_given_D, max_J_align): """Compute Pi_JinsDJ conditioned on D. This function returns the Pi array from the model factors of the J genomic contributions, P(D,J)*P(delJ|J), and the DJ (N2) insertions, first_nt_bias_insDJ(n_1)PinsDJ(\ell_{DJ})\prod_{i=2}^{\ell_{DJ}}Rdj(n_i|n_{i-1}) conditioned on D identity. This corresponds to {N^{x_3}}_{x_4}J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_J_given_D : ndarray List of (4, 3L) ndarrays corresponding to J(D)^{x_4}. max_J_align : int Maximum alignment of the CDR3_seq to any genomic J allele allowed by J_usage_mask. self.PinsDJ : ndarray Probability distribution of the DJ (N2) insertion sequence length self.first_nt_bias_insDJ : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the DJ junction. self.zero_nt_bias_insDJ : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the DJ insertion. Note, as the Markov model at the DJ junction goes 3' to 5' this is the position AFTER the insertions reading left to right. self.Tdj : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Sdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the DJ insertion ending in the first position. self.Ddj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.rTdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the DJ insertion starting in the first position. self.rDdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for DJ insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_JinsDJ_given_D : list List of (4, 3L) ndarrays corresponding to {N^{x_3}}_{x_4}J(D)^{x_4}. """ #max_insertions = 30 #len(PinsVD) - 1 should zeropad the last few spots max_insertions = len(self.PinsDJ) - 1 Pi_JinsDJ_given_D = [np.zeros((4, len(CDR3_seq)*3)) for i in range(len(Pi_J_given_D))] for D_in in range(len(Pi_J_given_D)): #start position is first nt in a codon for init_pos in range(-1, -(max_J_align+1), -3): #Zero insertions Pi_JinsDJ_given_D[D_in][:, init_pos] += self.PinsDJ[0]*Pi_J_given_D[D_in][:, init_pos] #One insertion Pi_JinsDJ_given_D[D_in][:, init_pos-1] += self.PinsDJ[1]*np.dot(self.rDdj[CDR3_seq[init_pos/3]], Pi_J_given_D[D_in][:, init_pos]) #Two insertions and compute the base nt vec for the standard loop current_base_nt_vec = np.dot(self.rTdj[CDR3_seq[init_pos/3]], Pi_J_given_D[D_in][:, init_pos]) Pi_JinsDJ_given_D[D_in][0, init_pos-2] += self.PinsDJ[2]*np.sum(current_base_nt_vec) base_ins = 2 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 - 1: init_pos/3 - max_insertions/3:-1]: Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-1] += self.PinsDJ[base_ins + 1]*np.dot(self.Sdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-2] += self.PinsDJ[base_ins + 2]*np.dot(self.Ddj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][0, init_pos-base_ins-3] += self.PinsDJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is second nt in a codon for init_pos in range(-2, -(max_J_align+1), -3): #Zero insertions Pi_JinsDJ_given_D[D_in][:, init_pos] += self.PinsDJ[0]*Pi_J_given_D[D_in][:, init_pos] #One insertion --- we first compute our p vec by pairwise mult with the ss distr current_base_nt_vec = np.multiply(Pi_J_given_D[D_in][:, init_pos], self.first_nt_bias_insDJ) Pi_JinsDJ_given_D[D_in][0, init_pos-1] += self.PinsDJ[1]*np.sum(current_base_nt_vec) base_ins = 1 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 - 1: init_pos/3 - max_insertions/3:-1]: Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-1] += self.PinsDJ[base_ins + 1]*np.dot(self.Sdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-2] += self.PinsDJ[base_ins + 2]*np.dot(self.Ddj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][0, init_pos-base_ins-3] += self.PinsDJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is last nt in a codon for init_pos in range(-3, -(max_J_align+1), -3): #Zero insertions Pi_JinsDJ_given_D[D_in][0, init_pos] += self.PinsDJ[0]*Pi_J_given_D[D_in][0, init_pos] #current_base_nt_vec = first_nt_bias_insDJ*Pi_J_given_D[D_in][0, init_pos] #Okay for steady state current_base_nt_vec = self.zero_nt_bias_insDJ*Pi_J_given_D[D_in][0, init_pos] base_ins = 0 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 - 1: init_pos/3 - max_insertions/3:-1]: Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-1] += self.PinsDJ[base_ins + 1]*np.dot(self.Sdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][:, init_pos-base_ins-2] += self.PinsDJ[base_ins + 2]*np.dot(self.Ddj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tdj[aa], current_base_nt_vec) Pi_JinsDJ_given_D[D_in][0, init_pos-base_ins-3] += self.PinsDJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 return Pi_JinsDJ_given_D
[ "def", "compute_Pi_JinsDJ_given_D", "(", "self", ",", "CDR3_seq", ",", "Pi_J_given_D", ",", "max_J_align", ")", ":", "#max_insertions = 30 #len(PinsVD) - 1 should zeropad the last few spots", "max_insertions", "=", "len", "(", "self", ".", "PinsDJ", ")", "-", "1", "Pi_JinsDJ_given_D", "=", "[", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "for", "i", "in", "range", "(", "len", "(", "Pi_J_given_D", ")", ")", "]", "for", "D_in", "in", "range", "(", "len", "(", "Pi_J_given_D", ")", ")", ":", "#start position is first nt in a codon", "for", "init_pos", "in", "range", "(", "-", "1", ",", "-", "(", "max_J_align", "+", "1", ")", ",", "-", "3", ")", ":", "#Zero insertions", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", "+=", "self", ".", "PinsDJ", "[", "0", "]", "*", "Pi_J_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", "#One insertion", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "1", "]", "+=", "self", ".", "PinsDJ", "[", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "rDdj", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ",", "Pi_J_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", ")", "#Two insertions and compute the base nt vec for the standard loop ", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "rTdj", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ",", "Pi_J_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "-", "2", "]", "+=", "self", ".", "PinsDJ", "[", "2", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "=", "2", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "-", "1", ":", "init_pos", "/", "3", "-", "max_insertions", "/", "3", ":", "-", "1", "]", ":", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "base_ins", "-", "1", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Sdj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "base_ins", "-", "2", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Ddj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tdj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "-", "base_ins", "-", "3", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "#start position is second nt in a codon", "for", "init_pos", "in", "range", "(", "-", "2", ",", "-", "(", "max_J_align", "+", "1", ")", ",", "-", "3", ")", ":", "#Zero insertions", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", "+=", "self", ".", "PinsDJ", "[", "0", "]", "*", "Pi_J_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", "#One insertion --- we first compute our p vec by pairwise mult with the ss distr", "current_base_nt_vec", "=", "np", ".", "multiply", "(", "Pi_J_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", ",", "self", ".", "first_nt_bias_insDJ", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "-", "1", "]", "+=", "self", ".", "PinsDJ", "[", "1", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "=", "1", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "-", "1", ":", "init_pos", "/", "3", "-", "max_insertions", "/", "3", ":", "-", "1", "]", ":", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "base_ins", "-", "1", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Sdj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "base_ins", "-", "2", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Ddj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tdj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "-", "base_ins", "-", "3", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "#start position is last nt in a codon ", "for", "init_pos", "in", "range", "(", "-", "3", ",", "-", "(", "max_J_align", "+", "1", ")", ",", "-", "3", ")", ":", "#Zero insertions", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "]", "+=", "self", ".", "PinsDJ", "[", "0", "]", "*", "Pi_J_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "]", "#current_base_nt_vec = first_nt_bias_insDJ*Pi_J_given_D[D_in][0, init_pos] #Okay for steady state", "current_base_nt_vec", "=", "self", ".", "zero_nt_bias_insDJ", "*", "Pi_J_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "]", "base_ins", "=", "0", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "-", "1", ":", "init_pos", "/", "3", "-", "max_insertions", "/", "3", ":", "-", "1", "]", ":", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "base_ins", "-", "1", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Sdj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "-", "base_ins", "-", "2", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Ddj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tdj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "-", "base_ins", "-", "3", "]", "+=", "self", ".", "PinsDJ", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "return", "Pi_JinsDJ_given_D" ]
Compute Pi_JinsDJ conditioned on D. This function returns the Pi array from the model factors of the J genomic contributions, P(D,J)*P(delJ|J), and the DJ (N2) insertions, first_nt_bias_insDJ(n_1)PinsDJ(\ell_{DJ})\prod_{i=2}^{\ell_{DJ}}Rdj(n_i|n_{i-1}) conditioned on D identity. This corresponds to {N^{x_3}}_{x_4}J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_J_given_D : ndarray List of (4, 3L) ndarrays corresponding to J(D)^{x_4}. max_J_align : int Maximum alignment of the CDR3_seq to any genomic J allele allowed by J_usage_mask. self.PinsDJ : ndarray Probability distribution of the DJ (N2) insertion sequence length self.first_nt_bias_insDJ : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the DJ junction. self.zero_nt_bias_insDJ : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the DJ insertion. Note, as the Markov model at the DJ junction goes 3' to 5' this is the position AFTER the insertions reading left to right. self.Tdj : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Sdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the DJ insertion ending in the first position. self.Ddj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.rTdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the DJ insertion starting in the first position. self.rDdj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for DJ insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_JinsDJ_given_D : list List of (4, 3L) ndarrays corresponding to {N^{x_3}}_{x_4}J(D)^{x_4}.
[ "Compute", "Pi_JinsDJ", "conditioned", "on", "D", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "J", "genomic", "contributions", "P", "(", "D", "J", ")", "*", "P", "(", "delJ|J", ")", "and", "the", "DJ", "(", "N2", ")", "insertions", "first_nt_bias_insDJ", "(", "n_1", ")", "PinsDJ", "(", "\\", "ell_", "{", "DJ", "}", ")", "\\", "prod_", "{", "i", "=", "2", "}", "^", "{", "\\", "ell_", "{", "DJ", "}}", "Rdj", "(", "n_i|n_", "{", "i", "-", "1", "}", ")", "conditioned", "on", "D", "identity", ".", "This", "corresponds", "to", "{", "N^", "{", "x_3", "}}", "_", "{", "x_4", "}", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "Pi_J_given_D", ":", "ndarray", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "max_J_align", ":", "int", "Maximum", "alignment", "of", "the", "CDR3_seq", "to", "any", "genomic", "J", "allele", "allowed", "by", "J_usage_mask", ".", "self", ".", "PinsDJ", ":", "ndarray", "Probability", "distribution", "of", "the", "DJ", "(", "N2", ")", "insertion", "sequence", "length", "self", ".", "first_nt_bias_insDJ", ":", "ndarray", "(", "4", ")", "array", "of", "the", "probability", "distribution", "of", "the", "indentity", "of", "the", "first", "nucleotide", "insertion", "for", "the", "DJ", "junction", ".", "self", ".", "zero_nt_bias_insDJ", ":", "ndarray", "(", "4", ")", "array", "of", "the", "probability", "distribution", "of", "the", "indentity", "of", "the", "the", "nucleotide", "BEFORE", "the", "DJ", "insertion", ".", "Note", "as", "the", "Markov", "model", "at", "the", "DJ", "junction", "goes", "3", "to", "5", "this", "is", "the", "position", "AFTER", "the", "insertions", "reading", "left", "to", "right", ".", "self", ".", "Tdj", ":", "dict", "Dictionary", "of", "full", "codon", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", ".", "self", ".", "Sdj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "DJ", "insertion", "ending", "in", "the", "first", "position", ".", "self", ".", "Ddj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "ending", "in", "the", "second", "position", ".", "self", ".", "rTdj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "DJ", "insertion", "starting", "in", "the", "first", "position", ".", "self", ".", "rDdj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "DJ", "insertion", "starting", "in", "the", "first", "position", "and", "ending", "in", "the", "second", "position", "of", "the", "same", "codon", ".", "Returns", "-------", "Pi_JinsDJ_given_D", ":", "list", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "{", "N^", "{", "x_3", "}}", "_", "{", "x_4", "}", "J", "(", "D", ")", "^", "{", "x_4", "}", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1162-L1282
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVDJ.compute_Pi_R
def compute_Pi_R(self, CDR3_seq, Pi_JinsDJ_given_D): """Compute Pi_R. This function returns the Pi array from the model factors of the D and J genomic contributions, P(D, J)*P(delJ|J)P(delDl, delDr |D) and the DJ (N2) insertions, first_nt_bias_insDJ(n_1)PinsDJ(\ell_{DJ})\prod_{i=2}^{\ell_{DJ}}Rdj(n_i|n_{i-1}). This corresponds to \sum_D {D^{x_2}}_{x_3}{N^{x_3}}_{x_4}J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_JinsDJ_given_D : list List of (4, 3L) ndarrays corresponding to {N^{x_3}}_{x_4}J(D)^{x_4}. self.cutD_genomic_CDR3_segs : list of strings List of all the D genomic nucleotide sequences with the maximum number of palindromic insertions appended on both ends. self.PD_given_J : ndarray Probability distribution of D conditioned on J, i.e. P(D|J). self.PD_nt_pos_vec : list of ndarrays For each D allele, format P(delDl, delDr|D) into the correct form for a Pi array as if each position were the first in a codon. self.PD_2nd_nt_pos_per_aa_vec : list of dicts For each D allele, and each 'amino acid', format P(delDl, delDr|D) for positions in the middle of a codon into the correct form for a Pi array as if each position were the middle of a codon corresponding to the 'amino acid'. self.min_delDl_given_DdelDr : list of lists minimum delDl for each delDr, D combination. self.max_delDl_given_DdelDr : list of lists maximum delDl for each delDr, D combination. self.PdelDldelDr_given_D : ndarray Joint probability distribution of the D deletions given the D allele, i.e. P(delDl, delDr |D) self.zeroD_given_D : list of floats The probability that a given D allele is fully deleted away. self.codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. self.sub_codons_right : dict Dictionary of the 1 and 2 nucleotide suffixes (read from 5') for each codon in an 'amino acid' grouping Returns ------- Pi_L : ndarray (4, 3L) array corresponding to \sum_D {D^{x_2}}_{x_3}{N^{x_3}}_{x_4}J(D)^{x_4}. """ #Need to consider all D alignments from all possible positions and right deletions. nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #n_aaseq = [aa_dict[aa] for aa in CDR3_seq] Pi_R = np.zeros((4, len(CDR3_seq)*3)) min_pos = -len(CDR3_seq)*3 num_dell_pos, num_delr_pos, num_D_genes = self.PdelDldelDr_given_D.shape for D_in, cutD_gen_seg in enumerate(self.cutD_genomic_CDR3_segs): l_D_seg = len(cutD_gen_seg) #start position is first nt in a codon for init_pos in range(-1,-len(CDR3_seq)*3-1,-3): Pi_R[:, init_pos] += Pi_JinsDJ_given_D[D_in][:, init_pos]*self.zeroD_given_D[D_in] second_pos_dict = {'A': np.zeros(4), 'C': np.zeros(4), 'G': np.zeros(4), 'T': np.zeros(4)} codon_prefix_dict = {} for last_nt in 'ACGT': for second_nt in 'ACGT': codon_prefix_dict[last_nt + second_nt] = np.zeros(4) #for first_nt in ['ACGT'[nt] for nt in range(4) if Pi_JinsDJ_given_D[D_in][nt, init_pos] > 0]: for first_nt in 'ACGT': if last_nt + second_nt + first_nt in self.codons_dict[CDR3_seq[init_pos/3]]: #possible allowed codon second_pos_dict[second_nt][nt2num[last_nt]] += Pi_JinsDJ_given_D[D_in][nt2num[first_nt], init_pos] #base weight for middle pos nt codon_prefix_dict[last_nt + second_nt][0] += Pi_JinsDJ_given_D[D_in][nt2num[first_nt], init_pos] #base weight for last pos nt for nt1 in 'ACGT': if np.sum(second_pos_dict[nt1]) == 0: second_pos_dict.pop(nt1, None) for nt2 in 'ACGT': if np.sum(codon_prefix_dict[nt1+nt2])== 0: codon_prefix_dict.pop(nt1+nt2, None) # if len(second_pos_dict)> 0: # print second_pos_dict # return -1 for delDr in range(num_delr_pos): if self.min_delDl_given_DdelDr[D_in][delDr] == -1: # P(delDr | D) = 0 for this delDr --> move to next continue #Check if first nt from the D segment is okay if cutD_gen_seg[l_D_seg - delDr - 1] in second_pos_dict.keys(): #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if l_D_seg - delDr - 1 <= self.max_delDl_given_DdelDr[D_in][delDr]: Pi_R[:, init_pos - 1] += self.PdelDldelDr_given_D[l_D_seg - delDr - 1, delDr, D_in]*second_pos_dict[cutD_gen_seg[l_D_seg - delDr - 1]] else: continue #not okay, reject the alignment #Check if the second nt from the D segment is okay if cutD_gen_seg[l_D_seg - delDr - 2:l_D_seg - delDr] in codon_prefix_dict.keys(): #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if l_D_seg - delDr - 2 <= self.max_delDl_given_DdelDr[D_in][delDr]: Pi_R[0, init_pos - 2] += self.PdelDldelDr_given_D[l_D_seg - delDr - 2, delDr, D_in]*codon_prefix_dict[cutD_gen_seg[l_D_seg - delDr - 2:l_D_seg - delDr]][0] base_prob = codon_prefix_dict[cutD_gen_seg[l_D_seg - delDr - 2:l_D_seg - delDr]][0] else: continue #no longer aligned, move to next delDr #Enter main loop for pos in range(init_pos - 3, max(init_pos - l_D_seg + delDr, min_pos)-1, -1): #note delDl = D_pos D_pos = l_D_seg - delDr - 1 - ((init_pos - 1) - pos) #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if D_pos > self.max_delDl_given_DdelDr[D_in][delDr]: current_PdelDldelDr = 0 else: current_PdelDldelDr = self.PdelDldelDr_given_D[D_pos, delDr, D_in] #Position is the first nt in codon if pos%3 == 2: #check alignment if cutD_gen_seg[D_pos] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_nt_pos_vec[D_in][:, D_pos] else: break #no longer aligned -- exit loop #Position is the second nt in codon elif pos%3 == 1: #check alignment if cutD_gen_seg[D_pos:D_pos + 2] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_2nd_nt_pos_per_aa_vec[D_in][CDR3_seq[pos/3]][ :, D_pos] else: break #no longer aligned --- exit loop #Position is the last nt in codon else: #check alignment if cutD_gen_seg[D_pos:D_pos + 3] in self.codons_dict[CDR3_seq[pos/3]]: Pi_R[0, pos] += current_PdelDldelDr*base_prob else: break #no longer aligned --- exit loop #start position is second nt in a codon for init_pos in range(-2,-len(CDR3_seq)*3-1,-3): Pi_R[:, init_pos] += Pi_JinsDJ_given_D[D_in][:, init_pos]*self.zeroD_given_D[D_in] allowed_final_nts = ['ACGT'[nt] for nt in range(4) if Pi_JinsDJ_given_D[D_in][nt, init_pos] > 0] for delDr in range(num_delr_pos): if self.min_delDl_given_DdelDr[D_in][delDr] == -1: # P(delDr | D) = 0 for this delDr --> move to next continue #check first nt of the D region (last in the codon) if cutD_gen_seg[l_D_seg - delDr - 1] in allowed_final_nts: #first nt match base_prob = Pi_JinsDJ_given_D[D_in][nt2num[cutD_gen_seg[l_D_seg - delDr - 1]], init_pos] #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if l_D_seg - delDr - 1 <= self.max_delDl_given_DdelDr[D_in][delDr]: Pi_R[0, init_pos-1] += self.PdelDldelDr_given_D[l_D_seg - delDr - 1, delDr, D_in]*base_prob else: continue #no alignment #Enter main loop for pos in range(init_pos - 2, max(init_pos - l_D_seg + delDr, min_pos)-1, -1): #note delDl = D_pos D_pos = l_D_seg - delDr - 1 - ((init_pos - 1) - pos) #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if D_pos > self.max_delDl_given_DdelDr[D_in][delDr]: current_PdelDldelDr = 0 else: current_PdelDldelDr = self.PdelDldelDr_given_D[D_pos, delDr, D_in] #Position is the first nt in codon if pos%3 == 2: #check alignment if cutD_gen_seg[D_pos] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_nt_pos_vec[D_in][:, D_pos] else: break #no longer aligned -- exit loop #Position is the second nt in codon elif pos%3 == 1: #check alignment if cutD_gen_seg[D_pos:D_pos + 2] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_2nd_nt_pos_per_aa_vec[D_in][CDR3_seq[pos/3]][ :, D_pos] else: break #no longer aligned --- exit loop #Position is the last nt in codon else: #check alignment if cutD_gen_seg[D_pos:D_pos + 3] in self.codons_dict[CDR3_seq[pos/3]]: Pi_R[0, pos] += current_PdelDldelDr*base_prob else: break #no longer aligned --- exit loop #start position is last nt in a codon for init_pos in range(-3,-len(CDR3_seq)*3-1,-3): Pi_R[0, init_pos] += Pi_JinsDJ_given_D[D_in][0, init_pos]*self.zeroD_given_D[D_in] for delDr in range(num_delr_pos): if self.min_delDl_given_DdelDr[D_in][delDr] == -1: # P(delDr | D) = 0 for this delDr --> move to next continue base_prob = Pi_JinsDJ_given_D[D_in][0, init_pos] for pos in range(init_pos - 1, max(init_pos - l_D_seg + delDr, min_pos)-1, -1): #note delDl = D_pos D_pos = l_D_seg - delDr - 1 - ((init_pos - 1) - pos) #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if D_pos > self.max_delDl_given_DdelDr[D_in][delDr]: current_PdelDldelDr = 0 else: current_PdelDldelDr = self.PdelDldelDr_given_D[D_pos, delDr, D_in] #Position is the first nt in codon if pos%3 == 2: #check alignment if cutD_gen_seg[D_pos] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_nt_pos_vec[D_in][:, D_pos] else: break #no longer aligned -- exit loop #Position is the second nt in codon elif pos%3 == 1: #check alignment if cutD_gen_seg[D_pos:D_pos + 2] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_2nd_nt_pos_per_aa_vec[D_in][CDR3_seq[pos/3]][ :, D_pos] else: break #no longer aligned --- exit loop #Position is the last nt in codon else: #check alignment if cutD_gen_seg[D_pos:D_pos + 3] in self.codons_dict[CDR3_seq[pos/3]]: Pi_R[0, pos] += current_PdelDldelDr*base_prob else: break #no longer aligned --- exit loop return Pi_R
python
def compute_Pi_R(self, CDR3_seq, Pi_JinsDJ_given_D): """Compute Pi_R. This function returns the Pi array from the model factors of the D and J genomic contributions, P(D, J)*P(delJ|J)P(delDl, delDr |D) and the DJ (N2) insertions, first_nt_bias_insDJ(n_1)PinsDJ(\ell_{DJ})\prod_{i=2}^{\ell_{DJ}}Rdj(n_i|n_{i-1}). This corresponds to \sum_D {D^{x_2}}_{x_3}{N^{x_3}}_{x_4}J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_JinsDJ_given_D : list List of (4, 3L) ndarrays corresponding to {N^{x_3}}_{x_4}J(D)^{x_4}. self.cutD_genomic_CDR3_segs : list of strings List of all the D genomic nucleotide sequences with the maximum number of palindromic insertions appended on both ends. self.PD_given_J : ndarray Probability distribution of D conditioned on J, i.e. P(D|J). self.PD_nt_pos_vec : list of ndarrays For each D allele, format P(delDl, delDr|D) into the correct form for a Pi array as if each position were the first in a codon. self.PD_2nd_nt_pos_per_aa_vec : list of dicts For each D allele, and each 'amino acid', format P(delDl, delDr|D) for positions in the middle of a codon into the correct form for a Pi array as if each position were the middle of a codon corresponding to the 'amino acid'. self.min_delDl_given_DdelDr : list of lists minimum delDl for each delDr, D combination. self.max_delDl_given_DdelDr : list of lists maximum delDl for each delDr, D combination. self.PdelDldelDr_given_D : ndarray Joint probability distribution of the D deletions given the D allele, i.e. P(delDl, delDr |D) self.zeroD_given_D : list of floats The probability that a given D allele is fully deleted away. self.codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. self.sub_codons_right : dict Dictionary of the 1 and 2 nucleotide suffixes (read from 5') for each codon in an 'amino acid' grouping Returns ------- Pi_L : ndarray (4, 3L) array corresponding to \sum_D {D^{x_2}}_{x_3}{N^{x_3}}_{x_4}J(D)^{x_4}. """ #Need to consider all D alignments from all possible positions and right deletions. nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #n_aaseq = [aa_dict[aa] for aa in CDR3_seq] Pi_R = np.zeros((4, len(CDR3_seq)*3)) min_pos = -len(CDR3_seq)*3 num_dell_pos, num_delr_pos, num_D_genes = self.PdelDldelDr_given_D.shape for D_in, cutD_gen_seg in enumerate(self.cutD_genomic_CDR3_segs): l_D_seg = len(cutD_gen_seg) #start position is first nt in a codon for init_pos in range(-1,-len(CDR3_seq)*3-1,-3): Pi_R[:, init_pos] += Pi_JinsDJ_given_D[D_in][:, init_pos]*self.zeroD_given_D[D_in] second_pos_dict = {'A': np.zeros(4), 'C': np.zeros(4), 'G': np.zeros(4), 'T': np.zeros(4)} codon_prefix_dict = {} for last_nt in 'ACGT': for second_nt in 'ACGT': codon_prefix_dict[last_nt + second_nt] = np.zeros(4) #for first_nt in ['ACGT'[nt] for nt in range(4) if Pi_JinsDJ_given_D[D_in][nt, init_pos] > 0]: for first_nt in 'ACGT': if last_nt + second_nt + first_nt in self.codons_dict[CDR3_seq[init_pos/3]]: #possible allowed codon second_pos_dict[second_nt][nt2num[last_nt]] += Pi_JinsDJ_given_D[D_in][nt2num[first_nt], init_pos] #base weight for middle pos nt codon_prefix_dict[last_nt + second_nt][0] += Pi_JinsDJ_given_D[D_in][nt2num[first_nt], init_pos] #base weight for last pos nt for nt1 in 'ACGT': if np.sum(second_pos_dict[nt1]) == 0: second_pos_dict.pop(nt1, None) for nt2 in 'ACGT': if np.sum(codon_prefix_dict[nt1+nt2])== 0: codon_prefix_dict.pop(nt1+nt2, None) # if len(second_pos_dict)> 0: # print second_pos_dict # return -1 for delDr in range(num_delr_pos): if self.min_delDl_given_DdelDr[D_in][delDr] == -1: # P(delDr | D) = 0 for this delDr --> move to next continue #Check if first nt from the D segment is okay if cutD_gen_seg[l_D_seg - delDr - 1] in second_pos_dict.keys(): #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if l_D_seg - delDr - 1 <= self.max_delDl_given_DdelDr[D_in][delDr]: Pi_R[:, init_pos - 1] += self.PdelDldelDr_given_D[l_D_seg - delDr - 1, delDr, D_in]*second_pos_dict[cutD_gen_seg[l_D_seg - delDr - 1]] else: continue #not okay, reject the alignment #Check if the second nt from the D segment is okay if cutD_gen_seg[l_D_seg - delDr - 2:l_D_seg - delDr] in codon_prefix_dict.keys(): #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if l_D_seg - delDr - 2 <= self.max_delDl_given_DdelDr[D_in][delDr]: Pi_R[0, init_pos - 2] += self.PdelDldelDr_given_D[l_D_seg - delDr - 2, delDr, D_in]*codon_prefix_dict[cutD_gen_seg[l_D_seg - delDr - 2:l_D_seg - delDr]][0] base_prob = codon_prefix_dict[cutD_gen_seg[l_D_seg - delDr - 2:l_D_seg - delDr]][0] else: continue #no longer aligned, move to next delDr #Enter main loop for pos in range(init_pos - 3, max(init_pos - l_D_seg + delDr, min_pos)-1, -1): #note delDl = D_pos D_pos = l_D_seg - delDr - 1 - ((init_pos - 1) - pos) #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if D_pos > self.max_delDl_given_DdelDr[D_in][delDr]: current_PdelDldelDr = 0 else: current_PdelDldelDr = self.PdelDldelDr_given_D[D_pos, delDr, D_in] #Position is the first nt in codon if pos%3 == 2: #check alignment if cutD_gen_seg[D_pos] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_nt_pos_vec[D_in][:, D_pos] else: break #no longer aligned -- exit loop #Position is the second nt in codon elif pos%3 == 1: #check alignment if cutD_gen_seg[D_pos:D_pos + 2] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_2nd_nt_pos_per_aa_vec[D_in][CDR3_seq[pos/3]][ :, D_pos] else: break #no longer aligned --- exit loop #Position is the last nt in codon else: #check alignment if cutD_gen_seg[D_pos:D_pos + 3] in self.codons_dict[CDR3_seq[pos/3]]: Pi_R[0, pos] += current_PdelDldelDr*base_prob else: break #no longer aligned --- exit loop #start position is second nt in a codon for init_pos in range(-2,-len(CDR3_seq)*3-1,-3): Pi_R[:, init_pos] += Pi_JinsDJ_given_D[D_in][:, init_pos]*self.zeroD_given_D[D_in] allowed_final_nts = ['ACGT'[nt] for nt in range(4) if Pi_JinsDJ_given_D[D_in][nt, init_pos] > 0] for delDr in range(num_delr_pos): if self.min_delDl_given_DdelDr[D_in][delDr] == -1: # P(delDr | D) = 0 for this delDr --> move to next continue #check first nt of the D region (last in the codon) if cutD_gen_seg[l_D_seg - delDr - 1] in allowed_final_nts: #first nt match base_prob = Pi_JinsDJ_given_D[D_in][nt2num[cutD_gen_seg[l_D_seg - delDr - 1]], init_pos] #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if l_D_seg - delDr - 1 <= self.max_delDl_given_DdelDr[D_in][delDr]: Pi_R[0, init_pos-1] += self.PdelDldelDr_given_D[l_D_seg - delDr - 1, delDr, D_in]*base_prob else: continue #no alignment #Enter main loop for pos in range(init_pos - 2, max(init_pos - l_D_seg + delDr, min_pos)-1, -1): #note delDl = D_pos D_pos = l_D_seg - delDr - 1 - ((init_pos - 1) - pos) #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if D_pos > self.max_delDl_given_DdelDr[D_in][delDr]: current_PdelDldelDr = 0 else: current_PdelDldelDr = self.PdelDldelDr_given_D[D_pos, delDr, D_in] #Position is the first nt in codon if pos%3 == 2: #check alignment if cutD_gen_seg[D_pos] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_nt_pos_vec[D_in][:, D_pos] else: break #no longer aligned -- exit loop #Position is the second nt in codon elif pos%3 == 1: #check alignment if cutD_gen_seg[D_pos:D_pos + 2] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_2nd_nt_pos_per_aa_vec[D_in][CDR3_seq[pos/3]][ :, D_pos] else: break #no longer aligned --- exit loop #Position is the last nt in codon else: #check alignment if cutD_gen_seg[D_pos:D_pos + 3] in self.codons_dict[CDR3_seq[pos/3]]: Pi_R[0, pos] += current_PdelDldelDr*base_prob else: break #no longer aligned --- exit loop #start position is last nt in a codon for init_pos in range(-3,-len(CDR3_seq)*3-1,-3): Pi_R[0, init_pos] += Pi_JinsDJ_given_D[D_in][0, init_pos]*self.zeroD_given_D[D_in] for delDr in range(num_delr_pos): if self.min_delDl_given_DdelDr[D_in][delDr] == -1: # P(delDr | D) = 0 for this delDr --> move to next continue base_prob = Pi_JinsDJ_given_D[D_in][0, init_pos] for pos in range(init_pos - 1, max(init_pos - l_D_seg + delDr, min_pos)-1, -1): #note delDl = D_pos D_pos = l_D_seg - delDr - 1 - ((init_pos - 1) - pos) #The dell pos may be out of range of the PdelDldelDr_given_D -- check! if D_pos > self.max_delDl_given_DdelDr[D_in][delDr]: current_PdelDldelDr = 0 else: current_PdelDldelDr = self.PdelDldelDr_given_D[D_pos, delDr, D_in] #Position is the first nt in codon if pos%3 == 2: #check alignment if cutD_gen_seg[D_pos] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_nt_pos_vec[D_in][:, D_pos] else: break #no longer aligned -- exit loop #Position is the second nt in codon elif pos%3 == 1: #check alignment if cutD_gen_seg[D_pos:D_pos + 2] in self.sub_codons_right[CDR3_seq[pos/3]]: Pi_R[:, pos] += current_PdelDldelDr*base_prob*self.PD_2nd_nt_pos_per_aa_vec[D_in][CDR3_seq[pos/3]][ :, D_pos] else: break #no longer aligned --- exit loop #Position is the last nt in codon else: #check alignment if cutD_gen_seg[D_pos:D_pos + 3] in self.codons_dict[CDR3_seq[pos/3]]: Pi_R[0, pos] += current_PdelDldelDr*base_prob else: break #no longer aligned --- exit loop return Pi_R
[ "def", "compute_Pi_R", "(", "self", ",", "CDR3_seq", ",", "Pi_JinsDJ_given_D", ")", ":", "#Need to consider all D alignments from all possible positions and right deletions.", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "#n_aaseq = [aa_dict[aa] for aa in CDR3_seq]", "Pi_R", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "min_pos", "=", "-", "len", "(", "CDR3_seq", ")", "*", "3", "num_dell_pos", ",", "num_delr_pos", ",", "num_D_genes", "=", "self", ".", "PdelDldelDr_given_D", ".", "shape", "for", "D_in", ",", "cutD_gen_seg", "in", "enumerate", "(", "self", ".", "cutD_genomic_CDR3_segs", ")", ":", "l_D_seg", "=", "len", "(", "cutD_gen_seg", ")", "#start position is first nt in a codon", "for", "init_pos", "in", "range", "(", "-", "1", ",", "-", "len", "(", "CDR3_seq", ")", "*", "3", "-", "1", ",", "-", "3", ")", ":", "Pi_R", "[", ":", ",", "init_pos", "]", "+=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", "*", "self", ".", "zeroD_given_D", "[", "D_in", "]", "second_pos_dict", "=", "{", "'A'", ":", "np", ".", "zeros", "(", "4", ")", ",", "'C'", ":", "np", ".", "zeros", "(", "4", ")", ",", "'G'", ":", "np", ".", "zeros", "(", "4", ")", ",", "'T'", ":", "np", ".", "zeros", "(", "4", ")", "}", "codon_prefix_dict", "=", "{", "}", "for", "last_nt", "in", "'ACGT'", ":", "for", "second_nt", "in", "'ACGT'", ":", "codon_prefix_dict", "[", "last_nt", "+", "second_nt", "]", "=", "np", ".", "zeros", "(", "4", ")", "#for first_nt in ['ACGT'[nt] for nt in range(4) if Pi_JinsDJ_given_D[D_in][nt, init_pos] > 0]:", "for", "first_nt", "in", "'ACGT'", ":", "if", "last_nt", "+", "second_nt", "+", "first_nt", "in", "self", ".", "codons_dict", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ":", "#possible allowed codon", "second_pos_dict", "[", "second_nt", "]", "[", "nt2num", "[", "last_nt", "]", "]", "+=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "nt2num", "[", "first_nt", "]", ",", "init_pos", "]", "#base weight for middle pos nt", "codon_prefix_dict", "[", "last_nt", "+", "second_nt", "]", "[", "0", "]", "+=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "nt2num", "[", "first_nt", "]", ",", "init_pos", "]", "#base weight for last pos nt", "for", "nt1", "in", "'ACGT'", ":", "if", "np", ".", "sum", "(", "second_pos_dict", "[", "nt1", "]", ")", "==", "0", ":", "second_pos_dict", ".", "pop", "(", "nt1", ",", "None", ")", "for", "nt2", "in", "'ACGT'", ":", "if", "np", ".", "sum", "(", "codon_prefix_dict", "[", "nt1", "+", "nt2", "]", ")", "==", "0", ":", "codon_prefix_dict", ".", "pop", "(", "nt1", "+", "nt2", ",", "None", ")", "# if len(second_pos_dict)> 0:", "# print second_pos_dict", "# return -1", "for", "delDr", "in", "range", "(", "num_delr_pos", ")", ":", "if", "self", ".", "min_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", "==", "-", "1", ":", "# P(delDr | D) = 0 for this delDr --> move to next", "continue", "#Check if first nt from the D segment is okay", "if", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "1", "]", "in", "second_pos_dict", ".", "keys", "(", ")", ":", "#The dell pos may be out of range of the PdelDldelDr_given_D -- check!", "if", "l_D_seg", "-", "delDr", "-", "1", "<=", "self", ".", "max_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", ":", "Pi_R", "[", ":", ",", "init_pos", "-", "1", "]", "+=", "self", ".", "PdelDldelDr_given_D", "[", "l_D_seg", "-", "delDr", "-", "1", ",", "delDr", ",", "D_in", "]", "*", "second_pos_dict", "[", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "1", "]", "]", "else", ":", "continue", "#not okay, reject the alignment", "#Check if the second nt from the D segment is okay", "if", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "2", ":", "l_D_seg", "-", "delDr", "]", "in", "codon_prefix_dict", ".", "keys", "(", ")", ":", "#The dell pos may be out of range of the PdelDldelDr_given_D -- check!", "if", "l_D_seg", "-", "delDr", "-", "2", "<=", "self", ".", "max_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", ":", "Pi_R", "[", "0", ",", "init_pos", "-", "2", "]", "+=", "self", ".", "PdelDldelDr_given_D", "[", "l_D_seg", "-", "delDr", "-", "2", ",", "delDr", ",", "D_in", "]", "*", "codon_prefix_dict", "[", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "2", ":", "l_D_seg", "-", "delDr", "]", "]", "[", "0", "]", "base_prob", "=", "codon_prefix_dict", "[", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "2", ":", "l_D_seg", "-", "delDr", "]", "]", "[", "0", "]", "else", ":", "continue", "#no longer aligned, move to next delDr", "#Enter main loop", "for", "pos", "in", "range", "(", "init_pos", "-", "3", ",", "max", "(", "init_pos", "-", "l_D_seg", "+", "delDr", ",", "min_pos", ")", "-", "1", ",", "-", "1", ")", ":", "#note delDl = D_pos", "D_pos", "=", "l_D_seg", "-", "delDr", "-", "1", "-", "(", "(", "init_pos", "-", "1", ")", "-", "pos", ")", "#The dell pos may be out of range of the PdelDldelDr_given_D -- check!", "if", "D_pos", ">", "self", ".", "max_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", ":", "current_PdelDldelDr", "=", "0", "else", ":", "current_PdelDldelDr", "=", "self", ".", "PdelDldelDr_given_D", "[", "D_pos", ",", "delDr", ",", "D_in", "]", "#Position is the first nt in codon", "if", "pos", "%", "3", "==", "2", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", "]", "in", "self", ".", "sub_codons_right", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", ":", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "*", "self", ".", "PD_nt_pos_vec", "[", "D_in", "]", "[", ":", ",", "D_pos", "]", "else", ":", "break", "#no longer aligned -- exit loop", "#Position is the second nt in codon", "elif", "pos", "%", "3", "==", "1", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", ":", "D_pos", "+", "2", "]", "in", "self", ".", "sub_codons_right", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", ":", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "*", "self", ".", "PD_2nd_nt_pos_per_aa_vec", "[", "D_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "D_pos", "]", "else", ":", "break", "#no longer aligned --- exit loop", "#Position is the last nt in codon", "else", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", ":", "D_pos", "+", "3", "]", "in", "self", ".", "codons_dict", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", "0", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "else", ":", "break", "#no longer aligned --- exit loop", "#start position is second nt in a codon", "for", "init_pos", "in", "range", "(", "-", "2", ",", "-", "len", "(", "CDR3_seq", ")", "*", "3", "-", "1", ",", "-", "3", ")", ":", "Pi_R", "[", ":", ",", "init_pos", "]", "+=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", ":", ",", "init_pos", "]", "*", "self", ".", "zeroD_given_D", "[", "D_in", "]", "allowed_final_nts", "=", "[", "'ACGT'", "[", "nt", "]", "for", "nt", "in", "range", "(", "4", ")", "if", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "nt", ",", "init_pos", "]", ">", "0", "]", "for", "delDr", "in", "range", "(", "num_delr_pos", ")", ":", "if", "self", ".", "min_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", "==", "-", "1", ":", "# P(delDr | D) = 0 for this delDr --> move to next", "continue", "#check first nt of the D region (last in the codon)", "if", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "1", "]", "in", "allowed_final_nts", ":", "#first nt match", "base_prob", "=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "nt2num", "[", "cutD_gen_seg", "[", "l_D_seg", "-", "delDr", "-", "1", "]", "]", ",", "init_pos", "]", "#The dell pos may be out of range of the PdelDldelDr_given_D -- check!", "if", "l_D_seg", "-", "delDr", "-", "1", "<=", "self", ".", "max_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", ":", "Pi_R", "[", "0", ",", "init_pos", "-", "1", "]", "+=", "self", ".", "PdelDldelDr_given_D", "[", "l_D_seg", "-", "delDr", "-", "1", ",", "delDr", ",", "D_in", "]", "*", "base_prob", "else", ":", "continue", "#no alignment", "#Enter main loop", "for", "pos", "in", "range", "(", "init_pos", "-", "2", ",", "max", "(", "init_pos", "-", "l_D_seg", "+", "delDr", ",", "min_pos", ")", "-", "1", ",", "-", "1", ")", ":", "#note delDl = D_pos", "D_pos", "=", "l_D_seg", "-", "delDr", "-", "1", "-", "(", "(", "init_pos", "-", "1", ")", "-", "pos", ")", "#The dell pos may be out of range of the PdelDldelDr_given_D -- check!", "if", "D_pos", ">", "self", ".", "max_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", ":", "current_PdelDldelDr", "=", "0", "else", ":", "current_PdelDldelDr", "=", "self", ".", "PdelDldelDr_given_D", "[", "D_pos", ",", "delDr", ",", "D_in", "]", "#Position is the first nt in codon", "if", "pos", "%", "3", "==", "2", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", "]", "in", "self", ".", "sub_codons_right", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", ":", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "*", "self", ".", "PD_nt_pos_vec", "[", "D_in", "]", "[", ":", ",", "D_pos", "]", "else", ":", "break", "#no longer aligned -- exit loop", "#Position is the second nt in codon", "elif", "pos", "%", "3", "==", "1", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", ":", "D_pos", "+", "2", "]", "in", "self", ".", "sub_codons_right", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", ":", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "*", "self", ".", "PD_2nd_nt_pos_per_aa_vec", "[", "D_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "D_pos", "]", "else", ":", "break", "#no longer aligned --- exit loop", "#Position is the last nt in codon", "else", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", ":", "D_pos", "+", "3", "]", "in", "self", ".", "codons_dict", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", "0", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "else", ":", "break", "#no longer aligned --- exit loop", "#start position is last nt in a codon", "for", "init_pos", "in", "range", "(", "-", "3", ",", "-", "len", "(", "CDR3_seq", ")", "*", "3", "-", "1", ",", "-", "3", ")", ":", "Pi_R", "[", "0", ",", "init_pos", "]", "+=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "]", "*", "self", ".", "zeroD_given_D", "[", "D_in", "]", "for", "delDr", "in", "range", "(", "num_delr_pos", ")", ":", "if", "self", ".", "min_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", "==", "-", "1", ":", "# P(delDr | D) = 0 for this delDr --> move to next", "continue", "base_prob", "=", "Pi_JinsDJ_given_D", "[", "D_in", "]", "[", "0", ",", "init_pos", "]", "for", "pos", "in", "range", "(", "init_pos", "-", "1", ",", "max", "(", "init_pos", "-", "l_D_seg", "+", "delDr", ",", "min_pos", ")", "-", "1", ",", "-", "1", ")", ":", "#note delDl = D_pos", "D_pos", "=", "l_D_seg", "-", "delDr", "-", "1", "-", "(", "(", "init_pos", "-", "1", ")", "-", "pos", ")", "#The dell pos may be out of range of the PdelDldelDr_given_D -- check!", "if", "D_pos", ">", "self", ".", "max_delDl_given_DdelDr", "[", "D_in", "]", "[", "delDr", "]", ":", "current_PdelDldelDr", "=", "0", "else", ":", "current_PdelDldelDr", "=", "self", ".", "PdelDldelDr_given_D", "[", "D_pos", ",", "delDr", ",", "D_in", "]", "#Position is the first nt in codon", "if", "pos", "%", "3", "==", "2", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", "]", "in", "self", ".", "sub_codons_right", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", ":", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "*", "self", ".", "PD_nt_pos_vec", "[", "D_in", "]", "[", ":", ",", "D_pos", "]", "else", ":", "break", "#no longer aligned -- exit loop", "#Position is the second nt in codon", "elif", "pos", "%", "3", "==", "1", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", ":", "D_pos", "+", "2", "]", "in", "self", ".", "sub_codons_right", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", ":", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "*", "self", ".", "PD_2nd_nt_pos_per_aa_vec", "[", "D_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "D_pos", "]", "else", ":", "break", "#no longer aligned --- exit loop", "#Position is the last nt in codon", "else", ":", "#check alignment", "if", "cutD_gen_seg", "[", "D_pos", ":", "D_pos", "+", "3", "]", "in", "self", ".", "codons_dict", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", ":", "Pi_R", "[", "0", ",", "pos", "]", "+=", "current_PdelDldelDr", "*", "base_prob", "else", ":", "break", "#no longer aligned --- exit loop", "return", "Pi_R" ]
Compute Pi_R. This function returns the Pi array from the model factors of the D and J genomic contributions, P(D, J)*P(delJ|J)P(delDl, delDr |D) and the DJ (N2) insertions, first_nt_bias_insDJ(n_1)PinsDJ(\ell_{DJ})\prod_{i=2}^{\ell_{DJ}}Rdj(n_i|n_{i-1}). This corresponds to \sum_D {D^{x_2}}_{x_3}{N^{x_3}}_{x_4}J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_JinsDJ_given_D : list List of (4, 3L) ndarrays corresponding to {N^{x_3}}_{x_4}J(D)^{x_4}. self.cutD_genomic_CDR3_segs : list of strings List of all the D genomic nucleotide sequences with the maximum number of palindromic insertions appended on both ends. self.PD_given_J : ndarray Probability distribution of D conditioned on J, i.e. P(D|J). self.PD_nt_pos_vec : list of ndarrays For each D allele, format P(delDl, delDr|D) into the correct form for a Pi array as if each position were the first in a codon. self.PD_2nd_nt_pos_per_aa_vec : list of dicts For each D allele, and each 'amino acid', format P(delDl, delDr|D) for positions in the middle of a codon into the correct form for a Pi array as if each position were the middle of a codon corresponding to the 'amino acid'. self.min_delDl_given_DdelDr : list of lists minimum delDl for each delDr, D combination. self.max_delDl_given_DdelDr : list of lists maximum delDl for each delDr, D combination. self.PdelDldelDr_given_D : ndarray Joint probability distribution of the D deletions given the D allele, i.e. P(delDl, delDr |D) self.zeroD_given_D : list of floats The probability that a given D allele is fully deleted away. self.codons_dict : dict Dictionary, keyed by the allowed 'amino acid' symbols with the values being lists of codons corresponding to the symbol. self.sub_codons_right : dict Dictionary of the 1 and 2 nucleotide suffixes (read from 5') for each codon in an 'amino acid' grouping Returns ------- Pi_L : ndarray (4, 3L) array corresponding to \sum_D {D^{x_2}}_{x_3}{N^{x_3}}_{x_4}J(D)^{x_4}.
[ "Compute", "Pi_R", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "D", "and", "J", "genomic", "contributions", "P", "(", "D", "J", ")", "*", "P", "(", "delJ|J", ")", "P", "(", "delDl", "delDr", "|D", ")", "and", "the", "DJ", "(", "N2", ")", "insertions", "first_nt_bias_insDJ", "(", "n_1", ")", "PinsDJ", "(", "\\", "ell_", "{", "DJ", "}", ")", "\\", "prod_", "{", "i", "=", "2", "}", "^", "{", "\\", "ell_", "{", "DJ", "}}", "Rdj", "(", "n_i|n_", "{", "i", "-", "1", "}", ")", ".", "This", "corresponds", "to", "\\", "sum_D", "{", "D^", "{", "x_2", "}}", "_", "{", "x_3", "}", "{", "N^", "{", "x_3", "}}", "_", "{", "x_4", "}", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "Pi_JinsDJ_given_D", ":", "list", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "{", "N^", "{", "x_3", "}}", "_", "{", "x_4", "}", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "self", ".", "cutD_genomic_CDR3_segs", ":", "list", "of", "strings", "List", "of", "all", "the", "D", "genomic", "nucleotide", "sequences", "with", "the", "maximum", "number", "of", "palindromic", "insertions", "appended", "on", "both", "ends", ".", "self", ".", "PD_given_J", ":", "ndarray", "Probability", "distribution", "of", "D", "conditioned", "on", "J", "i", ".", "e", ".", "P", "(", "D|J", ")", ".", "self", ".", "PD_nt_pos_vec", ":", "list", "of", "ndarrays", "For", "each", "D", "allele", "format", "P", "(", "delDl", "delDr|D", ")", "into", "the", "correct", "form", "for", "a", "Pi", "array", "as", "if", "each", "position", "were", "the", "first", "in", "a", "codon", ".", "self", ".", "PD_2nd_nt_pos_per_aa_vec", ":", "list", "of", "dicts", "For", "each", "D", "allele", "and", "each", "amino", "acid", "format", "P", "(", "delDl", "delDr|D", ")", "for", "positions", "in", "the", "middle", "of", "a", "codon", "into", "the", "correct", "form", "for", "a", "Pi", "array", "as", "if", "each", "position", "were", "the", "middle", "of", "a", "codon", "corresponding", "to", "the", "amino", "acid", ".", "self", ".", "min_delDl_given_DdelDr", ":", "list", "of", "lists", "minimum", "delDl", "for", "each", "delDr", "D", "combination", ".", "self", ".", "max_delDl_given_DdelDr", ":", "list", "of", "lists", "maximum", "delDl", "for", "each", "delDr", "D", "combination", ".", "self", ".", "PdelDldelDr_given_D", ":", "ndarray", "Joint", "probability", "distribution", "of", "the", "D", "deletions", "given", "the", "D", "allele", "i", ".", "e", ".", "P", "(", "delDl", "delDr", "|D", ")", "self", ".", "zeroD_given_D", ":", "list", "of", "floats", "The", "probability", "that", "a", "given", "D", "allele", "is", "fully", "deleted", "away", ".", "self", ".", "codons_dict", ":", "dict", "Dictionary", "keyed", "by", "the", "allowed", "amino", "acid", "symbols", "with", "the", "values", "being", "lists", "of", "codons", "corresponding", "to", "the", "symbol", ".", "self", ".", "sub_codons_right", ":", "dict", "Dictionary", "of", "the", "1", "and", "2", "nucleotide", "suffixes", "(", "read", "from", "5", ")", "for", "each", "codon", "in", "an", "amino", "acid", "grouping", "Returns", "-------", "Pi_L", ":", "ndarray", "(", "4", "3L", ")", "array", "corresponding", "to", "\\", "sum_D", "{", "D^", "{", "x_2", "}}", "_", "{", "x_3", "}", "{", "N^", "{", "x_3", "}}", "_", "{", "x_4", "}", "J", "(", "D", ")", "^", "{", "x_4", "}", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1285-L1521
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVJ.compute_CDR3_pgen
def compute_CDR3_pgen(self, CDR3_seq, V_usage_mask, J_usage_mask): """Compute Pgen for CDR3 'amino acid' sequence CDR3_seq from VJ model. Conditioned on the already formatted V genes/alleles indicated in V_usage_mask and the J genes/alleles in J_usage_mask. Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> compute_CDR3_pgen('CAVKIQGAQKLVF', ppp, [72], [56]) 4.1818202431143785e-07 >>> compute_CDR3_pgen(nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC'), ppp, [42], [1]) 1.3971676613008565e-08 >>> compute_CDR3_pgen('\xbb\xb6\xbe\x80\xbc\xa1\x8a\x96\xa1\xa0\xad\x8e\xbf', ppp, [72], [56]) 1.3971676613008565e-08 """ #Genomic J alignment/matching (contribution from P(delJ | J)), return Pi_J and reduced J_usage_mask Pi_J, r_J_usage_mask = self.compute_Pi_J(CDR3_seq, J_usage_mask) #Genomic V alignment/matching conditioned on J gene (contribution from P(V, J, delV)), return Pi_V_given_J Pi_V_given_J, max_V_align = self.compute_Pi_V_given_J(CDR3_seq, V_usage_mask, r_J_usage_mask) #Include insertions (R and PinsVJ) to get the total contribution from the left (3') side conditioned on J gene. Return Pi_V_insVJ_given_J Pi_V_insVJ_given_J = self.compute_Pi_V_insVJ_given_J(CDR3_seq, Pi_V_given_J, max_V_align) pgen = 0 #zip Pi_V_insVJ_given_J and Pi_J together for each J gene to get total pgen for j in range(len(r_J_usage_mask)): for pos in range(len(CDR3_seq)*3 - 1): pgen += np.dot(Pi_V_insVJ_given_J[j][:, pos], Pi_J[j][:, pos+1]) return pgen
python
def compute_CDR3_pgen(self, CDR3_seq, V_usage_mask, J_usage_mask): """Compute Pgen for CDR3 'amino acid' sequence CDR3_seq from VJ model. Conditioned on the already formatted V genes/alleles indicated in V_usage_mask and the J genes/alleles in J_usage_mask. Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> compute_CDR3_pgen('CAVKIQGAQKLVF', ppp, [72], [56]) 4.1818202431143785e-07 >>> compute_CDR3_pgen(nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC'), ppp, [42], [1]) 1.3971676613008565e-08 >>> compute_CDR3_pgen('\xbb\xb6\xbe\x80\xbc\xa1\x8a\x96\xa1\xa0\xad\x8e\xbf', ppp, [72], [56]) 1.3971676613008565e-08 """ #Genomic J alignment/matching (contribution from P(delJ | J)), return Pi_J and reduced J_usage_mask Pi_J, r_J_usage_mask = self.compute_Pi_J(CDR3_seq, J_usage_mask) #Genomic V alignment/matching conditioned on J gene (contribution from P(V, J, delV)), return Pi_V_given_J Pi_V_given_J, max_V_align = self.compute_Pi_V_given_J(CDR3_seq, V_usage_mask, r_J_usage_mask) #Include insertions (R and PinsVJ) to get the total contribution from the left (3') side conditioned on J gene. Return Pi_V_insVJ_given_J Pi_V_insVJ_given_J = self.compute_Pi_V_insVJ_given_J(CDR3_seq, Pi_V_given_J, max_V_align) pgen = 0 #zip Pi_V_insVJ_given_J and Pi_J together for each J gene to get total pgen for j in range(len(r_J_usage_mask)): for pos in range(len(CDR3_seq)*3 - 1): pgen += np.dot(Pi_V_insVJ_given_J[j][:, pos], Pi_J[j][:, pos+1]) return pgen
[ "def", "compute_CDR3_pgen", "(", "self", ",", "CDR3_seq", ",", "V_usage_mask", ",", "J_usage_mask", ")", ":", "#Genomic J alignment/matching (contribution from P(delJ | J)), return Pi_J and reduced J_usage_mask", "Pi_J", ",", "r_J_usage_mask", "=", "self", ".", "compute_Pi_J", "(", "CDR3_seq", ",", "J_usage_mask", ")", "#Genomic V alignment/matching conditioned on J gene (contribution from P(V, J, delV)), return Pi_V_given_J", "Pi_V_given_J", ",", "max_V_align", "=", "self", ".", "compute_Pi_V_given_J", "(", "CDR3_seq", ",", "V_usage_mask", ",", "r_J_usage_mask", ")", "#Include insertions (R and PinsVJ) to get the total contribution from the left (3') side conditioned on J gene. Return Pi_V_insVJ_given_J", "Pi_V_insVJ_given_J", "=", "self", ".", "compute_Pi_V_insVJ_given_J", "(", "CDR3_seq", ",", "Pi_V_given_J", ",", "max_V_align", ")", "pgen", "=", "0", "#zip Pi_V_insVJ_given_J and Pi_J together for each J gene to get total pgen", "for", "j", "in", "range", "(", "len", "(", "r_J_usage_mask", ")", ")", ":", "for", "pos", "in", "range", "(", "len", "(", "CDR3_seq", ")", "*", "3", "-", "1", ")", ":", "pgen", "+=", "np", ".", "dot", "(", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "pos", "]", ",", "Pi_J", "[", "j", "]", "[", ":", ",", "pos", "+", "1", "]", ")", "return", "pgen" ]
Compute Pgen for CDR3 'amino acid' sequence CDR3_seq from VJ model. Conditioned on the already formatted V genes/alleles indicated in V_usage_mask and the J genes/alleles in J_usage_mask. Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation Returns ------- pgen : float The generation probability (Pgen) of the sequence Examples -------- >>> compute_CDR3_pgen('CAVKIQGAQKLVF', ppp, [72], [56]) 4.1818202431143785e-07 >>> compute_CDR3_pgen(nt2codon_rep('TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC'), ppp, [42], [1]) 1.3971676613008565e-08 >>> compute_CDR3_pgen('\xbb\xb6\xbe\x80\xbc\xa1\x8a\x96\xa1\xa0\xad\x8e\xbf', ppp, [72], [56]) 1.3971676613008565e-08
[ "Compute", "Pgen", "for", "CDR3", "amino", "acid", "sequence", "CDR3_seq", "from", "VJ", "model", ".", "Conditioned", "on", "the", "already", "formatted", "V", "genes", "/", "alleles", "indicated", "in", "V_usage_mask", "and", "the", "J", "genes", "/", "alleles", "in", "J_usage_mask", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "V_usage_mask", ":", "list", "Indices", "of", "the", "V", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "J_usage_mask", ":", "list", "Indices", "of", "the", "J", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "Returns", "-------", "pgen", ":", "float", "The", "generation", "probability", "(", "Pgen", ")", "of", "the", "sequence", "Examples", "--------", ">>>", "compute_CDR3_pgen", "(", "CAVKIQGAQKLVF", "ppp", "[", "72", "]", "[", "56", "]", ")", "4", ".", "1818202431143785e", "-", "07", ">>>", "compute_CDR3_pgen", "(", "nt2codon_rep", "(", "TGTGCCTGGAGTGTAGCTCCGGACAGGGGTGGCTACACCTTC", ")", "ppp", "[", "42", "]", "[", "1", "]", ")", "1", ".", "3971676613008565e", "-", "08", ">>>", "compute_CDR3_pgen", "(", "\\", "xbb", "\\", "xb6", "\\", "xbe", "\\", "x80", "\\", "xbc", "\\", "xa1", "\\", "x8a", "\\", "x96", "\\", "xa1", "\\", "xa0", "\\", "xad", "\\", "x8e", "\\", "xbf", "ppp", "[", "72", "]", "[", "56", "]", ")", "1", ".", "3971676613008565e", "-", "08" ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1630-L1676
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVJ.compute_Pi_V_given_J
def compute_Pi_V_given_J(self, CDR3_seq, V_usage_mask, J_usage_mask): """Compute Pi_V conditioned on J. This function returns the Pi array from the model factors of the V genomic contributions, P(V, J)*P(delV|V). This corresponds to V(J)_{x_1}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation self.cutV_genomic_CDR3_segs : list of strings List of all the V genomic nucleotide sequences trimmed to begin at the conserved C residue and with the maximum number of palindromic insertions appended. self.PVdelV_nt_pos_vec : list of ndarrays For each V allele, format P(delV|V) into the correct form for a Pi array or V(J)_{x_1}. This is only done for the first and last position in each codon. self.PVdelV_2nd_nt_pos_per_aa_vec : list of dicts For each V allele, and each 'amino acid', format P(V)*P(delV|V) for positions in the middle of a codon into the correct form for a Pi array or V(J)_{x_1} given the 'amino acid'. self.PVJ : ndarray Joint probability distribution of V and J, P(V, J). Returns ------- Pi_V_given_J : list List of (4, 3L) ndarrays corresponding to V(J)_{x_1}. max_V_align: int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. """ #Note, the cutV_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at the conserved C Pi_V_given_J = [np.zeros((4, len(CDR3_seq)*3)) for i in J_usage_mask] #Holds the aggregate weight for each nt possiblity and position alignment_lengths = [] for V_in in V_usage_mask: try: cutV_gen_seg = self.cutV_genomic_CDR3_segs[V_in] except IndexError: print 'Check provided V usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_left(CDR3_seq, cutV_gen_seg) alignment_lengths += [current_alignment_length] current_Pi_V = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PVdelV_nt_pos_vec current_Pi_V[:, :current_alignment_length] = self.PVdelV_nt_pos_vec[V_in][:, :current_alignment_length] for pos in range(1, current_alignment_length, 3): #for middle nt use PVdelV_2nd_nt_pos_per_aa_vec current_Pi_V[:, pos] = self.PVdelV_2nd_nt_pos_per_aa_vec[V_in][CDR3_seq[pos/3]][:, pos] for j, J_in in enumerate(J_usage_mask): Pi_V_given_J[j][:, :current_alignment_length] += self.PVJ[V_in, J_in]*current_Pi_V[:, :current_alignment_length] return Pi_V_given_J, max(alignment_lengths)
python
def compute_Pi_V_given_J(self, CDR3_seq, V_usage_mask, J_usage_mask): """Compute Pi_V conditioned on J. This function returns the Pi array from the model factors of the V genomic contributions, P(V, J)*P(delV|V). This corresponds to V(J)_{x_1}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation self.cutV_genomic_CDR3_segs : list of strings List of all the V genomic nucleotide sequences trimmed to begin at the conserved C residue and with the maximum number of palindromic insertions appended. self.PVdelV_nt_pos_vec : list of ndarrays For each V allele, format P(delV|V) into the correct form for a Pi array or V(J)_{x_1}. This is only done for the first and last position in each codon. self.PVdelV_2nd_nt_pos_per_aa_vec : list of dicts For each V allele, and each 'amino acid', format P(V)*P(delV|V) for positions in the middle of a codon into the correct form for a Pi array or V(J)_{x_1} given the 'amino acid'. self.PVJ : ndarray Joint probability distribution of V and J, P(V, J). Returns ------- Pi_V_given_J : list List of (4, 3L) ndarrays corresponding to V(J)_{x_1}. max_V_align: int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. """ #Note, the cutV_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at the conserved C Pi_V_given_J = [np.zeros((4, len(CDR3_seq)*3)) for i in J_usage_mask] #Holds the aggregate weight for each nt possiblity and position alignment_lengths = [] for V_in in V_usage_mask: try: cutV_gen_seg = self.cutV_genomic_CDR3_segs[V_in] except IndexError: print 'Check provided V usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_left(CDR3_seq, cutV_gen_seg) alignment_lengths += [current_alignment_length] current_Pi_V = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PVdelV_nt_pos_vec current_Pi_V[:, :current_alignment_length] = self.PVdelV_nt_pos_vec[V_in][:, :current_alignment_length] for pos in range(1, current_alignment_length, 3): #for middle nt use PVdelV_2nd_nt_pos_per_aa_vec current_Pi_V[:, pos] = self.PVdelV_2nd_nt_pos_per_aa_vec[V_in][CDR3_seq[pos/3]][:, pos] for j, J_in in enumerate(J_usage_mask): Pi_V_given_J[j][:, :current_alignment_length] += self.PVJ[V_in, J_in]*current_Pi_V[:, :current_alignment_length] return Pi_V_given_J, max(alignment_lengths)
[ "def", "compute_Pi_V_given_J", "(", "self", ",", "CDR3_seq", ",", "V_usage_mask", ",", "J_usage_mask", ")", ":", "#Note, the cutV_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template.", "#furthermore, the genomic sequence should be pruned to start at the conserved C", "Pi_V_given_J", "=", "[", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "for", "i", "in", "J_usage_mask", "]", "#Holds the aggregate weight for each nt possiblity and position", "alignment_lengths", "=", "[", "]", "for", "V_in", "in", "V_usage_mask", ":", "try", ":", "cutV_gen_seg", "=", "self", ".", "cutV_genomic_CDR3_segs", "[", "V_in", "]", "except", "IndexError", ":", "print", "'Check provided V usage mask. Contains indicies out of allowed range.'", "continue", "current_alignment_length", "=", "self", ".", "max_nt_to_aa_alignment_left", "(", "CDR3_seq", ",", "cutV_gen_seg", ")", "alignment_lengths", "+=", "[", "current_alignment_length", "]", "current_Pi_V", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "if", "current_alignment_length", ">", "0", ":", "#For first and last nt in a codon use PVdelV_nt_pos_vec", "current_Pi_V", "[", ":", ",", ":", "current_alignment_length", "]", "=", "self", ".", "PVdelV_nt_pos_vec", "[", "V_in", "]", "[", ":", ",", ":", "current_alignment_length", "]", "for", "pos", "in", "range", "(", "1", ",", "current_alignment_length", ",", "3", ")", ":", "#for middle nt use PVdelV_2nd_nt_pos_per_aa_vec", "current_Pi_V", "[", ":", ",", "pos", "]", "=", "self", ".", "PVdelV_2nd_nt_pos_per_aa_vec", "[", "V_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "pos", "]", "for", "j", ",", "J_in", "in", "enumerate", "(", "J_usage_mask", ")", ":", "Pi_V_given_J", "[", "j", "]", "[", ":", ",", ":", "current_alignment_length", "]", "+=", "self", ".", "PVJ", "[", "V_in", ",", "J_in", "]", "*", "current_Pi_V", "[", ":", ",", ":", "current_alignment_length", "]", "return", "Pi_V_given_J", ",", "max", "(", "alignment_lengths", ")" ]
Compute Pi_V conditioned on J. This function returns the Pi array from the model factors of the V genomic contributions, P(V, J)*P(delV|V). This corresponds to V(J)_{x_1}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). V_usage_mask : list Indices of the V alleles to be considered in the Pgen computation J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation self.cutV_genomic_CDR3_segs : list of strings List of all the V genomic nucleotide sequences trimmed to begin at the conserved C residue and with the maximum number of palindromic insertions appended. self.PVdelV_nt_pos_vec : list of ndarrays For each V allele, format P(delV|V) into the correct form for a Pi array or V(J)_{x_1}. This is only done for the first and last position in each codon. self.PVdelV_2nd_nt_pos_per_aa_vec : list of dicts For each V allele, and each 'amino acid', format P(V)*P(delV|V) for positions in the middle of a codon into the correct form for a Pi array or V(J)_{x_1} given the 'amino acid'. self.PVJ : ndarray Joint probability distribution of V and J, P(V, J). Returns ------- Pi_V_given_J : list List of (4, 3L) ndarrays corresponding to V(J)_{x_1}. max_V_align: int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask.
[ "Compute", "Pi_V", "conditioned", "on", "J", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "V", "genomic", "contributions", "P", "(", "V", "J", ")", "*", "P", "(", "delV|V", ")", ".", "This", "corresponds", "to", "V", "(", "J", ")", "_", "{", "x_1", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "V_usage_mask", ":", "list", "Indices", "of", "the", "V", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "J_usage_mask", ":", "list", "Indices", "of", "the", "J", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "self", ".", "cutV_genomic_CDR3_segs", ":", "list", "of", "strings", "List", "of", "all", "the", "V", "genomic", "nucleotide", "sequences", "trimmed", "to", "begin", "at", "the", "conserved", "C", "residue", "and", "with", "the", "maximum", "number", "of", "palindromic", "insertions", "appended", ".", "self", ".", "PVdelV_nt_pos_vec", ":", "list", "of", "ndarrays", "For", "each", "V", "allele", "format", "P", "(", "delV|V", ")", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "V", "(", "J", ")", "_", "{", "x_1", "}", ".", "This", "is", "only", "done", "for", "the", "first", "and", "last", "position", "in", "each", "codon", ".", "self", ".", "PVdelV_2nd_nt_pos_per_aa_vec", ":", "list", "of", "dicts", "For", "each", "V", "allele", "and", "each", "amino", "acid", "format", "P", "(", "V", ")", "*", "P", "(", "delV|V", ")", "for", "positions", "in", "the", "middle", "of", "a", "codon", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "V", "(", "J", ")", "_", "{", "x_1", "}", "given", "the", "amino", "acid", ".", "self", ".", "PVJ", ":", "ndarray", "Joint", "probability", "distribution", "of", "V", "and", "J", "P", "(", "V", "J", ")", ".", "Returns", "-------", "Pi_V_given_J", ":", "list", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "V", "(", "J", ")", "_", "{", "x_1", "}", ".", "max_V_align", ":", "int", "Maximum", "alignment", "of", "the", "CDR3_seq", "to", "any", "genomic", "V", "allele", "allowed", "by", "V_usage_mask", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1679-L1746
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVJ.compute_Pi_V_insVJ_given_J
def compute_Pi_V_insVJ_given_J(self, CDR3_seq, Pi_V_given_J, max_V_align): """Compute Pi_V_insVJ conditioned on J. This function returns the Pi array from the model factors of the V genomic contributions, P(V, J)*P(delV|V), and the VJ (N) insertions, first_nt_bias_insVJ(m_1)PinsVJ(\ell_{VJ})\prod_{i=2}^{\ell_{VJ}}Rvj(m_i|m_{i-1}). This corresponds to V(J)_{x_1}{M^{x_1}}_{x_2}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_V_given_J : ndarray List of (4, 3L) ndarrays corresponding to V(J)_{x_1}. max_V_align : int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. self.PinsVJ : ndarray Probability distribution of the VJ insertion sequence length self.first_nt_bias_insVJ : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the VJ junction. self.zero_nt_bias_insVJ : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the VJ insertion. zero_nt_bias_insVJ = Rvj^{-1}first_nt_bias_insVJ self.Tvj : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Svj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the first position. self.Dvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.lTvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion starting in the first position. self.lDvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for VD insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_V_insVJ_given_J : list List of (4, 3L) ndarrays corresponding to V(J)_{x_1}{M^{x_1}}_{x_2}. """ #max_insertions = 30 #len(PinsVJ) - 1 should zeropad the last few spots max_insertions = len(self.PinsVJ) - 1 Pi_V_insVJ_given_J = [np.zeros((4, len(CDR3_seq)*3)) for i in range(len(Pi_V_given_J))] for j in range(len(Pi_V_given_J)): #start position is first nt in a codon for init_pos in range(0, max_V_align, 3): #Zero insertions Pi_V_insVJ_given_J[j][:, init_pos] += self.PinsVJ[0]*Pi_V_given_J[j][:, init_pos] #One insertion Pi_V_insVJ_given_J[j][:, init_pos+1] += self.PinsVJ[1]*np.dot(self.lDvj[CDR3_seq[init_pos/3]], Pi_V_given_J[j][:, init_pos]) #Two insertions and compute the base nt vec for the standard loop current_base_nt_vec = np.dot(self.lTvj[CDR3_seq[init_pos/3]], Pi_V_given_J[j][:, init_pos]) Pi_V_insVJ_given_J[j][0, init_pos+2] += self.PinsVJ[2]*np.sum(current_base_nt_vec) base_ins = 2 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_V_insVJ_given_J[j][:, init_pos+base_ins+1] += self.PinsVJ[base_ins + 1]*np.dot(self.Svj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][:, init_pos+base_ins+2] += self.PinsVJ[base_ins + 2]*np.dot(self.Dvj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][0, init_pos+base_ins+3] += self.PinsVJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is second nt in a codon for init_pos in range(1, max_V_align, 3): #Zero insertions Pi_V_insVJ_given_J[j][:, init_pos] += self.PinsVJ[0]*Pi_V_given_J[j][:, init_pos] #One insertion --- we first compute our p vec by pairwise mult with the ss distr current_base_nt_vec = np.multiply(Pi_V_given_J[j][:, init_pos], self.first_nt_bias_insVJ) Pi_V_insVJ_given_J[j][0, init_pos+1] += self.PinsVJ[1]*np.sum(current_base_nt_vec) base_ins = 1 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_V_insVJ_given_J[j][:, init_pos+base_ins+1] += self.PinsVJ[base_ins + 1]*np.dot(self.Svj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][:, init_pos+base_ins+2] += self.PinsVJ[base_ins + 2]*np.dot(self.Dvj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][0, init_pos+base_ins+3] += self.PinsVJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is last nt in a codon for init_pos in range(2, max_V_align, 3): #Zero insertions Pi_V_insVJ_given_J[j][0, init_pos] += self.PinsVJ[0]*Pi_V_given_J[j][0, init_pos] #current_base_nt_vec = first_nt_bias_insVJ*Pi_V_given_J[j][0, init_pos] #Okay for steady state current_base_nt_vec = self.zero_nt_bias_insVJ*Pi_V_given_J[j][0, init_pos] base_ins = 0 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_V_insVJ_given_J[j][:, init_pos+base_ins+1] += self.PinsVJ[base_ins + 1]*np.dot(self.Svj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][:, init_pos+base_ins+2] += self.PinsVJ[base_ins + 2]*np.dot(self.Dvj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][0, init_pos+base_ins+3] += self.PinsVJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 return Pi_V_insVJ_given_J
python
def compute_Pi_V_insVJ_given_J(self, CDR3_seq, Pi_V_given_J, max_V_align): """Compute Pi_V_insVJ conditioned on J. This function returns the Pi array from the model factors of the V genomic contributions, P(V, J)*P(delV|V), and the VJ (N) insertions, first_nt_bias_insVJ(m_1)PinsVJ(\ell_{VJ})\prod_{i=2}^{\ell_{VJ}}Rvj(m_i|m_{i-1}). This corresponds to V(J)_{x_1}{M^{x_1}}_{x_2}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_V_given_J : ndarray List of (4, 3L) ndarrays corresponding to V(J)_{x_1}. max_V_align : int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. self.PinsVJ : ndarray Probability distribution of the VJ insertion sequence length self.first_nt_bias_insVJ : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the VJ junction. self.zero_nt_bias_insVJ : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the VJ insertion. zero_nt_bias_insVJ = Rvj^{-1}first_nt_bias_insVJ self.Tvj : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Svj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the first position. self.Dvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.lTvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion starting in the first position. self.lDvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for VD insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_V_insVJ_given_J : list List of (4, 3L) ndarrays corresponding to V(J)_{x_1}{M^{x_1}}_{x_2}. """ #max_insertions = 30 #len(PinsVJ) - 1 should zeropad the last few spots max_insertions = len(self.PinsVJ) - 1 Pi_V_insVJ_given_J = [np.zeros((4, len(CDR3_seq)*3)) for i in range(len(Pi_V_given_J))] for j in range(len(Pi_V_given_J)): #start position is first nt in a codon for init_pos in range(0, max_V_align, 3): #Zero insertions Pi_V_insVJ_given_J[j][:, init_pos] += self.PinsVJ[0]*Pi_V_given_J[j][:, init_pos] #One insertion Pi_V_insVJ_given_J[j][:, init_pos+1] += self.PinsVJ[1]*np.dot(self.lDvj[CDR3_seq[init_pos/3]], Pi_V_given_J[j][:, init_pos]) #Two insertions and compute the base nt vec for the standard loop current_base_nt_vec = np.dot(self.lTvj[CDR3_seq[init_pos/3]], Pi_V_given_J[j][:, init_pos]) Pi_V_insVJ_given_J[j][0, init_pos+2] += self.PinsVJ[2]*np.sum(current_base_nt_vec) base_ins = 2 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_V_insVJ_given_J[j][:, init_pos+base_ins+1] += self.PinsVJ[base_ins + 1]*np.dot(self.Svj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][:, init_pos+base_ins+2] += self.PinsVJ[base_ins + 2]*np.dot(self.Dvj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][0, init_pos+base_ins+3] += self.PinsVJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is second nt in a codon for init_pos in range(1, max_V_align, 3): #Zero insertions Pi_V_insVJ_given_J[j][:, init_pos] += self.PinsVJ[0]*Pi_V_given_J[j][:, init_pos] #One insertion --- we first compute our p vec by pairwise mult with the ss distr current_base_nt_vec = np.multiply(Pi_V_given_J[j][:, init_pos], self.first_nt_bias_insVJ) Pi_V_insVJ_given_J[j][0, init_pos+1] += self.PinsVJ[1]*np.sum(current_base_nt_vec) base_ins = 1 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_V_insVJ_given_J[j][:, init_pos+base_ins+1] += self.PinsVJ[base_ins + 1]*np.dot(self.Svj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][:, init_pos+base_ins+2] += self.PinsVJ[base_ins + 2]*np.dot(self.Dvj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][0, init_pos+base_ins+3] += self.PinsVJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 #start position is last nt in a codon for init_pos in range(2, max_V_align, 3): #Zero insertions Pi_V_insVJ_given_J[j][0, init_pos] += self.PinsVJ[0]*Pi_V_given_J[j][0, init_pos] #current_base_nt_vec = first_nt_bias_insVJ*Pi_V_given_J[j][0, init_pos] #Okay for steady state current_base_nt_vec = self.zero_nt_bias_insVJ*Pi_V_given_J[j][0, init_pos] base_ins = 0 #Loop over all other insertions using base_nt_vec for aa in CDR3_seq[init_pos/3 + 1: init_pos/3 + max_insertions/3]: Pi_V_insVJ_given_J[j][:, init_pos+base_ins+1] += self.PinsVJ[base_ins + 1]*np.dot(self.Svj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][:, init_pos+base_ins+2] += self.PinsVJ[base_ins + 2]*np.dot(self.Dvj[aa], current_base_nt_vec) current_base_nt_vec = np.dot(self.Tvj[aa], current_base_nt_vec) Pi_V_insVJ_given_J[j][0, init_pos+base_ins+3] += self.PinsVJ[base_ins + 3]*np.sum(current_base_nt_vec) base_ins +=3 return Pi_V_insVJ_given_J
[ "def", "compute_Pi_V_insVJ_given_J", "(", "self", ",", "CDR3_seq", ",", "Pi_V_given_J", ",", "max_V_align", ")", ":", "#max_insertions = 30 #len(PinsVJ) - 1 should zeropad the last few spots", "max_insertions", "=", "len", "(", "self", ".", "PinsVJ", ")", "-", "1", "Pi_V_insVJ_given_J", "=", "[", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "for", "i", "in", "range", "(", "len", "(", "Pi_V_given_J", ")", ")", "]", "for", "j", "in", "range", "(", "len", "(", "Pi_V_given_J", ")", ")", ":", "#start position is first nt in a codon", "for", "init_pos", "in", "range", "(", "0", ",", "max_V_align", ",", "3", ")", ":", "#Zero insertions", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", "+=", "self", ".", "PinsVJ", "[", "0", "]", "*", "Pi_V_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", "#One insertion", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "1", "]", "+=", "self", ".", "PinsVJ", "[", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "lDvj", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ",", "Pi_V_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", ")", "#Two insertions and compute the base nt vec for the standard loop ", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "lTvj", "[", "CDR3_seq", "[", "init_pos", "/", "3", "]", "]", ",", "Pi_V_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "+", "2", "]", "+=", "self", ".", "PinsVJ", "[", "2", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "=", "2", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "+", "1", ":", "init_pos", "/", "3", "+", "max_insertions", "/", "3", "]", ":", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "base_ins", "+", "1", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Svj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "base_ins", "+", "2", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Dvj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tvj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "+", "base_ins", "+", "3", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "#start position is second nt in a codon", "for", "init_pos", "in", "range", "(", "1", ",", "max_V_align", ",", "3", ")", ":", "#Zero insertions", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", "+=", "self", ".", "PinsVJ", "[", "0", "]", "*", "Pi_V_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", "#One insertion --- we first compute our p vec by pairwise mult with the ss distr", "current_base_nt_vec", "=", "np", ".", "multiply", "(", "Pi_V_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "]", ",", "self", ".", "first_nt_bias_insVJ", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "+", "1", "]", "+=", "self", ".", "PinsVJ", "[", "1", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "=", "1", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "+", "1", ":", "init_pos", "/", "3", "+", "max_insertions", "/", "3", "]", ":", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "base_ins", "+", "1", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Svj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "base_ins", "+", "2", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Dvj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tvj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "+", "base_ins", "+", "3", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "#start position is last nt in a codon ", "for", "init_pos", "in", "range", "(", "2", ",", "max_V_align", ",", "3", ")", ":", "#Zero insertions", "Pi_V_insVJ_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "]", "+=", "self", ".", "PinsVJ", "[", "0", "]", "*", "Pi_V_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "]", "#current_base_nt_vec = first_nt_bias_insVJ*Pi_V_given_J[j][0, init_pos] #Okay for steady state", "current_base_nt_vec", "=", "self", ".", "zero_nt_bias_insVJ", "*", "Pi_V_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "]", "base_ins", "=", "0", "#Loop over all other insertions using base_nt_vec", "for", "aa", "in", "CDR3_seq", "[", "init_pos", "/", "3", "+", "1", ":", "init_pos", "/", "3", "+", "max_insertions", "/", "3", "]", ":", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "base_ins", "+", "1", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "1", "]", "*", "np", ".", "dot", "(", "self", ".", "Svj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", ":", ",", "init_pos", "+", "base_ins", "+", "2", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "2", "]", "*", "np", ".", "dot", "(", "self", ".", "Dvj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "current_base_nt_vec", "=", "np", ".", "dot", "(", "self", ".", "Tvj", "[", "aa", "]", ",", "current_base_nt_vec", ")", "Pi_V_insVJ_given_J", "[", "j", "]", "[", "0", ",", "init_pos", "+", "base_ins", "+", "3", "]", "+=", "self", ".", "PinsVJ", "[", "base_ins", "+", "3", "]", "*", "np", ".", "sum", "(", "current_base_nt_vec", ")", "base_ins", "+=", "3", "return", "Pi_V_insVJ_given_J" ]
Compute Pi_V_insVJ conditioned on J. This function returns the Pi array from the model factors of the V genomic contributions, P(V, J)*P(delV|V), and the VJ (N) insertions, first_nt_bias_insVJ(m_1)PinsVJ(\ell_{VJ})\prod_{i=2}^{\ell_{VJ}}Rvj(m_i|m_{i-1}). This corresponds to V(J)_{x_1}{M^{x_1}}_{x_2}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). Pi_V_given_J : ndarray List of (4, 3L) ndarrays corresponding to V(J)_{x_1}. max_V_align : int Maximum alignment of the CDR3_seq to any genomic V allele allowed by V_usage_mask. self.PinsVJ : ndarray Probability distribution of the VJ insertion sequence length self.first_nt_bias_insVJ : ndarray (4,) array of the probability distribution of the indentity of the first nucleotide insertion for the VJ junction. self.zero_nt_bias_insVJ : ndarray (4,) array of the probability distribution of the indentity of the the nucleotide BEFORE the VJ insertion. zero_nt_bias_insVJ = Rvj^{-1}first_nt_bias_insVJ self.Tvj : dict Dictionary of full codon transfer matrices ((4, 4) ndarrays) by 'amino acid'. self.Svj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the first position. self.Dvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion ending in the second position. self.lTvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for the VD insertion starting in the first position. self.lDvj : dict Dictionary of transfer matrices ((4, 4) ndarrays) by 'amino acid' for VD insertion starting in the first position and ending in the second position of the same codon. Returns ------- Pi_V_insVJ_given_J : list List of (4, 3L) ndarrays corresponding to V(J)_{x_1}{M^{x_1}}_{x_2}.
[ "Compute", "Pi_V_insVJ", "conditioned", "on", "J", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "V", "genomic", "contributions", "P", "(", "V", "J", ")", "*", "P", "(", "delV|V", ")", "and", "the", "VJ", "(", "N", ")", "insertions", "first_nt_bias_insVJ", "(", "m_1", ")", "PinsVJ", "(", "\\", "ell_", "{", "VJ", "}", ")", "\\", "prod_", "{", "i", "=", "2", "}", "^", "{", "\\", "ell_", "{", "VJ", "}}", "Rvj", "(", "m_i|m_", "{", "i", "-", "1", "}", ")", ".", "This", "corresponds", "to", "V", "(", "J", ")", "_", "{", "x_1", "}", "{", "M^", "{", "x_1", "}}", "_", "{", "x_2", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "Pi_V_given_J", ":", "ndarray", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "V", "(", "J", ")", "_", "{", "x_1", "}", ".", "max_V_align", ":", "int", "Maximum", "alignment", "of", "the", "CDR3_seq", "to", "any", "genomic", "V", "allele", "allowed", "by", "V_usage_mask", ".", "self", ".", "PinsVJ", ":", "ndarray", "Probability", "distribution", "of", "the", "VJ", "insertion", "sequence", "length", "self", ".", "first_nt_bias_insVJ", ":", "ndarray", "(", "4", ")", "array", "of", "the", "probability", "distribution", "of", "the", "indentity", "of", "the", "first", "nucleotide", "insertion", "for", "the", "VJ", "junction", ".", "self", ".", "zero_nt_bias_insVJ", ":", "ndarray", "(", "4", ")", "array", "of", "the", "probability", "distribution", "of", "the", "indentity", "of", "the", "the", "nucleotide", "BEFORE", "the", "VJ", "insertion", ".", "zero_nt_bias_insVJ", "=", "Rvj^", "{", "-", "1", "}", "first_nt_bias_insVJ", "self", ".", "Tvj", ":", "dict", "Dictionary", "of", "full", "codon", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", ".", "self", ".", "Svj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "ending", "in", "the", "first", "position", ".", "self", ".", "Dvj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "ending", "in", "the", "second", "position", ".", "self", ".", "lTvj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "the", "VD", "insertion", "starting", "in", "the", "first", "position", ".", "self", ".", "lDvj", ":", "dict", "Dictionary", "of", "transfer", "matrices", "((", "4", "4", ")", "ndarrays", ")", "by", "amino", "acid", "for", "VD", "insertion", "starting", "in", "the", "first", "position", "and", "ending", "in", "the", "second", "position", "of", "the", "same", "codon", ".", "Returns", "-------", "Pi_V_insVJ_given_J", ":", "list", "List", "of", "(", "4", "3L", ")", "ndarrays", "corresponding", "to", "V", "(", "J", ")", "_", "{", "x_1", "}", "{", "M^", "{", "x_1", "}}", "_", "{", "x_2", "}", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1749-L1868
train
zsethna/OLGA
olga/generation_probability.py
GenerationProbabilityVJ.compute_Pi_J
def compute_Pi_J(self, CDR3_seq, J_usage_mask): """Compute Pi_J. This function returns the Pi array from the model factors of the J genomic contributions, P(delJ|J). This corresponds to J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation self.cutJ_genomic_CDR3_segs : list of strings List of all the J genomic nucleotide sequences trimmed to begin at the conserved 3' residue (F/W) and with the maximum number of palindromic insertions appended. self.PJdelJ_nt_pos_vec : list of ndarrays For each J allele, format P(delJ|J) into the correct form for a Pi array or J^{x_2}. This is only done for the first and last position in each codon. self.PJdelJ_2nd_nt_pos_per_aa_vec : list of dicts For each J allele, and each 'amino acid', format P(delJ|J) for positions in the middle of a codon into the correct form for a Pi array or J^{x_2} given the 'amino acid'. Returns ------- Pi_J : ndarray (4, 3L) array corresponding to J^{x_4}. r_J_usage_mask: list Reduced J_usage mask. J genes/alleles with no contribution (bad alignment) are removed from the mask. This is done to speed up the computation on the V side (which must be done conditioned on the J). """ #Note, the cutJ_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at a conserved region on the J side Pi_J = [] #Holds the aggregate weight for each nt possiblity and position r_J_usage_mask = [] for j, J_in in enumerate(J_usage_mask): try: cutJ_gen_seg = self.cutJ_genomic_CDR3_segs[J_in] except IndexError: print 'Check provided J usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_right(CDR3_seq, cutJ_gen_seg) #alignment_lengths += [current_alignment_length] current_Pi_J = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PJdelJ_nt_pos_vec current_Pi_J[:, -current_alignment_length:] = self.PJdelJ_nt_pos_vec[J_in][:, -current_alignment_length:] for pos in range(-2, -current_alignment_length-1, -3): #for middle nt use PJdelJ_2nd_nt_pos_per_aa_vec current_Pi_J[:, pos] = self.PJdelJ_2nd_nt_pos_per_aa_vec[J_in][CDR3_seq[pos/3]][:, pos] if np.sum(current_Pi_J) > 0: Pi_J.append(current_Pi_J) r_J_usage_mask.append(J_in) return Pi_J, r_J_usage_mask
python
def compute_Pi_J(self, CDR3_seq, J_usage_mask): """Compute Pi_J. This function returns the Pi array from the model factors of the J genomic contributions, P(delJ|J). This corresponds to J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation self.cutJ_genomic_CDR3_segs : list of strings List of all the J genomic nucleotide sequences trimmed to begin at the conserved 3' residue (F/W) and with the maximum number of palindromic insertions appended. self.PJdelJ_nt_pos_vec : list of ndarrays For each J allele, format P(delJ|J) into the correct form for a Pi array or J^{x_2}. This is only done for the first and last position in each codon. self.PJdelJ_2nd_nt_pos_per_aa_vec : list of dicts For each J allele, and each 'amino acid', format P(delJ|J) for positions in the middle of a codon into the correct form for a Pi array or J^{x_2} given the 'amino acid'. Returns ------- Pi_J : ndarray (4, 3L) array corresponding to J^{x_4}. r_J_usage_mask: list Reduced J_usage mask. J genes/alleles with no contribution (bad alignment) are removed from the mask. This is done to speed up the computation on the V side (which must be done conditioned on the J). """ #Note, the cutJ_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template. #furthermore, the genomic sequence should be pruned to start at a conserved region on the J side Pi_J = [] #Holds the aggregate weight for each nt possiblity and position r_J_usage_mask = [] for j, J_in in enumerate(J_usage_mask): try: cutJ_gen_seg = self.cutJ_genomic_CDR3_segs[J_in] except IndexError: print 'Check provided J usage mask. Contains indicies out of allowed range.' continue current_alignment_length = self.max_nt_to_aa_alignment_right(CDR3_seq, cutJ_gen_seg) #alignment_lengths += [current_alignment_length] current_Pi_J = np.zeros((4, len(CDR3_seq)*3)) if current_alignment_length > 0: #For first and last nt in a codon use PJdelJ_nt_pos_vec current_Pi_J[:, -current_alignment_length:] = self.PJdelJ_nt_pos_vec[J_in][:, -current_alignment_length:] for pos in range(-2, -current_alignment_length-1, -3): #for middle nt use PJdelJ_2nd_nt_pos_per_aa_vec current_Pi_J[:, pos] = self.PJdelJ_2nd_nt_pos_per_aa_vec[J_in][CDR3_seq[pos/3]][:, pos] if np.sum(current_Pi_J) > 0: Pi_J.append(current_Pi_J) r_J_usage_mask.append(J_in) return Pi_J, r_J_usage_mask
[ "def", "compute_Pi_J", "(", "self", ",", "CDR3_seq", ",", "J_usage_mask", ")", ":", "#Note, the cutJ_genomic_CDR3_segs INCLUDE the palindromic insertions and thus are max_palindrome nts longer than the template.", "#furthermore, the genomic sequence should be pruned to start at a conserved region on the J side", "Pi_J", "=", "[", "]", "#Holds the aggregate weight for each nt possiblity and position", "r_J_usage_mask", "=", "[", "]", "for", "j", ",", "J_in", "in", "enumerate", "(", "J_usage_mask", ")", ":", "try", ":", "cutJ_gen_seg", "=", "self", ".", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", "except", "IndexError", ":", "print", "'Check provided J usage mask. Contains indicies out of allowed range.'", "continue", "current_alignment_length", "=", "self", ".", "max_nt_to_aa_alignment_right", "(", "CDR3_seq", ",", "cutJ_gen_seg", ")", "#alignment_lengths += [current_alignment_length]", "current_Pi_J", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "CDR3_seq", ")", "*", "3", ")", ")", "if", "current_alignment_length", ">", "0", ":", "#For first and last nt in a codon use PJdelJ_nt_pos_vec", "current_Pi_J", "[", ":", ",", "-", "current_alignment_length", ":", "]", "=", "self", ".", "PJdelJ_nt_pos_vec", "[", "J_in", "]", "[", ":", ",", "-", "current_alignment_length", ":", "]", "for", "pos", "in", "range", "(", "-", "2", ",", "-", "current_alignment_length", "-", "1", ",", "-", "3", ")", ":", "#for middle nt use PJdelJ_2nd_nt_pos_per_aa_vec", "current_Pi_J", "[", ":", ",", "pos", "]", "=", "self", ".", "PJdelJ_2nd_nt_pos_per_aa_vec", "[", "J_in", "]", "[", "CDR3_seq", "[", "pos", "/", "3", "]", "]", "[", ":", ",", "pos", "]", "if", "np", ".", "sum", "(", "current_Pi_J", ")", ">", "0", ":", "Pi_J", ".", "append", "(", "current_Pi_J", ")", "r_J_usage_mask", ".", "append", "(", "J_in", ")", "return", "Pi_J", ",", "r_J_usage_mask" ]
Compute Pi_J. This function returns the Pi array from the model factors of the J genomic contributions, P(delJ|J). This corresponds to J(D)^{x_4}. For clarity in parsing the algorithm implementation, we include which instance attributes are used in the method as 'parameters.' Parameters ---------- CDR3_seq : str CDR3 sequence composed of 'amino acids' (single character symbols each corresponding to a collection of codons as given by codons_dict). J_usage_mask : list Indices of the J alleles to be considered in the Pgen computation self.cutJ_genomic_CDR3_segs : list of strings List of all the J genomic nucleotide sequences trimmed to begin at the conserved 3' residue (F/W) and with the maximum number of palindromic insertions appended. self.PJdelJ_nt_pos_vec : list of ndarrays For each J allele, format P(delJ|J) into the correct form for a Pi array or J^{x_2}. This is only done for the first and last position in each codon. self.PJdelJ_2nd_nt_pos_per_aa_vec : list of dicts For each J allele, and each 'amino acid', format P(delJ|J) for positions in the middle of a codon into the correct form for a Pi array or J^{x_2} given the 'amino acid'. Returns ------- Pi_J : ndarray (4, 3L) array corresponding to J^{x_4}. r_J_usage_mask: list Reduced J_usage mask. J genes/alleles with no contribution (bad alignment) are removed from the mask. This is done to speed up the computation on the V side (which must be done conditioned on the J).
[ "Compute", "Pi_J", ".", "This", "function", "returns", "the", "Pi", "array", "from", "the", "model", "factors", "of", "the", "J", "genomic", "contributions", "P", "(", "delJ|J", ")", ".", "This", "corresponds", "to", "J", "(", "D", ")", "^", "{", "x_4", "}", ".", "For", "clarity", "in", "parsing", "the", "algorithm", "implementation", "we", "include", "which", "instance", "attributes", "are", "used", "in", "the", "method", "as", "parameters", ".", "Parameters", "----------", "CDR3_seq", ":", "str", "CDR3", "sequence", "composed", "of", "amino", "acids", "(", "single", "character", "symbols", "each", "corresponding", "to", "a", "collection", "of", "codons", "as", "given", "by", "codons_dict", ")", ".", "J_usage_mask", ":", "list", "Indices", "of", "the", "J", "alleles", "to", "be", "considered", "in", "the", "Pgen", "computation", "self", ".", "cutJ_genomic_CDR3_segs", ":", "list", "of", "strings", "List", "of", "all", "the", "J", "genomic", "nucleotide", "sequences", "trimmed", "to", "begin", "at", "the", "conserved", "3", "residue", "(", "F", "/", "W", ")", "and", "with", "the", "maximum", "number", "of", "palindromic", "insertions", "appended", ".", "self", ".", "PJdelJ_nt_pos_vec", ":", "list", "of", "ndarrays", "For", "each", "J", "allele", "format", "P", "(", "delJ|J", ")", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "J^", "{", "x_2", "}", ".", "This", "is", "only", "done", "for", "the", "first", "and", "last", "position", "in", "each", "codon", ".", "self", ".", "PJdelJ_2nd_nt_pos_per_aa_vec", ":", "list", "of", "dicts", "For", "each", "J", "allele", "and", "each", "amino", "acid", "format", "P", "(", "delJ|J", ")", "for", "positions", "in", "the", "middle", "of", "a", "codon", "into", "the", "correct", "form", "for", "a", "Pi", "array", "or", "J^", "{", "x_2", "}", "given", "the", "amino", "acid", ".", "Returns", "-------", "Pi_J", ":", "ndarray", "(", "4", "3L", ")", "array", "corresponding", "to", "J^", "{", "x_4", "}", ".", "r_J_usage_mask", ":", "list", "Reduced", "J_usage", "mask", ".", "J", "genes", "/", "alleles", "with", "no", "contribution", "(", "bad", "alignment", ")", "are", "removed", "from", "the", "mask", ".", "This", "is", "done", "to", "speed", "up", "the", "computation", "on", "the", "V", "side", "(", "which", "must", "be", "done", "conditioned", "on", "the", "J", ")", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/generation_probability.py#L1871-L1935
train
tansey/gfl
pygfl/trendfiltering.py
TrendFilteringSolver.solve
def solve(self, lam): '''Solves the GFL for a fixed value of lambda.''' s = weighted_graphtf(self.nnodes, self.y, self.weights, lam, self.Dk.shape[0], self.Dk.shape[1], self.Dk.nnz, self.Dk.row.astype('int32'), self.Dk.col.astype('int32'), self.Dk.data.astype('double'), self.maxsteps, self.converge, self.beta, self.u) self.steps.append(s) return self.beta
python
def solve(self, lam): '''Solves the GFL for a fixed value of lambda.''' s = weighted_graphtf(self.nnodes, self.y, self.weights, lam, self.Dk.shape[0], self.Dk.shape[1], self.Dk.nnz, self.Dk.row.astype('int32'), self.Dk.col.astype('int32'), self.Dk.data.astype('double'), self.maxsteps, self.converge, self.beta, self.u) self.steps.append(s) return self.beta
[ "def", "solve", "(", "self", ",", "lam", ")", ":", "s", "=", "weighted_graphtf", "(", "self", ".", "nnodes", ",", "self", ".", "y", ",", "self", ".", "weights", ",", "lam", ",", "self", ".", "Dk", ".", "shape", "[", "0", "]", ",", "self", ".", "Dk", ".", "shape", "[", "1", "]", ",", "self", ".", "Dk", ".", "nnz", ",", "self", ".", "Dk", ".", "row", ".", "astype", "(", "'int32'", ")", ",", "self", ".", "Dk", ".", "col", ".", "astype", "(", "'int32'", ")", ",", "self", ".", "Dk", ".", "data", ".", "astype", "(", "'double'", ")", ",", "self", ".", "maxsteps", ",", "self", ".", "converge", ",", "self", ".", "beta", ",", "self", ".", "u", ")", "self", ".", "steps", ".", "append", "(", "s", ")", "return", "self", ".", "beta" ]
Solves the GFL for a fixed value of lambda.
[ "Solves", "the", "GFL", "for", "a", "fixed", "value", "of", "lambda", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trendfiltering.py#L73-L81
train
tansey/gfl
pygfl/trendfiltering.py
TrendFilteringSolver.solution_path
def solution_path(self, min_lambda, max_lambda, lambda_bins, verbose=0): '''Follows the solution path to find the best lambda value.''' self.u = np.zeros(self.Dk.shape[0], dtype='double') lambda_grid = np.exp(np.linspace(np.log(max_lambda), np.log(min_lambda), lambda_bins)) aic_trace = np.zeros(lambda_grid.shape) # The AIC score for each lambda value aicc_trace = np.zeros(lambda_grid.shape) # The AICc score for each lambda value (correcting for finite sample size) bic_trace = np.zeros(lambda_grid.shape) # The BIC score for each lambda value dof_trace = np.zeros(lambda_grid.shape) # The degrees of freedom of each final solution log_likelihood_trace = np.zeros(lambda_grid.shape) beta_trace = [] best_idx = None best_plateaus = None if self.edges is None: self.edges = defaultdict(list) elist = csr_matrix(self.D).indices.reshape((self.D.shape[0], 2)) for n1, n2 in elist: self.edges[n1].append(n2) self.edges[n2].append(n1) # Solve the series of lambda values with warm starts at each point for i, lam in enumerate(lambda_grid): if verbose: print('#{0} Lambda = {1}'.format(i, lam)) # Fit to the final values beta = self.solve(lam) if verbose: print('Calculating degrees of freedom') # Count the number of free parameters in the grid (dof) -- TODO: the graph trend filtering paper seems to imply we shouldn't multiply by (k+1)? dof_vals = self.Dk_minus_one.dot(beta) if self.k > 0 else beta plateaus = calc_plateaus(dof_vals, self.edges, rel_tol=0.01) if (self.k % 2) == 0 else nearly_unique(dof_vals, rel_tol=0.03) dof_trace[i] = max(1,len(plateaus)) #* (k+1) if verbose: print('Calculating Information Criteria') # Get the negative log-likelihood log_likelihood_trace[i] = -0.5 * ((self.y - beta)**2).sum() # Calculate AIC = 2k - 2ln(L) aic_trace[i] = 2. * dof_trace[i] - 2. * log_likelihood_trace[i] # Calculate AICc = AIC + 2k * (k+1) / (n - k - 1) aicc_trace[i] = aic_trace[i] + 2 * dof_trace[i] * (dof_trace[i]+1) / (len(beta) - dof_trace[i] - 1.) # Calculate BIC = -2ln(L) + k * (ln(n) - ln(2pi)) bic_trace[i] = -2 * log_likelihood_trace[i] + dof_trace[i] * (np.log(len(beta)) - np.log(2 * np.pi)) # Track the best model thus far if best_idx is None or bic_trace[i] < bic_trace[best_idx]: best_idx = i best_plateaus = plateaus # Save the trace of all the resulting parameters beta_trace.append(np.array(beta)) if verbose: print('DoF: {0} AIC: {1} AICc: {2} BIC: {3}\n'.format(dof_trace[i], aic_trace[i], aicc_trace[i], bic_trace[i])) if verbose: print('Best setting (by BIC): lambda={0} [DoF: {1}, AIC: {2}, AICc: {3} BIC: {4}]'.format(lambda_grid[best_idx], dof_trace[best_idx], aic_trace[best_idx], aicc_trace[best_idx], bic_trace[best_idx])) return {'aic': aic_trace, 'aicc': aicc_trace, 'bic': bic_trace, 'dof': dof_trace, 'loglikelihood': log_likelihood_trace, 'beta': np.array(beta_trace), 'lambda': lambda_grid, 'best_idx': best_idx, 'best': beta_trace[best_idx], 'plateaus': best_plateaus}
python
def solution_path(self, min_lambda, max_lambda, lambda_bins, verbose=0): '''Follows the solution path to find the best lambda value.''' self.u = np.zeros(self.Dk.shape[0], dtype='double') lambda_grid = np.exp(np.linspace(np.log(max_lambda), np.log(min_lambda), lambda_bins)) aic_trace = np.zeros(lambda_grid.shape) # The AIC score for each lambda value aicc_trace = np.zeros(lambda_grid.shape) # The AICc score for each lambda value (correcting for finite sample size) bic_trace = np.zeros(lambda_grid.shape) # The BIC score for each lambda value dof_trace = np.zeros(lambda_grid.shape) # The degrees of freedom of each final solution log_likelihood_trace = np.zeros(lambda_grid.shape) beta_trace = [] best_idx = None best_plateaus = None if self.edges is None: self.edges = defaultdict(list) elist = csr_matrix(self.D).indices.reshape((self.D.shape[0], 2)) for n1, n2 in elist: self.edges[n1].append(n2) self.edges[n2].append(n1) # Solve the series of lambda values with warm starts at each point for i, lam in enumerate(lambda_grid): if verbose: print('#{0} Lambda = {1}'.format(i, lam)) # Fit to the final values beta = self.solve(lam) if verbose: print('Calculating degrees of freedom') # Count the number of free parameters in the grid (dof) -- TODO: the graph trend filtering paper seems to imply we shouldn't multiply by (k+1)? dof_vals = self.Dk_minus_one.dot(beta) if self.k > 0 else beta plateaus = calc_plateaus(dof_vals, self.edges, rel_tol=0.01) if (self.k % 2) == 0 else nearly_unique(dof_vals, rel_tol=0.03) dof_trace[i] = max(1,len(plateaus)) #* (k+1) if verbose: print('Calculating Information Criteria') # Get the negative log-likelihood log_likelihood_trace[i] = -0.5 * ((self.y - beta)**2).sum() # Calculate AIC = 2k - 2ln(L) aic_trace[i] = 2. * dof_trace[i] - 2. * log_likelihood_trace[i] # Calculate AICc = AIC + 2k * (k+1) / (n - k - 1) aicc_trace[i] = aic_trace[i] + 2 * dof_trace[i] * (dof_trace[i]+1) / (len(beta) - dof_trace[i] - 1.) # Calculate BIC = -2ln(L) + k * (ln(n) - ln(2pi)) bic_trace[i] = -2 * log_likelihood_trace[i] + dof_trace[i] * (np.log(len(beta)) - np.log(2 * np.pi)) # Track the best model thus far if best_idx is None or bic_trace[i] < bic_trace[best_idx]: best_idx = i best_plateaus = plateaus # Save the trace of all the resulting parameters beta_trace.append(np.array(beta)) if verbose: print('DoF: {0} AIC: {1} AICc: {2} BIC: {3}\n'.format(dof_trace[i], aic_trace[i], aicc_trace[i], bic_trace[i])) if verbose: print('Best setting (by BIC): lambda={0} [DoF: {1}, AIC: {2}, AICc: {3} BIC: {4}]'.format(lambda_grid[best_idx], dof_trace[best_idx], aic_trace[best_idx], aicc_trace[best_idx], bic_trace[best_idx])) return {'aic': aic_trace, 'aicc': aicc_trace, 'bic': bic_trace, 'dof': dof_trace, 'loglikelihood': log_likelihood_trace, 'beta': np.array(beta_trace), 'lambda': lambda_grid, 'best_idx': best_idx, 'best': beta_trace[best_idx], 'plateaus': best_plateaus}
[ "def", "solution_path", "(", "self", ",", "min_lambda", ",", "max_lambda", ",", "lambda_bins", ",", "verbose", "=", "0", ")", ":", "self", ".", "u", "=", "np", ".", "zeros", "(", "self", ".", "Dk", ".", "shape", "[", "0", "]", ",", "dtype", "=", "'double'", ")", "lambda_grid", "=", "np", ".", "exp", "(", "np", ".", "linspace", "(", "np", ".", "log", "(", "max_lambda", ")", ",", "np", ".", "log", "(", "min_lambda", ")", ",", "lambda_bins", ")", ")", "aic_trace", "=", "np", ".", "zeros", "(", "lambda_grid", ".", "shape", ")", "# The AIC score for each lambda value", "aicc_trace", "=", "np", ".", "zeros", "(", "lambda_grid", ".", "shape", ")", "# The AICc score for each lambda value (correcting for finite sample size)", "bic_trace", "=", "np", ".", "zeros", "(", "lambda_grid", ".", "shape", ")", "# The BIC score for each lambda value", "dof_trace", "=", "np", ".", "zeros", "(", "lambda_grid", ".", "shape", ")", "# The degrees of freedom of each final solution", "log_likelihood_trace", "=", "np", ".", "zeros", "(", "lambda_grid", ".", "shape", ")", "beta_trace", "=", "[", "]", "best_idx", "=", "None", "best_plateaus", "=", "None", "if", "self", ".", "edges", "is", "None", ":", "self", ".", "edges", "=", "defaultdict", "(", "list", ")", "elist", "=", "csr_matrix", "(", "self", ".", "D", ")", ".", "indices", ".", "reshape", "(", "(", "self", ".", "D", ".", "shape", "[", "0", "]", ",", "2", ")", ")", "for", "n1", ",", "n2", "in", "elist", ":", "self", ".", "edges", "[", "n1", "]", ".", "append", "(", "n2", ")", "self", ".", "edges", "[", "n2", "]", ".", "append", "(", "n1", ")", "# Solve the series of lambda values with warm starts at each point", "for", "i", ",", "lam", "in", "enumerate", "(", "lambda_grid", ")", ":", "if", "verbose", ":", "print", "(", "'#{0} Lambda = {1}'", ".", "format", "(", "i", ",", "lam", ")", ")", "# Fit to the final values", "beta", "=", "self", ".", "solve", "(", "lam", ")", "if", "verbose", ":", "print", "(", "'Calculating degrees of freedom'", ")", "# Count the number of free parameters in the grid (dof) -- TODO: the graph trend filtering paper seems to imply we shouldn't multiply by (k+1)?", "dof_vals", "=", "self", ".", "Dk_minus_one", ".", "dot", "(", "beta", ")", "if", "self", ".", "k", ">", "0", "else", "beta", "plateaus", "=", "calc_plateaus", "(", "dof_vals", ",", "self", ".", "edges", ",", "rel_tol", "=", "0.01", ")", "if", "(", "self", ".", "k", "%", "2", ")", "==", "0", "else", "nearly_unique", "(", "dof_vals", ",", "rel_tol", "=", "0.03", ")", "dof_trace", "[", "i", "]", "=", "max", "(", "1", ",", "len", "(", "plateaus", ")", ")", "#* (k+1)", "if", "verbose", ":", "print", "(", "'Calculating Information Criteria'", ")", "# Get the negative log-likelihood", "log_likelihood_trace", "[", "i", "]", "=", "-", "0.5", "*", "(", "(", "self", ".", "y", "-", "beta", ")", "**", "2", ")", ".", "sum", "(", ")", "# Calculate AIC = 2k - 2ln(L)", "aic_trace", "[", "i", "]", "=", "2.", "*", "dof_trace", "[", "i", "]", "-", "2.", "*", "log_likelihood_trace", "[", "i", "]", "# Calculate AICc = AIC + 2k * (k+1) / (n - k - 1)", "aicc_trace", "[", "i", "]", "=", "aic_trace", "[", "i", "]", "+", "2", "*", "dof_trace", "[", "i", "]", "*", "(", "dof_trace", "[", "i", "]", "+", "1", ")", "/", "(", "len", "(", "beta", ")", "-", "dof_trace", "[", "i", "]", "-", "1.", ")", "# Calculate BIC = -2ln(L) + k * (ln(n) - ln(2pi))", "bic_trace", "[", "i", "]", "=", "-", "2", "*", "log_likelihood_trace", "[", "i", "]", "+", "dof_trace", "[", "i", "]", "*", "(", "np", ".", "log", "(", "len", "(", "beta", ")", ")", "-", "np", ".", "log", "(", "2", "*", "np", ".", "pi", ")", ")", "# Track the best model thus far", "if", "best_idx", "is", "None", "or", "bic_trace", "[", "i", "]", "<", "bic_trace", "[", "best_idx", "]", ":", "best_idx", "=", "i", "best_plateaus", "=", "plateaus", "# Save the trace of all the resulting parameters", "beta_trace", ".", "append", "(", "np", ".", "array", "(", "beta", ")", ")", "if", "verbose", ":", "print", "(", "'DoF: {0} AIC: {1} AICc: {2} BIC: {3}\\n'", ".", "format", "(", "dof_trace", "[", "i", "]", ",", "aic_trace", "[", "i", "]", ",", "aicc_trace", "[", "i", "]", ",", "bic_trace", "[", "i", "]", ")", ")", "if", "verbose", ":", "print", "(", "'Best setting (by BIC): lambda={0} [DoF: {1}, AIC: {2}, AICc: {3} BIC: {4}]'", ".", "format", "(", "lambda_grid", "[", "best_idx", "]", ",", "dof_trace", "[", "best_idx", "]", ",", "aic_trace", "[", "best_idx", "]", ",", "aicc_trace", "[", "best_idx", "]", ",", "bic_trace", "[", "best_idx", "]", ")", ")", "return", "{", "'aic'", ":", "aic_trace", ",", "'aicc'", ":", "aicc_trace", ",", "'bic'", ":", "bic_trace", ",", "'dof'", ":", "dof_trace", ",", "'loglikelihood'", ":", "log_likelihood_trace", ",", "'beta'", ":", "np", ".", "array", "(", "beta_trace", ")", ",", "'lambda'", ":", "lambda_grid", ",", "'best_idx'", ":", "best_idx", ",", "'best'", ":", "beta_trace", "[", "best_idx", "]", ",", "'plateaus'", ":", "best_plateaus", "}" ]
Follows the solution path to find the best lambda value.
[ "Follows", "the", "solution", "path", "to", "find", "the", "best", "lambda", "value", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trendfiltering.py#L83-L157
train
tansey/gfl
pygfl/trendfiltering.py
LogitTrendFilteringSolver.solve
def solve(self, lam): '''Solves the GFL for a fixed value of lambda.''' s = weighted_graphtf_logit(self.nnodes, self.trials, self.successes, lam, self.Dk.shape[0], self.Dk.shape[1], self.Dk.nnz, self.Dk.row.astype('int32'), self.Dk.col.astype('int32'), self.Dk.data.astype('double'), self.maxsteps, self.converge, self.beta, self.u) self.steps.append(s) return self.beta
python
def solve(self, lam): '''Solves the GFL for a fixed value of lambda.''' s = weighted_graphtf_logit(self.nnodes, self.trials, self.successes, lam, self.Dk.shape[0], self.Dk.shape[1], self.Dk.nnz, self.Dk.row.astype('int32'), self.Dk.col.astype('int32'), self.Dk.data.astype('double'), self.maxsteps, self.converge, self.beta, self.u) self.steps.append(s) return self.beta
[ "def", "solve", "(", "self", ",", "lam", ")", ":", "s", "=", "weighted_graphtf_logit", "(", "self", ".", "nnodes", ",", "self", ".", "trials", ",", "self", ".", "successes", ",", "lam", ",", "self", ".", "Dk", ".", "shape", "[", "0", "]", ",", "self", ".", "Dk", ".", "shape", "[", "1", "]", ",", "self", ".", "Dk", ".", "nnz", ",", "self", ".", "Dk", ".", "row", ".", "astype", "(", "'int32'", ")", ",", "self", ".", "Dk", ".", "col", ".", "astype", "(", "'int32'", ")", ",", "self", ".", "Dk", ".", "data", ".", "astype", "(", "'double'", ")", ",", "self", ".", "maxsteps", ",", "self", ".", "converge", ",", "self", ".", "beta", ",", "self", ".", "u", ")", "self", ".", "steps", ".", "append", "(", "s", ")", "return", "self", ".", "beta" ]
Solves the GFL for a fixed value of lambda.
[ "Solves", "the", "GFL", "for", "a", "fixed", "value", "of", "lambda", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trendfiltering.py#L169-L177
train
tansey/gfl
pygfl/trendfiltering.py
PoissonTrendFilteringSolver.solve
def solve(self, lam): '''Solves the GFL for a fixed value of lambda.''' s = weighted_graphtf_poisson(self.nnodes, self.obs, lam, self.Dk.shape[0], self.Dk.shape[1], self.Dk.nnz, self.Dk.row.astype('int32'), self.Dk.col.astype('int32'), self.Dk.data.astype('double'), self.maxsteps, self.converge, self.beta, self.u) self.steps.append(s) return self.beta
python
def solve(self, lam): '''Solves the GFL for a fixed value of lambda.''' s = weighted_graphtf_poisson(self.nnodes, self.obs, lam, self.Dk.shape[0], self.Dk.shape[1], self.Dk.nnz, self.Dk.row.astype('int32'), self.Dk.col.astype('int32'), self.Dk.data.astype('double'), self.maxsteps, self.converge, self.beta, self.u) self.steps.append(s) return self.beta
[ "def", "solve", "(", "self", ",", "lam", ")", ":", "s", "=", "weighted_graphtf_poisson", "(", "self", ".", "nnodes", ",", "self", ".", "obs", ",", "lam", ",", "self", ".", "Dk", ".", "shape", "[", "0", "]", ",", "self", ".", "Dk", ".", "shape", "[", "1", "]", ",", "self", ".", "Dk", ".", "nnz", ",", "self", ".", "Dk", ".", "row", ".", "astype", "(", "'int32'", ")", ",", "self", ".", "Dk", ".", "col", ".", "astype", "(", "'int32'", ")", ",", "self", ".", "Dk", ".", "data", ".", "astype", "(", "'double'", ")", ",", "self", ".", "maxsteps", ",", "self", ".", "converge", ",", "self", ".", "beta", ",", "self", ".", "u", ")", "self", ".", "steps", ".", "append", "(", "s", ")", "return", "self", ".", "beta" ]
Solves the GFL for a fixed value of lambda.
[ "Solves", "the", "GFL", "for", "a", "fixed", "value", "of", "lambda", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trendfiltering.py#L188-L196
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParameters.make_V_and_J_mask_mapping
def make_V_and_J_mask_mapping(self, genV, genJ): """Constructs the V and J mask mapping dictionaries. Parameters ---------- genV : list List of genomic V information. genJ : list List of genomic J information. """ #construct mapping between allele/gene names and index for custom V_usage_masks V_allele_names = [V[0] for V in genV] V_mask_mapping = {} for v in set([x.split('*')[0] for x in V_allele_names]): V_mask_mapping[v] = [] for v in set(['V'.join((x.split('*')[0]).split('V')[1:]) for x in V_allele_names]): V_mask_mapping[v] = [] for i, v in enumerate(V_allele_names): V_mask_mapping[v] = [i] V_mask_mapping['V'.join((v.split('*')[0]).split('V')[1:])].append(i) V_mask_mapping[v.split('*')[0]].append(i) #construct mapping between allele/gene names and index for custom J_usage_masks J_allele_names = [J[0] for J in genJ] J_mask_mapping = {} for j in set([x.split('*')[0] for x in J_allele_names]): J_mask_mapping[j] = [] for j in set(['J'.join((x.split('*')[0]).split('J')[1:]) for x in J_allele_names]): J_mask_mapping[j] = [] for i, j in enumerate(J_allele_names): J_mask_mapping[j] = [i] J_mask_mapping['J'.join((j.split('*')[0]).split('J')[1:])].append(i) J_mask_mapping[j.split('*')[0]].append(i) self.V_allele_names = V_allele_names self.V_mask_mapping = V_mask_mapping self.J_allele_names = J_allele_names self.J_mask_mapping = J_mask_mapping
python
def make_V_and_J_mask_mapping(self, genV, genJ): """Constructs the V and J mask mapping dictionaries. Parameters ---------- genV : list List of genomic V information. genJ : list List of genomic J information. """ #construct mapping between allele/gene names and index for custom V_usage_masks V_allele_names = [V[0] for V in genV] V_mask_mapping = {} for v in set([x.split('*')[0] for x in V_allele_names]): V_mask_mapping[v] = [] for v in set(['V'.join((x.split('*')[0]).split('V')[1:]) for x in V_allele_names]): V_mask_mapping[v] = [] for i, v in enumerate(V_allele_names): V_mask_mapping[v] = [i] V_mask_mapping['V'.join((v.split('*')[0]).split('V')[1:])].append(i) V_mask_mapping[v.split('*')[0]].append(i) #construct mapping between allele/gene names and index for custom J_usage_masks J_allele_names = [J[0] for J in genJ] J_mask_mapping = {} for j in set([x.split('*')[0] for x in J_allele_names]): J_mask_mapping[j] = [] for j in set(['J'.join((x.split('*')[0]).split('J')[1:]) for x in J_allele_names]): J_mask_mapping[j] = [] for i, j in enumerate(J_allele_names): J_mask_mapping[j] = [i] J_mask_mapping['J'.join((j.split('*')[0]).split('J')[1:])].append(i) J_mask_mapping[j.split('*')[0]].append(i) self.V_allele_names = V_allele_names self.V_mask_mapping = V_mask_mapping self.J_allele_names = J_allele_names self.J_mask_mapping = J_mask_mapping
[ "def", "make_V_and_J_mask_mapping", "(", "self", ",", "genV", ",", "genJ", ")", ":", "#construct mapping between allele/gene names and index for custom V_usage_masks", "V_allele_names", "=", "[", "V", "[", "0", "]", "for", "V", "in", "genV", "]", "V_mask_mapping", "=", "{", "}", "for", "v", "in", "set", "(", "[", "x", ".", "split", "(", "'*'", ")", "[", "0", "]", "for", "x", "in", "V_allele_names", "]", ")", ":", "V_mask_mapping", "[", "v", "]", "=", "[", "]", "for", "v", "in", "set", "(", "[", "'V'", ".", "join", "(", "(", "x", ".", "split", "(", "'*'", ")", "[", "0", "]", ")", ".", "split", "(", "'V'", ")", "[", "1", ":", "]", ")", "for", "x", "in", "V_allele_names", "]", ")", ":", "V_mask_mapping", "[", "v", "]", "=", "[", "]", "for", "i", ",", "v", "in", "enumerate", "(", "V_allele_names", ")", ":", "V_mask_mapping", "[", "v", "]", "=", "[", "i", "]", "V_mask_mapping", "[", "'V'", ".", "join", "(", "(", "v", ".", "split", "(", "'*'", ")", "[", "0", "]", ")", ".", "split", "(", "'V'", ")", "[", "1", ":", "]", ")", "]", ".", "append", "(", "i", ")", "V_mask_mapping", "[", "v", ".", "split", "(", "'*'", ")", "[", "0", "]", "]", ".", "append", "(", "i", ")", "#construct mapping between allele/gene names and index for custom J_usage_masks", "J_allele_names", "=", "[", "J", "[", "0", "]", "for", "J", "in", "genJ", "]", "J_mask_mapping", "=", "{", "}", "for", "j", "in", "set", "(", "[", "x", ".", "split", "(", "'*'", ")", "[", "0", "]", "for", "x", "in", "J_allele_names", "]", ")", ":", "J_mask_mapping", "[", "j", "]", "=", "[", "]", "for", "j", "in", "set", "(", "[", "'J'", ".", "join", "(", "(", "x", ".", "split", "(", "'*'", ")", "[", "0", "]", ")", ".", "split", "(", "'J'", ")", "[", "1", ":", "]", ")", "for", "x", "in", "J_allele_names", "]", ")", ":", "J_mask_mapping", "[", "j", "]", "=", "[", "]", "for", "i", ",", "j", "in", "enumerate", "(", "J_allele_names", ")", ":", "J_mask_mapping", "[", "j", "]", "=", "[", "i", "]", "J_mask_mapping", "[", "'J'", ".", "join", "(", "(", "j", ".", "split", "(", "'*'", ")", "[", "0", "]", ")", ".", "split", "(", "'J'", ")", "[", "1", ":", "]", ")", "]", ".", "append", "(", "i", ")", "J_mask_mapping", "[", "j", ".", "split", "(", "'*'", ")", "[", "0", "]", "]", ".", "append", "(", "i", ")", "self", ".", "V_allele_names", "=", "V_allele_names", "self", ".", "V_mask_mapping", "=", "V_mask_mapping", "self", ".", "J_allele_names", "=", "J_allele_names", "self", ".", "J_mask_mapping", "=", "J_mask_mapping" ]
Constructs the V and J mask mapping dictionaries. Parameters ---------- genV : list List of genomic V information. genJ : list List of genomic J information.
[ "Constructs", "the", "V", "and", "J", "mask", "mapping", "dictionaries", ".", "Parameters", "----------", "genV", ":", "list", "List", "of", "genomic", "V", "information", ".", "genJ", ":", "list", "List", "of", "genomic", "J", "information", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L157-L195
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVDJ.preprocess_D_segs
def preprocess_D_segs(self, generative_model, genomic_data): """Process P(delDl, delDr|D) into Pi arrays. Sets the attributes PD_nt_pos_vec, PD_2nd_nt_pos_per_aa_vec, min_delDl_given_DdelDr, max_delDl_given_DdelDr, and zeroD_given_D. Parameters ---------- generative_model : GenerativeModelVDJ VDJ generative model class containing the model parameters. genomic_data : GenomicDataVDJ VDJ genomic data class containing the V, D, and J germline sequences and info. """ cutD_genomic_CDR3_segs = genomic_data.cutD_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_dell_pos, num_delr_pos, num_D_genes = generative_model.PdelDldelDr_given_D.shape #These arrays only include the nt identity information, not the PdelDldelDr_given_D info PD_nt_pos_vec = [[]]*num_D_genes PD_2nd_nt_pos_per_aa_vec = [[]]*num_D_genes for D_in in range(num_D_genes): current_PD_nt_pos_vec = np.zeros((4, len(cutD_genomic_CDR3_segs[D_in]))) current_PD_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PD_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutD_genomic_CDR3_segs[D_in]))) for pos, nt in enumerate(cutD_genomic_CDR3_segs[D_in]): current_PD_nt_pos_vec[nt2num[nt], pos] = 1 for ins_nt in 'ACGT': for aa in self.codons_dict.keys(): if ins_nt + cutD_genomic_CDR3_segs[D_in][pos:pos+2] in self.codons_dict[aa]: current_PD_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = 1 PD_nt_pos_vec[D_in] = current_PD_nt_pos_vec PD_2nd_nt_pos_per_aa_vec[D_in] = current_PD_2nd_nt_pos_per_aa_vec min_delDl_given_DdelDr = [[]]*num_D_genes max_delDl_given_DdelDr = [[]]*num_D_genes zeroD_given_D = [[]]*num_D_genes for D_in in range(num_D_genes): current_min_delDl_given_delDr = [0]*num_delr_pos current_max_delDl_given_delDr = [0]*num_delr_pos current_zeroD = 0 for delr in range(num_delr_pos): if num_dell_pos > len(cutD_genomic_CDR3_segs[D_in])-delr: current_zeroD += generative_model.PdelDldelDr_given_D[len(cutD_genomic_CDR3_segs[D_in])-delr, delr, D_in] dell = 0 while generative_model.PdelDldelDr_given_D[dell, delr, D_in]==0 and dell<num_dell_pos-1: dell+=1 if generative_model.PdelDldelDr_given_D[dell, delr, D_in] == 0: current_min_delDl_given_delDr[delr] = -1 else: current_min_delDl_given_delDr[delr] = dell if current_min_delDl_given_delDr[delr] == -1: current_max_delDl_given_delDr[delr] = -1 else: dell = num_dell_pos-1 while generative_model.PdelDldelDr_given_D[dell, delr, D_in]==0 and dell>=0: dell -= 1 if generative_model.PdelDldelDr_given_D[dell, delr, D_in] == 0: current_max_delDl_given_delDr[delr] = -1 else: current_max_delDl_given_delDr[delr] = dell min_delDl_given_DdelDr[D_in] = current_min_delDl_given_delDr max_delDl_given_DdelDr[D_in] = current_max_delDl_given_delDr zeroD_given_D[D_in] = current_zeroD self.PD_nt_pos_vec = PD_nt_pos_vec self.PD_2nd_nt_pos_per_aa_vec = PD_2nd_nt_pos_per_aa_vec self.min_delDl_given_DdelDr = min_delDl_given_DdelDr self.max_delDl_given_DdelDr = max_delDl_given_DdelDr self.zeroD_given_D = zeroD_given_D
python
def preprocess_D_segs(self, generative_model, genomic_data): """Process P(delDl, delDr|D) into Pi arrays. Sets the attributes PD_nt_pos_vec, PD_2nd_nt_pos_per_aa_vec, min_delDl_given_DdelDr, max_delDl_given_DdelDr, and zeroD_given_D. Parameters ---------- generative_model : GenerativeModelVDJ VDJ generative model class containing the model parameters. genomic_data : GenomicDataVDJ VDJ genomic data class containing the V, D, and J germline sequences and info. """ cutD_genomic_CDR3_segs = genomic_data.cutD_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_dell_pos, num_delr_pos, num_D_genes = generative_model.PdelDldelDr_given_D.shape #These arrays only include the nt identity information, not the PdelDldelDr_given_D info PD_nt_pos_vec = [[]]*num_D_genes PD_2nd_nt_pos_per_aa_vec = [[]]*num_D_genes for D_in in range(num_D_genes): current_PD_nt_pos_vec = np.zeros((4, len(cutD_genomic_CDR3_segs[D_in]))) current_PD_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PD_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutD_genomic_CDR3_segs[D_in]))) for pos, nt in enumerate(cutD_genomic_CDR3_segs[D_in]): current_PD_nt_pos_vec[nt2num[nt], pos] = 1 for ins_nt in 'ACGT': for aa in self.codons_dict.keys(): if ins_nt + cutD_genomic_CDR3_segs[D_in][pos:pos+2] in self.codons_dict[aa]: current_PD_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = 1 PD_nt_pos_vec[D_in] = current_PD_nt_pos_vec PD_2nd_nt_pos_per_aa_vec[D_in] = current_PD_2nd_nt_pos_per_aa_vec min_delDl_given_DdelDr = [[]]*num_D_genes max_delDl_given_DdelDr = [[]]*num_D_genes zeroD_given_D = [[]]*num_D_genes for D_in in range(num_D_genes): current_min_delDl_given_delDr = [0]*num_delr_pos current_max_delDl_given_delDr = [0]*num_delr_pos current_zeroD = 0 for delr in range(num_delr_pos): if num_dell_pos > len(cutD_genomic_CDR3_segs[D_in])-delr: current_zeroD += generative_model.PdelDldelDr_given_D[len(cutD_genomic_CDR3_segs[D_in])-delr, delr, D_in] dell = 0 while generative_model.PdelDldelDr_given_D[dell, delr, D_in]==0 and dell<num_dell_pos-1: dell+=1 if generative_model.PdelDldelDr_given_D[dell, delr, D_in] == 0: current_min_delDl_given_delDr[delr] = -1 else: current_min_delDl_given_delDr[delr] = dell if current_min_delDl_given_delDr[delr] == -1: current_max_delDl_given_delDr[delr] = -1 else: dell = num_dell_pos-1 while generative_model.PdelDldelDr_given_D[dell, delr, D_in]==0 and dell>=0: dell -= 1 if generative_model.PdelDldelDr_given_D[dell, delr, D_in] == 0: current_max_delDl_given_delDr[delr] = -1 else: current_max_delDl_given_delDr[delr] = dell min_delDl_given_DdelDr[D_in] = current_min_delDl_given_delDr max_delDl_given_DdelDr[D_in] = current_max_delDl_given_delDr zeroD_given_D[D_in] = current_zeroD self.PD_nt_pos_vec = PD_nt_pos_vec self.PD_2nd_nt_pos_per_aa_vec = PD_2nd_nt_pos_per_aa_vec self.min_delDl_given_DdelDr = min_delDl_given_DdelDr self.max_delDl_given_DdelDr = max_delDl_given_DdelDr self.zeroD_given_D = zeroD_given_D
[ "def", "preprocess_D_segs", "(", "self", ",", "generative_model", ",", "genomic_data", ")", ":", "cutD_genomic_CDR3_segs", "=", "genomic_data", ".", "cutD_genomic_CDR3_segs", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "num_dell_pos", ",", "num_delr_pos", ",", "num_D_genes", "=", "generative_model", ".", "PdelDldelDr_given_D", ".", "shape", "#These arrays only include the nt identity information, not the PdelDldelDr_given_D info", "PD_nt_pos_vec", "=", "[", "[", "]", "]", "*", "num_D_genes", "PD_2nd_nt_pos_per_aa_vec", "=", "[", "[", "]", "]", "*", "num_D_genes", "for", "D_in", "in", "range", "(", "num_D_genes", ")", ":", "current_PD_nt_pos_vec", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutD_genomic_CDR3_segs", "[", "D_in", "]", ")", ")", ")", "current_PD_2nd_nt_pos_per_aa_vec", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_PD_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutD_genomic_CDR3_segs", "[", "D_in", "]", ")", ")", ")", "for", "pos", ",", "nt", "in", "enumerate", "(", "cutD_genomic_CDR3_segs", "[", "D_in", "]", ")", ":", "current_PD_nt_pos_vec", "[", "nt2num", "[", "nt", "]", ",", "pos", "]", "=", "1", "for", "ins_nt", "in", "'ACGT'", ":", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "if", "ins_nt", "+", "cutD_genomic_CDR3_segs", "[", "D_in", "]", "[", "pos", ":", "pos", "+", "2", "]", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_PD_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "[", "nt2num", "[", "ins_nt", "]", ",", "pos", "]", "=", "1", "PD_nt_pos_vec", "[", "D_in", "]", "=", "current_PD_nt_pos_vec", "PD_2nd_nt_pos_per_aa_vec", "[", "D_in", "]", "=", "current_PD_2nd_nt_pos_per_aa_vec", "min_delDl_given_DdelDr", "=", "[", "[", "]", "]", "*", "num_D_genes", "max_delDl_given_DdelDr", "=", "[", "[", "]", "]", "*", "num_D_genes", "zeroD_given_D", "=", "[", "[", "]", "]", "*", "num_D_genes", "for", "D_in", "in", "range", "(", "num_D_genes", ")", ":", "current_min_delDl_given_delDr", "=", "[", "0", "]", "*", "num_delr_pos", "current_max_delDl_given_delDr", "=", "[", "0", "]", "*", "num_delr_pos", "current_zeroD", "=", "0", "for", "delr", "in", "range", "(", "num_delr_pos", ")", ":", "if", "num_dell_pos", ">", "len", "(", "cutD_genomic_CDR3_segs", "[", "D_in", "]", ")", "-", "delr", ":", "current_zeroD", "+=", "generative_model", ".", "PdelDldelDr_given_D", "[", "len", "(", "cutD_genomic_CDR3_segs", "[", "D_in", "]", ")", "-", "delr", ",", "delr", ",", "D_in", "]", "dell", "=", "0", "while", "generative_model", ".", "PdelDldelDr_given_D", "[", "dell", ",", "delr", ",", "D_in", "]", "==", "0", "and", "dell", "<", "num_dell_pos", "-", "1", ":", "dell", "+=", "1", "if", "generative_model", ".", "PdelDldelDr_given_D", "[", "dell", ",", "delr", ",", "D_in", "]", "==", "0", ":", "current_min_delDl_given_delDr", "[", "delr", "]", "=", "-", "1", "else", ":", "current_min_delDl_given_delDr", "[", "delr", "]", "=", "dell", "if", "current_min_delDl_given_delDr", "[", "delr", "]", "==", "-", "1", ":", "current_max_delDl_given_delDr", "[", "delr", "]", "=", "-", "1", "else", ":", "dell", "=", "num_dell_pos", "-", "1", "while", "generative_model", ".", "PdelDldelDr_given_D", "[", "dell", ",", "delr", ",", "D_in", "]", "==", "0", "and", "dell", ">=", "0", ":", "dell", "-=", "1", "if", "generative_model", ".", "PdelDldelDr_given_D", "[", "dell", ",", "delr", ",", "D_in", "]", "==", "0", ":", "current_max_delDl_given_delDr", "[", "delr", "]", "=", "-", "1", "else", ":", "current_max_delDl_given_delDr", "[", "delr", "]", "=", "dell", "min_delDl_given_DdelDr", "[", "D_in", "]", "=", "current_min_delDl_given_delDr", "max_delDl_given_DdelDr", "[", "D_in", "]", "=", "current_max_delDl_given_delDr", "zeroD_given_D", "[", "D_in", "]", "=", "current_zeroD", "self", ".", "PD_nt_pos_vec", "=", "PD_nt_pos_vec", "self", ".", "PD_2nd_nt_pos_per_aa_vec", "=", "PD_2nd_nt_pos_per_aa_vec", "self", ".", "min_delDl_given_DdelDr", "=", "min_delDl_given_DdelDr", "self", ".", "max_delDl_given_DdelDr", "=", "max_delDl_given_DdelDr", "self", ".", "zeroD_given_D", "=", "zeroD_given_D" ]
Process P(delDl, delDr|D) into Pi arrays. Sets the attributes PD_nt_pos_vec, PD_2nd_nt_pos_per_aa_vec, min_delDl_given_DdelDr, max_delDl_given_DdelDr, and zeroD_given_D. Parameters ---------- generative_model : GenerativeModelVDJ VDJ generative model class containing the model parameters. genomic_data : GenomicDataVDJ VDJ genomic data class containing the V, D, and J germline sequences and info.
[ "Process", "P", "(", "delDl", "delDr|D", ")", "into", "Pi", "arrays", ".", "Sets", "the", "attributes", "PD_nt_pos_vec", "PD_2nd_nt_pos_per_aa_vec", "min_delDl_given_DdelDr", "max_delDl_given_DdelDr", "and", "zeroD_given_D", ".", "Parameters", "----------", "generative_model", ":", "GenerativeModelVDJ", "VDJ", "generative", "model", "class", "containing", "the", "model", "parameters", ".", "genomic_data", ":", "GenomicDataVDJ", "VDJ", "genomic", "data", "class", "containing", "the", "V", "D", "and", "J", "germline", "sequences", "and", "info", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L463-L541
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVDJ.generate_PJdelJ_nt_pos_vecs
def generate_PJdelJ_nt_pos_vecs(self, generative_model, genomic_data): """Process P(J)*P(delJ|J) into Pi arrays. Sets the attributes PJdelJ_nt_pos_vec and PJdelJ_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVDJ VDJ generative model class containing the model parameters. genomic_data : GenomicDataVDJ VDJ genomic data class containing the V, D, and J germline sequences and info. """ cutJ_genomic_CDR3_segs = genomic_data.cutJ_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_del_pos = generative_model.PdelJ_given_J.shape[0] num_D_genes, num_J_genes = generative_model.PDJ.shape PJ = np.sum(generative_model.PDJ, axis = 0) PJdelJ_nt_pos_vec = [[]]*num_J_genes PJdelJ_2nd_nt_pos_per_aa_vec = [[]]*num_J_genes for J_in, pj in enumerate(PJ): #We include the marginal PJ here current_PJdelJ_nt_pos_vec = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) current_PJdelJ_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PJdelJ_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) for pos, nt in enumerate(cutJ_genomic_CDR3_segs[J_in]): if pos >= num_del_pos: continue if (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 1: #Start of a codon current_PJdelJ_nt_pos_vec[nt2num[nt], pos] = pj*generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 2: #Mid codon position for ins_nt in 'ACGT': #We need to find what possible codons are allowed for any aa (or motif) for aa in self.codons_dict.keys(): if ins_nt + cutJ_genomic_CDR3_segs[J_in][pos:pos+2] in self.codons_dict[aa]: current_PJdelJ_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = pj*generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 0: #End of codon current_PJdelJ_nt_pos_vec[0, pos] = pj*generative_model.PdelJ_given_J[pos, J_in] PJdelJ_nt_pos_vec[J_in] = current_PJdelJ_nt_pos_vec PJdelJ_2nd_nt_pos_per_aa_vec[J_in] = current_PJdelJ_2nd_nt_pos_per_aa_vec self.PJdelJ_nt_pos_vec = PJdelJ_nt_pos_vec self.PJdelJ_2nd_nt_pos_per_aa_vec = PJdelJ_2nd_nt_pos_per_aa_vec
python
def generate_PJdelJ_nt_pos_vecs(self, generative_model, genomic_data): """Process P(J)*P(delJ|J) into Pi arrays. Sets the attributes PJdelJ_nt_pos_vec and PJdelJ_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVDJ VDJ generative model class containing the model parameters. genomic_data : GenomicDataVDJ VDJ genomic data class containing the V, D, and J germline sequences and info. """ cutJ_genomic_CDR3_segs = genomic_data.cutJ_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_del_pos = generative_model.PdelJ_given_J.shape[0] num_D_genes, num_J_genes = generative_model.PDJ.shape PJ = np.sum(generative_model.PDJ, axis = 0) PJdelJ_nt_pos_vec = [[]]*num_J_genes PJdelJ_2nd_nt_pos_per_aa_vec = [[]]*num_J_genes for J_in, pj in enumerate(PJ): #We include the marginal PJ here current_PJdelJ_nt_pos_vec = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) current_PJdelJ_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PJdelJ_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) for pos, nt in enumerate(cutJ_genomic_CDR3_segs[J_in]): if pos >= num_del_pos: continue if (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 1: #Start of a codon current_PJdelJ_nt_pos_vec[nt2num[nt], pos] = pj*generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 2: #Mid codon position for ins_nt in 'ACGT': #We need to find what possible codons are allowed for any aa (or motif) for aa in self.codons_dict.keys(): if ins_nt + cutJ_genomic_CDR3_segs[J_in][pos:pos+2] in self.codons_dict[aa]: current_PJdelJ_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = pj*generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 0: #End of codon current_PJdelJ_nt_pos_vec[0, pos] = pj*generative_model.PdelJ_given_J[pos, J_in] PJdelJ_nt_pos_vec[J_in] = current_PJdelJ_nt_pos_vec PJdelJ_2nd_nt_pos_per_aa_vec[J_in] = current_PJdelJ_2nd_nt_pos_per_aa_vec self.PJdelJ_nt_pos_vec = PJdelJ_nt_pos_vec self.PJdelJ_2nd_nt_pos_per_aa_vec = PJdelJ_2nd_nt_pos_per_aa_vec
[ "def", "generate_PJdelJ_nt_pos_vecs", "(", "self", ",", "generative_model", ",", "genomic_data", ")", ":", "cutJ_genomic_CDR3_segs", "=", "genomic_data", ".", "cutJ_genomic_CDR3_segs", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "num_del_pos", "=", "generative_model", ".", "PdelJ_given_J", ".", "shape", "[", "0", "]", "num_D_genes", ",", "num_J_genes", "=", "generative_model", ".", "PDJ", ".", "shape", "PJ", "=", "np", ".", "sum", "(", "generative_model", ".", "PDJ", ",", "axis", "=", "0", ")", "PJdelJ_nt_pos_vec", "=", "[", "[", "]", "]", "*", "num_J_genes", "PJdelJ_2nd_nt_pos_per_aa_vec", "=", "[", "[", "]", "]", "*", "num_J_genes", "for", "J_in", ",", "pj", "in", "enumerate", "(", "PJ", ")", ":", "#We include the marginal PJ here", "current_PJdelJ_nt_pos_vec", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", ")", ")", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", ")", ")", "for", "pos", ",", "nt", "in", "enumerate", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", ":", "if", "pos", ">=", "num_del_pos", ":", "continue", "if", "(", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", "-", "pos", ")", "%", "3", "==", "1", ":", "#Start of a codon", "current_PJdelJ_nt_pos_vec", "[", "nt2num", "[", "nt", "]", ",", "pos", "]", "=", "pj", "*", "generative_model", ".", "PdelJ_given_J", "[", "pos", ",", "J_in", "]", "elif", "(", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", "-", "pos", ")", "%", "3", "==", "2", ":", "#Mid codon position", "for", "ins_nt", "in", "'ACGT'", ":", "#We need to find what possible codons are allowed for any aa (or motif)", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "if", "ins_nt", "+", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", "[", "pos", ":", "pos", "+", "2", "]", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "[", "nt2num", "[", "ins_nt", "]", ",", "pos", "]", "=", "pj", "*", "generative_model", ".", "PdelJ_given_J", "[", "pos", ",", "J_in", "]", "elif", "(", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", "-", "pos", ")", "%", "3", "==", "0", ":", "#End of codon", "current_PJdelJ_nt_pos_vec", "[", "0", ",", "pos", "]", "=", "pj", "*", "generative_model", ".", "PdelJ_given_J", "[", "pos", ",", "J_in", "]", "PJdelJ_nt_pos_vec", "[", "J_in", "]", "=", "current_PJdelJ_nt_pos_vec", "PJdelJ_2nd_nt_pos_per_aa_vec", "[", "J_in", "]", "=", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "self", ".", "PJdelJ_nt_pos_vec", "=", "PJdelJ_nt_pos_vec", "self", ".", "PJdelJ_2nd_nt_pos_per_aa_vec", "=", "PJdelJ_2nd_nt_pos_per_aa_vec" ]
Process P(J)*P(delJ|J) into Pi arrays. Sets the attributes PJdelJ_nt_pos_vec and PJdelJ_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVDJ VDJ generative model class containing the model parameters. genomic_data : GenomicDataVDJ VDJ genomic data class containing the V, D, and J germline sequences and info.
[ "Process", "P", "(", "J", ")", "*", "P", "(", "delJ|J", ")", "into", "Pi", "arrays", ".", "Sets", "the", "attributes", "PJdelJ_nt_pos_vec", "and", "PJdelJ_2nd_nt_pos_per_aa_vec", ".", "Parameters", "----------", "generative_model", ":", "GenerativeModelVDJ", "VDJ", "generative", "model", "class", "containing", "the", "model", "parameters", ".", "genomic_data", ":", "GenomicDataVDJ", "VDJ", "genomic", "data", "class", "containing", "the", "V", "D", "and", "J", "germline", "sequences", "and", "info", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L543-L592
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVDJ.generate_VD_junction_transfer_matrices
def generate_VD_junction_transfer_matrices(self): """Compute the transfer matrices for the VD junction. Sets the attributes Tvd, Svd, Dvd, lTvd, and lDvd. """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #Compute Tvd Tvd = {} for aa in self.codons_dict.keys(): current_Tvd = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Tvd[nt2num[codon[2]], nt2num[init_nt]] += self.Rvd[nt2num[codon[2]],nt2num[codon[1]]]*self.Rvd[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvd[nt2num[codon[0]],nt2num[init_nt]] Tvd[aa] = current_Tvd #Compute Svd Svd = {} for aa in self.codons_dict.keys(): current_Svd = np.zeros((4, 4)) for ins_nt in 'ACGT': if any([codon.startswith(ins_nt) for codon in self.codons_dict[aa]]): current_Svd[nt2num[ins_nt], :] = self.Rvd[nt2num[ins_nt], :] Svd[aa] = current_Svd #Compute Dvd Dvd = {} for aa in self.codons_dict.keys(): current_Dvd = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Dvd[nt2num[codon[2]], nt2num[init_nt]] += self.Rvd[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvd[nt2num[codon[0]],nt2num[init_nt]] Dvd[aa] = current_Dvd #Compute lTvd lTvd = {} for aa in self.codons_dict.keys(): current_lTvd = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lTvd[nt2num[codon[2]], nt2num[codon[0]]] += self.Rvd[nt2num[codon[2]],nt2num[codon[1]]]*self.first_nt_bias_insVD[nt2num[codon[1]]] lTvd[aa] = current_lTvd #Compute lDvd lDvd = {} for aa in self.codons_dict.keys(): current_lDvd = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lDvd[nt2num[codon[2]], nt2num[codon[0]]] += self.first_nt_bias_insVD[nt2num[codon[1]]] lDvd[aa] = current_lDvd #Set the attributes self.Tvd = Tvd self.Svd = Svd self.Dvd = Dvd self.lTvd = lTvd self.lDvd = lDvd
python
def generate_VD_junction_transfer_matrices(self): """Compute the transfer matrices for the VD junction. Sets the attributes Tvd, Svd, Dvd, lTvd, and lDvd. """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #Compute Tvd Tvd = {} for aa in self.codons_dict.keys(): current_Tvd = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Tvd[nt2num[codon[2]], nt2num[init_nt]] += self.Rvd[nt2num[codon[2]],nt2num[codon[1]]]*self.Rvd[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvd[nt2num[codon[0]],nt2num[init_nt]] Tvd[aa] = current_Tvd #Compute Svd Svd = {} for aa in self.codons_dict.keys(): current_Svd = np.zeros((4, 4)) for ins_nt in 'ACGT': if any([codon.startswith(ins_nt) for codon in self.codons_dict[aa]]): current_Svd[nt2num[ins_nt], :] = self.Rvd[nt2num[ins_nt], :] Svd[aa] = current_Svd #Compute Dvd Dvd = {} for aa in self.codons_dict.keys(): current_Dvd = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Dvd[nt2num[codon[2]], nt2num[init_nt]] += self.Rvd[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvd[nt2num[codon[0]],nt2num[init_nt]] Dvd[aa] = current_Dvd #Compute lTvd lTvd = {} for aa in self.codons_dict.keys(): current_lTvd = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lTvd[nt2num[codon[2]], nt2num[codon[0]]] += self.Rvd[nt2num[codon[2]],nt2num[codon[1]]]*self.first_nt_bias_insVD[nt2num[codon[1]]] lTvd[aa] = current_lTvd #Compute lDvd lDvd = {} for aa in self.codons_dict.keys(): current_lDvd = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lDvd[nt2num[codon[2]], nt2num[codon[0]]] += self.first_nt_bias_insVD[nt2num[codon[1]]] lDvd[aa] = current_lDvd #Set the attributes self.Tvd = Tvd self.Svd = Svd self.Dvd = Dvd self.lTvd = lTvd self.lDvd = lDvd
[ "def", "generate_VD_junction_transfer_matrices", "(", "self", ")", ":", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "#Compute Tvd", "Tvd", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Tvd", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "init_nt", "in", "'ACGT'", ":", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_Tvd", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "+=", "self", ".", "Rvd", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "*", "self", ".", "Rvd", "[", "nt2num", "[", "codon", "[", "1", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "*", "self", ".", "Rvd", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "Tvd", "[", "aa", "]", "=", "current_Tvd", "#Compute Svd", "Svd", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Svd", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "ins_nt", "in", "'ACGT'", ":", "if", "any", "(", "[", "codon", ".", "startswith", "(", "ins_nt", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", "]", ")", ":", "current_Svd", "[", "nt2num", "[", "ins_nt", "]", ",", ":", "]", "=", "self", ".", "Rvd", "[", "nt2num", "[", "ins_nt", "]", ",", ":", "]", "Svd", "[", "aa", "]", "=", "current_Svd", "#Compute Dvd ", "Dvd", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Dvd", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "init_nt", "in", "'ACGT'", ":", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_Dvd", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "+=", "self", ".", "Rvd", "[", "nt2num", "[", "codon", "[", "1", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "*", "self", ".", "Rvd", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "Dvd", "[", "aa", "]", "=", "current_Dvd", "#Compute lTvd", "lTvd", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_lTvd", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_lTvd", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "+=", "self", ".", "Rvd", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "*", "self", ".", "first_nt_bias_insVD", "[", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "lTvd", "[", "aa", "]", "=", "current_lTvd", "#Compute lDvd", "lDvd", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_lDvd", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_lDvd", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "+=", "self", ".", "first_nt_bias_insVD", "[", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "lDvd", "[", "aa", "]", "=", "current_lDvd", "#Set the attributes", "self", ".", "Tvd", "=", "Tvd", "self", ".", "Svd", "=", "Svd", "self", ".", "Dvd", "=", "Dvd", "self", ".", "lTvd", "=", "lTvd", "self", ".", "lDvd", "=", "lDvd" ]
Compute the transfer matrices for the VD junction. Sets the attributes Tvd, Svd, Dvd, lTvd, and lDvd.
[ "Compute", "the", "transfer", "matrices", "for", "the", "VD", "junction", ".", "Sets", "the", "attributes", "Tvd", "Svd", "Dvd", "lTvd", "and", "lDvd", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L594-L654
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVDJ.generate_DJ_junction_transfer_matrices
def generate_DJ_junction_transfer_matrices(self): """Compute the transfer matrices for the VD junction. Sets the attributes Tdj, Sdj, Ddj, rTdj, and rDdj. """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #Compute Tdj Tdj = {} for aa in self.codons_dict.keys(): current_Tdj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Tdj[nt2num[codon[0]], nt2num[init_nt]] += self.Rdj[nt2num[codon[0]],nt2num[codon[1]]]*self.Rdj[nt2num[codon[1]],nt2num[codon[2]]] * self.Rdj[nt2num[codon[2]],nt2num[init_nt]] Tdj[aa] = current_Tdj #Compute Sdj Sdj = {} for aa in self.codons_dict.keys(): current_Sdj = np.zeros((4, 4)) for ins_nt in 'ACGT': if any([codon.endswith(ins_nt) for codon in self.codons_dict[aa]]): current_Sdj[nt2num[ins_nt], :] = self.Rdj[nt2num[ins_nt], :] Sdj[aa] = current_Sdj #Compute Ddj Ddj = {} for aa in self.codons_dict.keys(): current_Ddj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Ddj[nt2num[codon[0]], nt2num[init_nt]] += self.Rdj[nt2num[codon[1]],nt2num[codon[2]]] * self.Rdj[nt2num[codon[2]],nt2num[init_nt]] Ddj[aa] = current_Ddj #Compute rTdj rTdj = {} for aa in self.codons_dict.keys(): current_lTdj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lTdj[nt2num[codon[0]], nt2num[codon[2]]] += self.Rdj[nt2num[codon[0]],nt2num[codon[1]]]*self.first_nt_bias_insDJ[nt2num[codon[1]]] rTdj[aa] = current_lTdj #Compute rDdj rDdj = {} for aa in self.codons_dict.keys(): current_rDdj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_rDdj[nt2num[codon[0]], nt2num[codon[2]]] += self.first_nt_bias_insDJ[nt2num[codon[1]]] rDdj[aa] = current_rDdj #Set the attributes self.Tdj = Tdj self.Sdj = Sdj self.Ddj = Ddj self.rTdj = rTdj self.rDdj = rDdj
python
def generate_DJ_junction_transfer_matrices(self): """Compute the transfer matrices for the VD junction. Sets the attributes Tdj, Sdj, Ddj, rTdj, and rDdj. """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #Compute Tdj Tdj = {} for aa in self.codons_dict.keys(): current_Tdj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Tdj[nt2num[codon[0]], nt2num[init_nt]] += self.Rdj[nt2num[codon[0]],nt2num[codon[1]]]*self.Rdj[nt2num[codon[1]],nt2num[codon[2]]] * self.Rdj[nt2num[codon[2]],nt2num[init_nt]] Tdj[aa] = current_Tdj #Compute Sdj Sdj = {} for aa in self.codons_dict.keys(): current_Sdj = np.zeros((4, 4)) for ins_nt in 'ACGT': if any([codon.endswith(ins_nt) for codon in self.codons_dict[aa]]): current_Sdj[nt2num[ins_nt], :] = self.Rdj[nt2num[ins_nt], :] Sdj[aa] = current_Sdj #Compute Ddj Ddj = {} for aa in self.codons_dict.keys(): current_Ddj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Ddj[nt2num[codon[0]], nt2num[init_nt]] += self.Rdj[nt2num[codon[1]],nt2num[codon[2]]] * self.Rdj[nt2num[codon[2]],nt2num[init_nt]] Ddj[aa] = current_Ddj #Compute rTdj rTdj = {} for aa in self.codons_dict.keys(): current_lTdj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lTdj[nt2num[codon[0]], nt2num[codon[2]]] += self.Rdj[nt2num[codon[0]],nt2num[codon[1]]]*self.first_nt_bias_insDJ[nt2num[codon[1]]] rTdj[aa] = current_lTdj #Compute rDdj rDdj = {} for aa in self.codons_dict.keys(): current_rDdj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_rDdj[nt2num[codon[0]], nt2num[codon[2]]] += self.first_nt_bias_insDJ[nt2num[codon[1]]] rDdj[aa] = current_rDdj #Set the attributes self.Tdj = Tdj self.Sdj = Sdj self.Ddj = Ddj self.rTdj = rTdj self.rDdj = rDdj
[ "def", "generate_DJ_junction_transfer_matrices", "(", "self", ")", ":", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "#Compute Tdj ", "Tdj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Tdj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "init_nt", "in", "'ACGT'", ":", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_Tdj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "+=", "self", ".", "Rdj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "*", "self", ".", "Rdj", "[", "nt2num", "[", "codon", "[", "1", "]", "]", ",", "nt2num", "[", "codon", "[", "2", "]", "]", "]", "*", "self", ".", "Rdj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "Tdj", "[", "aa", "]", "=", "current_Tdj", "#Compute Sdj", "Sdj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Sdj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "ins_nt", "in", "'ACGT'", ":", "if", "any", "(", "[", "codon", ".", "endswith", "(", "ins_nt", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", "]", ")", ":", "current_Sdj", "[", "nt2num", "[", "ins_nt", "]", ",", ":", "]", "=", "self", ".", "Rdj", "[", "nt2num", "[", "ins_nt", "]", ",", ":", "]", "Sdj", "[", "aa", "]", "=", "current_Sdj", "#Compute Ddj", "Ddj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Ddj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "init_nt", "in", "'ACGT'", ":", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_Ddj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "+=", "self", ".", "Rdj", "[", "nt2num", "[", "codon", "[", "1", "]", "]", ",", "nt2num", "[", "codon", "[", "2", "]", "]", "]", "*", "self", ".", "Rdj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "Ddj", "[", "aa", "]", "=", "current_Ddj", "#Compute rTdj", "rTdj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_lTdj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_lTdj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "codon", "[", "2", "]", "]", "]", "+=", "self", ".", "Rdj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "*", "self", ".", "first_nt_bias_insDJ", "[", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "rTdj", "[", "aa", "]", "=", "current_lTdj", "#Compute rDdj", "rDdj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_rDdj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_rDdj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "codon", "[", "2", "]", "]", "]", "+=", "self", ".", "first_nt_bias_insDJ", "[", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "rDdj", "[", "aa", "]", "=", "current_rDdj", "#Set the attributes", "self", ".", "Tdj", "=", "Tdj", "self", ".", "Sdj", "=", "Sdj", "self", ".", "Ddj", "=", "Ddj", "self", ".", "rTdj", "=", "rTdj", "self", ".", "rDdj", "=", "rDdj" ]
Compute the transfer matrices for the VD junction. Sets the attributes Tdj, Sdj, Ddj, rTdj, and rDdj.
[ "Compute", "the", "transfer", "matrices", "for", "the", "VD", "junction", ".", "Sets", "the", "attributes", "Tdj", "Sdj", "Ddj", "rTdj", "and", "rDdj", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L656-L713
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVJ.generate_PVdelV_nt_pos_vecs
def generate_PVdelV_nt_pos_vecs(self, generative_model, genomic_data): """Process P(delV|V) into Pi arrays. Set the attributes PVdelV_nt_pos_vec and PVdelV_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVJ VJ generative model class containing the model parameters. genomic_data : GenomicDataVJ VJ genomic data class containing the V and J germline sequences and info. """ cutV_genomic_CDR3_segs = genomic_data.cutV_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_del_pos = generative_model.PdelV_given_V.shape[0] num_V_genes = generative_model.PdelV_given_V.shape[1] PVdelV_nt_pos_vec = [[]]*num_V_genes PVdelV_2nd_nt_pos_per_aa_vec = [[]]*num_V_genes for V_in in range(num_V_genes): current_PVdelV_nt_pos_vec = np.zeros((4, len(cutV_genomic_CDR3_segs[V_in]))) current_PVdelV_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PVdelV_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutV_genomic_CDR3_segs[V_in]))) for pos, nt in enumerate(cutV_genomic_CDR3_segs[V_in]): if len(cutV_genomic_CDR3_segs[V_in]) - pos > num_del_pos: continue if pos%3 == 0: #Start of a codon current_PVdelV_nt_pos_vec[nt2num[nt], pos] = generative_model.PdelV_given_V[len(cutV_genomic_CDR3_segs[V_in])-pos-1, V_in] elif pos%3 == 1: #Mid codon position for ins_nt in 'ACGT': #We need to find what possible codons are allowed for any aa (or motif) for aa in self.codons_dict.keys(): if cutV_genomic_CDR3_segs[V_in][pos-1:pos+1]+ ins_nt in self.codons_dict[aa]: current_PVdelV_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = generative_model.PdelV_given_V[len(cutV_genomic_CDR3_segs[V_in])-pos-1, V_in] elif pos%3 == 2: #End of codon current_PVdelV_nt_pos_vec[0, pos] = generative_model.PdelV_given_V[len(cutV_genomic_CDR3_segs[V_in])-pos-1, V_in] PVdelV_nt_pos_vec[V_in] = current_PVdelV_nt_pos_vec PVdelV_2nd_nt_pos_per_aa_vec[V_in] = current_PVdelV_2nd_nt_pos_per_aa_vec self.PVdelV_nt_pos_vec = PVdelV_nt_pos_vec self.PVdelV_2nd_nt_pos_per_aa_vec = PVdelV_2nd_nt_pos_per_aa_vec
python
def generate_PVdelV_nt_pos_vecs(self, generative_model, genomic_data): """Process P(delV|V) into Pi arrays. Set the attributes PVdelV_nt_pos_vec and PVdelV_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVJ VJ generative model class containing the model parameters. genomic_data : GenomicDataVJ VJ genomic data class containing the V and J germline sequences and info. """ cutV_genomic_CDR3_segs = genomic_data.cutV_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_del_pos = generative_model.PdelV_given_V.shape[0] num_V_genes = generative_model.PdelV_given_V.shape[1] PVdelV_nt_pos_vec = [[]]*num_V_genes PVdelV_2nd_nt_pos_per_aa_vec = [[]]*num_V_genes for V_in in range(num_V_genes): current_PVdelV_nt_pos_vec = np.zeros((4, len(cutV_genomic_CDR3_segs[V_in]))) current_PVdelV_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PVdelV_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutV_genomic_CDR3_segs[V_in]))) for pos, nt in enumerate(cutV_genomic_CDR3_segs[V_in]): if len(cutV_genomic_CDR3_segs[V_in]) - pos > num_del_pos: continue if pos%3 == 0: #Start of a codon current_PVdelV_nt_pos_vec[nt2num[nt], pos] = generative_model.PdelV_given_V[len(cutV_genomic_CDR3_segs[V_in])-pos-1, V_in] elif pos%3 == 1: #Mid codon position for ins_nt in 'ACGT': #We need to find what possible codons are allowed for any aa (or motif) for aa in self.codons_dict.keys(): if cutV_genomic_CDR3_segs[V_in][pos-1:pos+1]+ ins_nt in self.codons_dict[aa]: current_PVdelV_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = generative_model.PdelV_given_V[len(cutV_genomic_CDR3_segs[V_in])-pos-1, V_in] elif pos%3 == 2: #End of codon current_PVdelV_nt_pos_vec[0, pos] = generative_model.PdelV_given_V[len(cutV_genomic_CDR3_segs[V_in])-pos-1, V_in] PVdelV_nt_pos_vec[V_in] = current_PVdelV_nt_pos_vec PVdelV_2nd_nt_pos_per_aa_vec[V_in] = current_PVdelV_2nd_nt_pos_per_aa_vec self.PVdelV_nt_pos_vec = PVdelV_nt_pos_vec self.PVdelV_2nd_nt_pos_per_aa_vec = PVdelV_2nd_nt_pos_per_aa_vec
[ "def", "generate_PVdelV_nt_pos_vecs", "(", "self", ",", "generative_model", ",", "genomic_data", ")", ":", "cutV_genomic_CDR3_segs", "=", "genomic_data", ".", "cutV_genomic_CDR3_segs", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "num_del_pos", "=", "generative_model", ".", "PdelV_given_V", ".", "shape", "[", "0", "]", "num_V_genes", "=", "generative_model", ".", "PdelV_given_V", ".", "shape", "[", "1", "]", "PVdelV_nt_pos_vec", "=", "[", "[", "]", "]", "*", "num_V_genes", "PVdelV_2nd_nt_pos_per_aa_vec", "=", "[", "[", "]", "]", "*", "num_V_genes", "for", "V_in", "in", "range", "(", "num_V_genes", ")", ":", "current_PVdelV_nt_pos_vec", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", ")", ")", "current_PVdelV_2nd_nt_pos_per_aa_vec", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_PVdelV_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", ")", ")", "for", "pos", ",", "nt", "in", "enumerate", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", ":", "if", "len", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", "-", "pos", ">", "num_del_pos", ":", "continue", "if", "pos", "%", "3", "==", "0", ":", "#Start of a codon", "current_PVdelV_nt_pos_vec", "[", "nt2num", "[", "nt", "]", ",", "pos", "]", "=", "generative_model", ".", "PdelV_given_V", "[", "len", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", "-", "pos", "-", "1", ",", "V_in", "]", "elif", "pos", "%", "3", "==", "1", ":", "#Mid codon position", "for", "ins_nt", "in", "'ACGT'", ":", "#We need to find what possible codons are allowed for any aa (or motif)", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "if", "cutV_genomic_CDR3_segs", "[", "V_in", "]", "[", "pos", "-", "1", ":", "pos", "+", "1", "]", "+", "ins_nt", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_PVdelV_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "[", "nt2num", "[", "ins_nt", "]", ",", "pos", "]", "=", "generative_model", ".", "PdelV_given_V", "[", "len", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", "-", "pos", "-", "1", ",", "V_in", "]", "elif", "pos", "%", "3", "==", "2", ":", "#End of codon", "current_PVdelV_nt_pos_vec", "[", "0", ",", "pos", "]", "=", "generative_model", ".", "PdelV_given_V", "[", "len", "(", "cutV_genomic_CDR3_segs", "[", "V_in", "]", ")", "-", "pos", "-", "1", ",", "V_in", "]", "PVdelV_nt_pos_vec", "[", "V_in", "]", "=", "current_PVdelV_nt_pos_vec", "PVdelV_2nd_nt_pos_per_aa_vec", "[", "V_in", "]", "=", "current_PVdelV_2nd_nt_pos_per_aa_vec", "self", ".", "PVdelV_nt_pos_vec", "=", "PVdelV_nt_pos_vec", "self", ".", "PVdelV_2nd_nt_pos_per_aa_vec", "=", "PVdelV_2nd_nt_pos_per_aa_vec" ]
Process P(delV|V) into Pi arrays. Set the attributes PVdelV_nt_pos_vec and PVdelV_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVJ VJ generative model class containing the model parameters. genomic_data : GenomicDataVJ VJ genomic data class containing the V and J germline sequences and info.
[ "Process", "P", "(", "delV|V", ")", "into", "Pi", "arrays", ".", "Set", "the", "attributes", "PVdelV_nt_pos_vec", "and", "PVdelV_2nd_nt_pos_per_aa_vec", ".", "Parameters", "----------", "generative_model", ":", "GenerativeModelVJ", "VJ", "generative", "model", "class", "containing", "the", "model", "parameters", ".", "genomic_data", ":", "GenomicDataVJ", "VJ", "genomic", "data", "class", "containing", "the", "V", "and", "J", "germline", "sequences", "and", "info", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L860-L904
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVJ.generate_PJdelJ_nt_pos_vecs
def generate_PJdelJ_nt_pos_vecs(self, generative_model, genomic_data): """Process P(delJ|J) into Pi arrays. Set the attributes PJdelJ_nt_pos_vec and PJdelJ_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVJ VJ generative model class containing the model parameters. genomic_data : GenomicDataVJ VJ genomic data class containing the V and J germline sequences and info. """ cutJ_genomic_CDR3_segs = genomic_data.cutJ_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_del_pos = generative_model.PdelJ_given_J.shape[0] num_D_genes, num_J_genes = generative_model.PVJ.shape PJdelJ_nt_pos_vec = [[]]*num_J_genes PJdelJ_2nd_nt_pos_per_aa_vec = [[]]*num_J_genes for J_in in range(num_J_genes): current_PJdelJ_nt_pos_vec = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) current_PJdelJ_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PJdelJ_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) for pos, nt in enumerate(cutJ_genomic_CDR3_segs[J_in]): if pos >= num_del_pos: continue if (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 1: #Start of a codon current_PJdelJ_nt_pos_vec[nt2num[nt], pos] = generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 2: #Mid codon position for ins_nt in 'ACGT': #We need to find what possible codons are allowed for any aa (or motif) for aa in self.codons_dict.keys(): if ins_nt + cutJ_genomic_CDR3_segs[J_in][pos:pos+2] in self.codons_dict[aa]: current_PJdelJ_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 0: #End of codon current_PJdelJ_nt_pos_vec[0, pos] = generative_model.PdelJ_given_J[pos, J_in] PJdelJ_nt_pos_vec[J_in] = current_PJdelJ_nt_pos_vec PJdelJ_2nd_nt_pos_per_aa_vec[J_in] = current_PJdelJ_2nd_nt_pos_per_aa_vec self.PJdelJ_nt_pos_vec = PJdelJ_nt_pos_vec self.PJdelJ_2nd_nt_pos_per_aa_vec = PJdelJ_2nd_nt_pos_per_aa_vec
python
def generate_PJdelJ_nt_pos_vecs(self, generative_model, genomic_data): """Process P(delJ|J) into Pi arrays. Set the attributes PJdelJ_nt_pos_vec and PJdelJ_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVJ VJ generative model class containing the model parameters. genomic_data : GenomicDataVJ VJ genomic data class containing the V and J germline sequences and info. """ cutJ_genomic_CDR3_segs = genomic_data.cutJ_genomic_CDR3_segs nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} num_del_pos = generative_model.PdelJ_given_J.shape[0] num_D_genes, num_J_genes = generative_model.PVJ.shape PJdelJ_nt_pos_vec = [[]]*num_J_genes PJdelJ_2nd_nt_pos_per_aa_vec = [[]]*num_J_genes for J_in in range(num_J_genes): current_PJdelJ_nt_pos_vec = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) current_PJdelJ_2nd_nt_pos_per_aa_vec = {} for aa in self.codons_dict.keys(): current_PJdelJ_2nd_nt_pos_per_aa_vec[aa] = np.zeros((4, len(cutJ_genomic_CDR3_segs[J_in]))) for pos, nt in enumerate(cutJ_genomic_CDR3_segs[J_in]): if pos >= num_del_pos: continue if (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 1: #Start of a codon current_PJdelJ_nt_pos_vec[nt2num[nt], pos] = generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 2: #Mid codon position for ins_nt in 'ACGT': #We need to find what possible codons are allowed for any aa (or motif) for aa in self.codons_dict.keys(): if ins_nt + cutJ_genomic_CDR3_segs[J_in][pos:pos+2] in self.codons_dict[aa]: current_PJdelJ_2nd_nt_pos_per_aa_vec[aa][nt2num[ins_nt], pos] = generative_model.PdelJ_given_J[pos, J_in] elif (len(cutJ_genomic_CDR3_segs[J_in]) - pos)%3 == 0: #End of codon current_PJdelJ_nt_pos_vec[0, pos] = generative_model.PdelJ_given_J[pos, J_in] PJdelJ_nt_pos_vec[J_in] = current_PJdelJ_nt_pos_vec PJdelJ_2nd_nt_pos_per_aa_vec[J_in] = current_PJdelJ_2nd_nt_pos_per_aa_vec self.PJdelJ_nt_pos_vec = PJdelJ_nt_pos_vec self.PJdelJ_2nd_nt_pos_per_aa_vec = PJdelJ_2nd_nt_pos_per_aa_vec
[ "def", "generate_PJdelJ_nt_pos_vecs", "(", "self", ",", "generative_model", ",", "genomic_data", ")", ":", "cutJ_genomic_CDR3_segs", "=", "genomic_data", ".", "cutJ_genomic_CDR3_segs", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "num_del_pos", "=", "generative_model", ".", "PdelJ_given_J", ".", "shape", "[", "0", "]", "num_D_genes", ",", "num_J_genes", "=", "generative_model", ".", "PVJ", ".", "shape", "PJdelJ_nt_pos_vec", "=", "[", "[", "]", "]", "*", "num_J_genes", "PJdelJ_2nd_nt_pos_per_aa_vec", "=", "[", "[", "]", "]", "*", "num_J_genes", "for", "J_in", "in", "range", "(", "num_J_genes", ")", ":", "current_PJdelJ_nt_pos_vec", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", ")", ")", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "=", "np", ".", "zeros", "(", "(", "4", ",", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", ")", ")", "for", "pos", ",", "nt", "in", "enumerate", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", ":", "if", "pos", ">=", "num_del_pos", ":", "continue", "if", "(", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", "-", "pos", ")", "%", "3", "==", "1", ":", "#Start of a codon", "current_PJdelJ_nt_pos_vec", "[", "nt2num", "[", "nt", "]", ",", "pos", "]", "=", "generative_model", ".", "PdelJ_given_J", "[", "pos", ",", "J_in", "]", "elif", "(", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", "-", "pos", ")", "%", "3", "==", "2", ":", "#Mid codon position", "for", "ins_nt", "in", "'ACGT'", ":", "#We need to find what possible codons are allowed for any aa (or motif)", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "if", "ins_nt", "+", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", "[", "pos", ":", "pos", "+", "2", "]", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "[", "aa", "]", "[", "nt2num", "[", "ins_nt", "]", ",", "pos", "]", "=", "generative_model", ".", "PdelJ_given_J", "[", "pos", ",", "J_in", "]", "elif", "(", "len", "(", "cutJ_genomic_CDR3_segs", "[", "J_in", "]", ")", "-", "pos", ")", "%", "3", "==", "0", ":", "#End of codon", "current_PJdelJ_nt_pos_vec", "[", "0", ",", "pos", "]", "=", "generative_model", ".", "PdelJ_given_J", "[", "pos", ",", "J_in", "]", "PJdelJ_nt_pos_vec", "[", "J_in", "]", "=", "current_PJdelJ_nt_pos_vec", "PJdelJ_2nd_nt_pos_per_aa_vec", "[", "J_in", "]", "=", "current_PJdelJ_2nd_nt_pos_per_aa_vec", "self", ".", "PJdelJ_nt_pos_vec", "=", "PJdelJ_nt_pos_vec", "self", ".", "PJdelJ_2nd_nt_pos_per_aa_vec", "=", "PJdelJ_2nd_nt_pos_per_aa_vec" ]
Process P(delJ|J) into Pi arrays. Set the attributes PJdelJ_nt_pos_vec and PJdelJ_2nd_nt_pos_per_aa_vec. Parameters ---------- generative_model : GenerativeModelVJ VJ generative model class containing the model parameters. genomic_data : GenomicDataVJ VJ genomic data class containing the V and J germline sequences and info.
[ "Process", "P", "(", "delJ|J", ")", "into", "Pi", "arrays", ".", "Set", "the", "attributes", "PJdelJ_nt_pos_vec", "and", "PJdelJ_2nd_nt_pos_per_aa_vec", ".", "Parameters", "----------", "generative_model", ":", "GenerativeModelVJ", "VJ", "generative", "model", "class", "containing", "the", "model", "parameters", ".", "genomic_data", ":", "GenomicDataVJ", "VJ", "genomic", "data", "class", "containing", "the", "V", "and", "J", "germline", "sequences", "and", "info", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L906-L953
train
zsethna/OLGA
olga/preprocess_generative_model_and_data.py
PreprocessedParametersVJ.generate_VJ_junction_transfer_matrices
def generate_VJ_junction_transfer_matrices(self): """Compute the transfer matrices for the VJ junction. Sets the attributes Tvj, Svj, Dvj, lTvj, and lDvj. """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #Compute Tvj Tvj = {} for aa in self.codons_dict.keys(): current_Tvj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Tvj[nt2num[codon[2]], nt2num[init_nt]] += self.Rvj[nt2num[codon[2]],nt2num[codon[1]]]*self.Rvj[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvj[nt2num[codon[0]],nt2num[init_nt]] Tvj[aa] = current_Tvj #Compute Svj Svj = {} for aa in self.codons_dict.keys(): current_Svj = np.zeros((4, 4)) for ins_nt in 'ACGT': if any([codon.startswith(ins_nt) for codon in self.codons_dict[aa]]): current_Svj[nt2num[ins_nt], :] = self.Rvj[nt2num[ins_nt], :] Svj[aa] = current_Svj #Compute Dvj Dvj = {} for aa in self.codons_dict.keys(): current_Dvj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Dvj[nt2num[codon[2]], nt2num[init_nt]] += self.Rvj[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvj[nt2num[codon[0]],nt2num[init_nt]] Dvj[aa] = current_Dvj #Compute lTvj lTvj = {} for aa in self.codons_dict.keys(): current_lTvj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lTvj[nt2num[codon[2]], nt2num[codon[0]]] += self.Rvj[nt2num[codon[2]],nt2num[codon[1]]]*self.first_nt_bias_insVJ[nt2num[codon[1]]] lTvj[aa] = current_lTvj #Compute lDvj lDvj = {} for aa in self.codons_dict.keys(): current_lDvj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lDvj[nt2num[codon[2]], nt2num[codon[0]]] += self.first_nt_bias_insVJ[nt2num[codon[1]]] lDvj[aa] = current_lDvj #Set the attributes self.Tvj = Tvj self.Svj = Svj self.Dvj = Dvj self.lTvj = lTvj self.lDvj = lDvj
python
def generate_VJ_junction_transfer_matrices(self): """Compute the transfer matrices for the VJ junction. Sets the attributes Tvj, Svj, Dvj, lTvj, and lDvj. """ nt2num = {'A': 0, 'C': 1, 'G': 2, 'T': 3} #Compute Tvj Tvj = {} for aa in self.codons_dict.keys(): current_Tvj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Tvj[nt2num[codon[2]], nt2num[init_nt]] += self.Rvj[nt2num[codon[2]],nt2num[codon[1]]]*self.Rvj[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvj[nt2num[codon[0]],nt2num[init_nt]] Tvj[aa] = current_Tvj #Compute Svj Svj = {} for aa in self.codons_dict.keys(): current_Svj = np.zeros((4, 4)) for ins_nt in 'ACGT': if any([codon.startswith(ins_nt) for codon in self.codons_dict[aa]]): current_Svj[nt2num[ins_nt], :] = self.Rvj[nt2num[ins_nt], :] Svj[aa] = current_Svj #Compute Dvj Dvj = {} for aa in self.codons_dict.keys(): current_Dvj = np.zeros((4, 4)) for init_nt in 'ACGT': for codon in self.codons_dict[aa]: current_Dvj[nt2num[codon[2]], nt2num[init_nt]] += self.Rvj[nt2num[codon[1]],nt2num[codon[0]]] * self.Rvj[nt2num[codon[0]],nt2num[init_nt]] Dvj[aa] = current_Dvj #Compute lTvj lTvj = {} for aa in self.codons_dict.keys(): current_lTvj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lTvj[nt2num[codon[2]], nt2num[codon[0]]] += self.Rvj[nt2num[codon[2]],nt2num[codon[1]]]*self.first_nt_bias_insVJ[nt2num[codon[1]]] lTvj[aa] = current_lTvj #Compute lDvj lDvj = {} for aa in self.codons_dict.keys(): current_lDvj = np.zeros((4, 4)) for codon in self.codons_dict[aa]: current_lDvj[nt2num[codon[2]], nt2num[codon[0]]] += self.first_nt_bias_insVJ[nt2num[codon[1]]] lDvj[aa] = current_lDvj #Set the attributes self.Tvj = Tvj self.Svj = Svj self.Dvj = Dvj self.lTvj = lTvj self.lDvj = lDvj
[ "def", "generate_VJ_junction_transfer_matrices", "(", "self", ")", ":", "nt2num", "=", "{", "'A'", ":", "0", ",", "'C'", ":", "1", ",", "'G'", ":", "2", ",", "'T'", ":", "3", "}", "#Compute Tvj", "Tvj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Tvj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "init_nt", "in", "'ACGT'", ":", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_Tvj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "+=", "self", ".", "Rvj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "*", "self", ".", "Rvj", "[", "nt2num", "[", "codon", "[", "1", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "*", "self", ".", "Rvj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "Tvj", "[", "aa", "]", "=", "current_Tvj", "#Compute Svj", "Svj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Svj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "ins_nt", "in", "'ACGT'", ":", "if", "any", "(", "[", "codon", ".", "startswith", "(", "ins_nt", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", "]", ")", ":", "current_Svj", "[", "nt2num", "[", "ins_nt", "]", ",", ":", "]", "=", "self", ".", "Rvj", "[", "nt2num", "[", "ins_nt", "]", ",", ":", "]", "Svj", "[", "aa", "]", "=", "current_Svj", "#Compute Dvj ", "Dvj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_Dvj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "init_nt", "in", "'ACGT'", ":", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_Dvj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "+=", "self", ".", "Rvj", "[", "nt2num", "[", "codon", "[", "1", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "*", "self", ".", "Rvj", "[", "nt2num", "[", "codon", "[", "0", "]", "]", ",", "nt2num", "[", "init_nt", "]", "]", "Dvj", "[", "aa", "]", "=", "current_Dvj", "#Compute lTvj", "lTvj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_lTvj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_lTvj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "+=", "self", ".", "Rvj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "*", "self", ".", "first_nt_bias_insVJ", "[", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "lTvj", "[", "aa", "]", "=", "current_lTvj", "#Compute lDvj ", "lDvj", "=", "{", "}", "for", "aa", "in", "self", ".", "codons_dict", ".", "keys", "(", ")", ":", "current_lDvj", "=", "np", ".", "zeros", "(", "(", "4", ",", "4", ")", ")", "for", "codon", "in", "self", ".", "codons_dict", "[", "aa", "]", ":", "current_lDvj", "[", "nt2num", "[", "codon", "[", "2", "]", "]", ",", "nt2num", "[", "codon", "[", "0", "]", "]", "]", "+=", "self", ".", "first_nt_bias_insVJ", "[", "nt2num", "[", "codon", "[", "1", "]", "]", "]", "lDvj", "[", "aa", "]", "=", "current_lDvj", "#Set the attributes", "self", ".", "Tvj", "=", "Tvj", "self", ".", "Svj", "=", "Svj", "self", ".", "Dvj", "=", "Dvj", "self", ".", "lTvj", "=", "lTvj", "self", ".", "lDvj", "=", "lDvj" ]
Compute the transfer matrices for the VJ junction. Sets the attributes Tvj, Svj, Dvj, lTvj, and lDvj.
[ "Compute", "the", "transfer", "matrices", "for", "the", "VJ", "junction", ".", "Sets", "the", "attributes", "Tvj", "Svj", "Dvj", "lTvj", "and", "lDvj", "." ]
e825c333f0f9a4eb02132e0bcf86f0dca9123114
https://github.com/zsethna/OLGA/blob/e825c333f0f9a4eb02132e0bcf86f0dca9123114/olga/preprocess_generative_model_and_data.py#L956-L1014
train
crackinglandia/pype32
tools/readpe.py
showDosHeaderData
def showDosHeaderData(peInstance): """ Prints IMAGE_DOS_HEADER fields. """ dosFields = peInstance.dosHeader.getFields() print "[+] IMAGE_DOS_HEADER values:\n" for field in dosFields: if isinstance(dosFields[field], datatypes.Array): print "--> %s - Array of length %d" % (field, len(dosFields[field])) counter = 0 for element in dosFields[field]: print "[%d] 0x%08x" % (counter, element.value) counter += 1 else: print "--> %s = 0x%08x" % (field, dosFields[field].value)
python
def showDosHeaderData(peInstance): """ Prints IMAGE_DOS_HEADER fields. """ dosFields = peInstance.dosHeader.getFields() print "[+] IMAGE_DOS_HEADER values:\n" for field in dosFields: if isinstance(dosFields[field], datatypes.Array): print "--> %s - Array of length %d" % (field, len(dosFields[field])) counter = 0 for element in dosFields[field]: print "[%d] 0x%08x" % (counter, element.value) counter += 1 else: print "--> %s = 0x%08x" % (field, dosFields[field].value)
[ "def", "showDosHeaderData", "(", "peInstance", ")", ":", "dosFields", "=", "peInstance", ".", "dosHeader", ".", "getFields", "(", ")", "print", "\"[+] IMAGE_DOS_HEADER values:\\n\"", "for", "field", "in", "dosFields", ":", "if", "isinstance", "(", "dosFields", "[", "field", "]", ",", "datatypes", ".", "Array", ")", ":", "print", "\"--> %s - Array of length %d\"", "%", "(", "field", ",", "len", "(", "dosFields", "[", "field", "]", ")", ")", "counter", "=", "0", "for", "element", "in", "dosFields", "[", "field", "]", ":", "print", "\"[%d] 0x%08x\"", "%", "(", "counter", ",", "element", ".", "value", ")", "counter", "+=", "1", "else", ":", "print", "\"--> %s = 0x%08x\"", "%", "(", "field", ",", "dosFields", "[", "field", "]", ".", "value", ")" ]
Prints IMAGE_DOS_HEADER fields.
[ "Prints", "IMAGE_DOS_HEADER", "fields", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L43-L56
train
crackinglandia/pype32
tools/readpe.py
showFileHeaderData
def showFileHeaderData(peInstance): """ Prints IMAGE_FILE_HEADER fields. """ fileHeaderFields = peInstance.ntHeaders.fileHeader.getFields() print "[+] IMAGE_FILE_HEADER values:\n" for field in fileHeaderFields: print "--> %s = 0x%08x" % (field, fileHeaderFields[field].value)
python
def showFileHeaderData(peInstance): """ Prints IMAGE_FILE_HEADER fields. """ fileHeaderFields = peInstance.ntHeaders.fileHeader.getFields() print "[+] IMAGE_FILE_HEADER values:\n" for field in fileHeaderFields: print "--> %s = 0x%08x" % (field, fileHeaderFields[field].value)
[ "def", "showFileHeaderData", "(", "peInstance", ")", ":", "fileHeaderFields", "=", "peInstance", ".", "ntHeaders", ".", "fileHeader", ".", "getFields", "(", ")", "print", "\"[+] IMAGE_FILE_HEADER values:\\n\"", "for", "field", "in", "fileHeaderFields", ":", "print", "\"--> %s = 0x%08x\"", "%", "(", "field", ",", "fileHeaderFields", "[", "field", "]", ".", "value", ")" ]
Prints IMAGE_FILE_HEADER fields.
[ "Prints", "IMAGE_FILE_HEADER", "fields", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L64-L70
train
crackinglandia/pype32
tools/readpe.py
showOptionalHeaderData
def showOptionalHeaderData(peInstance): """ Prints IMAGE_OPTIONAL_HEADER fields. """ print "[+] IMAGE_OPTIONAL_HEADER:\n" ohFields = peInstance.ntHeaders.optionalHeader.getFields() for field in ohFields: if not isinstance(ohFields[field], datadirs.DataDirectory): print "--> %s = 0x%08x" % (field, ohFields[field].value)
python
def showOptionalHeaderData(peInstance): """ Prints IMAGE_OPTIONAL_HEADER fields. """ print "[+] IMAGE_OPTIONAL_HEADER:\n" ohFields = peInstance.ntHeaders.optionalHeader.getFields() for field in ohFields: if not isinstance(ohFields[field], datadirs.DataDirectory): print "--> %s = 0x%08x" % (field, ohFields[field].value)
[ "def", "showOptionalHeaderData", "(", "peInstance", ")", ":", "print", "\"[+] IMAGE_OPTIONAL_HEADER:\\n\"", "ohFields", "=", "peInstance", ".", "ntHeaders", ".", "optionalHeader", ".", "getFields", "(", ")", "for", "field", "in", "ohFields", ":", "if", "not", "isinstance", "(", "ohFields", "[", "field", "]", ",", "datadirs", ".", "DataDirectory", ")", ":", "print", "\"--> %s = 0x%08x\"", "%", "(", "field", ",", "ohFields", "[", "field", "]", ".", "value", ")" ]
Prints IMAGE_OPTIONAL_HEADER fields.
[ "Prints", "IMAGE_OPTIONAL_HEADER", "fields", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L72-L79
train
crackinglandia/pype32
tools/readpe.py
showDataDirectoriesData
def showDataDirectoriesData(peInstance): """ Prints the DATA_DIRECTORY fields. """ print "[+] Data directories:\n" dirs = peInstance.ntHeaders.optionalHeader.dataDirectory counter = 1 for dir in dirs: print "[%d] --> Name: %s -- RVA: 0x%08x -- SIZE: 0x%08x" % (counter, dir.name.value, dir.rva.value, dir.size.value) counter += 1
python
def showDataDirectoriesData(peInstance): """ Prints the DATA_DIRECTORY fields. """ print "[+] Data directories:\n" dirs = peInstance.ntHeaders.optionalHeader.dataDirectory counter = 1 for dir in dirs: print "[%d] --> Name: %s -- RVA: 0x%08x -- SIZE: 0x%08x" % (counter, dir.name.value, dir.rva.value, dir.size.value) counter += 1
[ "def", "showDataDirectoriesData", "(", "peInstance", ")", ":", "print", "\"[+] Data directories:\\n\"", "dirs", "=", "peInstance", ".", "ntHeaders", ".", "optionalHeader", ".", "dataDirectory", "counter", "=", "1", "for", "dir", "in", "dirs", ":", "print", "\"[%d] --> Name: %s -- RVA: 0x%08x -- SIZE: 0x%08x\"", "%", "(", "counter", ",", "dir", ".", "name", ".", "value", ",", "dir", ".", "rva", ".", "value", ",", "dir", ".", "size", ".", "value", ")", "counter", "+=", "1" ]
Prints the DATA_DIRECTORY fields.
[ "Prints", "the", "DATA_DIRECTORY", "fields", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L81-L89
train
crackinglandia/pype32
tools/readpe.py
showSectionsHeaders
def showSectionsHeaders(peInstance): """ Prints IMAGE_SECTION_HEADER for every section present in the file. """ print "[+] Sections information:\n" print "--> NumberOfSections: %d\n" % peInstance.ntHeaders.fileHeader.numberOfSections.value for section in peInstance.sectionHeaders: fields = section.getFields() for field in fields: if isinstance(fields[field], datatypes.String): fmt = "%s = %s" else: fmt = "%s = 0x%08x" print fmt % (field, fields[field].value) print "\n"
python
def showSectionsHeaders(peInstance): """ Prints IMAGE_SECTION_HEADER for every section present in the file. """ print "[+] Sections information:\n" print "--> NumberOfSections: %d\n" % peInstance.ntHeaders.fileHeader.numberOfSections.value for section in peInstance.sectionHeaders: fields = section.getFields() for field in fields: if isinstance(fields[field], datatypes.String): fmt = "%s = %s" else: fmt = "%s = 0x%08x" print fmt % (field, fields[field].value) print "\n"
[ "def", "showSectionsHeaders", "(", "peInstance", ")", ":", "print", "\"[+] Sections information:\\n\"", "print", "\"--> NumberOfSections: %d\\n\"", "%", "peInstance", ".", "ntHeaders", ".", "fileHeader", ".", "numberOfSections", ".", "value", "for", "section", "in", "peInstance", ".", "sectionHeaders", ":", "fields", "=", "section", ".", "getFields", "(", ")", "for", "field", "in", "fields", ":", "if", "isinstance", "(", "fields", "[", "field", "]", ",", "datatypes", ".", "String", ")", ":", "fmt", "=", "\"%s = %s\"", "else", ":", "fmt", "=", "\"%s = 0x%08x\"", "print", "fmt", "%", "(", "field", ",", "fields", "[", "field", "]", ".", "value", ")", "print", "\"\\n\"" ]
Prints IMAGE_SECTION_HEADER for every section present in the file.
[ "Prints", "IMAGE_SECTION_HEADER", "for", "every", "section", "present", "in", "the", "file", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L91-L104
train
crackinglandia/pype32
tools/readpe.py
showImports
def showImports(peInstance): """ Shows imports information. """ iidEntries = peInstance.ntHeaders.optionalHeader.dataDirectory[consts.IMPORT_DIRECTORY].info if iidEntries: for iidEntry in iidEntries: fields = iidEntry.getFields() print "module: %s" % iidEntry.metaData.moduleName.value for field in fields: print "%s -> %x" % (field, fields[field].value) for iatEntry in iidEntry.iat: fields = iatEntry.getFields() for field in fields: print "%s - %r" % (field, fields[field].value) print "\n" else: print "The file does not have imported functions."
python
def showImports(peInstance): """ Shows imports information. """ iidEntries = peInstance.ntHeaders.optionalHeader.dataDirectory[consts.IMPORT_DIRECTORY].info if iidEntries: for iidEntry in iidEntries: fields = iidEntry.getFields() print "module: %s" % iidEntry.metaData.moduleName.value for field in fields: print "%s -> %x" % (field, fields[field].value) for iatEntry in iidEntry.iat: fields = iatEntry.getFields() for field in fields: print "%s - %r" % (field, fields[field].value) print "\n" else: print "The file does not have imported functions."
[ "def", "showImports", "(", "peInstance", ")", ":", "iidEntries", "=", "peInstance", ".", "ntHeaders", ".", "optionalHeader", ".", "dataDirectory", "[", "consts", ".", "IMPORT_DIRECTORY", "]", ".", "info", "if", "iidEntries", ":", "for", "iidEntry", "in", "iidEntries", ":", "fields", "=", "iidEntry", ".", "getFields", "(", ")", "print", "\"module: %s\"", "%", "iidEntry", ".", "metaData", ".", "moduleName", ".", "value", "for", "field", "in", "fields", ":", "print", "\"%s -> %x\"", "%", "(", "field", ",", "fields", "[", "field", "]", ".", "value", ")", "for", "iatEntry", "in", "iidEntry", ".", "iat", ":", "fields", "=", "iatEntry", ".", "getFields", "(", ")", "for", "field", "in", "fields", ":", "print", "\"%s - %r\"", "%", "(", "field", ",", "fields", "[", "field", "]", ".", "value", ")", "print", "\"\\n\"", "else", ":", "print", "\"The file does not have imported functions.\"" ]
Shows imports information.
[ "Shows", "imports", "information", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L106-L124
train
crackinglandia/pype32
tools/readpe.py
showExports
def showExports(peInstance): """ Show exports information """ exports = peInstance.ntHeaders.optionalHeader.dataDirectory[consts.EXPORT_DIRECTORY].info if exports: exp_fields = exports.getFields() for field in exp_fields: print "%s -> %x" % (field, exp_fields[field].value) for entry in exports.exportTable: entry_fields = entry.getFields() for field in entry_fields: print "%s -> %r" % (field, entry_fields[field].value) else: print "The file does not have exported functions."
python
def showExports(peInstance): """ Show exports information """ exports = peInstance.ntHeaders.optionalHeader.dataDirectory[consts.EXPORT_DIRECTORY].info if exports: exp_fields = exports.getFields() for field in exp_fields: print "%s -> %x" % (field, exp_fields[field].value) for entry in exports.exportTable: entry_fields = entry.getFields() for field in entry_fields: print "%s -> %r" % (field, entry_fields[field].value) else: print "The file does not have exported functions."
[ "def", "showExports", "(", "peInstance", ")", ":", "exports", "=", "peInstance", ".", "ntHeaders", ".", "optionalHeader", ".", "dataDirectory", "[", "consts", ".", "EXPORT_DIRECTORY", "]", ".", "info", "if", "exports", ":", "exp_fields", "=", "exports", ".", "getFields", "(", ")", "for", "field", "in", "exp_fields", ":", "print", "\"%s -> %x\"", "%", "(", "field", ",", "exp_fields", "[", "field", "]", ".", "value", ")", "for", "entry", "in", "exports", ".", "exportTable", ":", "entry_fields", "=", "entry", ".", "getFields", "(", ")", "for", "field", "in", "entry_fields", ":", "print", "\"%s -> %r\"", "%", "(", "field", ",", "entry_fields", "[", "field", "]", ".", "value", ")", "else", ":", "print", "\"The file does not have exported functions.\"" ]
Show exports information
[ "Show", "exports", "information" ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/tools/readpe.py#L126-L140
train
crackinglandia/pype32
pype32/baseclasses.py
BaseStructClass.getFields
def getFields(self): """ Returns all the class attributues. @rtype: dict @return: A dictionary containing all the class attributes. """ d = {} for i in self._attrsList: key = i value = getattr(self, i) d[key] = value return d
python
def getFields(self): """ Returns all the class attributues. @rtype: dict @return: A dictionary containing all the class attributes. """ d = {} for i in self._attrsList: key = i value = getattr(self, i) d[key] = value return d
[ "def", "getFields", "(", "self", ")", ":", "d", "=", "{", "}", "for", "i", "in", "self", ".", "_attrsList", ":", "key", "=", "i", "value", "=", "getattr", "(", "self", ",", "i", ")", "d", "[", "key", "]", "=", "value", "return", "d" ]
Returns all the class attributues. @rtype: dict @return: A dictionary containing all the class attributes.
[ "Returns", "all", "the", "class", "attributues", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/baseclasses.py#L65-L77
train
crackinglandia/pype32
pype32/datatypes.py
Array.parse
def parse(readDataInstance, arrayType, arrayLength): """ Returns a new L{Array} object. @type readDataInstance: L{ReadData} @param readDataInstance: The L{ReadData} object containing the array data. @type arrayType: int @param arrayType: The type of L{Array} to be built. @type arrayLength: int @param arrayLength: The length of the array passed as an argument. @rtype: L{Array} @return: New L{Array} object. """ newArray = Array(arrayType) dataLength = len(readDataInstance) if arrayType is TYPE_DWORD: toRead = arrayLength * 4 if dataLength >= toRead: for i in range(arrayLength): newArray.append(DWORD(readDataInstance.readDword())) else: raise excep.DataLengthException("Not enough bytes to read.") elif arrayType is TYPE_WORD: toRead = arrayLength * 2 if dataLength >= toRead: for i in range(arrayLength): newArray.append(DWORD(readDataInstance.readWord())) else: raise excep.DataLengthException("Not enough bytes to read.") elif arrayType is TYPE_QWORD: toRead = arrayLength * 8 if dataLength >= toRead: for i in range(arrayLength): newArray.append(QWORD(readDataInstance.readQword())) else: raise excep.DataLengthException("Not enough bytes to read.") elif arrayType is TYPE_BYTE: for i in range(arrayLength): newArray.append(BYTE(readDataInstance.readByte())) else: raise excep.ArrayTypeException("Could\'t create an array of type %d" % arrayType) return newArray
python
def parse(readDataInstance, arrayType, arrayLength): """ Returns a new L{Array} object. @type readDataInstance: L{ReadData} @param readDataInstance: The L{ReadData} object containing the array data. @type arrayType: int @param arrayType: The type of L{Array} to be built. @type arrayLength: int @param arrayLength: The length of the array passed as an argument. @rtype: L{Array} @return: New L{Array} object. """ newArray = Array(arrayType) dataLength = len(readDataInstance) if arrayType is TYPE_DWORD: toRead = arrayLength * 4 if dataLength >= toRead: for i in range(arrayLength): newArray.append(DWORD(readDataInstance.readDword())) else: raise excep.DataLengthException("Not enough bytes to read.") elif arrayType is TYPE_WORD: toRead = arrayLength * 2 if dataLength >= toRead: for i in range(arrayLength): newArray.append(DWORD(readDataInstance.readWord())) else: raise excep.DataLengthException("Not enough bytes to read.") elif arrayType is TYPE_QWORD: toRead = arrayLength * 8 if dataLength >= toRead: for i in range(arrayLength): newArray.append(QWORD(readDataInstance.readQword())) else: raise excep.DataLengthException("Not enough bytes to read.") elif arrayType is TYPE_BYTE: for i in range(arrayLength): newArray.append(BYTE(readDataInstance.readByte())) else: raise excep.ArrayTypeException("Could\'t create an array of type %d" % arrayType) return newArray
[ "def", "parse", "(", "readDataInstance", ",", "arrayType", ",", "arrayLength", ")", ":", "newArray", "=", "Array", "(", "arrayType", ")", "dataLength", "=", "len", "(", "readDataInstance", ")", "if", "arrayType", "is", "TYPE_DWORD", ":", "toRead", "=", "arrayLength", "*", "4", "if", "dataLength", ">=", "toRead", ":", "for", "i", "in", "range", "(", "arrayLength", ")", ":", "newArray", ".", "append", "(", "DWORD", "(", "readDataInstance", ".", "readDword", "(", ")", ")", ")", "else", ":", "raise", "excep", ".", "DataLengthException", "(", "\"Not enough bytes to read.\"", ")", "elif", "arrayType", "is", "TYPE_WORD", ":", "toRead", "=", "arrayLength", "*", "2", "if", "dataLength", ">=", "toRead", ":", "for", "i", "in", "range", "(", "arrayLength", ")", ":", "newArray", ".", "append", "(", "DWORD", "(", "readDataInstance", ".", "readWord", "(", ")", ")", ")", "else", ":", "raise", "excep", ".", "DataLengthException", "(", "\"Not enough bytes to read.\"", ")", "elif", "arrayType", "is", "TYPE_QWORD", ":", "toRead", "=", "arrayLength", "*", "8", "if", "dataLength", ">=", "toRead", ":", "for", "i", "in", "range", "(", "arrayLength", ")", ":", "newArray", ".", "append", "(", "QWORD", "(", "readDataInstance", ".", "readQword", "(", ")", ")", ")", "else", ":", "raise", "excep", ".", "DataLengthException", "(", "\"Not enough bytes to read.\"", ")", "elif", "arrayType", "is", "TYPE_BYTE", ":", "for", "i", "in", "range", "(", "arrayLength", ")", ":", "newArray", ".", "append", "(", "BYTE", "(", "readDataInstance", ".", "readByte", "(", ")", ")", ")", "else", ":", "raise", "excep", ".", "ArrayTypeException", "(", "\"Could\\'t create an array of type %d\"", "%", "arrayType", ")", "return", "newArray" ]
Returns a new L{Array} object. @type readDataInstance: L{ReadData} @param readDataInstance: The L{ReadData} object containing the array data. @type arrayType: int @param arrayType: The type of L{Array} to be built. @type arrayLength: int @param arrayLength: The length of the array passed as an argument. @rtype: L{Array} @return: New L{Array} object.
[ "Returns", "a", "new", "L", "{", "Array", "}", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/datatypes.py#L145-L196
train
tansey/gfl
pygfl/trails.py
calc_euler_tour
def calc_euler_tour(g, start, end): '''Calculates an Euler tour over the graph g from vertex start to vertex end. Assumes start and end are odd-degree vertices and that there are no other odd-degree vertices.''' even_g = nx.subgraph(g, g.nodes()).copy() if end in even_g.neighbors(start): # If start and end are neighbors, remove the edge even_g.remove_edge(start, end) comps = list(nx.connected_components(even_g)) # If the graph did not split, just find the euler circuit if len(comps) == 1: trail = list(nx.eulerian_circuit(even_g, start)) trail.append((start, end)) elif len(comps) == 2: subg1 = nx.subgraph(even_g, comps[0]) subg2 = nx.subgraph(even_g, comps[1]) start_subg, end_subg = (subg1, subg2) if start in subg1.nodes() else (subg2, subg1) trail = list(nx.eulerian_circuit(start_subg, start)) + [(start, end)] + list(nx.eulerian_circuit(end_subg, end)) else: raise Exception('Unknown edge case with connected components of size {0}:\n{1}'.format(len(comps), comps)) else: # If they are not neighbors, we add an imaginary edge and calculate the euler circuit even_g.add_edge(start, end) circ = list(nx.eulerian_circuit(even_g, start)) try: trail_start = circ.index((start, end)) except: trail_start = circ.index((end, start)) trail = circ[trail_start+1:] + circ[:trail_start] return trail
python
def calc_euler_tour(g, start, end): '''Calculates an Euler tour over the graph g from vertex start to vertex end. Assumes start and end are odd-degree vertices and that there are no other odd-degree vertices.''' even_g = nx.subgraph(g, g.nodes()).copy() if end in even_g.neighbors(start): # If start and end are neighbors, remove the edge even_g.remove_edge(start, end) comps = list(nx.connected_components(even_g)) # If the graph did not split, just find the euler circuit if len(comps) == 1: trail = list(nx.eulerian_circuit(even_g, start)) trail.append((start, end)) elif len(comps) == 2: subg1 = nx.subgraph(even_g, comps[0]) subg2 = nx.subgraph(even_g, comps[1]) start_subg, end_subg = (subg1, subg2) if start in subg1.nodes() else (subg2, subg1) trail = list(nx.eulerian_circuit(start_subg, start)) + [(start, end)] + list(nx.eulerian_circuit(end_subg, end)) else: raise Exception('Unknown edge case with connected components of size {0}:\n{1}'.format(len(comps), comps)) else: # If they are not neighbors, we add an imaginary edge and calculate the euler circuit even_g.add_edge(start, end) circ = list(nx.eulerian_circuit(even_g, start)) try: trail_start = circ.index((start, end)) except: trail_start = circ.index((end, start)) trail = circ[trail_start+1:] + circ[:trail_start] return trail
[ "def", "calc_euler_tour", "(", "g", ",", "start", ",", "end", ")", ":", "even_g", "=", "nx", ".", "subgraph", "(", "g", ",", "g", ".", "nodes", "(", ")", ")", ".", "copy", "(", ")", "if", "end", "in", "even_g", ".", "neighbors", "(", "start", ")", ":", "# If start and end are neighbors, remove the edge", "even_g", ".", "remove_edge", "(", "start", ",", "end", ")", "comps", "=", "list", "(", "nx", ".", "connected_components", "(", "even_g", ")", ")", "# If the graph did not split, just find the euler circuit", "if", "len", "(", "comps", ")", "==", "1", ":", "trail", "=", "list", "(", "nx", ".", "eulerian_circuit", "(", "even_g", ",", "start", ")", ")", "trail", ".", "append", "(", "(", "start", ",", "end", ")", ")", "elif", "len", "(", "comps", ")", "==", "2", ":", "subg1", "=", "nx", ".", "subgraph", "(", "even_g", ",", "comps", "[", "0", "]", ")", "subg2", "=", "nx", ".", "subgraph", "(", "even_g", ",", "comps", "[", "1", "]", ")", "start_subg", ",", "end_subg", "=", "(", "subg1", ",", "subg2", ")", "if", "start", "in", "subg1", ".", "nodes", "(", ")", "else", "(", "subg2", ",", "subg1", ")", "trail", "=", "list", "(", "nx", ".", "eulerian_circuit", "(", "start_subg", ",", "start", ")", ")", "+", "[", "(", "start", ",", "end", ")", "]", "+", "list", "(", "nx", ".", "eulerian_circuit", "(", "end_subg", ",", "end", ")", ")", "else", ":", "raise", "Exception", "(", "'Unknown edge case with connected components of size {0}:\\n{1}'", ".", "format", "(", "len", "(", "comps", ")", ",", "comps", ")", ")", "else", ":", "# If they are not neighbors, we add an imaginary edge and calculate the euler circuit", "even_g", ".", "add_edge", "(", "start", ",", "end", ")", "circ", "=", "list", "(", "nx", ".", "eulerian_circuit", "(", "even_g", ",", "start", ")", ")", "try", ":", "trail_start", "=", "circ", ".", "index", "(", "(", "start", ",", "end", ")", ")", "except", ":", "trail_start", "=", "circ", ".", "index", "(", "(", "end", ",", "start", ")", ")", "trail", "=", "circ", "[", "trail_start", "+", "1", ":", "]", "+", "circ", "[", ":", "trail_start", "]", "return", "trail" ]
Calculates an Euler tour over the graph g from vertex start to vertex end. Assumes start and end are odd-degree vertices and that there are no other odd-degree vertices.
[ "Calculates", "an", "Euler", "tour", "over", "the", "graph", "g", "from", "vertex", "start", "to", "vertex", "end", ".", "Assumes", "start", "and", "end", "are", "odd", "-", "degree", "vertices", "and", "that", "there", "are", "no", "other", "odd", "-", "degree", "vertices", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trails.py#L50-L79
train
tansey/gfl
pygfl/trails.py
greedy_trails
def greedy_trails(subg, odds, verbose): '''Greedily select trails by making the longest you can until the end''' if verbose: print('\tCreating edge map') edges = defaultdict(list) for x,y in subg.edges(): edges[x].append(y) edges[y].append(x) if verbose: print('\tSelecting trails') trails = [] for x in subg.nodes(): if verbose > 2: print('\t\tNode {0}'.format(x)) while len(edges[x]) > 0: y = edges[x][0] trail = [(x,y)] edges[x].remove(y) edges[y].remove(x) while len(edges[y]) > 0: x = y y = edges[y][0] trail.append((x,y)) edges[x].remove(y) edges[y].remove(x) trails.append(trail) return trails
python
def greedy_trails(subg, odds, verbose): '''Greedily select trails by making the longest you can until the end''' if verbose: print('\tCreating edge map') edges = defaultdict(list) for x,y in subg.edges(): edges[x].append(y) edges[y].append(x) if verbose: print('\tSelecting trails') trails = [] for x in subg.nodes(): if verbose > 2: print('\t\tNode {0}'.format(x)) while len(edges[x]) > 0: y = edges[x][0] trail = [(x,y)] edges[x].remove(y) edges[y].remove(x) while len(edges[y]) > 0: x = y y = edges[y][0] trail.append((x,y)) edges[x].remove(y) edges[y].remove(x) trails.append(trail) return trails
[ "def", "greedy_trails", "(", "subg", ",", "odds", ",", "verbose", ")", ":", "if", "verbose", ":", "print", "(", "'\\tCreating edge map'", ")", "edges", "=", "defaultdict", "(", "list", ")", "for", "x", ",", "y", "in", "subg", ".", "edges", "(", ")", ":", "edges", "[", "x", "]", ".", "append", "(", "y", ")", "edges", "[", "y", "]", ".", "append", "(", "x", ")", "if", "verbose", ":", "print", "(", "'\\tSelecting trails'", ")", "trails", "=", "[", "]", "for", "x", "in", "subg", ".", "nodes", "(", ")", ":", "if", "verbose", ">", "2", ":", "print", "(", "'\\t\\tNode {0}'", ".", "format", "(", "x", ")", ")", "while", "len", "(", "edges", "[", "x", "]", ")", ">", "0", ":", "y", "=", "edges", "[", "x", "]", "[", "0", "]", "trail", "=", "[", "(", "x", ",", "y", ")", "]", "edges", "[", "x", "]", ".", "remove", "(", "y", ")", "edges", "[", "y", "]", ".", "remove", "(", "x", ")", "while", "len", "(", "edges", "[", "y", "]", ")", ">", "0", ":", "x", "=", "y", "y", "=", "edges", "[", "y", "]", "[", "0", "]", "trail", ".", "append", "(", "(", "x", ",", "y", ")", ")", "edges", "[", "x", "]", ".", "remove", "(", "y", ")", "edges", "[", "y", "]", ".", "remove", "(", "x", ")", "trails", ".", "append", "(", "trail", ")", "return", "trails" ]
Greedily select trails by making the longest you can until the end
[ "Greedily", "select", "trails", "by", "making", "the", "longest", "you", "can", "until", "the", "end" ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trails.py#L307-L338
train
tansey/gfl
pygfl/trails.py
decompose_graph
def decompose_graph(g, heuristic='tour', max_odds=20, verbose=0): '''Decompose a graph into a set of non-overlapping trails.''' # Get the connected subgraphs subgraphs = [nx.subgraph(g, x).copy() for x in nx.connected_components(g)] chains = [] num_subgraphs = len(subgraphs) step = 0 while num_subgraphs > 0: if verbose: print('Step #{0} ({1} subgraphs)'.format(step, num_subgraphs)) for i in range(num_subgraphs-1, -1, -1): subg = subgraphs[i] # Get all odd-degree nodes odds = [x for x,y in dict(nx.degree(subg)).items() if y % 2 == 1] if verbose > 1: if len(odds) == 0: print('\t\tNo odds') elif len(odds) == 2: print('\t\tExactly 2 odds') else: print('\t\t{0} odds'.format(len(odds))) # If there are no odd-degree edges, we can find an euler circuit if len(odds) == 0: trails = [list(nx.eulerian_circuit(subg))] elif len(odds) == 2: # If there are only two odd-degree edges, we can find an euler tour trails = [calc_euler_tour(subg, odds[0], odds[1])] elif heuristic in ['min', 'max', 'median', 'any']: trails = select_odd_degree_trail(subg, odds, max_odds, heuristic, verbose) elif heuristic == 'random': trails = select_random_trail(subg, verbose) elif heuristic == 'mindegree': trails = select_min_degree_trail(subg, max_odds, verbose) elif heuristic == 'ones': trails = select_single_edge_trails(subg, verbose) elif heuristic == 'tour': trails = pseudo_tour_trails(subg, odds, verbose) elif heuristic == 'greedy': trails = greedy_trails(subg, odds, verbose) if verbose > 2: print('\t\tTrails: {0}'.format(len(trails))) # Remove the trail for trail in trails: subg.remove_edges_from(trail) # Add it to the list of chains chains.extend(trails) # If the subgraph is empty, remove it from the list if subg.number_of_edges() == 0: del subgraphs[i] else: comps = list(nx.connected_components(subg)) # If the last edge split the graph, add the new subgraphs to the list of subgraphs if len(comps) > 1: for x in comps: compg = nx.subgraph(subg, x) if compg.number_of_edges() > 0: subgraphs.append(compg) del subgraphs[i] # Update the count of connected subgraphs num_subgraphs = len(subgraphs) step += 1 return chains
python
def decompose_graph(g, heuristic='tour', max_odds=20, verbose=0): '''Decompose a graph into a set of non-overlapping trails.''' # Get the connected subgraphs subgraphs = [nx.subgraph(g, x).copy() for x in nx.connected_components(g)] chains = [] num_subgraphs = len(subgraphs) step = 0 while num_subgraphs > 0: if verbose: print('Step #{0} ({1} subgraphs)'.format(step, num_subgraphs)) for i in range(num_subgraphs-1, -1, -1): subg = subgraphs[i] # Get all odd-degree nodes odds = [x for x,y in dict(nx.degree(subg)).items() if y % 2 == 1] if verbose > 1: if len(odds) == 0: print('\t\tNo odds') elif len(odds) == 2: print('\t\tExactly 2 odds') else: print('\t\t{0} odds'.format(len(odds))) # If there are no odd-degree edges, we can find an euler circuit if len(odds) == 0: trails = [list(nx.eulerian_circuit(subg))] elif len(odds) == 2: # If there are only two odd-degree edges, we can find an euler tour trails = [calc_euler_tour(subg, odds[0], odds[1])] elif heuristic in ['min', 'max', 'median', 'any']: trails = select_odd_degree_trail(subg, odds, max_odds, heuristic, verbose) elif heuristic == 'random': trails = select_random_trail(subg, verbose) elif heuristic == 'mindegree': trails = select_min_degree_trail(subg, max_odds, verbose) elif heuristic == 'ones': trails = select_single_edge_trails(subg, verbose) elif heuristic == 'tour': trails = pseudo_tour_trails(subg, odds, verbose) elif heuristic == 'greedy': trails = greedy_trails(subg, odds, verbose) if verbose > 2: print('\t\tTrails: {0}'.format(len(trails))) # Remove the trail for trail in trails: subg.remove_edges_from(trail) # Add it to the list of chains chains.extend(trails) # If the subgraph is empty, remove it from the list if subg.number_of_edges() == 0: del subgraphs[i] else: comps = list(nx.connected_components(subg)) # If the last edge split the graph, add the new subgraphs to the list of subgraphs if len(comps) > 1: for x in comps: compg = nx.subgraph(subg, x) if compg.number_of_edges() > 0: subgraphs.append(compg) del subgraphs[i] # Update the count of connected subgraphs num_subgraphs = len(subgraphs) step += 1 return chains
[ "def", "decompose_graph", "(", "g", ",", "heuristic", "=", "'tour'", ",", "max_odds", "=", "20", ",", "verbose", "=", "0", ")", ":", "# Get the connected subgraphs", "subgraphs", "=", "[", "nx", ".", "subgraph", "(", "g", ",", "x", ")", ".", "copy", "(", ")", "for", "x", "in", "nx", ".", "connected_components", "(", "g", ")", "]", "chains", "=", "[", "]", "num_subgraphs", "=", "len", "(", "subgraphs", ")", "step", "=", "0", "while", "num_subgraphs", ">", "0", ":", "if", "verbose", ":", "print", "(", "'Step #{0} ({1} subgraphs)'", ".", "format", "(", "step", ",", "num_subgraphs", ")", ")", "for", "i", "in", "range", "(", "num_subgraphs", "-", "1", ",", "-", "1", ",", "-", "1", ")", ":", "subg", "=", "subgraphs", "[", "i", "]", "# Get all odd-degree nodes", "odds", "=", "[", "x", "for", "x", ",", "y", "in", "dict", "(", "nx", ".", "degree", "(", "subg", ")", ")", ".", "items", "(", ")", "if", "y", "%", "2", "==", "1", "]", "if", "verbose", ">", "1", ":", "if", "len", "(", "odds", ")", "==", "0", ":", "print", "(", "'\\t\\tNo odds'", ")", "elif", "len", "(", "odds", ")", "==", "2", ":", "print", "(", "'\\t\\tExactly 2 odds'", ")", "else", ":", "print", "(", "'\\t\\t{0} odds'", ".", "format", "(", "len", "(", "odds", ")", ")", ")", "# If there are no odd-degree edges, we can find an euler circuit", "if", "len", "(", "odds", ")", "==", "0", ":", "trails", "=", "[", "list", "(", "nx", ".", "eulerian_circuit", "(", "subg", ")", ")", "]", "elif", "len", "(", "odds", ")", "==", "2", ":", "# If there are only two odd-degree edges, we can find an euler tour", "trails", "=", "[", "calc_euler_tour", "(", "subg", ",", "odds", "[", "0", "]", ",", "odds", "[", "1", "]", ")", "]", "elif", "heuristic", "in", "[", "'min'", ",", "'max'", ",", "'median'", ",", "'any'", "]", ":", "trails", "=", "select_odd_degree_trail", "(", "subg", ",", "odds", ",", "max_odds", ",", "heuristic", ",", "verbose", ")", "elif", "heuristic", "==", "'random'", ":", "trails", "=", "select_random_trail", "(", "subg", ",", "verbose", ")", "elif", "heuristic", "==", "'mindegree'", ":", "trails", "=", "select_min_degree_trail", "(", "subg", ",", "max_odds", ",", "verbose", ")", "elif", "heuristic", "==", "'ones'", ":", "trails", "=", "select_single_edge_trails", "(", "subg", ",", "verbose", ")", "elif", "heuristic", "==", "'tour'", ":", "trails", "=", "pseudo_tour_trails", "(", "subg", ",", "odds", ",", "verbose", ")", "elif", "heuristic", "==", "'greedy'", ":", "trails", "=", "greedy_trails", "(", "subg", ",", "odds", ",", "verbose", ")", "if", "verbose", ">", "2", ":", "print", "(", "'\\t\\tTrails: {0}'", ".", "format", "(", "len", "(", "trails", ")", ")", ")", "# Remove the trail", "for", "trail", "in", "trails", ":", "subg", ".", "remove_edges_from", "(", "trail", ")", "# Add it to the list of chains", "chains", ".", "extend", "(", "trails", ")", "# If the subgraph is empty, remove it from the list", "if", "subg", ".", "number_of_edges", "(", ")", "==", "0", ":", "del", "subgraphs", "[", "i", "]", "else", ":", "comps", "=", "list", "(", "nx", ".", "connected_components", "(", "subg", ")", ")", "# If the last edge split the graph, add the new subgraphs to the list of subgraphs", "if", "len", "(", "comps", ")", ">", "1", ":", "for", "x", "in", "comps", ":", "compg", "=", "nx", ".", "subgraph", "(", "subg", ",", "x", ")", "if", "compg", ".", "number_of_edges", "(", ")", ">", "0", ":", "subgraphs", ".", "append", "(", "compg", ")", "del", "subgraphs", "[", "i", "]", "# Update the count of connected subgraphs", "num_subgraphs", "=", "len", "(", "subgraphs", ")", "step", "+=", "1", "return", "chains" ]
Decompose a graph into a set of non-overlapping trails.
[ "Decompose", "a", "graph", "into", "a", "set", "of", "non", "-", "overlapping", "trails", "." ]
ae0f078bab57aba9e827ed6162f247ff9dc2aa19
https://github.com/tansey/gfl/blob/ae0f078bab57aba9e827ed6162f247ff9dc2aa19/pygfl/trails.py#L341-L414
train
timknip/pycsg
csg/geom.py
Vector.plus
def plus(self, a): """ Add. """ return Vector(self.x+a.x, self.y+a.y, self.z+a.z)
python
def plus(self, a): """ Add. """ return Vector(self.x+a.x, self.y+a.y, self.z+a.z)
[ "def", "plus", "(", "self", ",", "a", ")", ":", "return", "Vector", "(", "self", ".", "x", "+", "a", ".", "x", ",", "self", ".", "y", "+", "a", ".", "y", ",", "self", ".", "z", "+", "a", ".", "z", ")" ]
Add.
[ "Add", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L50-L52
train
timknip/pycsg
csg/geom.py
Vector.minus
def minus(self, a): """ Subtract. """ return Vector(self.x-a.x, self.y-a.y, self.z-a.z)
python
def minus(self, a): """ Subtract. """ return Vector(self.x-a.x, self.y-a.y, self.z-a.z)
[ "def", "minus", "(", "self", ",", "a", ")", ":", "return", "Vector", "(", "self", ".", "x", "-", "a", ".", "x", ",", "self", ".", "y", "-", "a", ".", "y", ",", "self", ".", "z", "-", "a", ".", "z", ")" ]
Subtract.
[ "Subtract", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L57-L59
train
timknip/pycsg
csg/geom.py
Vector.times
def times(self, a): """ Multiply. """ return Vector(self.x*a, self.y*a, self.z*a)
python
def times(self, a): """ Multiply. """ return Vector(self.x*a, self.y*a, self.z*a)
[ "def", "times", "(", "self", ",", "a", ")", ":", "return", "Vector", "(", "self", ".", "x", "*", "a", ",", "self", ".", "y", "*", "a", ",", "self", ".", "z", "*", "a", ")" ]
Multiply.
[ "Multiply", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L64-L66
train
timknip/pycsg
csg/geom.py
Vector.dividedBy
def dividedBy(self, a): """ Divide. """ return Vector(self.x/a, self.y/a, self.z/a)
python
def dividedBy(self, a): """ Divide. """ return Vector(self.x/a, self.y/a, self.z/a)
[ "def", "dividedBy", "(", "self", ",", "a", ")", ":", "return", "Vector", "(", "self", ".", "x", "/", "a", ",", "self", ".", "y", "/", "a", ",", "self", ".", "z", "/", "a", ")" ]
Divide.
[ "Divide", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L71-L73
train
timknip/pycsg
csg/geom.py
Vector.lerp
def lerp(self, a, t): """ Lerp. Linear interpolation from self to a""" return self.plus(a.minus(self).times(t));
python
def lerp(self, a, t): """ Lerp. Linear interpolation from self to a""" return self.plus(a.minus(self).times(t));
[ "def", "lerp", "(", "self", ",", "a", ",", "t", ")", ":", "return", "self", ".", "plus", "(", "a", ".", "minus", "(", "self", ")", ".", "times", "(", "t", ")", ")" ]
Lerp. Linear interpolation from self to a
[ "Lerp", ".", "Linear", "interpolation", "from", "self", "to", "a" ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L85-L87
train
timknip/pycsg
csg/geom.py
Vertex.interpolate
def interpolate(self, other, t): """ Create a new vertex between this vertex and `other` by linearly interpolating all properties using a parameter of `t`. Subclasses should override this to interpolate additional properties. """ return Vertex(self.pos.lerp(other.pos, t), self.normal.lerp(other.normal, t))
python
def interpolate(self, other, t): """ Create a new vertex between this vertex and `other` by linearly interpolating all properties using a parameter of `t`. Subclasses should override this to interpolate additional properties. """ return Vertex(self.pos.lerp(other.pos, t), self.normal.lerp(other.normal, t))
[ "def", "interpolate", "(", "self", ",", "other", ",", "t", ")", ":", "return", "Vertex", "(", "self", ".", "pos", ".", "lerp", "(", "other", ".", "pos", ",", "t", ")", ",", "self", ".", "normal", ".", "lerp", "(", "other", ".", "normal", ",", "t", ")", ")" ]
Create a new vertex between this vertex and `other` by linearly interpolating all properties using a parameter of `t`. Subclasses should override this to interpolate additional properties.
[ "Create", "a", "new", "vertex", "between", "this", "vertex", "and", "other", "by", "linearly", "interpolating", "all", "properties", "using", "a", "parameter", "of", "t", ".", "Subclasses", "should", "override", "this", "to", "interpolate", "additional", "properties", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L147-L154
train
timknip/pycsg
csg/geom.py
Plane.splitPolygon
def splitPolygon(self, polygon, coplanarFront, coplanarBack, front, back): """ Split `polygon` by this plane if needed, then put the polygon or polygon fragments in the appropriate lists. Coplanar polygons go into either `coplanarFront` or `coplanarBack` depending on their orientation with respect to this plane. Polygons in front or in back of this plane go into either `front` or `back` """ COPLANAR = 0 # all the vertices are within EPSILON distance from plane FRONT = 1 # all the vertices are in front of the plane BACK = 2 # all the vertices are at the back of the plane SPANNING = 3 # some vertices are in front, some in the back # Classify each point as well as the entire polygon into one of the above # four classes. polygonType = 0 vertexLocs = [] numVertices = len(polygon.vertices) for i in range(numVertices): t = self.normal.dot(polygon.vertices[i].pos) - self.w loc = -1 if t < -Plane.EPSILON: loc = BACK elif t > Plane.EPSILON: loc = FRONT else: loc = COPLANAR polygonType |= loc vertexLocs.append(loc) # Put the polygon in the correct list, splitting it when necessary. if polygonType == COPLANAR: normalDotPlaneNormal = self.normal.dot(polygon.plane.normal) if normalDotPlaneNormal > 0: coplanarFront.append(polygon) else: coplanarBack.append(polygon) elif polygonType == FRONT: front.append(polygon) elif polygonType == BACK: back.append(polygon) elif polygonType == SPANNING: f = [] b = [] for i in range(numVertices): j = (i+1) % numVertices ti = vertexLocs[i] tj = vertexLocs[j] vi = polygon.vertices[i] vj = polygon.vertices[j] if ti != BACK: f.append(vi) if ti != FRONT: if ti != BACK: b.append(vi.clone()) else: b.append(vi) if (ti | tj) == SPANNING: # interpolation weight at the intersection point t = (self.w - self.normal.dot(vi.pos)) / self.normal.dot(vj.pos.minus(vi.pos)) # intersection point on the plane v = vi.interpolate(vj, t) f.append(v) b.append(v.clone()) if len(f) >= 3: front.append(Polygon(f, polygon.shared)) if len(b) >= 3: back.append(Polygon(b, polygon.shared))
python
def splitPolygon(self, polygon, coplanarFront, coplanarBack, front, back): """ Split `polygon` by this plane if needed, then put the polygon or polygon fragments in the appropriate lists. Coplanar polygons go into either `coplanarFront` or `coplanarBack` depending on their orientation with respect to this plane. Polygons in front or in back of this plane go into either `front` or `back` """ COPLANAR = 0 # all the vertices are within EPSILON distance from plane FRONT = 1 # all the vertices are in front of the plane BACK = 2 # all the vertices are at the back of the plane SPANNING = 3 # some vertices are in front, some in the back # Classify each point as well as the entire polygon into one of the above # four classes. polygonType = 0 vertexLocs = [] numVertices = len(polygon.vertices) for i in range(numVertices): t = self.normal.dot(polygon.vertices[i].pos) - self.w loc = -1 if t < -Plane.EPSILON: loc = BACK elif t > Plane.EPSILON: loc = FRONT else: loc = COPLANAR polygonType |= loc vertexLocs.append(loc) # Put the polygon in the correct list, splitting it when necessary. if polygonType == COPLANAR: normalDotPlaneNormal = self.normal.dot(polygon.plane.normal) if normalDotPlaneNormal > 0: coplanarFront.append(polygon) else: coplanarBack.append(polygon) elif polygonType == FRONT: front.append(polygon) elif polygonType == BACK: back.append(polygon) elif polygonType == SPANNING: f = [] b = [] for i in range(numVertices): j = (i+1) % numVertices ti = vertexLocs[i] tj = vertexLocs[j] vi = polygon.vertices[i] vj = polygon.vertices[j] if ti != BACK: f.append(vi) if ti != FRONT: if ti != BACK: b.append(vi.clone()) else: b.append(vi) if (ti | tj) == SPANNING: # interpolation weight at the intersection point t = (self.w - self.normal.dot(vi.pos)) / self.normal.dot(vj.pos.minus(vi.pos)) # intersection point on the plane v = vi.interpolate(vj, t) f.append(v) b.append(v.clone()) if len(f) >= 3: front.append(Polygon(f, polygon.shared)) if len(b) >= 3: back.append(Polygon(b, polygon.shared))
[ "def", "splitPolygon", "(", "self", ",", "polygon", ",", "coplanarFront", ",", "coplanarBack", ",", "front", ",", "back", ")", ":", "COPLANAR", "=", "0", "# all the vertices are within EPSILON distance from plane", "FRONT", "=", "1", "# all the vertices are in front of the plane", "BACK", "=", "2", "# all the vertices are at the back of the plane", "SPANNING", "=", "3", "# some vertices are in front, some in the back", "# Classify each point as well as the entire polygon into one of the above", "# four classes.", "polygonType", "=", "0", "vertexLocs", "=", "[", "]", "numVertices", "=", "len", "(", "polygon", ".", "vertices", ")", "for", "i", "in", "range", "(", "numVertices", ")", ":", "t", "=", "self", ".", "normal", ".", "dot", "(", "polygon", ".", "vertices", "[", "i", "]", ".", "pos", ")", "-", "self", ".", "w", "loc", "=", "-", "1", "if", "t", "<", "-", "Plane", ".", "EPSILON", ":", "loc", "=", "BACK", "elif", "t", ">", "Plane", ".", "EPSILON", ":", "loc", "=", "FRONT", "else", ":", "loc", "=", "COPLANAR", "polygonType", "|=", "loc", "vertexLocs", ".", "append", "(", "loc", ")", "# Put the polygon in the correct list, splitting it when necessary.", "if", "polygonType", "==", "COPLANAR", ":", "normalDotPlaneNormal", "=", "self", ".", "normal", ".", "dot", "(", "polygon", ".", "plane", ".", "normal", ")", "if", "normalDotPlaneNormal", ">", "0", ":", "coplanarFront", ".", "append", "(", "polygon", ")", "else", ":", "coplanarBack", ".", "append", "(", "polygon", ")", "elif", "polygonType", "==", "FRONT", ":", "front", ".", "append", "(", "polygon", ")", "elif", "polygonType", "==", "BACK", ":", "back", ".", "append", "(", "polygon", ")", "elif", "polygonType", "==", "SPANNING", ":", "f", "=", "[", "]", "b", "=", "[", "]", "for", "i", "in", "range", "(", "numVertices", ")", ":", "j", "=", "(", "i", "+", "1", ")", "%", "numVertices", "ti", "=", "vertexLocs", "[", "i", "]", "tj", "=", "vertexLocs", "[", "j", "]", "vi", "=", "polygon", ".", "vertices", "[", "i", "]", "vj", "=", "polygon", ".", "vertices", "[", "j", "]", "if", "ti", "!=", "BACK", ":", "f", ".", "append", "(", "vi", ")", "if", "ti", "!=", "FRONT", ":", "if", "ti", "!=", "BACK", ":", "b", ".", "append", "(", "vi", ".", "clone", "(", ")", ")", "else", ":", "b", ".", "append", "(", "vi", ")", "if", "(", "ti", "|", "tj", ")", "==", "SPANNING", ":", "# interpolation weight at the intersection point", "t", "=", "(", "self", ".", "w", "-", "self", ".", "normal", ".", "dot", "(", "vi", ".", "pos", ")", ")", "/", "self", ".", "normal", ".", "dot", "(", "vj", ".", "pos", ".", "minus", "(", "vi", ".", "pos", ")", ")", "# intersection point on the plane", "v", "=", "vi", ".", "interpolate", "(", "vj", ",", "t", ")", "f", ".", "append", "(", "v", ")", "b", ".", "append", "(", "v", ".", "clone", "(", ")", ")", "if", "len", "(", "f", ")", ">=", "3", ":", "front", ".", "append", "(", "Polygon", "(", "f", ",", "polygon", ".", "shared", ")", ")", "if", "len", "(", "b", ")", ">=", "3", ":", "back", ".", "append", "(", "Polygon", "(", "b", ",", "polygon", ".", "shared", ")", ")" ]
Split `polygon` by this plane if needed, then put the polygon or polygon fragments in the appropriate lists. Coplanar polygons go into either `coplanarFront` or `coplanarBack` depending on their orientation with respect to this plane. Polygons in front or in back of this plane go into either `front` or `back`
[ "Split", "polygon", "by", "this", "plane", "if", "needed", "then", "put", "the", "polygon", "or", "polygon", "fragments", "in", "the", "appropriate", "lists", ".", "Coplanar", "polygons", "go", "into", "either", "coplanarFront", "or", "coplanarBack", "depending", "on", "their", "orientation", "with", "respect", "to", "this", "plane", ".", "Polygons", "in", "front", "or", "in", "back", "of", "this", "plane", "go", "into", "either", "front", "or", "back" ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L192-L260
train
timknip/pycsg
csg/geom.py
BSPNode.invert
def invert(self): """ Convert solid space to empty space and empty space to solid space. """ for poly in self.polygons: poly.flip() self.plane.flip() if self.front: self.front.invert() if self.back: self.back.invert() temp = self.front self.front = self.back self.back = temp
python
def invert(self): """ Convert solid space to empty space and empty space to solid space. """ for poly in self.polygons: poly.flip() self.plane.flip() if self.front: self.front.invert() if self.back: self.back.invert() temp = self.front self.front = self.back self.back = temp
[ "def", "invert", "(", "self", ")", ":", "for", "poly", "in", "self", ".", "polygons", ":", "poly", ".", "flip", "(", ")", "self", ".", "plane", ".", "flip", "(", ")", "if", "self", ".", "front", ":", "self", ".", "front", ".", "invert", "(", ")", "if", "self", ".", "back", ":", "self", ".", "back", ".", "invert", "(", ")", "temp", "=", "self", ".", "front", "self", ".", "front", "=", "self", ".", "back", "self", ".", "back", "=", "temp" ]
Convert solid space to empty space and empty space to solid space.
[ "Convert", "solid", "space", "to", "empty", "space", "and", "empty", "space", "to", "solid", "space", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L323-L336
train
timknip/pycsg
csg/geom.py
BSPNode.clipPolygons
def clipPolygons(self, polygons): """ Recursively remove all polygons in `polygons` that are inside this BSP tree. """ if not self.plane: return polygons[:] front = [] back = [] for poly in polygons: self.plane.splitPolygon(poly, front, back, front, back) if self.front: front = self.front.clipPolygons(front) if self.back: back = self.back.clipPolygons(back) else: back = [] front.extend(back) return front
python
def clipPolygons(self, polygons): """ Recursively remove all polygons in `polygons` that are inside this BSP tree. """ if not self.plane: return polygons[:] front = [] back = [] for poly in polygons: self.plane.splitPolygon(poly, front, back, front, back) if self.front: front = self.front.clipPolygons(front) if self.back: back = self.back.clipPolygons(back) else: back = [] front.extend(back) return front
[ "def", "clipPolygons", "(", "self", ",", "polygons", ")", ":", "if", "not", "self", ".", "plane", ":", "return", "polygons", "[", ":", "]", "front", "=", "[", "]", "back", "=", "[", "]", "for", "poly", "in", "polygons", ":", "self", ".", "plane", ".", "splitPolygon", "(", "poly", ",", "front", ",", "back", ",", "front", ",", "back", ")", "if", "self", ".", "front", ":", "front", "=", "self", ".", "front", ".", "clipPolygons", "(", "front", ")", "if", "self", ".", "back", ":", "back", "=", "self", ".", "back", ".", "clipPolygons", "(", "back", ")", "else", ":", "back", "=", "[", "]", "front", ".", "extend", "(", "back", ")", "return", "front" ]
Recursively remove all polygons in `polygons` that are inside this BSP tree.
[ "Recursively", "remove", "all", "polygons", "in", "polygons", "that", "are", "inside", "this", "BSP", "tree", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L338-L360
train
timknip/pycsg
csg/geom.py
BSPNode.clipTo
def clipTo(self, bsp): """ Remove all polygons in this BSP tree that are inside the other BSP tree `bsp`. """ self.polygons = bsp.clipPolygons(self.polygons) if self.front: self.front.clipTo(bsp) if self.back: self.back.clipTo(bsp)
python
def clipTo(self, bsp): """ Remove all polygons in this BSP tree that are inside the other BSP tree `bsp`. """ self.polygons = bsp.clipPolygons(self.polygons) if self.front: self.front.clipTo(bsp) if self.back: self.back.clipTo(bsp)
[ "def", "clipTo", "(", "self", ",", "bsp", ")", ":", "self", ".", "polygons", "=", "bsp", ".", "clipPolygons", "(", "self", ".", "polygons", ")", "if", "self", ".", "front", ":", "self", ".", "front", ".", "clipTo", "(", "bsp", ")", "if", "self", ".", "back", ":", "self", ".", "back", ".", "clipTo", "(", "bsp", ")" ]
Remove all polygons in this BSP tree that are inside the other BSP tree `bsp`.
[ "Remove", "all", "polygons", "in", "this", "BSP", "tree", "that", "are", "inside", "the", "other", "BSP", "tree", "bsp", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L362-L371
train
timknip/pycsg
csg/geom.py
BSPNode.allPolygons
def allPolygons(self): """ Return a list of all polygons in this BSP tree. """ polygons = self.polygons[:] if self.front: polygons.extend(self.front.allPolygons()) if self.back: polygons.extend(self.back.allPolygons()) return polygons
python
def allPolygons(self): """ Return a list of all polygons in this BSP tree. """ polygons = self.polygons[:] if self.front: polygons.extend(self.front.allPolygons()) if self.back: polygons.extend(self.back.allPolygons()) return polygons
[ "def", "allPolygons", "(", "self", ")", ":", "polygons", "=", "self", ".", "polygons", "[", ":", "]", "if", "self", ".", "front", ":", "polygons", ".", "extend", "(", "self", ".", "front", ".", "allPolygons", "(", ")", ")", "if", "self", ".", "back", ":", "polygons", ".", "extend", "(", "self", ".", "back", ".", "allPolygons", "(", ")", ")", "return", "polygons" ]
Return a list of all polygons in this BSP tree.
[ "Return", "a", "list", "of", "all", "polygons", "in", "this", "BSP", "tree", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L373-L382
train
timknip/pycsg
csg/geom.py
BSPNode.build
def build(self, polygons): """ Build a BSP tree out of `polygons`. When called on an existing tree, the new polygons are filtered down to the bottom of the tree and become new nodes there. Each set of polygons is partitioned using the first polygon (no heuristic is used to pick a good split). """ if len(polygons) == 0: return if not self.plane: self.plane = polygons[0].plane.clone() # add polygon to this node self.polygons.append(polygons[0]) front = [] back = [] # split all other polygons using the first polygon's plane for poly in polygons[1:]: # coplanar front and back polygons go into self.polygons self.plane.splitPolygon(poly, self.polygons, self.polygons, front, back) # recursively build the BSP tree if len(front) > 0: if not self.front: self.front = BSPNode() self.front.build(front) if len(back) > 0: if not self.back: self.back = BSPNode() self.back.build(back)
python
def build(self, polygons): """ Build a BSP tree out of `polygons`. When called on an existing tree, the new polygons are filtered down to the bottom of the tree and become new nodes there. Each set of polygons is partitioned using the first polygon (no heuristic is used to pick a good split). """ if len(polygons) == 0: return if not self.plane: self.plane = polygons[0].plane.clone() # add polygon to this node self.polygons.append(polygons[0]) front = [] back = [] # split all other polygons using the first polygon's plane for poly in polygons[1:]: # coplanar front and back polygons go into self.polygons self.plane.splitPolygon(poly, self.polygons, self.polygons, front, back) # recursively build the BSP tree if len(front) > 0: if not self.front: self.front = BSPNode() self.front.build(front) if len(back) > 0: if not self.back: self.back = BSPNode() self.back.build(back)
[ "def", "build", "(", "self", ",", "polygons", ")", ":", "if", "len", "(", "polygons", ")", "==", "0", ":", "return", "if", "not", "self", ".", "plane", ":", "self", ".", "plane", "=", "polygons", "[", "0", "]", ".", "plane", ".", "clone", "(", ")", "# add polygon to this node", "self", ".", "polygons", ".", "append", "(", "polygons", "[", "0", "]", ")", "front", "=", "[", "]", "back", "=", "[", "]", "# split all other polygons using the first polygon's plane", "for", "poly", "in", "polygons", "[", "1", ":", "]", ":", "# coplanar front and back polygons go into self.polygons", "self", ".", "plane", ".", "splitPolygon", "(", "poly", ",", "self", ".", "polygons", ",", "self", ".", "polygons", ",", "front", ",", "back", ")", "# recursively build the BSP tree", "if", "len", "(", "front", ")", ">", "0", ":", "if", "not", "self", ".", "front", ":", "self", ".", "front", "=", "BSPNode", "(", ")", "self", ".", "front", ".", "build", "(", "front", ")", "if", "len", "(", "back", ")", ">", "0", ":", "if", "not", "self", ".", "back", ":", "self", ".", "back", "=", "BSPNode", "(", ")", "self", ".", "back", ".", "build", "(", "back", ")" ]
Build a BSP tree out of `polygons`. When called on an existing tree, the new polygons are filtered down to the bottom of the tree and become new nodes there. Each set of polygons is partitioned using the first polygon (no heuristic is used to pick a good split).
[ "Build", "a", "BSP", "tree", "out", "of", "polygons", ".", "When", "called", "on", "an", "existing", "tree", "the", "new", "polygons", "are", "filtered", "down", "to", "the", "bottom", "of", "the", "tree", "and", "become", "new", "nodes", "there", ".", "Each", "set", "of", "polygons", "is", "partitioned", "using", "the", "first", "polygon", "(", "no", "heuristic", "is", "used", "to", "pick", "a", "good", "split", ")", "." ]
b8f9710fd15c38dcc275d56a2108f604af38dcc8
https://github.com/timknip/pycsg/blob/b8f9710fd15c38dcc275d56a2108f604af38dcc8/csg/geom.py#L384-L412
train
evonove/django-money-rates
djmoney_rates/utils.py
get_rate
def get_rate(currency): """Returns the rate from the default currency to `currency`.""" source = get_rate_source() try: return Rate.objects.get(source=source, currency=currency).value except Rate.DoesNotExist: raise CurrencyConversionException( "Rate for %s in %s do not exists. " "Please run python manage.py update_rates" % ( currency, source.name))
python
def get_rate(currency): """Returns the rate from the default currency to `currency`.""" source = get_rate_source() try: return Rate.objects.get(source=source, currency=currency).value except Rate.DoesNotExist: raise CurrencyConversionException( "Rate for %s in %s do not exists. " "Please run python manage.py update_rates" % ( currency, source.name))
[ "def", "get_rate", "(", "currency", ")", ":", "source", "=", "get_rate_source", "(", ")", "try", ":", "return", "Rate", ".", "objects", ".", "get", "(", "source", "=", "source", ",", "currency", "=", "currency", ")", ".", "value", "except", "Rate", ".", "DoesNotExist", ":", "raise", "CurrencyConversionException", "(", "\"Rate for %s in %s do not exists. \"", "\"Please run python manage.py update_rates\"", "%", "(", "currency", ",", "source", ".", "name", ")", ")" ]
Returns the rate from the default currency to `currency`.
[ "Returns", "the", "rate", "from", "the", "default", "currency", "to", "currency", "." ]
ac1f7636b9a38d3e153eb833019342c4d88634c2
https://github.com/evonove/django-money-rates/blob/ac1f7636b9a38d3e153eb833019342c4d88634c2/djmoney_rates/utils.py#L12-L21
train
evonove/django-money-rates
djmoney_rates/utils.py
get_rate_source
def get_rate_source(): """Get the default Rate Source and return it.""" backend = money_rates_settings.DEFAULT_BACKEND() try: return RateSource.objects.get(name=backend.get_source_name()) except RateSource.DoesNotExist: raise CurrencyConversionException( "Rate for %s source do not exists. " "Please run python manage.py update_rates" % backend.get_source_name())
python
def get_rate_source(): """Get the default Rate Source and return it.""" backend = money_rates_settings.DEFAULT_BACKEND() try: return RateSource.objects.get(name=backend.get_source_name()) except RateSource.DoesNotExist: raise CurrencyConversionException( "Rate for %s source do not exists. " "Please run python manage.py update_rates" % backend.get_source_name())
[ "def", "get_rate_source", "(", ")", ":", "backend", "=", "money_rates_settings", ".", "DEFAULT_BACKEND", "(", ")", "try", ":", "return", "RateSource", ".", "objects", ".", "get", "(", "name", "=", "backend", ".", "get_source_name", "(", ")", ")", "except", "RateSource", ".", "DoesNotExist", ":", "raise", "CurrencyConversionException", "(", "\"Rate for %s source do not exists. \"", "\"Please run python manage.py update_rates\"", "%", "backend", ".", "get_source_name", "(", ")", ")" ]
Get the default Rate Source and return it.
[ "Get", "the", "default", "Rate", "Source", "and", "return", "it", "." ]
ac1f7636b9a38d3e153eb833019342c4d88634c2
https://github.com/evonove/django-money-rates/blob/ac1f7636b9a38d3e153eb833019342c4d88634c2/djmoney_rates/utils.py#L24-L32
train
evonove/django-money-rates
djmoney_rates/utils.py
base_convert_money
def base_convert_money(amount, currency_from, currency_to): """ Convert 'amount' from 'currency_from' to 'currency_to' """ source = get_rate_source() # Get rate for currency_from. if source.base_currency != currency_from: rate_from = get_rate(currency_from) else: # If currency from is the same as base currency its rate is 1. rate_from = Decimal(1) # Get rate for currency_to. rate_to = get_rate(currency_to) if isinstance(amount, float): amount = Decimal(amount).quantize(Decimal('.000001')) # After finishing the operation, quantize down final amount to two points. return ((amount / rate_from) * rate_to).quantize(Decimal("1.00"))
python
def base_convert_money(amount, currency_from, currency_to): """ Convert 'amount' from 'currency_from' to 'currency_to' """ source = get_rate_source() # Get rate for currency_from. if source.base_currency != currency_from: rate_from = get_rate(currency_from) else: # If currency from is the same as base currency its rate is 1. rate_from = Decimal(1) # Get rate for currency_to. rate_to = get_rate(currency_to) if isinstance(amount, float): amount = Decimal(amount).quantize(Decimal('.000001')) # After finishing the operation, quantize down final amount to two points. return ((amount / rate_from) * rate_to).quantize(Decimal("1.00"))
[ "def", "base_convert_money", "(", "amount", ",", "currency_from", ",", "currency_to", ")", ":", "source", "=", "get_rate_source", "(", ")", "# Get rate for currency_from.", "if", "source", ".", "base_currency", "!=", "currency_from", ":", "rate_from", "=", "get_rate", "(", "currency_from", ")", "else", ":", "# If currency from is the same as base currency its rate is 1.", "rate_from", "=", "Decimal", "(", "1", ")", "# Get rate for currency_to.", "rate_to", "=", "get_rate", "(", "currency_to", ")", "if", "isinstance", "(", "amount", ",", "float", ")", ":", "amount", "=", "Decimal", "(", "amount", ")", ".", "quantize", "(", "Decimal", "(", "'.000001'", ")", ")", "# After finishing the operation, quantize down final amount to two points.", "return", "(", "(", "amount", "/", "rate_from", ")", "*", "rate_to", ")", ".", "quantize", "(", "Decimal", "(", "\"1.00\"", ")", ")" ]
Convert 'amount' from 'currency_from' to 'currency_to'
[ "Convert", "amount", "from", "currency_from", "to", "currency_to" ]
ac1f7636b9a38d3e153eb833019342c4d88634c2
https://github.com/evonove/django-money-rates/blob/ac1f7636b9a38d3e153eb833019342c4d88634c2/djmoney_rates/utils.py#L35-L55
train
evonove/django-money-rates
djmoney_rates/utils.py
convert_money
def convert_money(amount, currency_from, currency_to): """ Convert 'amount' from 'currency_from' to 'currency_to' and return a Money instance of the converted amount. """ new_amount = base_convert_money(amount, currency_from, currency_to) return moneyed.Money(new_amount, currency_to)
python
def convert_money(amount, currency_from, currency_to): """ Convert 'amount' from 'currency_from' to 'currency_to' and return a Money instance of the converted amount. """ new_amount = base_convert_money(amount, currency_from, currency_to) return moneyed.Money(new_amount, currency_to)
[ "def", "convert_money", "(", "amount", ",", "currency_from", ",", "currency_to", ")", ":", "new_amount", "=", "base_convert_money", "(", "amount", ",", "currency_from", ",", "currency_to", ")", "return", "moneyed", ".", "Money", "(", "new_amount", ",", "currency_to", ")" ]
Convert 'amount' from 'currency_from' to 'currency_to' and return a Money instance of the converted amount.
[ "Convert", "amount", "from", "currency_from", "to", "currency_to", "and", "return", "a", "Money", "instance", "of", "the", "converted", "amount", "." ]
ac1f7636b9a38d3e153eb833019342c4d88634c2
https://github.com/evonove/django-money-rates/blob/ac1f7636b9a38d3e153eb833019342c4d88634c2/djmoney_rates/utils.py#L58-L64
train
Blazemeter/apiritif
apiritif/utilities.py
format_date
def format_date(format_string=None, datetime_obj=None): """ Format a datetime object with Java SimpleDateFormat's-like string. If datetime_obj is not given - use current datetime. If format_string is not given - return number of millisecond since epoch. :param format_string: :param datetime_obj: :return: :rtype string """ datetime_obj = datetime_obj or datetime.now() if format_string is None: seconds = int(datetime_obj.strftime("%s")) milliseconds = datetime_obj.microsecond // 1000 return str(seconds * 1000 + milliseconds) else: formatter = SimpleDateFormat(format_string) return formatter.format_datetime(datetime_obj)
python
def format_date(format_string=None, datetime_obj=None): """ Format a datetime object with Java SimpleDateFormat's-like string. If datetime_obj is not given - use current datetime. If format_string is not given - return number of millisecond since epoch. :param format_string: :param datetime_obj: :return: :rtype string """ datetime_obj = datetime_obj or datetime.now() if format_string is None: seconds = int(datetime_obj.strftime("%s")) milliseconds = datetime_obj.microsecond // 1000 return str(seconds * 1000 + milliseconds) else: formatter = SimpleDateFormat(format_string) return formatter.format_datetime(datetime_obj)
[ "def", "format_date", "(", "format_string", "=", "None", ",", "datetime_obj", "=", "None", ")", ":", "datetime_obj", "=", "datetime_obj", "or", "datetime", ".", "now", "(", ")", "if", "format_string", "is", "None", ":", "seconds", "=", "int", "(", "datetime_obj", ".", "strftime", "(", "\"%s\"", ")", ")", "milliseconds", "=", "datetime_obj", ".", "microsecond", "//", "1000", "return", "str", "(", "seconds", "*", "1000", "+", "milliseconds", ")", "else", ":", "formatter", "=", "SimpleDateFormat", "(", "format_string", ")", "return", "formatter", ".", "format_datetime", "(", "datetime_obj", ")" ]
Format a datetime object with Java SimpleDateFormat's-like string. If datetime_obj is not given - use current datetime. If format_string is not given - return number of millisecond since epoch. :param format_string: :param datetime_obj: :return: :rtype string
[ "Format", "a", "datetime", "object", "with", "Java", "SimpleDateFormat", "s", "-", "like", "string", "." ]
27b48a68425949998c2254e5e1e0226882d9eee8
https://github.com/Blazemeter/apiritif/blob/27b48a68425949998c2254e5e1e0226882d9eee8/apiritif/utilities.py#L93-L112
train
crackinglandia/pype32
pype32/utils.py
allZero
def allZero(buffer): """ Tries to determine if a buffer is empty. @type buffer: str @param buffer: Buffer to test if it is empty. @rtype: bool @return: C{True} if the given buffer is empty, i.e. full of zeros, C{False} if it doesn't. """ allZero = True for byte in buffer: if byte != "\x00": allZero = False break return allZero
python
def allZero(buffer): """ Tries to determine if a buffer is empty. @type buffer: str @param buffer: Buffer to test if it is empty. @rtype: bool @return: C{True} if the given buffer is empty, i.e. full of zeros, C{False} if it doesn't. """ allZero = True for byte in buffer: if byte != "\x00": allZero = False break return allZero
[ "def", "allZero", "(", "buffer", ")", ":", "allZero", "=", "True", "for", "byte", "in", "buffer", ":", "if", "byte", "!=", "\"\\x00\"", ":", "allZero", "=", "False", "break", "return", "allZero" ]
Tries to determine if a buffer is empty. @type buffer: str @param buffer: Buffer to test if it is empty. @rtype: bool @return: C{True} if the given buffer is empty, i.e. full of zeros, C{False} if it doesn't.
[ "Tries", "to", "determine", "if", "a", "buffer", "is", "empty", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L65-L81
train
crackinglandia/pype32
pype32/utils.py
WriteData.writeByte
def writeByte(self, byte): """ Writes a byte into the L{WriteData} stream object. @type byte: int @param byte: Byte value to write into the stream. """ self.data.write(pack("B" if not self.signed else "b", byte))
python
def writeByte(self, byte): """ Writes a byte into the L{WriteData} stream object. @type byte: int @param byte: Byte value to write into the stream. """ self.data.write(pack("B" if not self.signed else "b", byte))
[ "def", "writeByte", "(", "self", ",", "byte", ")", ":", "self", ".", "data", ".", "write", "(", "pack", "(", "\"B\"", "if", "not", "self", ".", "signed", "else", "\"b\"", ",", "byte", ")", ")" ]
Writes a byte into the L{WriteData} stream object. @type byte: int @param byte: Byte value to write into the stream.
[ "Writes", "a", "byte", "into", "the", "L", "{", "WriteData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L106-L113
train
crackinglandia/pype32
pype32/utils.py
WriteData.writeWord
def writeWord(self, word): """ Writes a word value into the L{WriteData} stream object. @type word: int @param word: Word value to write into the stream. """ self.data.write(pack(self.endianness + ("H" if not self.signed else "h"), word))
python
def writeWord(self, word): """ Writes a word value into the L{WriteData} stream object. @type word: int @param word: Word value to write into the stream. """ self.data.write(pack(self.endianness + ("H" if not self.signed else "h"), word))
[ "def", "writeWord", "(", "self", ",", "word", ")", ":", "self", ".", "data", ".", "write", "(", "pack", "(", "self", ".", "endianness", "+", "(", "\"H\"", "if", "not", "self", ".", "signed", "else", "\"h\"", ")", ",", "word", ")", ")" ]
Writes a word value into the L{WriteData} stream object. @type word: int @param word: Word value to write into the stream.
[ "Writes", "a", "word", "value", "into", "the", "L", "{", "WriteData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L115-L122
train
crackinglandia/pype32
pype32/utils.py
WriteData.writeDword
def writeDword(self, dword): """ Writes a dword value into the L{WriteData} stream object. @type dword: int @param dword: Dword value to write into the stream. """ self.data.write(pack(self.endianness + ("L" if not self.signed else "l"), dword))
python
def writeDword(self, dword): """ Writes a dword value into the L{WriteData} stream object. @type dword: int @param dword: Dword value to write into the stream. """ self.data.write(pack(self.endianness + ("L" if not self.signed else "l"), dword))
[ "def", "writeDword", "(", "self", ",", "dword", ")", ":", "self", ".", "data", ".", "write", "(", "pack", "(", "self", ".", "endianness", "+", "(", "\"L\"", "if", "not", "self", ".", "signed", "else", "\"l\"", ")", ",", "dword", ")", ")" ]
Writes a dword value into the L{WriteData} stream object. @type dword: int @param dword: Dword value to write into the stream.
[ "Writes", "a", "dword", "value", "into", "the", "L", "{", "WriteData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L124-L131
train
crackinglandia/pype32
pype32/utils.py
WriteData.writeQword
def writeQword(self, qword): """ Writes a qword value into the L{WriteData} stream object. @type qword: int @param qword: Qword value to write into the stream. """ self.data.write(pack(self.endianness + ("Q" if not self.signed else "q"), qword))
python
def writeQword(self, qword): """ Writes a qword value into the L{WriteData} stream object. @type qword: int @param qword: Qword value to write into the stream. """ self.data.write(pack(self.endianness + ("Q" if not self.signed else "q"), qword))
[ "def", "writeQword", "(", "self", ",", "qword", ")", ":", "self", ".", "data", ".", "write", "(", "pack", "(", "self", ".", "endianness", "+", "(", "\"Q\"", "if", "not", "self", ".", "signed", "else", "\"q\"", ")", ",", "qword", ")", ")" ]
Writes a qword value into the L{WriteData} stream object. @type qword: int @param qword: Qword value to write into the stream.
[ "Writes", "a", "qword", "value", "into", "the", "L", "{", "WriteData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L133-L140
train
crackinglandia/pype32
pype32/utils.py
WriteData.setOffset
def setOffset(self, value): """ Sets the offset of the L{WriteData} stream object in wich the data is written. @type value: int @param value: Integer value that represent the offset we want to start writing in the L{WriteData} stream. @raise WrongOffsetValueException: The value is beyond the total length of the data. """ if value >= len(self.data.getvalue()): raise excep.WrongOffsetValueException("Wrong offset value. Must be less than %d" % len(self.data)) self.data.seek(value)
python
def setOffset(self, value): """ Sets the offset of the L{WriteData} stream object in wich the data is written. @type value: int @param value: Integer value that represent the offset we want to start writing in the L{WriteData} stream. @raise WrongOffsetValueException: The value is beyond the total length of the data. """ if value >= len(self.data.getvalue()): raise excep.WrongOffsetValueException("Wrong offset value. Must be less than %d" % len(self.data)) self.data.seek(value)
[ "def", "setOffset", "(", "self", ",", "value", ")", ":", "if", "value", ">=", "len", "(", "self", ".", "data", ".", "getvalue", "(", ")", ")", ":", "raise", "excep", ".", "WrongOffsetValueException", "(", "\"Wrong offset value. Must be less than %d\"", "%", "len", "(", "self", ".", "data", ")", ")", "self", ".", "data", ".", "seek", "(", "value", ")" ]
Sets the offset of the L{WriteData} stream object in wich the data is written. @type value: int @param value: Integer value that represent the offset we want to start writing in the L{WriteData} stream. @raise WrongOffsetValueException: The value is beyond the total length of the data.
[ "Sets", "the", "offset", "of", "the", "L", "{", "WriteData", "}", "stream", "object", "in", "wich", "the", "data", "is", "written", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L151-L162
train
crackinglandia/pype32
pype32/utils.py
WriteData.skipBytes
def skipBytes(self, nroBytes): """ Skips the specified number as parameter to the current value of the L{WriteData} stream. @type nroBytes: int @param nroBytes: The number of bytes to skip. """ self.data.seek(nroBytes + self.data.tell())
python
def skipBytes(self, nroBytes): """ Skips the specified number as parameter to the current value of the L{WriteData} stream. @type nroBytes: int @param nroBytes: The number of bytes to skip. """ self.data.seek(nroBytes + self.data.tell())
[ "def", "skipBytes", "(", "self", ",", "nroBytes", ")", ":", "self", ".", "data", ".", "seek", "(", "nroBytes", "+", "self", ".", "data", ".", "tell", "(", ")", ")" ]
Skips the specified number as parameter to the current value of the L{WriteData} stream. @type nroBytes: int @param nroBytes: The number of bytes to skip.
[ "Skips", "the", "specified", "number", "as", "parameter", "to", "the", "current", "value", "of", "the", "L", "{", "WriteData", "}", "stream", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L164-L171
train
crackinglandia/pype32
pype32/utils.py
ReadData.readDword
def readDword(self): """ Reads a dword value from the L{ReadData} stream object. @rtype: int @return: The dword value read from the L{ReadData} stream. """ dword = unpack(self.endianness + ('L' if not self.signed else 'l'), self.readAt(self.offset, 4))[0] self.offset += 4 return dword
python
def readDword(self): """ Reads a dword value from the L{ReadData} stream object. @rtype: int @return: The dword value read from the L{ReadData} stream. """ dword = unpack(self.endianness + ('L' if not self.signed else 'l'), self.readAt(self.offset, 4))[0] self.offset += 4 return dword
[ "def", "readDword", "(", "self", ")", ":", "dword", "=", "unpack", "(", "self", ".", "endianness", "+", "(", "'L'", "if", "not", "self", ".", "signed", "else", "'l'", ")", ",", "self", ".", "readAt", "(", "self", ".", "offset", ",", "4", ")", ")", "[", "0", "]", "self", ".", "offset", "+=", "4", "return", "dword" ]
Reads a dword value from the L{ReadData} stream object. @rtype: int @return: The dword value read from the L{ReadData} stream.
[ "Reads", "a", "dword", "value", "from", "the", "L", "{", "ReadData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L209-L218
train
crackinglandia/pype32
pype32/utils.py
ReadData.readWord
def readWord(self): """ Reads a word value from the L{ReadData} stream object. @rtype: int @return: The word value read from the L{ReadData} stream. """ word = unpack(self.endianness + ('H' if not self.signed else 'h'), self.readAt(self.offset, 2))[0] self.offset += 2 return word
python
def readWord(self): """ Reads a word value from the L{ReadData} stream object. @rtype: int @return: The word value read from the L{ReadData} stream. """ word = unpack(self.endianness + ('H' if not self.signed else 'h'), self.readAt(self.offset, 2))[0] self.offset += 2 return word
[ "def", "readWord", "(", "self", ")", ":", "word", "=", "unpack", "(", "self", ".", "endianness", "+", "(", "'H'", "if", "not", "self", ".", "signed", "else", "'h'", ")", ",", "self", ".", "readAt", "(", "self", ".", "offset", ",", "2", ")", ")", "[", "0", "]", "self", ".", "offset", "+=", "2", "return", "word" ]
Reads a word value from the L{ReadData} stream object. @rtype: int @return: The word value read from the L{ReadData} stream.
[ "Reads", "a", "word", "value", "from", "the", "L", "{", "ReadData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L220-L229
train
crackinglandia/pype32
pype32/utils.py
ReadData.readByte
def readByte(self): """ Reads a byte value from the L{ReadData} stream object. @rtype: int @return: The byte value read from the L{ReadData} stream. """ byte = unpack('B' if not self.signed else 'b', self.readAt(self.offset, 1))[0] self.offset += 1 return byte
python
def readByte(self): """ Reads a byte value from the L{ReadData} stream object. @rtype: int @return: The byte value read from the L{ReadData} stream. """ byte = unpack('B' if not self.signed else 'b', self.readAt(self.offset, 1))[0] self.offset += 1 return byte
[ "def", "readByte", "(", "self", ")", ":", "byte", "=", "unpack", "(", "'B'", "if", "not", "self", ".", "signed", "else", "'b'", ",", "self", ".", "readAt", "(", "self", ".", "offset", ",", "1", ")", ")", "[", "0", "]", "self", ".", "offset", "+=", "1", "return", "byte" ]
Reads a byte value from the L{ReadData} stream object. @rtype: int @return: The byte value read from the L{ReadData} stream.
[ "Reads", "a", "byte", "value", "from", "the", "L", "{", "ReadData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L231-L240
train
crackinglandia/pype32
pype32/utils.py
ReadData.readQword
def readQword(self): """ Reads a qword value from the L{ReadData} stream object. @rtype: int @return: The qword value read from the L{ReadData} stream. """ qword = unpack(self.endianness + ('Q' if not self.signed else 'b'), self.readAt(self.offset, 8))[0] self.offset += 8 return qword
python
def readQword(self): """ Reads a qword value from the L{ReadData} stream object. @rtype: int @return: The qword value read from the L{ReadData} stream. """ qword = unpack(self.endianness + ('Q' if not self.signed else 'b'), self.readAt(self.offset, 8))[0] self.offset += 8 return qword
[ "def", "readQword", "(", "self", ")", ":", "qword", "=", "unpack", "(", "self", ".", "endianness", "+", "(", "'Q'", "if", "not", "self", ".", "signed", "else", "'b'", ")", ",", "self", ".", "readAt", "(", "self", ".", "offset", ",", "8", ")", ")", "[", "0", "]", "self", ".", "offset", "+=", "8", "return", "qword" ]
Reads a qword value from the L{ReadData} stream object. @rtype: int @return: The qword value read from the L{ReadData} stream.
[ "Reads", "a", "qword", "value", "from", "the", "L", "{", "ReadData", "}", "stream", "object", "." ]
192fd14dfc0dd36d953739a81c17fbaf5e3d6076
https://github.com/crackinglandia/pype32/blob/192fd14dfc0dd36d953739a81c17fbaf5e3d6076/pype32/utils.py#L242-L251
train