repo
stringlengths
2
99
file
stringlengths
13
225
code
stringlengths
0
18.3M
file_length
int64
0
18.3M
avg_line_length
float64
0
1.36M
max_line_length
int64
0
4.26M
extension_type
stringclasses
1 value
lingua-py
lingua-py-main/lingua/_model.py
# # Copyright © 2022-present Peter M. Stahl pemistahl@gmail.com # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either expressed or implied. # See the License for the specific language governing permissions and # limitations under the License. import math import numpy as np import regex from collections import Counter from dataclasses import dataclass from fractions import Fraction from pathlib import Path from typing import Counter as TypedCounter, Dict, List, Optional from .language import Language from ._ngram import _get_ngram_name_by_length, _NgramRange @dataclass class _TrainingDataLanguageModel: language: Language absolute_frequencies: Optional[Dict[str, int]] relative_frequencies: Optional[Dict[str, Fraction]] @classmethod def from_text( cls, text: List[str], language: Language, ngram_length: int, char_class: str, lower_ngram_absolute_frequencies: Optional[Dict[str, int]], ) -> "_TrainingDataLanguageModel": absolute_frequencies = cls.compute_absolute_frequencies( text, ngram_length, char_class ) relative_frequencies = cls.compute_relative_frequencies( ngram_length, absolute_frequencies, lower_ngram_absolute_frequencies ) return _TrainingDataLanguageModel( language=language, absolute_frequencies=absolute_frequencies, relative_frequencies=relative_frequencies, ) @classmethod def from_numpy_binary_file( cls, language: Language, ngram_length: int ) -> Optional[np.ndarray]: ngram_name = _get_ngram_name_by_length(ngram_length) iso_code = language.iso_code_639_1.name.lower() relative_file_path = f"./language-models/{iso_code}/{ngram_name}s.npz" absolute_file_path = Path(__file__).parent / relative_file_path try: with np.load(absolute_file_path) as data: return data["arr"] except OSError: return None def to_numpy_binary_file(self, file_path: Path, ngram_length: int): frequencies = [] if self.relative_frequencies is not None: for ngram, fraction in self.relative_frequencies.items(): frequency = math.log(fraction.numerator / fraction.denominator) frequencies.append((ngram, frequency)) dtype = [("ngram", f"U{ngram_length}"), ("frequency", "f2")] arr = np.array(frequencies, dtype=dtype) arr.sort() np.savez_compressed(file_path, arr=arr) @classmethod def compute_absolute_frequencies( cls, text: List[str], ngram_length: int, char_class: str ) -> Dict[str, int]: absolute_frequencies: TypedCounter[str] = Counter() regexp = regex.compile(r"^[{}]+$".format(char_class)) for line in text: lowercased_line = line.lower() for i in range(0, len(lowercased_line) - ngram_length + 1): substr = lowercased_line[i : i + ngram_length] if regexp.match(substr) is not None: absolute_frequencies.update([substr]) return absolute_frequencies @classmethod def compute_relative_frequencies( cls, ngram_length: int, absolute_frequencies: Dict[str, int], lower_ngram_absolute_frequencies: Optional[Dict[str, int]], ) -> Dict[str, Fraction]: if lower_ngram_absolute_frequencies is None: return {} ngram_probabilities = {} total_ngram_frequency = sum(absolute_frequencies.values()) for ngram, frequency in absolute_frequencies.items(): if ngram_length == 1 or len(lower_ngram_absolute_frequencies) == 0: denominator = total_ngram_frequency else: substr = ngram[: ngram_length - 1] denominator = lower_ngram_absolute_frequencies[substr] ngram_probabilities[ngram] = Fraction(frequency, denominator) return ngram_probabilities @dataclass class _TestDataLanguageModel: ngrams: List[List[str]] @classmethod def from_text(cls, words: List[str], ngram_length: int) -> "_TestDataLanguageModel": if ngram_length not in range(1, 6): raise ValueError(f"ngram length {ngram_length} is not in range 1..6") ngrams = set() for word in words: chars_count = len(word) if chars_count >= ngram_length: for i in range(0, chars_count - ngram_length + 1): substr = word[i : i + ngram_length] ngrams.add(substr) lower_order_ngrams = [list(_NgramRange(ngram)) for ngram in ngrams] return _TestDataLanguageModel(lower_order_ngrams)
5,137
36.50365
88
py
lingua-py
lingua-py-main/lingua/language.py
# # Copyright © 2022-present Peter M. Stahl pemistahl@gmail.com # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either expressed or implied. # See the License for the specific language governing permissions and # limitations under the License. import regex from enum import Enum from functools import total_ordering from typing import Dict, FrozenSet, List, Optional, Pattern from .isocode import IsoCode639_1, IsoCode639_3 def _pattern(char_class: str) -> Pattern[str]: return regex.compile(r"^\p{{Is{}}}+$".format(char_class)) class _Alphabet(Enum): ARABIC = (1, _pattern("Arabic")) ARMENIAN = (2, _pattern("Armenian")) BENGALI = (3, _pattern("Bengali")) CYRILLIC = (4, _pattern("Cyrillic")) DEVANAGARI = (5, _pattern("Devanagari")) GEORGIAN = (6, _pattern("Georgian")) GREEK = (7, _pattern("Greek")) GUJARATI = (8, _pattern("Gujarati")) GURMUKHI = (9, _pattern("Gurmukhi")) HAN = (10, _pattern("Han")) HANGUL = (11, _pattern("Hangul")) HEBREW = (12, _pattern("Hebrew")) HIRAGANA = (13, _pattern("Hiragana")) KATAKANA = (14, _pattern("Katakana")) LATIN = (15, _pattern("Latin")) TAMIL = (16, _pattern("Tamil")) TELUGU = (17, _pattern("Telugu")) THAI = (18, _pattern("Thai")) def __new__(cls, *args, **kwargs): obj = object.__new__(cls) obj._value_ = args[0] return obj def __init__(self, _: int, pattern: Pattern[str]): self._pattern = pattern def matches(self, text: str) -> bool: return self._pattern.match(text) is not None @classmethod def all_supporting_single_language(cls) -> Dict["_Alphabet", "Language"]: alphabets = {} for alphabet in _Alphabet: supported_languages = alphabet._supported_languages() if len(supported_languages) == 1: alphabets[alphabet] = supported_languages[0] return alphabets def _supported_languages(self) -> List["Language"]: languages = [] for language in Language: if self in language._alphabets: languages.append(language) return languages @total_ordering class Language(Enum): """This enum specifies the so far 75 supported languages which can be detected by *Lingua*. """ AFRIKAANS = ( 1, IsoCode639_1.AF, IsoCode639_3.AFR, frozenset([_Alphabet.LATIN]), ) ALBANIAN = (2, IsoCode639_1.SQ, IsoCode639_3.SQI, frozenset([_Alphabet.LATIN])) ARABIC = (3, IsoCode639_1.AR, IsoCode639_3.ARA, frozenset([_Alphabet.ARABIC])) ARMENIAN = ( 4, IsoCode639_1.HY, IsoCode639_3.HYE, frozenset([_Alphabet.ARMENIAN]), ) AZERBAIJANI = ( 5, IsoCode639_1.AZ, IsoCode639_3.AZE, frozenset([_Alphabet.LATIN]), "Əə", ) BASQUE = (6, IsoCode639_1.EU, IsoCode639_3.EUS, frozenset([_Alphabet.LATIN])) BELARUSIAN = ( 7, IsoCode639_1.BE, IsoCode639_3.BEL, frozenset([_Alphabet.CYRILLIC]), ) BENGALI = ( 8, IsoCode639_1.BN, IsoCode639_3.BEN, frozenset([_Alphabet.BENGALI]), ) BOKMAL = (9, IsoCode639_1.NB, IsoCode639_3.NOB, frozenset([_Alphabet.LATIN])) BOSNIAN = (10, IsoCode639_1.BS, IsoCode639_3.BOS, frozenset([_Alphabet.LATIN])) BULGARIAN = ( 11, IsoCode639_1.BG, IsoCode639_3.BUL, frozenset([_Alphabet.CYRILLIC]), ) CATALAN = ( 12, IsoCode639_1.CA, IsoCode639_3.CAT, frozenset([_Alphabet.LATIN]), "Ïï", ) CHINESE = (13, IsoCode639_1.ZH, IsoCode639_3.ZHO, frozenset([_Alphabet.HAN])) CROATIAN = (14, IsoCode639_1.HR, IsoCode639_3.HRV, frozenset([_Alphabet.LATIN])) CZECH = ( 15, IsoCode639_1.CS, IsoCode639_3.CES, frozenset([_Alphabet.LATIN]), "ĚěŘřŮů", ) DANISH = (16, IsoCode639_1.DA, IsoCode639_3.DAN, frozenset([_Alphabet.LATIN])) DUTCH = (17, IsoCode639_1.NL, IsoCode639_3.NLD, frozenset([_Alphabet.LATIN])) ENGLISH = (18, IsoCode639_1.EN, IsoCode639_3.ENG, frozenset([_Alphabet.LATIN])) ESPERANTO = ( 19, IsoCode639_1.EO, IsoCode639_3.EPO, frozenset([_Alphabet.LATIN]), "ĈĉĜĝĤĥĴĵŜŝŬŭ", ) ESTONIAN = (20, IsoCode639_1.ET, IsoCode639_3.EST, frozenset([_Alphabet.LATIN])) FINNISH = (21, IsoCode639_1.FI, IsoCode639_3.FIN, frozenset([_Alphabet.LATIN])) FRENCH = (22, IsoCode639_1.FR, IsoCode639_3.FRA, frozenset([_Alphabet.LATIN])) GANDA = (23, IsoCode639_1.LG, IsoCode639_3.LUG, frozenset([_Alphabet.LATIN])) GEORGIAN = ( 24, IsoCode639_1.KA, IsoCode639_3.KAT, frozenset([_Alphabet.GEORGIAN]), ) GERMAN = ( 25, IsoCode639_1.DE, IsoCode639_3.DEU, frozenset([_Alphabet.LATIN]), "ß", ) GREEK = (26, IsoCode639_1.EL, IsoCode639_3.ELL, frozenset([_Alphabet.GREEK])) GUJARATI = ( 27, IsoCode639_1.GU, IsoCode639_3.GUJ, frozenset([_Alphabet.GUJARATI]), ) HEBREW = (28, IsoCode639_1.HE, IsoCode639_3.HEB, frozenset([_Alphabet.HEBREW])) HINDI = ( 29, IsoCode639_1.HI, IsoCode639_3.HIN, frozenset([_Alphabet.DEVANAGARI]), ) HUNGARIAN = ( 30, IsoCode639_1.HU, IsoCode639_3.HUN, frozenset([_Alphabet.LATIN]), "ŐőŰű", ) ICELANDIC = ( 31, IsoCode639_1.IS, IsoCode639_3.ISL, frozenset([_Alphabet.LATIN]), ) INDONESIAN = ( 32, IsoCode639_1.ID, IsoCode639_3.IND, frozenset([_Alphabet.LATIN]), ) IRISH = (33, IsoCode639_1.GA, IsoCode639_3.GLE, frozenset([_Alphabet.LATIN])) ITALIAN = (34, IsoCode639_1.IT, IsoCode639_3.ITA, frozenset([_Alphabet.LATIN])) JAPANESE = ( 35, IsoCode639_1.JA, IsoCode639_3.JPN, frozenset([_Alphabet.HIRAGANA, _Alphabet.KATAKANA, _Alphabet.HAN]), ) KAZAKH = ( 36, IsoCode639_1.KK, IsoCode639_3.KAZ, frozenset([_Alphabet.CYRILLIC]), "ӘәҒғҚқҢңҰұ", ) KOREAN = (37, IsoCode639_1.KO, IsoCode639_3.KOR, frozenset([_Alphabet.HANGUL])) LATIN = (38, IsoCode639_1.LA, IsoCode639_3.LAT, frozenset([_Alphabet.LATIN])) LATVIAN = ( 39, IsoCode639_1.LV, IsoCode639_3.LAV, frozenset([_Alphabet.LATIN]), "ĢģĶķĻļŅņ", ) LITHUANIAN = ( 40, IsoCode639_1.LT, IsoCode639_3.LIT, frozenset([_Alphabet.LATIN]), "ĖėĮįŲų", ) MACEDONIAN = ( 41, IsoCode639_1.MK, IsoCode639_3.MKD, frozenset([_Alphabet.CYRILLIC]), "ЃѓЅѕЌќЏџ", ) MALAY = (42, IsoCode639_1.MS, IsoCode639_3.MSA, frozenset([_Alphabet.LATIN])) MAORI = (43, IsoCode639_1.MI, IsoCode639_3.MRI, frozenset([_Alphabet.LATIN])) MARATHI = ( 44, IsoCode639_1.MR, IsoCode639_3.MAR, frozenset([_Alphabet.DEVANAGARI]), "ळ", ) MONGOLIAN = ( 45, IsoCode639_1.MN, IsoCode639_3.MON, frozenset([_Alphabet.CYRILLIC]), "ӨөҮү", ) NYNORSK = (46, IsoCode639_1.NN, IsoCode639_3.NNO, frozenset([_Alphabet.LATIN])) PERSIAN = (47, IsoCode639_1.FA, IsoCode639_3.FAS, frozenset([_Alphabet.ARABIC])) POLISH = ( 48, IsoCode639_1.PL, IsoCode639_3.POL, frozenset([_Alphabet.LATIN]), "ŁłŃńŚśŹź", ) PORTUGUESE = ( 49, IsoCode639_1.PT, IsoCode639_3.POR, frozenset([_Alphabet.LATIN]), ) PUNJABI = ( 50, IsoCode639_1.PA, IsoCode639_3.PAN, frozenset([_Alphabet.GURMUKHI]), ) ROMANIAN = ( 51, IsoCode639_1.RO, IsoCode639_3.RON, frozenset([_Alphabet.LATIN]), "Țţ", ) RUSSIAN = ( 52, IsoCode639_1.RU, IsoCode639_3.RUS, frozenset([_Alphabet.CYRILLIC]), ) SERBIAN = ( 53, IsoCode639_1.SR, IsoCode639_3.SRP, frozenset([_Alphabet.CYRILLIC]), "ЂђЋћ", ) SHONA = (54, IsoCode639_1.SN, IsoCode639_3.SNA, frozenset([_Alphabet.LATIN])) SLOVAK = ( 55, IsoCode639_1.SK, IsoCode639_3.SLK, frozenset([_Alphabet.LATIN]), "Ĺ弾Ŕŕ", ) SLOVENE = (56, IsoCode639_1.SL, IsoCode639_3.SLV, frozenset([_Alphabet.LATIN])) SOMALI = (57, IsoCode639_1.SO, IsoCode639_3.SOM, frozenset([_Alphabet.LATIN])) SOTHO = (58, IsoCode639_1.ST, IsoCode639_3.SOT, frozenset([_Alphabet.LATIN])) SPANISH = ( 59, IsoCode639_1.ES, IsoCode639_3.SPA, frozenset([_Alphabet.LATIN]), "¿¡", ) SWAHILI = (60, IsoCode639_1.SW, IsoCode639_3.SWA, frozenset([_Alphabet.LATIN])) SWEDISH = (61, IsoCode639_1.SV, IsoCode639_3.SWE, frozenset([_Alphabet.LATIN])) TAGALOG = (62, IsoCode639_1.TL, IsoCode639_3.TGL, frozenset([_Alphabet.LATIN])) TAMIL = (63, IsoCode639_1.TA, IsoCode639_3.TAM, frozenset([_Alphabet.TAMIL])) TELUGU = (64, IsoCode639_1.TE, IsoCode639_3.TEL, frozenset([_Alphabet.TELUGU])) THAI = (65, IsoCode639_1.TH, IsoCode639_3.THA, frozenset([_Alphabet.THAI])) TSONGA = (66, IsoCode639_1.TS, IsoCode639_3.TSO, frozenset([_Alphabet.LATIN])) TSWANA = (67, IsoCode639_1.TN, IsoCode639_3.TSN, frozenset([_Alphabet.LATIN])) TURKISH = (68, IsoCode639_1.TR, IsoCode639_3.TUR, frozenset([_Alphabet.LATIN])) UKRAINIAN = ( 69, IsoCode639_1.UK, IsoCode639_3.UKR, frozenset([_Alphabet.CYRILLIC]), "ҐґЄєЇї", ) URDU = (70, IsoCode639_1.UR, IsoCode639_3.URD, frozenset([_Alphabet.ARABIC])) VIETNAMESE = ( 71, IsoCode639_1.VI, IsoCode639_3.VIE, frozenset([_Alphabet.LATIN]), "ẰằẦầẲẳẨẩẴẵẪẫẮắẤấẠạẶặẬậỀềẺẻỂểẼẽỄễẾếỆệỈỉĨĩỊị" + "ƠơỒồỜờỎỏỔổỞởỖỗỠỡỐốỚớỘộỢợƯưỪừỦủỬửŨũỮữỨứỤụỰựỲỳỶỷỸỹỴỵ", ) WELSH = (72, IsoCode639_1.CY, IsoCode639_3.CYM, frozenset([_Alphabet.LATIN])) XHOSA = (73, IsoCode639_1.XH, IsoCode639_3.XHO, frozenset([_Alphabet.LATIN])) YORUBA = ( 74, IsoCode639_1.YO, IsoCode639_3.YOR, frozenset([_Alphabet.LATIN]), "Ṣṣ", ) ZULU = (75, IsoCode639_1.ZU, IsoCode639_3.ZUL, frozenset([_Alphabet.LATIN])) def __new__(cls, *args, **kwargs): obj = object.__new__(cls) obj._value_ = args[0] return obj def __init__( self, _: int, iso_code639_1: IsoCode639_1, iso_code639_3: IsoCode639_3, alphabets: FrozenSet[_Alphabet], unique_characters: Optional[str] = None, ): self.iso_code_639_1 = iso_code639_1 self.iso_code_639_3 = iso_code639_3 self._alphabets = alphabets self._unique_characters = unique_characters def __lt__(self, other): if self is None and other is not None: return False if self is not None and other is None: return True if not isinstance(other, Language): return NotImplemented return self.value < other.value def __repr__(self): return str(self) @classmethod def all(cls) -> FrozenSet["Language"]: """Return a set of all supported languages.""" return frozenset(language for language in Language) @classmethod def all_spoken_ones(cls) -> FrozenSet["Language"]: """Return a set of all supported spoken languages.""" return frozenset( language for language in Language if language is not Language.LATIN ) @classmethod def all_with_arabic_script(cls) -> FrozenSet["Language"]: """Return a set of all languages supporting the Arabic script.""" return frozenset( language for language in Language if _Alphabet.ARABIC in language._alphabets ) @classmethod def all_with_cyrillic_script(cls) -> FrozenSet["Language"]: """Return a set of all languages supporting the Cyrillic script.""" return frozenset( language for language in Language if _Alphabet.CYRILLIC in language._alphabets ) @classmethod def all_with_devanagari_script(cls) -> FrozenSet["Language"]: """Return a set of all languages supporting the Devanagari script.""" return frozenset( language for language in Language if _Alphabet.DEVANAGARI in language._alphabets ) @classmethod def all_with_latin_script(cls) -> FrozenSet["Language"]: """Return a set of all languages supporting the Latin script.""" return frozenset( language for language in Language if _Alphabet.LATIN in language._alphabets ) @classmethod def from_iso_code_639_1(cls, iso_code: IsoCode639_1) -> "Language": """Return the language associated with the ISO 639-1 code passed to this method. Raises: ValueError: if there is no language for the given ISO code """ for language in Language: if language.iso_code_639_1 == iso_code: return language raise ValueError(f"There is no language for ISO 639-1 code {iso_code}") @classmethod def from_iso_code_639_3(cls, iso_code: IsoCode639_3) -> "Language": """Return the language associated with the ISO 639-3 code passed to this method. Raises: ValueError: if there is no language for the given ISO code """ for language in Language: if language.iso_code_639_3 == iso_code: return language raise ValueError(f"There is no language for ISO 639-3 code {iso_code}")
14,230
30.694878
88
py
lingua-py
lingua-py-main/lingua/detector.py
# # Copyright © 2022-present Peter M. Stahl pemistahl@gmail.com # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either expressed or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np from collections import Counter from dataclasses import dataclass from decimal import Decimal from math import exp from typing import ( Counter as TypedCounter, Dict, FrozenSet, NamedTuple, Optional, List, ) from ._constant import ( CHARS_TO_LANGUAGES_MAPPING, JAPANESE_CHARACTER_SET, LETTERS, TOKENS_WITHOUT_WHITESPACE, TOKENS_WITH_OPTIONAL_WHITESPACE, ) from .language import Language, _Alphabet from ._model import _TrainingDataLanguageModel, _TestDataLanguageModel _UNIGRAM_MODELS: Dict[Language, np.ndarray] = {} _BIGRAM_MODELS: Dict[Language, np.ndarray] = {} _TRIGRAM_MODELS: Dict[Language, np.ndarray] = {} _QUADRIGRAM_MODELS: Dict[Language, np.ndarray] = {} _FIVEGRAM_MODELS: Dict[Language, np.ndarray] = {} _CACHE: Dict[Language, Dict[str, Optional[float]]] = {} _HIGH_ACCURACY_MODE_MAX_TEXT_LENGTH = 120 def _split_text_into_words(text: str) -> List[str]: return LETTERS.findall(text.lower()) def _load_language_models( language: Language, ngram_length: int, ) -> Optional[Dict[Language, np.ndarray]]: loaded_model = _TrainingDataLanguageModel.from_numpy_binary_file( language, ngram_length ) if loaded_model is None: return None return {language: loaded_model} def _sum_up_probabilities( probabilities: List[Dict[Language, float]], unigram_counts: Optional[TypedCounter[Language]], filtered_languages: FrozenSet[Language], ) -> Dict[Language, Decimal]: summed_up_probabilities = {} for language in filtered_languages: result = 0.0 for dct in probabilities: if language in dct: result += dct[language] if unigram_counts is not None and language in unigram_counts: result /= unigram_counts[language] if result != 0: # Use Decimal instead of float to prevent numerical underflow summed_up_probabilities[language] = _compute_exponent(result) return summed_up_probabilities def _compute_exponent(value: float) -> Decimal: exponent = exp(value) if exponent > 0: return Decimal(exponent) return Decimal(value).exp() def _sort_confidence_values(values: List["ConfidenceValue"]): values.sort(key=lambda tup: (-tup[1], tup[0])) def _collect_languages_with_unique_characters( languages: FrozenSet[Language], ) -> FrozenSet[Language]: return frozenset( {language for language in languages if language._unique_characters is not None} ) def _collect_one_language_alphabets( languages: FrozenSet[Language], ) -> Dict[_Alphabet, Language]: return { alphabet: language for alphabet, language in _Alphabet.all_supporting_single_language().items() if language in languages } def _merge_adjacent_results( results: List["DetectionResult"], mergeable_result_indices: List[int] ): mergeable_result_indices.sort(reverse=True) for i in mergeable_result_indices: if i == 0: results[i + 1] = DetectionResult( start_index=results[i].start_index, end_index=results[i + 1].end_index, word_count=results[i + 1].word_count, language=results[i + 1].language, ) else: results[i - 1] = DetectionResult( start_index=results[i - 1].start_index, end_index=results[i].end_index, word_count=results[i - 1].word_count, language=results[i - 1].language, ) del results[i] if len(results) == 1: break class ConfidenceValue(NamedTuple): """This class describes a language's confidence value. Attributes: language (Language): The language associated with this confidence value. value (float): The language's confidence value which lies between 0.0 and 1.0. """ language: Language value: float class DetectionResult(NamedTuple): """This class describes a contiguous single-language text section within a possibly mixed-language text. Attributes: start_index (int): The start index of the identified single-language substring. end_index (int): The end index of the identified single-language substring. word_count (int): The number of words being part of the identified single-language substring. language (Language): The detected language of the identified single-language substring. """ start_index: int end_index: int word_count: int language: Language @dataclass class LanguageDetector: """This class detects the language of text.""" _languages: FrozenSet[Language] _minimum_relative_distance: float _is_low_accuracy_mode_enabled: bool _languages_with_unique_characters: FrozenSet[Language] _one_language_alphabets: Dict[_Alphabet, Language] _unigram_language_models: Dict[Language, np.ndarray] _bigram_language_models: Dict[Language, np.ndarray] _trigram_language_models: Dict[Language, np.ndarray] _quadrigram_language_models: Dict[Language, np.ndarray] _fivegram_language_models: Dict[Language, np.ndarray] _cache: Dict[Language, Dict[str, Optional[float]]] def __repr__(self): languages = sorted([language.name for language in self._languages]) return ( "LanguageDetector(" f"_languages={languages}, " f"_minimum_relative_distance={self._minimum_relative_distance})" ) @classmethod def _from( cls, languages: FrozenSet[Language], minimum_relative_distance: float, is_every_language_model_preloaded: bool, is_low_accuracy_mode_enabled: bool, ) -> "LanguageDetector": languages_with_unique_characters = _collect_languages_with_unique_characters( languages ) one_language_alphabets = _collect_one_language_alphabets(languages) detector = LanguageDetector( languages, minimum_relative_distance, is_low_accuracy_mode_enabled, languages_with_unique_characters, one_language_alphabets, _UNIGRAM_MODELS, _BIGRAM_MODELS, _TRIGRAM_MODELS, _QUADRIGRAM_MODELS, _FIVEGRAM_MODELS, _CACHE, ) if is_every_language_model_preloaded: detector._preload_language_models() return detector def _preload_language_models(self): trigram_models = [ _load_language_models(language, 3) for language in self._languages ] for trigram_model in trigram_models: if trigram_model is not None: self._trigram_language_models.update(trigram_model) if not self._is_low_accuracy_mode_enabled: (unigram_models, bigram_models, quadrigram_models, fivegram_models,) = [ [ _load_language_models(language, ngram_length) for language in self._languages ] for ngram_length in (1, 2, 4, 5) ] for unigram_model in unigram_models: if unigram_model is not None: self._unigram_language_models.update(unigram_model) for bigram_model in bigram_models: if bigram_model is not None: self._bigram_language_models.update(bigram_model) for quadrigram_model in quadrigram_models: if quadrigram_model is not None: self._quadrigram_language_models.update(quadrigram_model) for fivegram_model in fivegram_models: if fivegram_model is not None: self._fivegram_language_models.update(fivegram_model) def detect_language_of(self, text: str) -> Optional[Language]: """Detect the language of text. Args: text (str): The text whose language should be identified. Returns: The identified language. If the language cannot be reliably detected, None is returned. """ confidence_values = self.compute_language_confidence_values(text) if len(confidence_values) == 0: return None most_likely_language, most_likely_language_probability = confidence_values[0] if len(confidence_values) == 1: return most_likely_language second_most_likely_language_probability = confidence_values[1].value if most_likely_language_probability == second_most_likely_language_probability: return None if ( most_likely_language_probability - second_most_likely_language_probability < self._minimum_relative_distance ): return None return most_likely_language def detect_multiple_languages_of(self, text: str) -> List[DetectionResult]: """Attempt to detect multiple languages in mixed-language text. This feature is experimental and under continuous development. A list of DetectionResult is returned containing an entry for each contiguous single-language text section as identified by the library. Each entry consists of the identified language, a start index and an end index. The indices denote the substring that has been identified as a contiguous single-language text section. Args: text (str): The text whose language should be identified. Returns: A list of detection results. Each result contains the identified language, the start index and end index of the identified single-language substring. """ if len(text) == 0: return [] tokens_without_whitespace = TOKENS_WITHOUT_WHITESPACE.findall(text) if len(tokens_without_whitespace) == 0: return [] results = [] language_counts: TypedCounter[Language] = Counter() language = self.detect_language_of(text) if language is not None: language_counts[language] += 1 for word in tokens_without_whitespace: if len(word) < 5: continue language = self.detect_language_of(word) if language is not None: language_counts[language] += 1 languages = frozenset(language_counts.keys()) if len(languages) == 1: result = DetectionResult( start_index=0, end_index=len(text), word_count=len(tokens_without_whitespace), language=next(iter(languages)), ) results.append(result) else: previous_detector_languages = self._languages.copy() self._languages = languages current_start_index = 0 current_end_index = 0 word_count = 0 current_language = None token_matches = list(TOKENS_WITH_OPTIONAL_WHITESPACE.finditer(text)) last_index = len(token_matches) - 1 for i, token_match in enumerate(token_matches): word = token_match.group(0) language = self.detect_language_of(word) if i == 0: current_language = language if language != current_language and current_language is not None: result = DetectionResult( start_index=current_start_index, end_index=current_end_index, word_count=word_count, language=current_language, ) results.append(result) current_start_index = current_end_index current_language = language word_count = 0 current_end_index = token_match.end() word_count += 1 if i == last_index and current_language is not None: result = DetectionResult( start_index=current_start_index, end_index=current_end_index, word_count=word_count, language=current_language, ) results.append(result) if len(results) > 1: mergeable_result_indices = [] for i, result in enumerate(results): if result.word_count == 1: mergeable_result_indices.append(i) _merge_adjacent_results(results, mergeable_result_indices) if len(results) > 1: mergeable_result_indices.clear() for i in range(len(results) - 1): if results[i].language == results[i + 1].language: mergeable_result_indices.append(i + 1) _merge_adjacent_results(results, mergeable_result_indices) self._languages = previous_detector_languages return results def compute_language_confidence_values(self, text: str) -> List[ConfidenceValue]: """Compute confidence values for each language supported by this detector for the given text. The confidence values denote how likely it is that the given text has been written in any of the languages supported by this detector. A list is returned containing those languages which the calling instance of LanguageDetector has been built from. The entries are sorted by their confidence value in descending order. Each value is a probability between 0.0 and 1.0. The probabilities of all languages will sum to 1.0. If the language is unambiguously identified by the rule engine, the value 1.0 will always be returned for this language. The other languages will receive a value of 0.0. Args: text (str): The text for which to compute confidence values. Returns: A list of 2-element tuples. Each tuple contains a language and the associated confidence value. """ values = [ConfidenceValue(language, 0.0) for language in self._languages] words = _split_text_into_words(text) if len(words) == 0: _sort_confidence_values(values) return values language_detected_by_rules = self._detect_language_with_rules(words) if language_detected_by_rules is not None: for i in range(len(values)): if values[i].language == language_detected_by_rules: values[i] = ConfidenceValue(language_detected_by_rules, 1.0) break _sort_confidence_values(values) return values filtered_languages = self._filter_languages_by_rules(words) if len(filtered_languages) == 1: language_detected_by_filter = next(iter(filtered_languages)) for i in range(len(values)): if values[i].language == language_detected_by_filter: values[i] = ConfidenceValue(language_detected_by_filter, 1.0) break _sort_confidence_values(values) return values character_count = sum(len(word) for word in words) if self._is_low_accuracy_mode_enabled and character_count < 3: _sort_confidence_values(values) return values ngram_length_range = ( range(3, 4) if character_count >= _HIGH_ACCURACY_MODE_MAX_TEXT_LENGTH or self._is_low_accuracy_mode_enabled else range(1, 6) ) unigram_counts = None all_probabilities = [] for ngram_length in ngram_length_range: if character_count >= ngram_length: ngram_model = _TestDataLanguageModel.from_text(words, ngram_length) if ngram_length == 1: unigram_counts = self._count_unigrams( ngram_model, filtered_languages ) probabilities = self._compute_language_probabilities( ngram_model, filtered_languages ) all_probabilities.append(probabilities) summed_up_probabilities = _sum_up_probabilities( all_probabilities, unigram_counts, filtered_languages ) if len(summed_up_probabilities) == 0: _sort_confidence_values(values) return values denominator = sum(summed_up_probabilities.values()) for language, probability in summed_up_probabilities.items(): for i in range(len(values)): if values[i].language == language: # apply softmax function normalized_probability = probability / denominator values[i] = ConfidenceValue(language, float(normalized_probability)) break _sort_confidence_values(values) return values def compute_language_confidence(self, text: str, language: Language) -> float: """Compute the confidence value for the given language and input text. The confidence value denotes how likely it is that the given text has been written in the given language. The value that this method computes is a number between 0.0 and 1.0. If the language is unambiguously identified by the rule engine, the value 1.0 will always be returned. If the given language is not supported by this detector instance, the value 0.0 will always be returned. Args: text (str): The text for which to compute the confidence value. language (Language): The language for which to compute the confidence value. Returns: A float value between 0.0 and 1.0. """ confidence_values = self.compute_language_confidence_values(text) for value in confidence_values: if value.language == language: return value.value return 0.0 def _detect_language_with_rules(self, words: List[str]) -> Optional[Language]: total_language_counts: TypedCounter[Optional[Language]] = Counter() half_word_count = len(words) * 0.5 for word in words: word_language_counts: TypedCounter[Language] = Counter() for char in word: is_match = False for alphabet, language in self._one_language_alphabets.items(): if alphabet.matches(char): word_language_counts[language] += 1 is_match = True break if not is_match: if _Alphabet.HAN.matches(char): word_language_counts[Language.CHINESE] += 1 elif JAPANESE_CHARACTER_SET.fullmatch(char) is not None: word_language_counts[Language.JAPANESE] += 1 elif ( _Alphabet.LATIN.matches(char) or _Alphabet.CYRILLIC.matches(char) or _Alphabet.DEVANAGARI.matches(char) ): for language in self._languages_with_unique_characters: if ( language._unique_characters is not None and char in language._unique_characters ): word_language_counts[language] += 1 if len(word_language_counts) == 0: total_language_counts[None] += 1 elif len(word_language_counts) == 1: language = list(word_language_counts.elements())[0] if language in self._languages: total_language_counts[language] += 1 else: total_language_counts[None] += 1 elif ( Language.CHINESE in word_language_counts and Language.JAPANESE in word_language_counts ): total_language_counts[Language.JAPANESE] += 1 else: most_frequent_word_languages = word_language_counts.most_common(2) ( most_frequent_word_language, first_count, ) = most_frequent_word_languages[0] (_, second_count) = most_frequent_word_languages[1] if ( first_count > second_count and most_frequent_word_language in self._languages ): total_language_counts[most_frequent_word_language] += 1 else: total_language_counts[None] += 1 if total_language_counts[None] < half_word_count: del total_language_counts[None] if len(total_language_counts) == 0: return None if len(total_language_counts) == 1: return list(total_language_counts)[0] if ( len(total_language_counts) == 2 and Language.CHINESE in total_language_counts and Language.JAPANESE in total_language_counts ): return Language.JAPANESE most_frequent_total_languages = total_language_counts.most_common(2) (most_frequent_total_language, first_count) = most_frequent_total_languages[0] (_, second_count) = most_frequent_total_languages[1] if first_count == second_count: return None return most_frequent_total_language def _filter_languages_by_rules(self, words: List[str]) -> FrozenSet[Language]: detected_alphabets: TypedCounter[_Alphabet] = Counter() half_word_count = len(words) * 0.5 for word in words: for alphabet in _Alphabet: if alphabet.matches(word): detected_alphabets[alphabet] += 1 break if len(detected_alphabets) == 0: return self._languages if len(detected_alphabets) > 1: distinct_alphabets = {count for count in detected_alphabets.values()} if len(distinct_alphabets) == 1: return self._languages most_frequent_alphabet = detected_alphabets.most_common(1)[0][0] filtered_languages = { language for language in self._languages if most_frequent_alphabet in language._alphabets } language_counts: TypedCounter[Language] = Counter() for characters, languages in CHARS_TO_LANGUAGES_MAPPING.items(): relevant_languages = languages.intersection(filtered_languages) for word in words: for character in characters: if character in word: for language in relevant_languages: language_counts[language] += 1 languages_subset = { language for language, count in language_counts.items() if count >= half_word_count } if len(languages_subset) > 0: return frozenset(languages_subset) return frozenset(filtered_languages) def _compute_language_probabilities( self, ngram_model: _TestDataLanguageModel, filtered_languages: FrozenSet[Language], ) -> Dict[Language, float]: probabilities = {} for language in filtered_languages: result = self._compute_sum_of_ngram_probabilities(language, ngram_model) if result < 0: probabilities[language] = result return probabilities def _compute_sum_of_ngram_probabilities( self, language: Language, ngram_model: _TestDataLanguageModel ) -> float: result = 0.0 for ngrams in ngram_model.ngrams: for ngram in ngrams: probability = self._look_up_ngram_probability(language, ngram) if probability is not None: result += probability break return result def _look_up_ngram_probability( self, language: Language, ngram: str ) -> Optional[float]: if language not in self._cache: self._cache[language] = {} if ngram in self._cache[language]: return self._cache[language][ngram] ngram_length = len(ngram) if ngram_length == 5: language_models = self._fivegram_language_models elif ngram_length == 4: language_models = self._quadrigram_language_models elif ngram_length == 3: language_models = self._trigram_language_models elif ngram_length == 2: language_models = self._bigram_language_models elif ngram_length == 1: language_models = self._unigram_language_models elif ngram_length == 0: raise ValueError("zerogram detected") else: raise ValueError(f"unsupported ngram length detected: {ngram_length}") probability = None if language not in language_models: models = _load_language_models(language, ngram_length) if models is None: self._cache[language][ngram] = probability return probability language_models.update(models) mask = np.isin(language_models[language]["ngram"], ngram) try: probability = language_models[language]["frequency"][mask][0] except IndexError: pass self._cache[language][ngram] = probability return probability def _count_unigrams( self, unigram_model: _TestDataLanguageModel, filtered_languages: FrozenSet[Language], ) -> TypedCounter[Language]: unigram_counts: TypedCounter[Language] = Counter() for language in filtered_languages: for unigrams in unigram_model.ngrams: if self._look_up_ngram_probability(language, unigrams[0]) is not None: unigram_counts[language] += 1 return unigram_counts
27,116
35.301205
88
py
lingua-py
lingua-py-main/lingua/__init__.py
# # Copyright © 2022-present Peter M. Stahl pemistahl@gmail.com # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either expressed or implied. # See the License for the specific language governing permissions and # limitations under the License. """ ## 1. What does this library do? Its task is simple: It tells you which language some provided textual data is written in. This is very useful as a preprocessing step for linguistic data in natural language processing applications such as text classification and spell checking. Other use cases, for instance, might include routing e-mails to the right geographically located customer service department, based on the e-mails' languages. ## 2. Why does this library exist? Language detection is often done as part of large machine learning frameworks or natural language processing applications. In cases where you don't need the full-fledged functionality of those systems or don't want to learn the ropes of those, a small flexible library comes in handy. Python is widely used in natural language processing, so there are a couple of comprehensive open source libraries for this task, such as Google's [*CLD 2*](https://github.com/CLD2Owners/cld2) and [*CLD 3*](https://github.com/google/cld3), [*langid*](https://github.com/saffsd/langid.py), [*fastText*](https://fasttext.cc/docs/en/language-identification.html) and [*langdetect*](https://github.com/Mimino666/langdetect). Unfortunately, except for the last one they have two major drawbacks: 1. Detection only works with quite lengthy text fragments. For very short text snippets such as Twitter messages, it does not provide adequate results. 2. The more languages take part in the decision process, the less accurate are the detection results. *Lingua* aims at eliminating these problems. It nearly does not need any configuration and yields pretty accurate results on both long and short text, even on single words and phrases. It draws on both rule-based and statistical methods but does not use any dictionaries of words. It does not need a connection to any external API or service either. Once the library has been downloaded, it can be used completely offline. ## 3. Which languages are supported? Compared to other language detection libraries, *Lingua's* focus is on *quality over quantity*, that is, getting detection right for a small set of languages first before adding new ones. Currently, the following 75 languages are supported: - Afrikaans - Albanian - Arabic - Armenian - Azerbaijani - Basque - Belarusian - Bengali - Norwegian Bokmal - Bosnian - Bulgarian - Catalan - Chinese - Croatian - Czech - Danish - Dutch - English - Esperanto - Estonian - Finnish - French - Ganda - Georgian - German - Greek - Gujarati - Hebrew - Hindi - Hungarian - Icelandic - Indonesian - Irish - Italian - Japanese - Kazakh - Korean - Latin - Latvian - Lithuanian - Macedonian - Malay - Maori - Marathi - Mongolian - Norwegian Nynorsk - Persian - Polish - Portuguese - Punjabi - Romanian - Russian - Serbian - Shona - Slovak - Slovene - Somali - Sotho - Spanish - Swahili - Swedish - Tagalog - Tamil - Telugu - Thai - Tsonga - Tswana - Turkish - Ukrainian - Urdu - Vietnamese - Welsh - Xhosa - Yoruba - Zulu ## 4. How good is it? *Lingua* is able to report accuracy statistics for some bundled test data available for each supported language. The test data for each language is split into three parts: 1. a list of single words with a minimum length of 5 characters 2. a list of word pairs with a minimum length of 10 characters 3. a list of complete grammatical sentences of various lengths Both the language models and the test data have been created from separate documents of the [Wortschatz corpora](https://wortschatz.uni-leipzig.de) offered by Leipzig University, Germany. Data crawled from various news websites have been used for training, each corpus comprising one million sentences. For testing, corpora made of arbitrarily chosen websites have been used, each comprising ten thousand sentences. From each test corpus, a random unsorted subset of 1000 single words, 1000 word pairs and 1000 sentences has been extracted, respectively. Given the generated test data, I have compared the detection results of *Lingua*, *fastText*, *langdetect*, *langid*, *CLD 2* and *CLD 3* running over the data of *Lingua's* supported 75 languages. Languages that are not supported by the other detectors are simply ignored for them during the detection process. Each of the following sections contains two plots. The bar plot shows the detailed accuracy results for each supported language. The box plot illustrates the distributions of the accuracy values for each classifier. The boxes themselves represent the areas which the middle 50 % of data lie within. Within the colored boxes, the horizontal lines mark the median of the distributions. ### 4.1 Single word detection <br/> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/boxplot-single-words.png" alt="Single Word Detection Performance" /> <br/> <details> <summary>Bar plot</summary> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/barplot-single-words.png" alt="Single Word Detection Performance" /> </details> <br/><br/> ### 4.2 Word pair detection <br/> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/boxplot-word-pairs.png" alt="Word Pair Detection Performance" /> <br/> <details> <summary>Bar plot</summary> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/barplot-word-pairs.png" alt="Word Pair Detection Performance" /> </details> <br/><br/> ### 4.3 Sentence detection <br/> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/boxplot-sentences.png" alt="Sentence Detection Performance" /> <br/> <details> <summary>Bar plot</summary> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/barplot-sentences.png" alt="Sentence Detection Performance" /> </details> <br/><br/> ### 4.4 Average detection <br/> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/boxplot-average.png" alt="Average Detection Performance" /> <br/> <details> <summary>Bar plot</summary> <img src="https://raw.githubusercontent.com/pemistahl/lingua-py/main/images/plots/barplot-average.png" alt="Average Detection Performance" /> </details> ## 5. Why is it better than other libraries? Every language detector uses a probabilistic [n-gram](https://en.wikipedia.org/wiki/N-gram) model trained on the character distribution in some training corpus. Most libraries only use n-grams of size 3 (trigrams) which is satisfactory for detecting the language of longer text fragments consisting of multiple sentences. For short phrases or single words, however, trigrams are not enough. The shorter the input text is, the less n-grams are available. The probabilities estimated from such few n-grams are not reliable. This is why *Lingua* makes use of n-grams of sizes 1 up to 5 which results in much more accurate prediction of the correct language. A second important difference is that *Lingua* does not only use such a statistical model, but also a rule-based engine. This engine first determines the alphabet of the input text and searches for characters which are unique in one or more languages. If exactly one language can be reliably chosen this way, the statistical model is not necessary anymore. In any case, the rule-based engine filters out languages that do not satisfy the conditions of the input text. Only then, in a second step, the probabilistic n-gram model is taken into consideration. This makes sense because loading less language models means less memory consumption and better runtime performance. In general, it is always a good idea to restrict the set of languages to be considered in the classification process using the respective api methods. If you know beforehand that certain languages are never to occur in an input text, do not let those take part in the classifcation process. The filtering mechanism of the rule-based engine is quite good, however, filtering based on your own knowledge of the input text is always preferable. ## 6. How to add it to your project? *Lingua* is available in the [Python Package Index](https://pypi.org/project/lingua-language-detector) and can be installed with: ```shell pip install lingua-language-detector ``` ## 7. How to use? ### 7.1 Basic usage ```python >>> from lingua import Language, LanguageDetectorBuilder >>> languages = [Language.ENGLISH, Language.FRENCH, Language.GERMAN, Language.SPANISH] >>> detector = LanguageDetectorBuilder.from_languages(*languages).build() >>> detector.detect_language_of("languages are awesome") Language.ENGLISH ``` ### 7.2 Minimum relative distance By default, *Lingua* returns the most likely language for a given input text. However, there are certain words that are spelled the same in more than one language. The word *prologue*, for instance, is both a valid English and French word. *Lingua* would output either English or French which might be wrong in the given context. For cases like that, it is possible to specify a minimum relative distance that the logarithmized and summed up probabilities for each possible language have to satisfy. It can be stated in the following way: ```python >>> from lingua import Language, LanguageDetectorBuilder >>> languages = [Language.ENGLISH, Language.FRENCH, Language.GERMAN, Language.SPANISH] >>> detector = LanguageDetectorBuilder.from_languages(*languages)\ .with_minimum_relative_distance(0.9)\ .build() >>> print(detector.detect_language_of("languages are awesome")) None ``` Be aware that the distance between the language probabilities is dependent on the length of the input text. The longer the input text, the larger the distance between the languages. So if you want to classify very short text phrases, do not set the minimum relative distance too high. Otherwise, `None` will be returned most of the time as in the example above. This is the return value for cases where language detection is not reliably possible. ### 7.3 Confidence values Knowing about the most likely language is nice but how reliable is the computed likelihood? And how less likely are the other examined languages in comparison to the most likely one? These questions can be answered as well: ```python >>> from lingua import Language, LanguageDetectorBuilder >>> languages = [Language.ENGLISH, Language.FRENCH, Language.GERMAN, Language.SPANISH] >>> detector = LanguageDetectorBuilder.from_languages(*languages).build() >>> confidence_values = detector.compute_language_confidence_values("languages are awesome") >>> for language, value in confidence_values: ... print(f"{language.name}: {value:.2f}") ENGLISH: 0.93 FRENCH: 0.04 GERMAN: 0.02 SPANISH: 0.01 ``` In the example above, a list is returned containing those languages which the calling instance of LanguageDetector has been built from, sorted by their confidence value in descending order. Each value is a probability between 0.0 and 1.0. The probabilities of all languages will sum to 1.0. If the language is unambiguously identified by the rule engine, the value 1.0 will always be returned for this language. The other languages will receive a value of 0.0. There is also a method for returning the confidence value for one specific language only: ```python >>> from lingua import Language, LanguageDetectorBuilder >>> languages = [Language.ENGLISH, Language.FRENCH, Language.GERMAN, Language.SPANISH] >>> detector = LanguageDetectorBuilder.from_languages(*languages).build() >>> confidence_value = detector.compute_language_confidence("languages are awesome", Language.FRENCH) >>> print(f"{confidence_value:.2f}") 0.04 ``` The value that this method computes is a number between 0.0 and 1.0. If the language is unambiguously identified by the rule engine, the value 1.0 will always be returned. If the given language is not supported by this detector instance, the value 0.0 will always be returned. ### 7.4 Eager loading versus lazy loading By default, *Lingua* uses lazy-loading to load only those language models on demand which are considered relevant by the rule-based filter engine. For web services, for instance, it is rather beneficial to preload all language models into memory to avoid unexpected latency while waiting for the service response. If you want to enable the eager-loading mode, you can do it like this: ```python LanguageDetectorBuilder.from_all_languages().with_preloaded_language_models().build() ``` Multiple instances of `LanguageDetector` share the same language models in memory which are accessed asynchronously by the instances. ### 7.5 Low accuracy mode versus high accuracy mode *Lingua's* high detection accuracy comes at the cost of being noticeably slower than other language detectors. The large language models also consume significant amounts of memory. These requirements might not be feasible for systems running low on resources. If you want to classify mostly long texts or need to save resources, you can enable a *low accuracy mode* that loads only a small subset of the language models into memory: ```python LanguageDetectorBuilder.from_all_languages().with_low_accuracy_mode().build() ``` The downside of this approach is that detection accuracy for short texts consisting of less than 120 characters will drop significantly. However, detection accuracy for texts which are longer than 120 characters will remain mostly unaffected. In high accuracy mode (the default), the language detector consumes approximately 800 MB of memory if all language models are loaded. In low accuracy mode, memory consumption is reduced to approximately 60 MB. An alternative for a smaller memory footprint and faster performance is to reduce the set of languages when building the language detector. In most cases, it is not advisable to build the detector from all supported languages. When you have knowledge about the texts you want to classify you can almost always rule out certain languages as impossible or unlikely to occur. ### 7.6 Detection of multiple languages in mixed-language texts In contrast to most other language detectors, *Lingua* is able to detect multiple languages in mixed-language texts. This feature can yield quite reasonable results but it is still in an experimental state and therefore the detection result is highly dependent on the input text. It works best in high-accuracy mode with multiple long words for each language. The shorter the phrases and their words are, the less accurate are the results. Reducing the set of languages when building the language detector can also improve accuracy for this task if the languages occurring in the text are equal to the languages supported by the respective language detector instance. ```python >>> from lingua import Language, LanguageDetectorBuilder >>> languages = [Language.ENGLISH, Language.FRENCH, Language.GERMAN] >>> detector = LanguageDetectorBuilder.from_languages(*languages).build() >>> sentence = "Parlez-vous français? " + \\ ... "Ich spreche Französisch nur ein bisschen. " + \\ ... "A little bit is better than nothing." >>> for result in detector.detect_multiple_languages_of(sentence): ... print(f"{result.language.name}: '{sentence[result.start_index:result.end_index]}'") FRENCH: 'Parlez-vous français? ' GERMAN: 'Ich spreche Französisch nur ein bisschen. ' ENGLISH: 'A little bit is better than nothing.' ``` In the example above, a list of [`DetectionResult`](https://github.com/pemistahl/lingua-py/blob/main/lingua/detector.py#L144) is returned. Each entry in the list describes a contiguous single-language text section, providing start and end indices of the respective substring. ### 7.7 Methods to build the LanguageDetector There might be classification tasks where you know beforehand that your language data is definitely not written in Latin, for instance. The detection accuracy can become better in such cases if you exclude certain languages from the decision process or just explicitly include relevant languages: ```python from lingua import LanguageDetectorBuilder, Language, IsoCode639_1, IsoCode639_3 # Include all languages available in the library. LanguageDetectorBuilder.from_all_languages() # Include only languages that are not yet extinct (= currently excludes Latin). LanguageDetectorBuilder.from_all_spoken_languages() # Include only languages written with Cyrillic script. LanguageDetectorBuilder.from_all_languages_with_cyrillic_script() # Exclude only the Spanish language from the decision algorithm. LanguageDetectorBuilder.from_all_languages_without(Language.SPANISH) # Only decide between English and German. LanguageDetectorBuilder.from_languages(Language.ENGLISH, Language.GERMAN) # Select languages by ISO 639-1 code. LanguageDetectorBuilder.from_iso_codes_639_1(IsoCode639_1.EN, IsoCode639_1.DE) # Select languages by ISO 639-3 code. LanguageDetectorBuilder.from_iso_codes_639_3(IsoCode639_3.ENG, IsoCode639_3.DEU) ``` """ __all__ = ( "ConfidenceValue", "DetectionResult", "LanguageDetectorBuilder", "LanguageDetector", "IsoCode639_1", "IsoCode639_3", "Language", "LanguageModelFilesWriter", "TestDataFilesWriter", ) from .builder import LanguageDetectorBuilder from .detector import ConfidenceValue, DetectionResult, LanguageDetector from .isocode import IsoCode639_1, IsoCode639_3 from .language import Language from .writer import LanguageModelFilesWriter, TestDataFilesWriter
18,069
37.122363
154
py
lingua-py
lingua-py-main/lingua/builder.py
# # Copyright © 2022-present Peter M. Stahl pemistahl@gmail.com # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either expressed or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import FrozenSet, Iterable from .detector import LanguageDetector from .isocode import IsoCode639_1, IsoCode639_3 from .language import Language _MISSING_LANGUAGE_MESSAGE: str = ( "LanguageDetector needs at least 2 languages to choose from" ) class LanguageDetectorBuilder: """This class configures and creates an instance of LanguageDetector.""" def __init__(self, languages: FrozenSet[Language]): self._languages = languages self._minimum_relative_distance = 0.0 self._is_every_language_model_preloaded = False self._is_low_accuracy_mode_enabled = False def __repr__(self): languages = sorted([language.name for language in self._languages]) return ( "LanguageDetectorBuilder(" f"_languages={languages}, " f"_minimum_relative_distance={self._minimum_relative_distance}, " f"_is_every_language_model_preloaded={self._is_every_language_model_preloaded}), " f"_is_low_accuracy_mode_enabled={self._is_low_accuracy_mode_enabled}" ) @classmethod def from_all_languages(cls) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in languages. """ return cls._from(Language.all()) @classmethod def from_all_spoken_languages(cls) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in spoken languages. """ return cls._from(Language.all_spoken_ones()) @classmethod def from_all_languages_with_arabic_script(cls) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in languages supporting the Arabic script. """ return cls._from(Language.all_with_arabic_script()) @classmethod def from_all_languages_with_cyrillic_script(cls) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in languages supporting the Cyrillic script. """ return cls._from(Language.all_with_cyrillic_script()) @classmethod def from_all_languages_with_devanagari_script(cls) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in languages supporting the Devanagari script. """ return cls._from(Language.all_with_devanagari_script()) @classmethod def from_all_languages_with_latin_script(cls) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in languages supporting the Latin script. """ return cls._from(Language.all_with_latin_script()) @classmethod def from_all_languages_without( cls, *languages: Language ) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with all built-in languages except those passed to this method. """ languages_to_load = Language.all().difference(languages) if len(languages_to_load) < 2: raise ValueError(_MISSING_LANGUAGE_MESSAGE) return cls._from(languages_to_load) @classmethod def from_languages(cls, *languages: Language) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with the languages passed to this method. """ if len(languages) < 2: raise ValueError(_MISSING_LANGUAGE_MESSAGE) return cls._from(languages) @classmethod def from_iso_codes_639_1( cls, *iso_codes: IsoCode639_1 ) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with the languages specified by the ISO 639-1 codes passed to this method. Raises: ValueError: if less than two ISO codes are specified """ if len(iso_codes) < 2: raise ValueError(_MISSING_LANGUAGE_MESSAGE) languages = set() for iso_code in iso_codes: language = Language.from_iso_code_639_1(iso_code) languages.add(language) return cls._from(languages) @classmethod def from_iso_codes_639_3( cls, *iso_codes: IsoCode639_3 ) -> "LanguageDetectorBuilder": """Create and return an instance of LanguageDetectorBuilder with the languages specified by the ISO 639-3 codes passed to this method. Raises: ValueError: if less than two ISO codes are specified """ if len(iso_codes) < 2: raise ValueError(_MISSING_LANGUAGE_MESSAGE) languages = set() for iso_code in iso_codes: language = Language.from_iso_code_639_3(iso_code) languages.add(language) return cls._from(languages) def with_minimum_relative_distance( self, distance: float ) -> "LanguageDetectorBuilder": """Set the desired value for the minimum relative distance measure. By default, Lingua returns the most likely language for a given input text. However, there are certain words that are spelled the same in more than one language. The word 'prologue', for instance, is both a valid English and French word. Lingua would output either English or French which might be wrong in the given context. For cases like that, it is possible to specify a minimum relative distance that the logarithmized and summed up probabilities for each possible language have to satisfy. Be aware that the distance between the language probabilities is dependent on the length of the input text. The longer the input text, the larger the distance between the languages. So if you want to classify very short text phrases, do not set the minimum relative distance too high. Otherwise you will get most results returned as None which is the return value for cases where language detection is not reliably possible. Raises: ValueError: if distance is smaller than 0.0 or greater than 0.99 """ if not 0 <= distance < 1: raise ValueError( "Minimum relative distance must lie in between 0.0 and 0.99" ) self._minimum_relative_distance = distance return self def with_preloaded_language_models(self) -> "LanguageDetectorBuilder": """Preload all language models when creating the LanguageDetector instance. By default, Lingua uses lazy-loading to load only those language models on demand which are considered relevant by the rule-based filter engine. For web services, for instance, it is rather beneficial to preload all language models into memory to avoid unexpected latency while waiting for the service response. This method allows to switch between these two loading modes. """ self._is_every_language_model_preloaded = True return self def with_low_accuracy_mode(self) -> "LanguageDetectorBuilder": """Disables the high accuracy mode in order to save memory and increase performance. By default, Lingua's high detection accuracy comes at the cost of loading large language models into memory which might not be feasible for systems running low on resources. This method disables the high accuracy mode so that only a small subset of language models is loaded into memory. The downside of this approach is that detection accuracy for short texts consisting of less than 120 characters will drop significantly. However, detection accuracy for texts which are longer than 120 characters will remain mostly unaffected. """ self._is_low_accuracy_mode_enabled = True return self def build(self) -> LanguageDetector: """Create and return the configured LanguageDetector instance.""" return LanguageDetector._from( self._languages, self._minimum_relative_distance, self._is_every_language_model_preloaded, self._is_low_accuracy_mode_enabled, ) @classmethod def _from(cls, languages: Iterable[Language]) -> "LanguageDetectorBuilder": if not isinstance(languages, frozenset): return LanguageDetectorBuilder(frozenset(languages)) return LanguageDetectorBuilder(languages)
9,264
40.177778
94
py
taming-transformers
taming-transformers-master/main.py
import argparse, os, sys, datetime, glob, importlib from omegaconf import OmegaConf import numpy as np from PIL import Image import torch import torchvision from torch.utils.data import random_split, DataLoader, Dataset import pytorch_lightning as pl from pytorch_lightning import seed_everything from pytorch_lightning.trainer import Trainer from pytorch_lightning.callbacks import ModelCheckpoint, Callback, LearningRateMonitor from pytorch_lightning.utilities import rank_zero_only from taming.data.utils import custom_collate def get_obj_from_str(string, reload=False): module, cls = string.rsplit(".", 1) if reload: module_imp = importlib.import_module(module) importlib.reload(module_imp) return getattr(importlib.import_module(module, package=None), cls) def get_parser(**parser_kwargs): def str2bool(v): if isinstance(v, bool): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("Boolean value expected.") parser = argparse.ArgumentParser(**parser_kwargs) parser.add_argument( "-n", "--name", type=str, const=True, default="", nargs="?", help="postfix for logdir", ) parser.add_argument( "-r", "--resume", type=str, const=True, default="", nargs="?", help="resume from logdir or checkpoint in logdir", ) parser.add_argument( "-b", "--base", nargs="*", metavar="base_config.yaml", help="paths to base configs. Loaded from left-to-right. " "Parameters can be overwritten or added with command-line options of the form `--key value`.", default=list(), ) parser.add_argument( "-t", "--train", type=str2bool, const=True, default=False, nargs="?", help="train", ) parser.add_argument( "--no-test", type=str2bool, const=True, default=False, nargs="?", help="disable test", ) parser.add_argument("-p", "--project", help="name of new or path to existing project") parser.add_argument( "-d", "--debug", type=str2bool, nargs="?", const=True, default=False, help="enable post-mortem debugging", ) parser.add_argument( "-s", "--seed", type=int, default=23, help="seed for seed_everything", ) parser.add_argument( "-f", "--postfix", type=str, default="", help="post-postfix for default name", ) return parser def nondefault_trainer_args(opt): parser = argparse.ArgumentParser() parser = Trainer.add_argparse_args(parser) args = parser.parse_args([]) return sorted(k for k in vars(args) if getattr(opt, k) != getattr(args, k)) def instantiate_from_config(config): if not "target" in config: raise KeyError("Expected key `target` to instantiate.") return get_obj_from_str(config["target"])(**config.get("params", dict())) class WrappedDataset(Dataset): """Wraps an arbitrary object with __len__ and __getitem__ into a pytorch dataset""" def __init__(self, dataset): self.data = dataset def __len__(self): return len(self.data) def __getitem__(self, idx): return self.data[idx] class DataModuleFromConfig(pl.LightningDataModule): def __init__(self, batch_size, train=None, validation=None, test=None, wrap=False, num_workers=None): super().__init__() self.batch_size = batch_size self.dataset_configs = dict() self.num_workers = num_workers if num_workers is not None else batch_size*2 if train is not None: self.dataset_configs["train"] = train self.train_dataloader = self._train_dataloader if validation is not None: self.dataset_configs["validation"] = validation self.val_dataloader = self._val_dataloader if test is not None: self.dataset_configs["test"] = test self.test_dataloader = self._test_dataloader self.wrap = wrap def prepare_data(self): for data_cfg in self.dataset_configs.values(): instantiate_from_config(data_cfg) def setup(self, stage=None): self.datasets = dict( (k, instantiate_from_config(self.dataset_configs[k])) for k in self.dataset_configs) if self.wrap: for k in self.datasets: self.datasets[k] = WrappedDataset(self.datasets[k]) def _train_dataloader(self): return DataLoader(self.datasets["train"], batch_size=self.batch_size, num_workers=self.num_workers, shuffle=True, collate_fn=custom_collate) def _val_dataloader(self): return DataLoader(self.datasets["validation"], batch_size=self.batch_size, num_workers=self.num_workers, collate_fn=custom_collate) def _test_dataloader(self): return DataLoader(self.datasets["test"], batch_size=self.batch_size, num_workers=self.num_workers, collate_fn=custom_collate) class SetupCallback(Callback): def __init__(self, resume, now, logdir, ckptdir, cfgdir, config, lightning_config): super().__init__() self.resume = resume self.now = now self.logdir = logdir self.ckptdir = ckptdir self.cfgdir = cfgdir self.config = config self.lightning_config = lightning_config def on_pretrain_routine_start(self, trainer, pl_module): if trainer.global_rank == 0: # Create logdirs and save configs os.makedirs(self.logdir, exist_ok=True) os.makedirs(self.ckptdir, exist_ok=True) os.makedirs(self.cfgdir, exist_ok=True) print("Project config") print(self.config.pretty()) OmegaConf.save(self.config, os.path.join(self.cfgdir, "{}-project.yaml".format(self.now))) print("Lightning config") print(self.lightning_config.pretty()) OmegaConf.save(OmegaConf.create({"lightning": self.lightning_config}), os.path.join(self.cfgdir, "{}-lightning.yaml".format(self.now))) else: # ModelCheckpoint callback created log directory --- remove it if not self.resume and os.path.exists(self.logdir): dst, name = os.path.split(self.logdir) dst = os.path.join(dst, "child_runs", name) os.makedirs(os.path.split(dst)[0], exist_ok=True) try: os.rename(self.logdir, dst) except FileNotFoundError: pass class ImageLogger(Callback): def __init__(self, batch_frequency, max_images, clamp=True, increase_log_steps=True): super().__init__() self.batch_freq = batch_frequency self.max_images = max_images self.logger_log_images = { pl.loggers.WandbLogger: self._wandb, pl.loggers.TestTubeLogger: self._testtube, } self.log_steps = [2 ** n for n in range(int(np.log2(self.batch_freq)) + 1)] if not increase_log_steps: self.log_steps = [self.batch_freq] self.clamp = clamp @rank_zero_only def _wandb(self, pl_module, images, batch_idx, split): raise ValueError("No way wandb") grids = dict() for k in images: grid = torchvision.utils.make_grid(images[k]) grids[f"{split}/{k}"] = wandb.Image(grid) pl_module.logger.experiment.log(grids) @rank_zero_only def _testtube(self, pl_module, images, batch_idx, split): for k in images: grid = torchvision.utils.make_grid(images[k]) grid = (grid+1.0)/2.0 # -1,1 -> 0,1; c,h,w tag = f"{split}/{k}" pl_module.logger.experiment.add_image( tag, grid, global_step=pl_module.global_step) @rank_zero_only def log_local(self, save_dir, split, images, global_step, current_epoch, batch_idx): root = os.path.join(save_dir, "images", split) for k in images: grid = torchvision.utils.make_grid(images[k], nrow=4) grid = (grid+1.0)/2.0 # -1,1 -> 0,1; c,h,w grid = grid.transpose(0,1).transpose(1,2).squeeze(-1) grid = grid.numpy() grid = (grid*255).astype(np.uint8) filename = "{}_gs-{:06}_e-{:06}_b-{:06}.png".format( k, global_step, current_epoch, batch_idx) path = os.path.join(root, filename) os.makedirs(os.path.split(path)[0], exist_ok=True) Image.fromarray(grid).save(path) def log_img(self, pl_module, batch, batch_idx, split="train"): if (self.check_frequency(batch_idx) and # batch_idx % self.batch_freq == 0 hasattr(pl_module, "log_images") and callable(pl_module.log_images) and self.max_images > 0): logger = type(pl_module.logger) is_train = pl_module.training if is_train: pl_module.eval() with torch.no_grad(): images = pl_module.log_images(batch, split=split, pl_module=pl_module) for k in images: N = min(images[k].shape[0], self.max_images) images[k] = images[k][:N] if isinstance(images[k], torch.Tensor): images[k] = images[k].detach().cpu() if self.clamp: images[k] = torch.clamp(images[k], -1., 1.) self.log_local(pl_module.logger.save_dir, split, images, pl_module.global_step, pl_module.current_epoch, batch_idx) logger_log_images = self.logger_log_images.get(logger, lambda *args, **kwargs: None) logger_log_images(pl_module, images, pl_module.global_step, split) if is_train: pl_module.train() def check_frequency(self, batch_idx): if (batch_idx % self.batch_freq) == 0 or (batch_idx in self.log_steps): try: self.log_steps.pop(0) except IndexError: pass return True return False def on_train_batch_end(self, trainer, pl_module, outputs, batch, batch_idx, dataloader_idx): self.log_img(pl_module, batch, batch_idx, split="train") def on_validation_batch_end(self, trainer, pl_module, outputs, batch, batch_idx, dataloader_idx): self.log_img(pl_module, batch, batch_idx, split="val") if __name__ == "__main__": # custom parser to specify config files, train, test and debug mode, # postfix, resume. # `--key value` arguments are interpreted as arguments to the trainer. # `nested.key=value` arguments are interpreted as config parameters. # configs are merged from left-to-right followed by command line parameters. # model: # base_learning_rate: float # target: path to lightning module # params: # key: value # data: # target: main.DataModuleFromConfig # params: # batch_size: int # wrap: bool # train: # target: path to train dataset # params: # key: value # validation: # target: path to validation dataset # params: # key: value # test: # target: path to test dataset # params: # key: value # lightning: (optional, has sane defaults and can be specified on cmdline) # trainer: # additional arguments to trainer # logger: # logger to instantiate # modelcheckpoint: # modelcheckpoint to instantiate # callbacks: # callback1: # target: importpath # params: # key: value now = datetime.datetime.now().strftime("%Y-%m-%dT%H-%M-%S") # add cwd for convenience and to make classes in this file available when # running as `python main.py` # (in particular `main.DataModuleFromConfig`) sys.path.append(os.getcwd()) parser = get_parser() parser = Trainer.add_argparse_args(parser) opt, unknown = parser.parse_known_args() if opt.name and opt.resume: raise ValueError( "-n/--name and -r/--resume cannot be specified both." "If you want to resume training in a new log folder, " "use -n/--name in combination with --resume_from_checkpoint" ) if opt.resume: if not os.path.exists(opt.resume): raise ValueError("Cannot find {}".format(opt.resume)) if os.path.isfile(opt.resume): paths = opt.resume.split("/") idx = len(paths)-paths[::-1].index("logs")+1 logdir = "/".join(paths[:idx]) ckpt = opt.resume else: assert os.path.isdir(opt.resume), opt.resume logdir = opt.resume.rstrip("/") ckpt = os.path.join(logdir, "checkpoints", "last.ckpt") opt.resume_from_checkpoint = ckpt base_configs = sorted(glob.glob(os.path.join(logdir, "configs/*.yaml"))) opt.base = base_configs+opt.base _tmp = logdir.split("/") nowname = _tmp[_tmp.index("logs")+1] else: if opt.name: name = "_"+opt.name elif opt.base: cfg_fname = os.path.split(opt.base[0])[-1] cfg_name = os.path.splitext(cfg_fname)[0] name = "_"+cfg_name else: name = "" nowname = now+name+opt.postfix logdir = os.path.join("logs", nowname) ckptdir = os.path.join(logdir, "checkpoints") cfgdir = os.path.join(logdir, "configs") seed_everything(opt.seed) try: # init and save configs configs = [OmegaConf.load(cfg) for cfg in opt.base] cli = OmegaConf.from_dotlist(unknown) config = OmegaConf.merge(*configs, cli) lightning_config = config.pop("lightning", OmegaConf.create()) # merge trainer cli with config trainer_config = lightning_config.get("trainer", OmegaConf.create()) # default to ddp trainer_config["distributed_backend"] = "ddp" for k in nondefault_trainer_args(opt): trainer_config[k] = getattr(opt, k) if not "gpus" in trainer_config: del trainer_config["distributed_backend"] cpu = True else: gpuinfo = trainer_config["gpus"] print(f"Running on GPUs {gpuinfo}") cpu = False trainer_opt = argparse.Namespace(**trainer_config) lightning_config.trainer = trainer_config # model model = instantiate_from_config(config.model) # trainer and callbacks trainer_kwargs = dict() # default logger configs # NOTE wandb < 0.10.0 interferes with shutdown # wandb >= 0.10.0 seems to fix it but still interferes with pudb # debugging (wrongly sized pudb ui) # thus prefer testtube for now default_logger_cfgs = { "wandb": { "target": "pytorch_lightning.loggers.WandbLogger", "params": { "name": nowname, "save_dir": logdir, "offline": opt.debug, "id": nowname, } }, "testtube": { "target": "pytorch_lightning.loggers.TestTubeLogger", "params": { "name": "testtube", "save_dir": logdir, } }, } default_logger_cfg = default_logger_cfgs["testtube"] logger_cfg = lightning_config.logger or OmegaConf.create() logger_cfg = OmegaConf.merge(default_logger_cfg, logger_cfg) trainer_kwargs["logger"] = instantiate_from_config(logger_cfg) # modelcheckpoint - use TrainResult/EvalResult(checkpoint_on=metric) to # specify which metric is used to determine best models default_modelckpt_cfg = { "target": "pytorch_lightning.callbacks.ModelCheckpoint", "params": { "dirpath": ckptdir, "filename": "{epoch:06}", "verbose": True, "save_last": True, } } if hasattr(model, "monitor"): print(f"Monitoring {model.monitor} as checkpoint metric.") default_modelckpt_cfg["params"]["monitor"] = model.monitor default_modelckpt_cfg["params"]["save_top_k"] = 3 modelckpt_cfg = lightning_config.modelcheckpoint or OmegaConf.create() modelckpt_cfg = OmegaConf.merge(default_modelckpt_cfg, modelckpt_cfg) trainer_kwargs["checkpoint_callback"] = instantiate_from_config(modelckpt_cfg) # add callback which sets up log directory default_callbacks_cfg = { "setup_callback": { "target": "main.SetupCallback", "params": { "resume": opt.resume, "now": now, "logdir": logdir, "ckptdir": ckptdir, "cfgdir": cfgdir, "config": config, "lightning_config": lightning_config, } }, "image_logger": { "target": "main.ImageLogger", "params": { "batch_frequency": 750, "max_images": 4, "clamp": True } }, "learning_rate_logger": { "target": "main.LearningRateMonitor", "params": { "logging_interval": "step", #"log_momentum": True } }, } callbacks_cfg = lightning_config.callbacks or OmegaConf.create() callbacks_cfg = OmegaConf.merge(default_callbacks_cfg, callbacks_cfg) trainer_kwargs["callbacks"] = [instantiate_from_config(callbacks_cfg[k]) for k in callbacks_cfg] trainer = Trainer.from_argparse_args(trainer_opt, **trainer_kwargs) # data data = instantiate_from_config(config.data) # NOTE according to https://pytorch-lightning.readthedocs.io/en/latest/datamodules.html # calling these ourselves should not be necessary but it is. # lightning still takes care of proper multiprocessing though data.prepare_data() data.setup() # configure learning rate bs, base_lr = config.data.params.batch_size, config.model.base_learning_rate if not cpu: ngpu = len(lightning_config.trainer.gpus.strip(",").split(',')) else: ngpu = 1 accumulate_grad_batches = lightning_config.trainer.accumulate_grad_batches or 1 print(f"accumulate_grad_batches = {accumulate_grad_batches}") lightning_config.trainer.accumulate_grad_batches = accumulate_grad_batches model.learning_rate = accumulate_grad_batches * ngpu * bs * base_lr print("Setting learning rate to {:.2e} = {} (accumulate_grad_batches) * {} (num_gpus) * {} (batchsize) * {:.2e} (base_lr)".format( model.learning_rate, accumulate_grad_batches, ngpu, bs, base_lr)) # allow checkpointing via USR1 def melk(*args, **kwargs): # run all checkpoint hooks if trainer.global_rank == 0: print("Summoning checkpoint.") ckpt_path = os.path.join(ckptdir, "last.ckpt") trainer.save_checkpoint(ckpt_path) def divein(*args, **kwargs): if trainer.global_rank == 0: import pudb; pudb.set_trace() import signal signal.signal(signal.SIGUSR1, melk) signal.signal(signal.SIGUSR2, divein) # run if opt.train: try: trainer.fit(model, data) except Exception: melk() raise if not opt.no_test and not trainer.interrupted: trainer.test(model, data) except Exception: if opt.debug and trainer.global_rank==0: try: import pudb as debugger except ImportError: import pdb as debugger debugger.post_mortem() raise finally: # move newly created debug project to debug_runs if opt.debug and not opt.resume and trainer.global_rank==0: dst, name = os.path.split(logdir) dst = os.path.join(dst, "debug_runs", name) os.makedirs(os.path.split(dst)[0], exist_ok=True) os.rename(logdir, dst)
21,251
35.266212
138
py
taming-transformers
taming-transformers-master/setup.py
from setuptools import setup, find_packages setup( name='taming-transformers', version='0.0.1', description='Taming Transformers for High-Resolution Image Synthesis', packages=find_packages(), install_requires=[ 'torch', 'numpy', 'tqdm', ], )
292
19.928571
74
py
taming-transformers
taming-transformers-master/scripts/extract_segmentation.py
import sys, os import numpy as np import scipy import torch import torch.nn as nn from scipy import ndimage from tqdm import tqdm, trange from PIL import Image import torch.hub import torchvision import torch.nn.functional as F # download deeplabv2_resnet101_msc-cocostuff164k-100000.pth from # https://github.com/kazuto1011/deeplab-pytorch/releases/download/v1.0/deeplabv2_resnet101_msc-cocostuff164k-100000.pth # and put the path here CKPT_PATH = "TODO" rescale = lambda x: (x + 1.) / 2. def rescale_bgr(x): x = (x+1)*127.5 x = torch.flip(x, dims=[0]) return x class COCOStuffSegmenter(nn.Module): def __init__(self, config): super().__init__() self.config = config self.n_labels = 182 model = torch.hub.load("kazuto1011/deeplab-pytorch", "deeplabv2_resnet101", n_classes=self.n_labels) ckpt_path = CKPT_PATH model.load_state_dict(torch.load(ckpt_path)) self.model = model normalize = torchvision.transforms.Normalize(mean=self.mean, std=self.std) self.image_transform = torchvision.transforms.Compose([ torchvision.transforms.Lambda(lambda image: torch.stack( [normalize(rescale_bgr(x)) for x in image])) ]) def forward(self, x, upsample=None): x = self._pre_process(x) x = self.model(x) if upsample is not None: x = torch.nn.functional.upsample_bilinear(x, size=upsample) return x def _pre_process(self, x): x = self.image_transform(x) return x @property def mean(self): # bgr return [104.008, 116.669, 122.675] @property def std(self): return [1.0, 1.0, 1.0] @property def input_size(self): return [3, 224, 224] def run_model(img, model): model = model.eval() with torch.no_grad(): segmentation = model(img, upsample=(img.shape[2], img.shape[3])) segmentation = torch.argmax(segmentation, dim=1, keepdim=True) return segmentation.detach().cpu() def get_input(batch, k): x = batch[k] if len(x.shape) == 3: x = x[..., None] x = x.permute(0, 3, 1, 2).to(memory_format=torch.contiguous_format) return x.float() def save_segmentation(segmentation, path): # --> class label to uint8, save as png os.makedirs(os.path.dirname(path), exist_ok=True) assert len(segmentation.shape)==4 assert segmentation.shape[0]==1 for seg in segmentation: seg = seg.permute(1,2,0).numpy().squeeze().astype(np.uint8) seg = Image.fromarray(seg) seg.save(path) def iterate_dataset(dataloader, destpath, model): os.makedirs(destpath, exist_ok=True) num_processed = 0 for i, batch in tqdm(enumerate(dataloader), desc="Data"): try: img = get_input(batch, "image") img = img.cuda() seg = run_model(img, model) path = batch["relative_file_path_"][0] path = os.path.splitext(path)[0] path = os.path.join(destpath, path + ".png") save_segmentation(seg, path) num_processed += 1 except Exception as e: print(e) print("but anyhow..") print("Processed {} files. Bye.".format(num_processed)) from taming.data.sflckr import Examples from torch.utils.data import DataLoader if __name__ == "__main__": dest = sys.argv[1] batchsize = 1 print("Running with batch-size {}, saving to {}...".format(batchsize, dest)) model = COCOStuffSegmenter({}).cuda() print("Instantiated model.") dataset = Examples() dloader = DataLoader(dataset, batch_size=batchsize) iterate_dataset(dataloader=dloader, destpath=dest, model=model) print("done.")
3,753
27.656489
119
py
taming-transformers
taming-transformers-master/scripts/sample_conditional.py
import argparse, os, sys, glob, math, time import torch import numpy as np from omegaconf import OmegaConf import streamlit as st from streamlit import caching from PIL import Image from main import instantiate_from_config, DataModuleFromConfig from torch.utils.data import DataLoader from torch.utils.data.dataloader import default_collate rescale = lambda x: (x + 1.) / 2. def bchw_to_st(x): return rescale(x.detach().cpu().numpy().transpose(0,2,3,1)) def save_img(xstart, fname): I = (xstart.clip(0,1)[0]*255).astype(np.uint8) Image.fromarray(I).save(fname) def get_interactive_image(resize=False): image = st.file_uploader("Input", type=["jpg", "JPEG", "png"]) if image is not None: image = Image.open(image) if not image.mode == "RGB": image = image.convert("RGB") image = np.array(image).astype(np.uint8) print("upload image shape: {}".format(image.shape)) img = Image.fromarray(image) if resize: img = img.resize((256, 256)) image = np.array(img) return image def single_image_to_torch(x, permute=True): assert x is not None, "Please provide an image through the upload function" x = np.array(x) x = torch.FloatTensor(x/255.*2. - 1.)[None,...] if permute: x = x.permute(0, 3, 1, 2) return x def pad_to_M(x, M): hp = math.ceil(x.shape[2]/M)*M-x.shape[2] wp = math.ceil(x.shape[3]/M)*M-x.shape[3] x = torch.nn.functional.pad(x, (0,wp,0,hp,0,0,0,0)) return x @torch.no_grad() def run_conditional(model, dsets): if len(dsets.datasets) > 1: split = st.sidebar.radio("Split", sorted(dsets.datasets.keys())) dset = dsets.datasets[split] else: dset = next(iter(dsets.datasets.values())) batch_size = 1 start_index = st.sidebar.number_input("Example Index (Size: {})".format(len(dset)), value=0, min_value=0, max_value=len(dset)-batch_size) indices = list(range(start_index, start_index+batch_size)) example = default_collate([dset[i] for i in indices]) x = model.get_input("image", example).to(model.device) cond_key = model.cond_stage_key c = model.get_input(cond_key, example).to(model.device) scale_factor = st.sidebar.slider("Scale Factor", min_value=0.5, max_value=4.0, step=0.25, value=1.00) if scale_factor != 1.0: x = torch.nn.functional.interpolate(x, scale_factor=scale_factor, mode="bicubic") c = torch.nn.functional.interpolate(c, scale_factor=scale_factor, mode="bicubic") quant_z, z_indices = model.encode_to_z(x) quant_c, c_indices = model.encode_to_c(c) cshape = quant_z.shape xrec = model.first_stage_model.decode(quant_z) st.write("image: {}".format(x.shape)) st.image(bchw_to_st(x), clamp=True, output_format="PNG") st.write("image reconstruction: {}".format(xrec.shape)) st.image(bchw_to_st(xrec), clamp=True, output_format="PNG") if cond_key == "segmentation": # get image from segmentation mask num_classes = c.shape[1] c = torch.argmax(c, dim=1, keepdim=True) c = torch.nn.functional.one_hot(c, num_classes=num_classes) c = c.squeeze(1).permute(0, 3, 1, 2).float() c = model.cond_stage_model.to_rgb(c) st.write(f"{cond_key}: {tuple(c.shape)}") st.image(bchw_to_st(c), clamp=True, output_format="PNG") idx = z_indices half_sample = st.sidebar.checkbox("Image Completion", value=False) if half_sample: start = idx.shape[1]//2 else: start = 0 idx[:,start:] = 0 idx = idx.reshape(cshape[0],cshape[2],cshape[3]) start_i = start//cshape[3] start_j = start %cshape[3] if not half_sample and quant_z.shape == quant_c.shape: st.info("Setting idx to c_indices") idx = c_indices.clone().reshape(cshape[0],cshape[2],cshape[3]) cidx = c_indices cidx = cidx.reshape(quant_c.shape[0],quant_c.shape[2],quant_c.shape[3]) xstart = model.decode_to_img(idx[:,:cshape[2],:cshape[3]], cshape) st.image(bchw_to_st(xstart), clamp=True, output_format="PNG") temperature = st.number_input("Temperature", value=1.0) top_k = st.number_input("Top k", value=100) sample = st.checkbox("Sample", value=True) update_every = st.number_input("Update every", value=75) st.text(f"Sampling shape ({cshape[2]},{cshape[3]})") animate = st.checkbox("animate") if animate: import imageio outvid = "sampling.mp4" writer = imageio.get_writer(outvid, fps=25) elapsed_t = st.empty() info = st.empty() st.text("Sampled") if st.button("Sample"): output = st.empty() start_t = time.time() for i in range(start_i,cshape[2]-0): if i <= 8: local_i = i elif cshape[2]-i < 8: local_i = 16-(cshape[2]-i) else: local_i = 8 for j in range(start_j,cshape[3]-0): if j <= 8: local_j = j elif cshape[3]-j < 8: local_j = 16-(cshape[3]-j) else: local_j = 8 i_start = i-local_i i_end = i_start+16 j_start = j-local_j j_end = j_start+16 elapsed_t.text(f"Time: {time.time() - start_t} seconds") info.text(f"Step: ({i},{j}) | Local: ({local_i},{local_j}) | Crop: ({i_start}:{i_end},{j_start}:{j_end})") patch = idx[:,i_start:i_end,j_start:j_end] patch = patch.reshape(patch.shape[0],-1) cpatch = cidx[:, i_start:i_end, j_start:j_end] cpatch = cpatch.reshape(cpatch.shape[0], -1) patch = torch.cat((cpatch, patch), dim=1) logits,_ = model.transformer(patch[:,:-1]) logits = logits[:, -256:, :] logits = logits.reshape(cshape[0],16,16,-1) logits = logits[:,local_i,local_j,:] logits = logits/temperature if top_k is not None: logits = model.top_k_logits(logits, top_k) # apply softmax to convert to probabilities probs = torch.nn.functional.softmax(logits, dim=-1) # sample from the distribution or take the most likely if sample: ix = torch.multinomial(probs, num_samples=1) else: _, ix = torch.topk(probs, k=1, dim=-1) idx[:,i,j] = ix if (i*cshape[3]+j)%update_every==0: xstart = model.decode_to_img(idx[:, :cshape[2], :cshape[3]], cshape,) xstart = bchw_to_st(xstart) output.image(xstart, clamp=True, output_format="PNG") if animate: writer.append_data((xstart[0]*255).clip(0, 255).astype(np.uint8)) xstart = model.decode_to_img(idx[:,:cshape[2],:cshape[3]], cshape) xstart = bchw_to_st(xstart) output.image(xstart, clamp=True, output_format="PNG") #save_img(xstart, "full_res_sample.png") if animate: writer.close() st.video(outvid) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument( "-r", "--resume", type=str, nargs="?", help="load from logdir or checkpoint in logdir", ) parser.add_argument( "-b", "--base", nargs="*", metavar="base_config.yaml", help="paths to base configs. Loaded from left-to-right. " "Parameters can be overwritten or added with command-line options of the form `--key value`.", default=list(), ) parser.add_argument( "-c", "--config", nargs="?", metavar="single_config.yaml", help="path to single config. If specified, base configs will be ignored " "(except for the last one if left unspecified).", const=True, default="", ) parser.add_argument( "--ignore_base_data", action="store_true", help="Ignore data specification from base configs. Useful if you want " "to specify a custom datasets on the command line.", ) return parser def load_model_from_config(config, sd, gpu=True, eval_mode=True): if "ckpt_path" in config.params: st.warning("Deleting the restore-ckpt path from the config...") config.params.ckpt_path = None if "downsample_cond_size" in config.params: st.warning("Deleting downsample-cond-size from the config and setting factor=0.5 instead...") config.params.downsample_cond_size = -1 config.params["downsample_cond_factor"] = 0.5 try: if "ckpt_path" in config.params.first_stage_config.params: config.params.first_stage_config.params.ckpt_path = None st.warning("Deleting the first-stage restore-ckpt path from the config...") if "ckpt_path" in config.params.cond_stage_config.params: config.params.cond_stage_config.params.ckpt_path = None st.warning("Deleting the cond-stage restore-ckpt path from the config...") except: pass model = instantiate_from_config(config) if sd is not None: missing, unexpected = model.load_state_dict(sd, strict=False) st.info(f"Missing Keys in State Dict: {missing}") st.info(f"Unexpected Keys in State Dict: {unexpected}") if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def get_data(config): # get data data = instantiate_from_config(config.data) data.prepare_data() data.setup() return data @st.cache(allow_output_mutation=True, suppress_st_warning=True) def load_model_and_dset(config, ckpt, gpu, eval_mode): # get data dsets = get_data(config) # calls data.config ... # now load the specified checkpoint if ckpt: pl_sd = torch.load(ckpt, map_location="cpu") global_step = pl_sd["global_step"] else: pl_sd = {"state_dict": None} global_step = None model = load_model_from_config(config.model, pl_sd["state_dict"], gpu=gpu, eval_mode=eval_mode)["model"] return dsets, model, global_step if __name__ == "__main__": sys.path.append(os.getcwd()) parser = get_parser() opt, unknown = parser.parse_known_args() ckpt = None if opt.resume: if not os.path.exists(opt.resume): raise ValueError("Cannot find {}".format(opt.resume)) if os.path.isfile(opt.resume): paths = opt.resume.split("/") try: idx = len(paths)-paths[::-1].index("logs")+1 except ValueError: idx = -2 # take a guess: path/to/logdir/checkpoints/model.ckpt logdir = "/".join(paths[:idx]) ckpt = opt.resume else: assert os.path.isdir(opt.resume), opt.resume logdir = opt.resume.rstrip("/") ckpt = os.path.join(logdir, "checkpoints", "last.ckpt") print(f"logdir:{logdir}") base_configs = sorted(glob.glob(os.path.join(logdir, "configs/*-project.yaml"))) opt.base = base_configs+opt.base if opt.config: if type(opt.config) == str: opt.base = [opt.config] else: opt.base = [opt.base[-1]] configs = [OmegaConf.load(cfg) for cfg in opt.base] cli = OmegaConf.from_dotlist(unknown) if opt.ignore_base_data: for config in configs: if hasattr(config, "data"): del config["data"] config = OmegaConf.merge(*configs, cli) st.sidebar.text(ckpt) gs = st.sidebar.empty() gs.text(f"Global step: ?") st.sidebar.text("Options") #gpu = st.sidebar.checkbox("GPU", value=True) gpu = True #eval_mode = st.sidebar.checkbox("Eval Mode", value=True) eval_mode = True #show_config = st.sidebar.checkbox("Show Config", value=False) show_config = False if show_config: st.info("Checkpoint: {}".format(ckpt)) st.json(OmegaConf.to_container(config)) dsets, model, global_step = load_model_and_dset(config, ckpt, gpu, eval_mode) gs.text(f"Global step: {global_step}") run_conditional(model, dsets)
12,535
34.213483
122
py
taming-transformers
taming-transformers-master/scripts/sample_fast.py
import argparse, os, sys, glob import torch import time import numpy as np from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from einops import repeat from main import instantiate_from_config from taming.modules.transformer.mingpt import sample_with_past rescale = lambda x: (x + 1.) / 2. def chw_to_pillow(x): return Image.fromarray((255*rescale(x.detach().cpu().numpy().transpose(1,2,0))).clip(0,255).astype(np.uint8)) @torch.no_grad() def sample_classconditional(model, batch_size, class_label, steps=256, temperature=None, top_k=None, callback=None, dim_z=256, h=16, w=16, verbose_time=False, top_p=None): log = dict() assert type(class_label) == int, f'expecting type int but type is {type(class_label)}' qzshape = [batch_size, dim_z, h, w] assert not model.be_unconditional, 'Expecting a class-conditional Net2NetTransformer.' c_indices = repeat(torch.tensor([class_label]), '1 -> b 1', b=batch_size).to(model.device) # class token t1 = time.time() index_sample = sample_with_past(c_indices, model.transformer, steps=steps, sample_logits=True, top_k=top_k, callback=callback, temperature=temperature, top_p=top_p) if verbose_time: sampling_time = time.time() - t1 print(f"Full sampling takes about {sampling_time:.2f} seconds.") x_sample = model.decode_to_img(index_sample, qzshape) log["samples"] = x_sample log["class_label"] = c_indices return log @torch.no_grad() def sample_unconditional(model, batch_size, steps=256, temperature=None, top_k=None, top_p=None, callback=None, dim_z=256, h=16, w=16, verbose_time=False): log = dict() qzshape = [batch_size, dim_z, h, w] assert model.be_unconditional, 'Expecting an unconditional model.' c_indices = repeat(torch.tensor([model.sos_token]), '1 -> b 1', b=batch_size).to(model.device) # sos token t1 = time.time() index_sample = sample_with_past(c_indices, model.transformer, steps=steps, sample_logits=True, top_k=top_k, callback=callback, temperature=temperature, top_p=top_p) if verbose_time: sampling_time = time.time() - t1 print(f"Full sampling takes about {sampling_time:.2f} seconds.") x_sample = model.decode_to_img(index_sample, qzshape) log["samples"] = x_sample return log @torch.no_grad() def run(logdir, model, batch_size, temperature, top_k, unconditional=True, num_samples=50000, given_classes=None, top_p=None): batches = [batch_size for _ in range(num_samples//batch_size)] + [num_samples % batch_size] if not unconditional: assert given_classes is not None print("Running in pure class-conditional sampling mode. I will produce " f"{num_samples} samples for each of the {len(given_classes)} classes, " f"i.e. {num_samples*len(given_classes)} in total.") for class_label in tqdm(given_classes, desc="Classes"): for n, bs in tqdm(enumerate(batches), desc="Sampling Class"): if bs == 0: break logs = sample_classconditional(model, batch_size=bs, class_label=class_label, temperature=temperature, top_k=top_k, top_p=top_p) save_from_logs(logs, logdir, base_count=n * batch_size, cond_key=logs["class_label"]) else: print(f"Running in unconditional sampling mode, producing {num_samples} samples.") for n, bs in tqdm(enumerate(batches), desc="Sampling"): if bs == 0: break logs = sample_unconditional(model, batch_size=bs, temperature=temperature, top_k=top_k, top_p=top_p) save_from_logs(logs, logdir, base_count=n * batch_size) def save_from_logs(logs, logdir, base_count, key="samples", cond_key=None): xx = logs[key] for i, x in enumerate(xx): x = chw_to_pillow(x) count = base_count + i if cond_key is None: x.save(os.path.join(logdir, f"{count:06}.png")) else: condlabel = cond_key[i] if type(condlabel) == torch.Tensor: condlabel = condlabel.item() os.makedirs(os.path.join(logdir, str(condlabel)), exist_ok=True) x.save(os.path.join(logdir, str(condlabel), f"{count:06}.png")) def get_parser(): def str2bool(v): if isinstance(v, bool): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("Boolean value expected.") parser = argparse.ArgumentParser() parser.add_argument( "-r", "--resume", type=str, nargs="?", help="load from logdir or checkpoint in logdir", ) parser.add_argument( "-o", "--outdir", type=str, nargs="?", help="path where the samples will be logged to.", default="" ) parser.add_argument( "-b", "--base", nargs="*", metavar="base_config.yaml", help="paths to base configs. Loaded from left-to-right. " "Parameters can be overwritten or added with command-line options of the form `--key value`.", default=list(), ) parser.add_argument( "-n", "--num_samples", type=int, nargs="?", help="num_samples to draw", default=50000 ) parser.add_argument( "--batch_size", type=int, nargs="?", help="the batch size", default=25 ) parser.add_argument( "-k", "--top_k", type=int, nargs="?", help="top-k value to sample with", default=250, ) parser.add_argument( "-t", "--temperature", type=float, nargs="?", help="temperature value to sample with", default=1.0 ) parser.add_argument( "-p", "--top_p", type=float, nargs="?", help="top-p value to sample with", default=1.0 ) parser.add_argument( "--classes", type=str, nargs="?", help="specify comma-separated classes to sample from. Uses 1000 classes per default.", default="imagenet" ) return parser def load_model_from_config(config, sd, gpu=True, eval_mode=True): model = instantiate_from_config(config) if sd is not None: model.load_state_dict(sd) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def load_model(config, ckpt, gpu, eval_mode): # load the specified checkpoint if ckpt: pl_sd = torch.load(ckpt, map_location="cpu") global_step = pl_sd["global_step"] print(f"loaded model from global step {global_step}.") else: pl_sd = {"state_dict": None} global_step = None model = load_model_from_config(config.model, pl_sd["state_dict"], gpu=gpu, eval_mode=eval_mode)["model"] return model, global_step if __name__ == "__main__": sys.path.append(os.getcwd()) parser = get_parser() opt, unknown = parser.parse_known_args() assert opt.resume ckpt = None if not os.path.exists(opt.resume): raise ValueError("Cannot find {}".format(opt.resume)) if os.path.isfile(opt.resume): paths = opt.resume.split("/") try: idx = len(paths)-paths[::-1].index("logs")+1 except ValueError: idx = -2 # take a guess: path/to/logdir/checkpoints/model.ckpt logdir = "/".join(paths[:idx]) ckpt = opt.resume else: assert os.path.isdir(opt.resume), opt.resume logdir = opt.resume.rstrip("/") ckpt = os.path.join(logdir, "checkpoints", "last.ckpt") base_configs = sorted(glob.glob(os.path.join(logdir, "configs/*-project.yaml"))) opt.base = base_configs+opt.base configs = [OmegaConf.load(cfg) for cfg in opt.base] cli = OmegaConf.from_dotlist(unknown) config = OmegaConf.merge(*configs, cli) model, global_step = load_model(config, ckpt, gpu=True, eval_mode=True) if opt.outdir: print(f"Switching logdir from '{logdir}' to '{opt.outdir}'") logdir = opt.outdir if opt.classes == "imagenet": given_classes = [i for i in range(1000)] else: cls_str = opt.classes assert not cls_str.endswith(","), 'class string should not end with a ","' given_classes = [int(c) for c in cls_str.split(",")] logdir = os.path.join(logdir, "samples", f"top_k_{opt.top_k}_temp_{opt.temperature:.2f}_top_p_{opt.top_p}", f"{global_step}") print(f"Logging to {logdir}") os.makedirs(logdir, exist_ok=True) run(logdir, model, opt.batch_size, opt.temperature, opt.top_k, unconditional=model.be_unconditional, given_classes=given_classes, num_samples=opt.num_samples, top_p=opt.top_p) print("done.")
9,191
34.218391
115
py
taming-transformers
taming-transformers-master/scripts/make_scene_samples.py
import glob import os import sys from itertools import product from pathlib import Path from typing import Literal, List, Optional, Tuple import numpy as np import torch from omegaconf import OmegaConf from pytorch_lightning import seed_everything from torch import Tensor from torchvision.utils import save_image from tqdm import tqdm from scripts.make_samples import get_parser, load_model_and_dset from taming.data.conditional_builder.objects_center_points import ObjectsCenterPointsConditionalBuilder from taming.data.helper_types import BoundingBox, Annotation from taming.data.annotated_objects_dataset import AnnotatedObjectsDataset from taming.models.cond_transformer import Net2NetTransformer seed_everything(42424242) device: Literal['cuda', 'cpu'] = 'cuda' first_stage_factor = 16 trained_on_res = 256 def _helper(coord: int, coord_max: int, coord_window: int) -> (int, int): assert 0 <= coord < coord_max coord_desired_center = (coord_window - 1) // 2 return np.clip(coord - coord_desired_center, 0, coord_max - coord_window) def get_crop_coordinates(x: int, y: int) -> BoundingBox: WIDTH, HEIGHT = desired_z_shape[1], desired_z_shape[0] x0 = _helper(x, WIDTH, first_stage_factor) / WIDTH y0 = _helper(y, HEIGHT, first_stage_factor) / HEIGHT w = first_stage_factor / WIDTH h = first_stage_factor / HEIGHT return x0, y0, w, h def get_z_indices_crop_out(z_indices: Tensor, predict_x: int, predict_y: int) -> Tensor: WIDTH, HEIGHT = desired_z_shape[1], desired_z_shape[0] x0 = _helper(predict_x, WIDTH, first_stage_factor) y0 = _helper(predict_y, HEIGHT, first_stage_factor) no_images = z_indices.shape[0] cut_out_1 = z_indices[:, y0:predict_y, x0:x0+first_stage_factor].reshape((no_images, -1)) cut_out_2 = z_indices[:, predict_y, x0:predict_x] return torch.cat((cut_out_1, cut_out_2), dim=1) @torch.no_grad() def sample(model: Net2NetTransformer, annotations: List[Annotation], dataset: AnnotatedObjectsDataset, conditional_builder: ObjectsCenterPointsConditionalBuilder, no_samples: int, temperature: float, top_k: int) -> Tensor: x_max, y_max = desired_z_shape[1], desired_z_shape[0] annotations = [a._replace(category_no=dataset.get_category_number(a.category_id)) for a in annotations] recompute_conditional = any((desired_resolution[0] > trained_on_res, desired_resolution[1] > trained_on_res)) if not recompute_conditional: crop_coordinates = get_crop_coordinates(0, 0) conditional_indices = conditional_builder.build(annotations, crop_coordinates) c_indices = conditional_indices.to(device).repeat(no_samples, 1) z_indices = torch.zeros((no_samples, 0), device=device).long() output_indices = model.sample(z_indices, c_indices, steps=x_max*y_max, temperature=temperature, sample=True, top_k=top_k) else: output_indices = torch.zeros((no_samples, y_max, x_max), device=device).long() for predict_y, predict_x in tqdm(product(range(y_max), range(x_max)), desc='sampling_image', total=x_max*y_max): crop_coordinates = get_crop_coordinates(predict_x, predict_y) z_indices = get_z_indices_crop_out(output_indices, predict_x, predict_y) conditional_indices = conditional_builder.build(annotations, crop_coordinates) c_indices = conditional_indices.to(device).repeat(no_samples, 1) new_index = model.sample(z_indices, c_indices, steps=1, temperature=temperature, sample=True, top_k=top_k) output_indices[:, predict_y, predict_x] = new_index[:, -1] z_shape = ( no_samples, model.first_stage_model.quantize.e_dim, # codebook embed_dim desired_z_shape[0], # z_height desired_z_shape[1] # z_width ) x_sample = model.decode_to_img(output_indices, z_shape) * 0.5 + 0.5 x_sample = x_sample.to('cpu') plotter = conditional_builder.plot figure_size = (x_sample.shape[2], x_sample.shape[3]) scene_graph = conditional_builder.build(annotations, (0., 0., 1., 1.)) plot = plotter(scene_graph, dataset.get_textual_label_for_category_no, figure_size) return torch.cat((x_sample, plot.unsqueeze(0))) def get_resolution(resolution_str: str) -> (Tuple[int, int], Tuple[int, int]): if not resolution_str.count(',') == 1: raise ValueError("Give resolution as in 'height,width'") res_h, res_w = resolution_str.split(',') res_h = max(int(res_h), trained_on_res) res_w = max(int(res_w), trained_on_res) z_h = int(round(res_h/first_stage_factor)) z_w = int(round(res_w/first_stage_factor)) return (z_h, z_w), (z_h*first_stage_factor, z_w*first_stage_factor) def add_arg_to_parser(parser): parser.add_argument( "-R", "--resolution", type=str, default='256,256', help=f"give resolution in multiples of {first_stage_factor}, default is '256,256'", ) parser.add_argument( "-C", "--conditional", type=str, default='objects_bbox', help=f"objects_bbox or objects_center_points", ) parser.add_argument( "-N", "--n_samples_per_layout", type=int, default=4, help=f"how many samples to generate per layout", ) return parser if __name__ == "__main__": sys.path.append(os.getcwd()) parser = get_parser() parser = add_arg_to_parser(parser) opt, unknown = parser.parse_known_args() ckpt = None if opt.resume: if not os.path.exists(opt.resume): raise ValueError("Cannot find {}".format(opt.resume)) if os.path.isfile(opt.resume): paths = opt.resume.split("/") try: idx = len(paths)-paths[::-1].index("logs")+1 except ValueError: idx = -2 # take a guess: path/to/logdir/checkpoints/model.ckpt logdir = "/".join(paths[:idx]) ckpt = opt.resume else: assert os.path.isdir(opt.resume), opt.resume logdir = opt.resume.rstrip("/") ckpt = os.path.join(logdir, "checkpoints", "last.ckpt") print(f"logdir:{logdir}") base_configs = sorted(glob.glob(os.path.join(logdir, "configs/*-project.yaml"))) opt.base = base_configs+opt.base if opt.config: if type(opt.config) == str: opt.base = [opt.config] else: opt.base = [opt.base[-1]] configs = [OmegaConf.load(cfg) for cfg in opt.base] cli = OmegaConf.from_dotlist(unknown) if opt.ignore_base_data: for config in configs: if hasattr(config, "data"): del config["data"] config = OmegaConf.merge(*configs, cli) desired_z_shape, desired_resolution = get_resolution(opt.resolution) conditional = opt.conditional print(ckpt) gpu = True eval_mode = True show_config = False if show_config: print(OmegaConf.to_container(config)) dsets, model, global_step = load_model_and_dset(config, ckpt, gpu, eval_mode) print(f"Global step: {global_step}") data_loader = dsets.val_dataloader() print(dsets.datasets["validation"].conditional_builders) conditional_builder = dsets.datasets["validation"].conditional_builders[conditional] outdir = Path(opt.outdir).joinpath(f"{global_step:06}_{opt.top_k}_{opt.temperature}") outdir.mkdir(exist_ok=True, parents=True) print("Writing samples to ", outdir) p_bar_1 = tqdm(enumerate(iter(data_loader)), desc='batch', total=len(data_loader)) for batch_no, batch in p_bar_1: save_img: Optional[Tensor] = None for i, annotations in tqdm(enumerate(batch['annotations']), desc='within_batch', total=data_loader.batch_size): imgs = sample(model, annotations, dsets.datasets["validation"], conditional_builder, opt.n_samples_per_layout, opt.temperature, opt.top_k) save_image(imgs, outdir.joinpath(f'{batch_no:04}_{i:02}.png'), n_row=opt.n_samples_per_layout+1)
8,092
39.668342
120
py
taming-transformers
taming-transformers-master/scripts/extract_depth.py
import os import torch import numpy as np from tqdm import trange from PIL import Image def get_state(gpu): import torch midas = torch.hub.load("intel-isl/MiDaS", "MiDaS") if gpu: midas.cuda() midas.eval() midas_transforms = torch.hub.load("intel-isl/MiDaS", "transforms") transform = midas_transforms.default_transform state = {"model": midas, "transform": transform} return state def depth_to_rgba(x): assert x.dtype == np.float32 assert len(x.shape) == 2 y = x.copy() y.dtype = np.uint8 y = y.reshape(x.shape+(4,)) return np.ascontiguousarray(y) def rgba_to_depth(x): assert x.dtype == np.uint8 assert len(x.shape) == 3 and x.shape[2] == 4 y = x.copy() y.dtype = np.float32 y = y.reshape(x.shape[:2]) return np.ascontiguousarray(y) def run(x, state): model = state["model"] transform = state["transform"] hw = x.shape[:2] with torch.no_grad(): prediction = model(transform((x + 1.0) * 127.5).cuda()) prediction = torch.nn.functional.interpolate( prediction.unsqueeze(1), size=hw, mode="bicubic", align_corners=False, ).squeeze() output = prediction.cpu().numpy() return output def get_filename(relpath, level=-2): # save class folder structure and filename: fn = relpath.split(os.sep)[level:] folder = fn[-2] file = fn[-1].split('.')[0] return folder, file def save_depth(dataset, path, debug=False): os.makedirs(path) N = len(dset) if debug: N = 10 state = get_state(gpu=True) for idx in trange(N, desc="Data"): ex = dataset[idx] image, relpath = ex["image"], ex["relpath"] folder, filename = get_filename(relpath) # prepare folderabspath = os.path.join(path, folder) os.makedirs(folderabspath, exist_ok=True) savepath = os.path.join(folderabspath, filename) # run model xout = run(image, state) I = depth_to_rgba(xout) Image.fromarray(I).save("{}.png".format(savepath)) if __name__ == "__main__": from taming.data.imagenet import ImageNetTrain, ImageNetValidation out = "data/imagenet_depth" if not os.path.exists(out): print("Please create a folder or symlink '{}' to extract depth data ".format(out) + "(be prepared that the output size will be larger than ImageNet itself).") exit(1) # go dset = ImageNetValidation() abspath = os.path.join(out, "val") if os.path.exists(abspath): print("{} exists - not doing anything.".format(abspath)) else: print("preparing {}".format(abspath)) save_depth(dset, abspath) print("done with validation split") dset = ImageNetTrain() abspath = os.path.join(out, "train") if os.path.exists(abspath): print("{} exists - not doing anything.".format(abspath)) else: print("preparing {}".format(abspath)) save_depth(dset, abspath) print("done with train split") print("done done.")
3,121
26.628319
91
py
taming-transformers
taming-transformers-master/scripts/extract_submodel.py
import torch import sys if __name__ == "__main__": inpath = sys.argv[1] outpath = sys.argv[2] submodel = "cond_stage_model" if len(sys.argv) > 3: submodel = sys.argv[3] print("Extracting {} from {} to {}.".format(submodel, inpath, outpath)) sd = torch.load(inpath, map_location="cpu") new_sd = {"state_dict": dict((k.split(".", 1)[-1],v) for k,v in sd["state_dict"].items() if k.startswith("cond_stage_model"))} torch.save(new_sd, outpath)
549
29.555556
75
py
taming-transformers
taming-transformers-master/scripts/make_samples.py
import argparse, os, sys, glob, math, time import torch import numpy as np from omegaconf import OmegaConf from PIL import Image from main import instantiate_from_config, DataModuleFromConfig from torch.utils.data import DataLoader from torch.utils.data.dataloader import default_collate from tqdm import trange def save_image(x, path): c,h,w = x.shape assert c==3 x = ((x.detach().cpu().numpy().transpose(1,2,0)+1.0)*127.5).clip(0,255).astype(np.uint8) Image.fromarray(x).save(path) @torch.no_grad() def run_conditional(model, dsets, outdir, top_k, temperature, batch_size=1): if len(dsets.datasets) > 1: split = sorted(dsets.datasets.keys())[0] dset = dsets.datasets[split] else: dset = next(iter(dsets.datasets.values())) print("Dataset: ", dset.__class__.__name__) for start_idx in trange(0,len(dset)-batch_size+1,batch_size): indices = list(range(start_idx, start_idx+batch_size)) example = default_collate([dset[i] for i in indices]) x = model.get_input("image", example).to(model.device) for i in range(x.shape[0]): save_image(x[i], os.path.join(outdir, "originals", "{:06}.png".format(indices[i]))) cond_key = model.cond_stage_key c = model.get_input(cond_key, example).to(model.device) scale_factor = 1.0 quant_z, z_indices = model.encode_to_z(x) quant_c, c_indices = model.encode_to_c(c) cshape = quant_z.shape xrec = model.first_stage_model.decode(quant_z) for i in range(xrec.shape[0]): save_image(xrec[i], os.path.join(outdir, "reconstructions", "{:06}.png".format(indices[i]))) if cond_key == "segmentation": # get image from segmentation mask num_classes = c.shape[1] c = torch.argmax(c, dim=1, keepdim=True) c = torch.nn.functional.one_hot(c, num_classes=num_classes) c = c.squeeze(1).permute(0, 3, 1, 2).float() c = model.cond_stage_model.to_rgb(c) idx = z_indices half_sample = False if half_sample: start = idx.shape[1]//2 else: start = 0 idx[:,start:] = 0 idx = idx.reshape(cshape[0],cshape[2],cshape[3]) start_i = start//cshape[3] start_j = start %cshape[3] cidx = c_indices cidx = cidx.reshape(quant_c.shape[0],quant_c.shape[2],quant_c.shape[3]) sample = True for i in range(start_i,cshape[2]-0): if i <= 8: local_i = i elif cshape[2]-i < 8: local_i = 16-(cshape[2]-i) else: local_i = 8 for j in range(start_j,cshape[3]-0): if j <= 8: local_j = j elif cshape[3]-j < 8: local_j = 16-(cshape[3]-j) else: local_j = 8 i_start = i-local_i i_end = i_start+16 j_start = j-local_j j_end = j_start+16 patch = idx[:,i_start:i_end,j_start:j_end] patch = patch.reshape(patch.shape[0],-1) cpatch = cidx[:, i_start:i_end, j_start:j_end] cpatch = cpatch.reshape(cpatch.shape[0], -1) patch = torch.cat((cpatch, patch), dim=1) logits,_ = model.transformer(patch[:,:-1]) logits = logits[:, -256:, :] logits = logits.reshape(cshape[0],16,16,-1) logits = logits[:,local_i,local_j,:] logits = logits/temperature if top_k is not None: logits = model.top_k_logits(logits, top_k) # apply softmax to convert to probabilities probs = torch.nn.functional.softmax(logits, dim=-1) # sample from the distribution or take the most likely if sample: ix = torch.multinomial(probs, num_samples=1) else: _, ix = torch.topk(probs, k=1, dim=-1) idx[:,i,j] = ix xsample = model.decode_to_img(idx[:,:cshape[2],:cshape[3]], cshape) for i in range(xsample.shape[0]): save_image(xsample[i], os.path.join(outdir, "samples", "{:06}.png".format(indices[i]))) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument( "-r", "--resume", type=str, nargs="?", help="load from logdir or checkpoint in logdir", ) parser.add_argument( "-b", "--base", nargs="*", metavar="base_config.yaml", help="paths to base configs. Loaded from left-to-right. " "Parameters can be overwritten or added with command-line options of the form `--key value`.", default=list(), ) parser.add_argument( "-c", "--config", nargs="?", metavar="single_config.yaml", help="path to single config. If specified, base configs will be ignored " "(except for the last one if left unspecified).", const=True, default="", ) parser.add_argument( "--ignore_base_data", action="store_true", help="Ignore data specification from base configs. Useful if you want " "to specify a custom datasets on the command line.", ) parser.add_argument( "--outdir", required=True, type=str, help="Where to write outputs to.", ) parser.add_argument( "--top_k", type=int, default=100, help="Sample from among top-k predictions.", ) parser.add_argument( "--temperature", type=float, default=1.0, help="Sampling temperature.", ) return parser def load_model_from_config(config, sd, gpu=True, eval_mode=True): if "ckpt_path" in config.params: print("Deleting the restore-ckpt path from the config...") config.params.ckpt_path = None if "downsample_cond_size" in config.params: print("Deleting downsample-cond-size from the config and setting factor=0.5 instead...") config.params.downsample_cond_size = -1 config.params["downsample_cond_factor"] = 0.5 try: if "ckpt_path" in config.params.first_stage_config.params: config.params.first_stage_config.params.ckpt_path = None print("Deleting the first-stage restore-ckpt path from the config...") if "ckpt_path" in config.params.cond_stage_config.params: config.params.cond_stage_config.params.ckpt_path = None print("Deleting the cond-stage restore-ckpt path from the config...") except: pass model = instantiate_from_config(config) if sd is not None: missing, unexpected = model.load_state_dict(sd, strict=False) print(f"Missing Keys in State Dict: {missing}") print(f"Unexpected Keys in State Dict: {unexpected}") if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def get_data(config): # get data data = instantiate_from_config(config.data) data.prepare_data() data.setup() return data def load_model_and_dset(config, ckpt, gpu, eval_mode): # get data dsets = get_data(config) # calls data.config ... # now load the specified checkpoint if ckpt: pl_sd = torch.load(ckpt, map_location="cpu") global_step = pl_sd["global_step"] else: pl_sd = {"state_dict": None} global_step = None model = load_model_from_config(config.model, pl_sd["state_dict"], gpu=gpu, eval_mode=eval_mode)["model"] return dsets, model, global_step if __name__ == "__main__": sys.path.append(os.getcwd()) parser = get_parser() opt, unknown = parser.parse_known_args() ckpt = None if opt.resume: if not os.path.exists(opt.resume): raise ValueError("Cannot find {}".format(opt.resume)) if os.path.isfile(opt.resume): paths = opt.resume.split("/") try: idx = len(paths)-paths[::-1].index("logs")+1 except ValueError: idx = -2 # take a guess: path/to/logdir/checkpoints/model.ckpt logdir = "/".join(paths[:idx]) ckpt = opt.resume else: assert os.path.isdir(opt.resume), opt.resume logdir = opt.resume.rstrip("/") ckpt = os.path.join(logdir, "checkpoints", "last.ckpt") print(f"logdir:{logdir}") base_configs = sorted(glob.glob(os.path.join(logdir, "configs/*-project.yaml"))) opt.base = base_configs+opt.base if opt.config: if type(opt.config) == str: opt.base = [opt.config] else: opt.base = [opt.base[-1]] configs = [OmegaConf.load(cfg) for cfg in opt.base] cli = OmegaConf.from_dotlist(unknown) if opt.ignore_base_data: for config in configs: if hasattr(config, "data"): del config["data"] config = OmegaConf.merge(*configs, cli) print(ckpt) gpu = True eval_mode = True show_config = False if show_config: print(OmegaConf.to_container(config)) dsets, model, global_step = load_model_and_dset(config, ckpt, gpu, eval_mode) print(f"Global step: {global_step}") outdir = os.path.join(opt.outdir, "{:06}_{}_{}".format(global_step, opt.top_k, opt.temperature)) os.makedirs(outdir, exist_ok=True) print("Writing samples to ", outdir) for k in ["originals", "reconstructions", "samples"]: os.makedirs(os.path.join(outdir, k), exist_ok=True) run_conditional(model, dsets, outdir, opt.top_k, opt.temperature)
10,146
33.631399
102
py
taming-transformers
taming-transformers-master/taming/lr_scheduler.py
import numpy as np class LambdaWarmUpCosineScheduler: """ note: use with a base_lr of 1.0 """ def __init__(self, warm_up_steps, lr_min, lr_max, lr_start, max_decay_steps, verbosity_interval=0): self.lr_warm_up_steps = warm_up_steps self.lr_start = lr_start self.lr_min = lr_min self.lr_max = lr_max self.lr_max_decay_steps = max_decay_steps self.last_lr = 0. self.verbosity_interval = verbosity_interval def schedule(self, n): if self.verbosity_interval > 0: if n % self.verbosity_interval == 0: print(f"current step: {n}, recent lr-multiplier: {self.last_lr}") if n < self.lr_warm_up_steps: lr = (self.lr_max - self.lr_start) / self.lr_warm_up_steps * n + self.lr_start self.last_lr = lr return lr else: t = (n - self.lr_warm_up_steps) / (self.lr_max_decay_steps - self.lr_warm_up_steps) t = min(t, 1.0) lr = self.lr_min + 0.5 * (self.lr_max - self.lr_min) * ( 1 + np.cos(t * np.pi)) self.last_lr = lr return lr def __call__(self, n): return self.schedule(n)
1,205
33.457143
114
py
taming-transformers
taming-transformers-master/taming/util.py
import os, hashlib import requests from tqdm import tqdm URL_MAP = { "vgg_lpips": "https://heibox.uni-heidelberg.de/f/607503859c864bc1b30b/?dl=1" } CKPT_MAP = { "vgg_lpips": "vgg.pth" } MD5_MAP = { "vgg_lpips": "d507d7349b931f0638a25a48a722f98a" } def download(url, local_path, chunk_size=1024): os.makedirs(os.path.split(local_path)[0], exist_ok=True) with requests.get(url, stream=True) as r: total_size = int(r.headers.get("content-length", 0)) with tqdm(total=total_size, unit="B", unit_scale=True) as pbar: with open(local_path, "wb") as f: for data in r.iter_content(chunk_size=chunk_size): if data: f.write(data) pbar.update(chunk_size) def md5_hash(path): with open(path, "rb") as f: content = f.read() return hashlib.md5(content).hexdigest() def get_ckpt_path(name, root, check=False): assert name in URL_MAP path = os.path.join(root, CKPT_MAP[name]) if not os.path.exists(path) or (check and not md5_hash(path) == MD5_MAP[name]): print("Downloading {} model from {} to {}".format(name, URL_MAP[name], path)) download(URL_MAP[name], path) md5 = md5_hash(path) assert md5 == MD5_MAP[name], md5 return path class KeyNotFoundError(Exception): def __init__(self, cause, keys=None, visited=None): self.cause = cause self.keys = keys self.visited = visited messages = list() if keys is not None: messages.append("Key not found: {}".format(keys)) if visited is not None: messages.append("Visited: {}".format(visited)) messages.append("Cause:\n{}".format(cause)) message = "\n".join(messages) super().__init__(message) def retrieve( list_or_dict, key, splitval="/", default=None, expand=True, pass_success=False ): """Given a nested list or dict return the desired value at key expanding callable nodes if necessary and :attr:`expand` is ``True``. The expansion is done in-place. Parameters ---------- list_or_dict : list or dict Possibly nested list or dictionary. key : str key/to/value, path like string describing all keys necessary to consider to get to the desired value. List indices can also be passed here. splitval : str String that defines the delimiter between keys of the different depth levels in `key`. default : obj Value returned if :attr:`key` is not found. expand : bool Whether to expand callable nodes on the path or not. Returns ------- The desired value or if :attr:`default` is not ``None`` and the :attr:`key` is not found returns ``default``. Raises ------ Exception if ``key`` not in ``list_or_dict`` and :attr:`default` is ``None``. """ keys = key.split(splitval) success = True try: visited = [] parent = None last_key = None for key in keys: if callable(list_or_dict): if not expand: raise KeyNotFoundError( ValueError( "Trying to get past callable node with expand=False." ), keys=keys, visited=visited, ) list_or_dict = list_or_dict() parent[last_key] = list_or_dict last_key = key parent = list_or_dict try: if isinstance(list_or_dict, dict): list_or_dict = list_or_dict[key] else: list_or_dict = list_or_dict[int(key)] except (KeyError, IndexError, ValueError) as e: raise KeyNotFoundError(e, keys=keys, visited=visited) visited += [key] # final expansion of retrieved value if expand and callable(list_or_dict): list_or_dict = list_or_dict() parent[last_key] = list_or_dict except KeyNotFoundError as e: if default is None: raise e else: list_or_dict = default success = False if not pass_success: return list_or_dict else: return list_or_dict, success if __name__ == "__main__": config = {"keya": "a", "keyb": "b", "keyc": {"cc1": 1, "cc2": 2, } } from omegaconf import OmegaConf config = OmegaConf.create(config) print(config) retrieve(config, "keya")
4,777
29.240506
85
py
taming-transformers
taming-transformers-master/taming/modules/util.py
import torch import torch.nn as nn def count_params(model): total_params = sum(p.numel() for p in model.parameters()) return total_params class ActNorm(nn.Module): def __init__(self, num_features, logdet=False, affine=True, allow_reverse_init=False): assert affine super().__init__() self.logdet = logdet self.loc = nn.Parameter(torch.zeros(1, num_features, 1, 1)) self.scale = nn.Parameter(torch.ones(1, num_features, 1, 1)) self.allow_reverse_init = allow_reverse_init self.register_buffer('initialized', torch.tensor(0, dtype=torch.uint8)) def initialize(self, input): with torch.no_grad(): flatten = input.permute(1, 0, 2, 3).contiguous().view(input.shape[1], -1) mean = ( flatten.mean(1) .unsqueeze(1) .unsqueeze(2) .unsqueeze(3) .permute(1, 0, 2, 3) ) std = ( flatten.std(1) .unsqueeze(1) .unsqueeze(2) .unsqueeze(3) .permute(1, 0, 2, 3) ) self.loc.data.copy_(-mean) self.scale.data.copy_(1 / (std + 1e-6)) def forward(self, input, reverse=False): if reverse: return self.reverse(input) if len(input.shape) == 2: input = input[:,:,None,None] squeeze = True else: squeeze = False _, _, height, width = input.shape if self.training and self.initialized.item() == 0: self.initialize(input) self.initialized.fill_(1) h = self.scale * (input + self.loc) if squeeze: h = h.squeeze(-1).squeeze(-1) if self.logdet: log_abs = torch.log(torch.abs(self.scale)) logdet = height*width*torch.sum(log_abs) logdet = logdet * torch.ones(input.shape[0]).to(input) return h, logdet return h def reverse(self, output): if self.training and self.initialized.item() == 0: if not self.allow_reverse_init: raise RuntimeError( "Initializing ActNorm in reverse direction is " "disabled by default. Use allow_reverse_init=True to enable." ) else: self.initialize(output) self.initialized.fill_(1) if len(output.shape) == 2: output = output[:,:,None,None] squeeze = True else: squeeze = False h = output / self.scale - self.loc if squeeze: h = h.squeeze(-1).squeeze(-1) return h class AbstractEncoder(nn.Module): def __init__(self): super().__init__() def encode(self, *args, **kwargs): raise NotImplementedError class Labelator(AbstractEncoder): """Net2Net Interface for Class-Conditional Model""" def __init__(self, n_classes, quantize_interface=True): super().__init__() self.n_classes = n_classes self.quantize_interface = quantize_interface def encode(self, c): c = c[:,None] if self.quantize_interface: return c, None, [None, None, c.long()] return c class SOSProvider(AbstractEncoder): # for unconditional training def __init__(self, sos_token, quantize_interface=True): super().__init__() self.sos_token = sos_token self.quantize_interface = quantize_interface def encode(self, x): # get batch size from data and replicate sos_token c = torch.ones(x.shape[0], 1)*self.sos_token c = c.long().to(x.device) if self.quantize_interface: return c, None, [None, None, c] return c
3,847
28.374046
85
py
taming-transformers
taming-transformers-master/taming/modules/vqvae/quantize.py
import torch import torch.nn as nn import torch.nn.functional as F import numpy as np from torch import einsum from einops import rearrange class VectorQuantizer(nn.Module): """ see https://github.com/MishaLaskin/vqvae/blob/d761a999e2267766400dc646d82d3ac3657771d4/models/quantizer.py ____________________________________________ Discretization bottleneck part of the VQ-VAE. Inputs: - n_e : number of embeddings - e_dim : dimension of embedding - beta : commitment cost used in loss term, beta * ||z_e(x)-sg[e]||^2 _____________________________________________ """ # NOTE: this class contains a bug regarding beta; see VectorQuantizer2 for # a fix and use legacy=False to apply that fix. VectorQuantizer2 can be # used wherever VectorQuantizer has been used before and is additionally # more efficient. def __init__(self, n_e, e_dim, beta): super(VectorQuantizer, self).__init__() self.n_e = n_e self.e_dim = e_dim self.beta = beta self.embedding = nn.Embedding(self.n_e, self.e_dim) self.embedding.weight.data.uniform_(-1.0 / self.n_e, 1.0 / self.n_e) def forward(self, z): """ Inputs the output of the encoder network z and maps it to a discrete one-hot vector that is the index of the closest embedding vector e_j z (continuous) -> z_q (discrete) z.shape = (batch, channel, height, width) quantization pipeline: 1. get encoder input (B,C,H,W) 2. flatten input to (B*H*W,C) """ # reshape z -> (batch, height, width, channel) and flatten z = z.permute(0, 2, 3, 1).contiguous() z_flattened = z.view(-1, self.e_dim) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z d = torch.sum(z_flattened ** 2, dim=1, keepdim=True) + \ torch.sum(self.embedding.weight**2, dim=1) - 2 * \ torch.matmul(z_flattened, self.embedding.weight.t()) ## could possible replace this here # #\start... # find closest encodings min_encoding_indices = torch.argmin(d, dim=1).unsqueeze(1) min_encodings = torch.zeros( min_encoding_indices.shape[0], self.n_e).to(z) min_encodings.scatter_(1, min_encoding_indices, 1) # dtype min encodings: torch.float32 # min_encodings shape: torch.Size([2048, 512]) # min_encoding_indices.shape: torch.Size([2048, 1]) # get quantized latent vectors z_q = torch.matmul(min_encodings, self.embedding.weight).view(z.shape) #.........\end # with: # .........\start #min_encoding_indices = torch.argmin(d, dim=1) #z_q = self.embedding(min_encoding_indices) # ......\end......... (TODO) # compute loss for embedding loss = torch.mean((z_q.detach()-z)**2) + self.beta * \ torch.mean((z_q - z.detach()) ** 2) # preserve gradients z_q = z + (z_q - z).detach() # perplexity e_mean = torch.mean(min_encodings, dim=0) perplexity = torch.exp(-torch.sum(e_mean * torch.log(e_mean + 1e-10))) # reshape back to match original input shape z_q = z_q.permute(0, 3, 1, 2).contiguous() return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def get_codebook_entry(self, indices, shape): # shape specifying (batch, height, width, channel) # TODO: check for more easy handling with nn.Embedding min_encodings = torch.zeros(indices.shape[0], self.n_e).to(indices) min_encodings.scatter_(1, indices[:,None], 1) # get quantized latent vectors z_q = torch.matmul(min_encodings.float(), self.embedding.weight) if shape is not None: z_q = z_q.view(shape) # reshape back to match original input shape z_q = z_q.permute(0, 3, 1, 2).contiguous() return z_q class GumbelQuantize(nn.Module): """ credit to @karpathy: https://github.com/karpathy/deep-vector-quantization/blob/main/model.py (thanks!) Gumbel Softmax trick quantizer Categorical Reparameterization with Gumbel-Softmax, Jang et al. 2016 https://arxiv.org/abs/1611.01144 """ def __init__(self, num_hiddens, embedding_dim, n_embed, straight_through=True, kl_weight=5e-4, temp_init=1.0, use_vqinterface=True, remap=None, unknown_index="random"): super().__init__() self.embedding_dim = embedding_dim self.n_embed = n_embed self.straight_through = straight_through self.temperature = temp_init self.kl_weight = kl_weight self.proj = nn.Conv2d(num_hiddens, n_embed, 1) self.embed = nn.Embedding(n_embed, embedding_dim) self.use_vqinterface = use_vqinterface self.remap = remap if self.remap is not None: self.register_buffer("used", torch.tensor(np.load(self.remap))) self.re_embed = self.used.shape[0] self.unknown_index = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": self.unknown_index = self.re_embed self.re_embed = self.re_embed+1 print(f"Remapping {self.n_embed} indices to {self.re_embed} indices. " f"Using {self.unknown_index} for unknown indices.") else: self.re_embed = n_embed def remap_to_used(self, inds): ishape = inds.shape assert len(ishape)>1 inds = inds.reshape(ishape[0],-1) used = self.used.to(inds) match = (inds[:,:,None]==used[None,None,...]).long() new = match.argmax(-1) unknown = match.sum(2)<1 if self.unknown_index == "random": new[unknown]=torch.randint(0,self.re_embed,size=new[unknown].shape).to(device=new.device) else: new[unknown] = self.unknown_index return new.reshape(ishape) def unmap_to_all(self, inds): ishape = inds.shape assert len(ishape)>1 inds = inds.reshape(ishape[0],-1) used = self.used.to(inds) if self.re_embed > self.used.shape[0]: # extra token inds[inds>=self.used.shape[0]] = 0 # simply set to zero back=torch.gather(used[None,:][inds.shape[0]*[0],:], 1, inds) return back.reshape(ishape) def forward(self, z, temp=None, return_logits=False): # force hard = True when we are in eval mode, as we must quantize. actually, always true seems to work hard = self.straight_through if self.training else True temp = self.temperature if temp is None else temp logits = self.proj(z) if self.remap is not None: # continue only with used logits full_zeros = torch.zeros_like(logits) logits = logits[:,self.used,...] soft_one_hot = F.gumbel_softmax(logits, tau=temp, dim=1, hard=hard) if self.remap is not None: # go back to all entries but unused set to zero full_zeros[:,self.used,...] = soft_one_hot soft_one_hot = full_zeros z_q = einsum('b n h w, n d -> b d h w', soft_one_hot, self.embed.weight) # + kl divergence to the prior loss qy = F.softmax(logits, dim=1) diff = self.kl_weight * torch.sum(qy * torch.log(qy * self.n_embed + 1e-10), dim=1).mean() ind = soft_one_hot.argmax(dim=1) if self.remap is not None: ind = self.remap_to_used(ind) if self.use_vqinterface: if return_logits: return z_q, diff, (None, None, ind), logits return z_q, diff, (None, None, ind) return z_q, diff, ind def get_codebook_entry(self, indices, shape): b, h, w, c = shape assert b*h*w == indices.shape[0] indices = rearrange(indices, '(b h w) -> b h w', b=b, h=h, w=w) if self.remap is not None: indices = self.unmap_to_all(indices) one_hot = F.one_hot(indices, num_classes=self.n_embed).permute(0, 3, 1, 2).float() z_q = einsum('b n h w, n d -> b d h w', one_hot, self.embed.weight) return z_q class VectorQuantizer2(nn.Module): """ Improved version over VectorQuantizer, can be used as a drop-in replacement. Mostly avoids costly matrix multiplications and allows for post-hoc remapping of indices. """ # NOTE: due to a bug the beta term was applied to the wrong term. for # backwards compatibility we use the buggy version by default, but you can # specify legacy=False to fix it. def __init__(self, n_e, e_dim, beta, remap=None, unknown_index="random", sane_index_shape=False, legacy=True): super().__init__() self.n_e = n_e self.e_dim = e_dim self.beta = beta self.legacy = legacy self.embedding = nn.Embedding(self.n_e, self.e_dim) self.embedding.weight.data.uniform_(-1.0 / self.n_e, 1.0 / self.n_e) self.remap = remap if self.remap is not None: self.register_buffer("used", torch.tensor(np.load(self.remap))) self.re_embed = self.used.shape[0] self.unknown_index = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": self.unknown_index = self.re_embed self.re_embed = self.re_embed+1 print(f"Remapping {self.n_e} indices to {self.re_embed} indices. " f"Using {self.unknown_index} for unknown indices.") else: self.re_embed = n_e self.sane_index_shape = sane_index_shape def remap_to_used(self, inds): ishape = inds.shape assert len(ishape)>1 inds = inds.reshape(ishape[0],-1) used = self.used.to(inds) match = (inds[:,:,None]==used[None,None,...]).long() new = match.argmax(-1) unknown = match.sum(2)<1 if self.unknown_index == "random": new[unknown]=torch.randint(0,self.re_embed,size=new[unknown].shape).to(device=new.device) else: new[unknown] = self.unknown_index return new.reshape(ishape) def unmap_to_all(self, inds): ishape = inds.shape assert len(ishape)>1 inds = inds.reshape(ishape[0],-1) used = self.used.to(inds) if self.re_embed > self.used.shape[0]: # extra token inds[inds>=self.used.shape[0]] = 0 # simply set to zero back=torch.gather(used[None,:][inds.shape[0]*[0],:], 1, inds) return back.reshape(ishape) def forward(self, z, temp=None, rescale_logits=False, return_logits=False): assert temp is None or temp==1.0, "Only for interface compatible with Gumbel" assert rescale_logits==False, "Only for interface compatible with Gumbel" assert return_logits==False, "Only for interface compatible with Gumbel" # reshape z -> (batch, height, width, channel) and flatten z = rearrange(z, 'b c h w -> b h w c').contiguous() z_flattened = z.view(-1, self.e_dim) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z d = torch.sum(z_flattened ** 2, dim=1, keepdim=True) + \ torch.sum(self.embedding.weight**2, dim=1) - 2 * \ torch.einsum('bd,dn->bn', z_flattened, rearrange(self.embedding.weight, 'n d -> d n')) min_encoding_indices = torch.argmin(d, dim=1) z_q = self.embedding(min_encoding_indices).view(z.shape) perplexity = None min_encodings = None # compute loss for embedding if not self.legacy: loss = self.beta * torch.mean((z_q.detach()-z)**2) + \ torch.mean((z_q - z.detach()) ** 2) else: loss = torch.mean((z_q.detach()-z)**2) + self.beta * \ torch.mean((z_q - z.detach()) ** 2) # preserve gradients z_q = z + (z_q - z).detach() # reshape back to match original input shape z_q = rearrange(z_q, 'b h w c -> b c h w').contiguous() if self.remap is not None: min_encoding_indices = min_encoding_indices.reshape(z.shape[0],-1) # add batch axis min_encoding_indices = self.remap_to_used(min_encoding_indices) min_encoding_indices = min_encoding_indices.reshape(-1,1) # flatten if self.sane_index_shape: min_encoding_indices = min_encoding_indices.reshape( z_q.shape[0], z_q.shape[2], z_q.shape[3]) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def get_codebook_entry(self, indices, shape): # shape specifying (batch, height, width, channel) if self.remap is not None: indices = indices.reshape(shape[0],-1) # add batch axis indices = self.unmap_to_all(indices) indices = indices.reshape(-1) # flatten again # get quantized latent vectors z_q = self.embedding(indices) if shape is not None: z_q = z_q.view(shape) # reshape back to match original input shape z_q = z_q.permute(0, 3, 1, 2).contiguous() return z_q class EmbeddingEMA(nn.Module): def __init__(self, num_tokens, codebook_dim, decay=0.99, eps=1e-5): super().__init__() self.decay = decay self.eps = eps weight = torch.randn(num_tokens, codebook_dim) self.weight = nn.Parameter(weight, requires_grad = False) self.cluster_size = nn.Parameter(torch.zeros(num_tokens), requires_grad = False) self.embed_avg = nn.Parameter(weight.clone(), requires_grad = False) self.update = True def forward(self, embed_id): return F.embedding(embed_id, self.weight) def cluster_size_ema_update(self, new_cluster_size): self.cluster_size.data.mul_(self.decay).add_(new_cluster_size, alpha=1 - self.decay) def embed_avg_ema_update(self, new_embed_avg): self.embed_avg.data.mul_(self.decay).add_(new_embed_avg, alpha=1 - self.decay) def weight_update(self, num_tokens): n = self.cluster_size.sum() smoothed_cluster_size = ( (self.cluster_size + self.eps) / (n + num_tokens * self.eps) * n ) #normalize embedding average with smoothed cluster size embed_normalized = self.embed_avg / smoothed_cluster_size.unsqueeze(1) self.weight.data.copy_(embed_normalized) class EMAVectorQuantizer(nn.Module): def __init__(self, n_embed, embedding_dim, beta, decay=0.99, eps=1e-5, remap=None, unknown_index="random"): super().__init__() self.codebook_dim = codebook_dim self.num_tokens = num_tokens self.beta = beta self.embedding = EmbeddingEMA(self.num_tokens, self.codebook_dim, decay, eps) self.remap = remap if self.remap is not None: self.register_buffer("used", torch.tensor(np.load(self.remap))) self.re_embed = self.used.shape[0] self.unknown_index = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": self.unknown_index = self.re_embed self.re_embed = self.re_embed+1 print(f"Remapping {self.n_embed} indices to {self.re_embed} indices. " f"Using {self.unknown_index} for unknown indices.") else: self.re_embed = n_embed def remap_to_used(self, inds): ishape = inds.shape assert len(ishape)>1 inds = inds.reshape(ishape[0],-1) used = self.used.to(inds) match = (inds[:,:,None]==used[None,None,...]).long() new = match.argmax(-1) unknown = match.sum(2)<1 if self.unknown_index == "random": new[unknown]=torch.randint(0,self.re_embed,size=new[unknown].shape).to(device=new.device) else: new[unknown] = self.unknown_index return new.reshape(ishape) def unmap_to_all(self, inds): ishape = inds.shape assert len(ishape)>1 inds = inds.reshape(ishape[0],-1) used = self.used.to(inds) if self.re_embed > self.used.shape[0]: # extra token inds[inds>=self.used.shape[0]] = 0 # simply set to zero back=torch.gather(used[None,:][inds.shape[0]*[0],:], 1, inds) return back.reshape(ishape) def forward(self, z): # reshape z -> (batch, height, width, channel) and flatten #z, 'b c h w -> b h w c' z = rearrange(z, 'b c h w -> b h w c') z_flattened = z.reshape(-1, self.codebook_dim) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z d = z_flattened.pow(2).sum(dim=1, keepdim=True) + \ self.embedding.weight.pow(2).sum(dim=1) - 2 * \ torch.einsum('bd,nd->bn', z_flattened, self.embedding.weight) # 'n d -> d n' encoding_indices = torch.argmin(d, dim=1) z_q = self.embedding(encoding_indices).view(z.shape) encodings = F.one_hot(encoding_indices, self.num_tokens).type(z.dtype) avg_probs = torch.mean(encodings, dim=0) perplexity = torch.exp(-torch.sum(avg_probs * torch.log(avg_probs + 1e-10))) if self.training and self.embedding.update: #EMA cluster size encodings_sum = encodings.sum(0) self.embedding.cluster_size_ema_update(encodings_sum) #EMA embedding average embed_sum = encodings.transpose(0,1) @ z_flattened self.embedding.embed_avg_ema_update(embed_sum) #normalize embed_avg and update weight self.embedding.weight_update(self.num_tokens) # compute loss for embedding loss = self.beta * F.mse_loss(z_q.detach(), z) # preserve gradients z_q = z + (z_q - z).detach() # reshape back to match original input shape #z_q, 'b h w c -> b c h w' z_q = rearrange(z_q, 'b h w c -> b c h w') return z_q, loss, (perplexity, encodings, encoding_indices)
18,182
39.769058
110
py
taming-transformers
taming-transformers-master/taming/modules/discriminator/model.py
import functools import torch.nn as nn from taming.modules.util import ActNorm def weights_init(m): classname = m.__class__.__name__ if classname.find('Conv') != -1: nn.init.normal_(m.weight.data, 0.0, 0.02) elif classname.find('BatchNorm') != -1: nn.init.normal_(m.weight.data, 1.0, 0.02) nn.init.constant_(m.bias.data, 0) class NLayerDiscriminator(nn.Module): """Defines a PatchGAN discriminator as in Pix2Pix --> see https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/blob/master/models/networks.py """ def __init__(self, input_nc=3, ndf=64, n_layers=3, use_actnorm=False): """Construct a PatchGAN discriminator Parameters: input_nc (int) -- the number of channels in input images ndf (int) -- the number of filters in the last conv layer n_layers (int) -- the number of conv layers in the discriminator norm_layer -- normalization layer """ super(NLayerDiscriminator, self).__init__() if not use_actnorm: norm_layer = nn.BatchNorm2d else: norm_layer = ActNorm if type(norm_layer) == functools.partial: # no need to use bias as BatchNorm2d has affine parameters use_bias = norm_layer.func != nn.BatchNorm2d else: use_bias = norm_layer != nn.BatchNorm2d kw = 4 padw = 1 sequence = [nn.Conv2d(input_nc, ndf, kernel_size=kw, stride=2, padding=padw), nn.LeakyReLU(0.2, True)] nf_mult = 1 nf_mult_prev = 1 for n in range(1, n_layers): # gradually increase the number of filters nf_mult_prev = nf_mult nf_mult = min(2 ** n, 8) sequence += [ nn.Conv2d(ndf * nf_mult_prev, ndf * nf_mult, kernel_size=kw, stride=2, padding=padw, bias=use_bias), norm_layer(ndf * nf_mult), nn.LeakyReLU(0.2, True) ] nf_mult_prev = nf_mult nf_mult = min(2 ** n_layers, 8) sequence += [ nn.Conv2d(ndf * nf_mult_prev, ndf * nf_mult, kernel_size=kw, stride=1, padding=padw, bias=use_bias), norm_layer(ndf * nf_mult), nn.LeakyReLU(0.2, True) ] sequence += [ nn.Conv2d(ndf * nf_mult, 1, kernel_size=kw, stride=1, padding=padw)] # output 1 channel prediction map self.main = nn.Sequential(*sequence) def forward(self, input): """Standard forward.""" return self.main(input)
2,550
36.514706
116
py
taming-transformers
taming-transformers-master/taming/modules/misc/coord.py
import torch class CoordStage(object): def __init__(self, n_embed, down_factor): self.n_embed = n_embed self.down_factor = down_factor def eval(self): return self def encode(self, c): """fake vqmodel interface""" assert 0.0 <= c.min() and c.max() <= 1.0 b,ch,h,w = c.shape assert ch == 1 c = torch.nn.functional.interpolate(c, scale_factor=1/self.down_factor, mode="area") c = c.clamp(0.0, 1.0) c = self.n_embed*c c_quant = c.round() c_ind = c_quant.to(dtype=torch.long) info = None, None, c_ind return c_quant, None, info def decode(self, c): c = c/self.n_embed c = torch.nn.functional.interpolate(c, scale_factor=self.down_factor, mode="nearest") return c
904
27.28125
79
py
taming-transformers
taming-transformers-master/taming/modules/diffusionmodules/model.py
# pytorch_diffusion + derived encoder decoder import math import torch import torch.nn as nn import numpy as np def get_timestep_embedding(timesteps, embedding_dim): """ This matches the implementation in Denoising Diffusion Probabilistic Models: From Fairseq. Build sinusoidal embeddings. This matches the implementation in tensor2tensor, but differs slightly from the description in Section 3.5 of "Attention Is All You Need". """ assert len(timesteps.shape) == 1 half_dim = embedding_dim // 2 emb = math.log(10000) / (half_dim - 1) emb = torch.exp(torch.arange(half_dim, dtype=torch.float32) * -emb) emb = emb.to(device=timesteps.device) emb = timesteps.float()[:, None] * emb[None, :] emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=1) if embedding_dim % 2 == 1: # zero pad emb = torch.nn.functional.pad(emb, (0,1,0,0)) return emb def nonlinearity(x): # swish return x*torch.sigmoid(x) def Normalize(in_channels): return torch.nn.GroupNorm(num_groups=32, num_channels=in_channels, eps=1e-6, affine=True) class Upsample(nn.Module): def __init__(self, in_channels, with_conv): super().__init__() self.with_conv = with_conv if self.with_conv: self.conv = torch.nn.Conv2d(in_channels, in_channels, kernel_size=3, stride=1, padding=1) def forward(self, x): x = torch.nn.functional.interpolate(x, scale_factor=2.0, mode="nearest") if self.with_conv: x = self.conv(x) return x class Downsample(nn.Module): def __init__(self, in_channels, with_conv): super().__init__() self.with_conv = with_conv if self.with_conv: # no asymmetric padding in torch conv, must do it ourselves self.conv = torch.nn.Conv2d(in_channels, in_channels, kernel_size=3, stride=2, padding=0) def forward(self, x): if self.with_conv: pad = (0,1,0,1) x = torch.nn.functional.pad(x, pad, mode="constant", value=0) x = self.conv(x) else: x = torch.nn.functional.avg_pool2d(x, kernel_size=2, stride=2) return x class ResnetBlock(nn.Module): def __init__(self, *, in_channels, out_channels=None, conv_shortcut=False, dropout, temb_channels=512): super().__init__() self.in_channels = in_channels out_channels = in_channels if out_channels is None else out_channels self.out_channels = out_channels self.use_conv_shortcut = conv_shortcut self.norm1 = Normalize(in_channels) self.conv1 = torch.nn.Conv2d(in_channels, out_channels, kernel_size=3, stride=1, padding=1) if temb_channels > 0: self.temb_proj = torch.nn.Linear(temb_channels, out_channels) self.norm2 = Normalize(out_channels) self.dropout = torch.nn.Dropout(dropout) self.conv2 = torch.nn.Conv2d(out_channels, out_channels, kernel_size=3, stride=1, padding=1) if self.in_channels != self.out_channels: if self.use_conv_shortcut: self.conv_shortcut = torch.nn.Conv2d(in_channels, out_channels, kernel_size=3, stride=1, padding=1) else: self.nin_shortcut = torch.nn.Conv2d(in_channels, out_channels, kernel_size=1, stride=1, padding=0) def forward(self, x, temb): h = x h = self.norm1(h) h = nonlinearity(h) h = self.conv1(h) if temb is not None: h = h + self.temb_proj(nonlinearity(temb))[:,:,None,None] h = self.norm2(h) h = nonlinearity(h) h = self.dropout(h) h = self.conv2(h) if self.in_channels != self.out_channels: if self.use_conv_shortcut: x = self.conv_shortcut(x) else: x = self.nin_shortcut(x) return x+h class AttnBlock(nn.Module): def __init__(self, in_channels): super().__init__() self.in_channels = in_channels self.norm = Normalize(in_channels) self.q = torch.nn.Conv2d(in_channels, in_channels, kernel_size=1, stride=1, padding=0) self.k = torch.nn.Conv2d(in_channels, in_channels, kernel_size=1, stride=1, padding=0) self.v = torch.nn.Conv2d(in_channels, in_channels, kernel_size=1, stride=1, padding=0) self.proj_out = torch.nn.Conv2d(in_channels, in_channels, kernel_size=1, stride=1, padding=0) def forward(self, x): h_ = x h_ = self.norm(h_) q = self.q(h_) k = self.k(h_) v = self.v(h_) # compute attention b,c,h,w = q.shape q = q.reshape(b,c,h*w) q = q.permute(0,2,1) # b,hw,c k = k.reshape(b,c,h*w) # b,c,hw w_ = torch.bmm(q,k) # b,hw,hw w[b,i,j]=sum_c q[b,i,c]k[b,c,j] w_ = w_ * (int(c)**(-0.5)) w_ = torch.nn.functional.softmax(w_, dim=2) # attend to values v = v.reshape(b,c,h*w) w_ = w_.permute(0,2,1) # b,hw,hw (first hw of k, second of q) h_ = torch.bmm(v,w_) # b, c,hw (hw of q) h_[b,c,j] = sum_i v[b,c,i] w_[b,i,j] h_ = h_.reshape(b,c,h,w) h_ = self.proj_out(h_) return x+h_ class Model(nn.Module): def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, attn_resolutions, dropout=0.0, resamp_with_conv=True, in_channels, resolution, use_timestep=True): super().__init__() self.ch = ch self.temb_ch = self.ch*4 self.num_resolutions = len(ch_mult) self.num_res_blocks = num_res_blocks self.resolution = resolution self.in_channels = in_channels self.use_timestep = use_timestep if self.use_timestep: # timestep embedding self.temb = nn.Module() self.temb.dense = nn.ModuleList([ torch.nn.Linear(self.ch, self.temb_ch), torch.nn.Linear(self.temb_ch, self.temb_ch), ]) # downsampling self.conv_in = torch.nn.Conv2d(in_channels, self.ch, kernel_size=3, stride=1, padding=1) curr_res = resolution in_ch_mult = (1,)+tuple(ch_mult) self.down = nn.ModuleList() for i_level in range(self.num_resolutions): block = nn.ModuleList() attn = nn.ModuleList() block_in = ch*in_ch_mult[i_level] block_out = ch*ch_mult[i_level] for i_block in range(self.num_res_blocks): block.append(ResnetBlock(in_channels=block_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out if curr_res in attn_resolutions: attn.append(AttnBlock(block_in)) down = nn.Module() down.block = block down.attn = attn if i_level != self.num_resolutions-1: down.downsample = Downsample(block_in, resamp_with_conv) curr_res = curr_res // 2 self.down.append(down) # middle self.mid = nn.Module() self.mid.block_1 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) self.mid.attn_1 = AttnBlock(block_in) self.mid.block_2 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) # upsampling self.up = nn.ModuleList() for i_level in reversed(range(self.num_resolutions)): block = nn.ModuleList() attn = nn.ModuleList() block_out = ch*ch_mult[i_level] skip_in = ch*ch_mult[i_level] for i_block in range(self.num_res_blocks+1): if i_block == self.num_res_blocks: skip_in = ch*in_ch_mult[i_level] block.append(ResnetBlock(in_channels=block_in+skip_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out if curr_res in attn_resolutions: attn.append(AttnBlock(block_in)) up = nn.Module() up.block = block up.attn = attn if i_level != 0: up.upsample = Upsample(block_in, resamp_with_conv) curr_res = curr_res * 2 self.up.insert(0, up) # prepend to get consistent order # end self.norm_out = Normalize(block_in) self.conv_out = torch.nn.Conv2d(block_in, out_ch, kernel_size=3, stride=1, padding=1) def forward(self, x, t=None): #assert x.shape[2] == x.shape[3] == self.resolution if self.use_timestep: # timestep embedding assert t is not None temb = get_timestep_embedding(t, self.ch) temb = self.temb.dense[0](temb) temb = nonlinearity(temb) temb = self.temb.dense[1](temb) else: temb = None # downsampling hs = [self.conv_in(x)] for i_level in range(self.num_resolutions): for i_block in range(self.num_res_blocks): h = self.down[i_level].block[i_block](hs[-1], temb) if len(self.down[i_level].attn) > 0: h = self.down[i_level].attn[i_block](h) hs.append(h) if i_level != self.num_resolutions-1: hs.append(self.down[i_level].downsample(hs[-1])) # middle h = hs[-1] h = self.mid.block_1(h, temb) h = self.mid.attn_1(h) h = self.mid.block_2(h, temb) # upsampling for i_level in reversed(range(self.num_resolutions)): for i_block in range(self.num_res_blocks+1): h = self.up[i_level].block[i_block]( torch.cat([h, hs.pop()], dim=1), temb) if len(self.up[i_level].attn) > 0: h = self.up[i_level].attn[i_block](h) if i_level != 0: h = self.up[i_level].upsample(h) # end h = self.norm_out(h) h = nonlinearity(h) h = self.conv_out(h) return h class Encoder(nn.Module): def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, attn_resolutions, dropout=0.0, resamp_with_conv=True, in_channels, resolution, z_channels, double_z=True, **ignore_kwargs): super().__init__() self.ch = ch self.temb_ch = 0 self.num_resolutions = len(ch_mult) self.num_res_blocks = num_res_blocks self.resolution = resolution self.in_channels = in_channels # downsampling self.conv_in = torch.nn.Conv2d(in_channels, self.ch, kernel_size=3, stride=1, padding=1) curr_res = resolution in_ch_mult = (1,)+tuple(ch_mult) self.down = nn.ModuleList() for i_level in range(self.num_resolutions): block = nn.ModuleList() attn = nn.ModuleList() block_in = ch*in_ch_mult[i_level] block_out = ch*ch_mult[i_level] for i_block in range(self.num_res_blocks): block.append(ResnetBlock(in_channels=block_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out if curr_res in attn_resolutions: attn.append(AttnBlock(block_in)) down = nn.Module() down.block = block down.attn = attn if i_level != self.num_resolutions-1: down.downsample = Downsample(block_in, resamp_with_conv) curr_res = curr_res // 2 self.down.append(down) # middle self.mid = nn.Module() self.mid.block_1 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) self.mid.attn_1 = AttnBlock(block_in) self.mid.block_2 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) # end self.norm_out = Normalize(block_in) self.conv_out = torch.nn.Conv2d(block_in, 2*z_channels if double_z else z_channels, kernel_size=3, stride=1, padding=1) def forward(self, x): #assert x.shape[2] == x.shape[3] == self.resolution, "{}, {}, {}".format(x.shape[2], x.shape[3], self.resolution) # timestep embedding temb = None # downsampling hs = [self.conv_in(x)] for i_level in range(self.num_resolutions): for i_block in range(self.num_res_blocks): h = self.down[i_level].block[i_block](hs[-1], temb) if len(self.down[i_level].attn) > 0: h = self.down[i_level].attn[i_block](h) hs.append(h) if i_level != self.num_resolutions-1: hs.append(self.down[i_level].downsample(hs[-1])) # middle h = hs[-1] h = self.mid.block_1(h, temb) h = self.mid.attn_1(h) h = self.mid.block_2(h, temb) # end h = self.norm_out(h) h = nonlinearity(h) h = self.conv_out(h) return h class Decoder(nn.Module): def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, attn_resolutions, dropout=0.0, resamp_with_conv=True, in_channels, resolution, z_channels, give_pre_end=False, **ignorekwargs): super().__init__() self.ch = ch self.temb_ch = 0 self.num_resolutions = len(ch_mult) self.num_res_blocks = num_res_blocks self.resolution = resolution self.in_channels = in_channels self.give_pre_end = give_pre_end # compute in_ch_mult, block_in and curr_res at lowest res in_ch_mult = (1,)+tuple(ch_mult) block_in = ch*ch_mult[self.num_resolutions-1] curr_res = resolution // 2**(self.num_resolutions-1) self.z_shape = (1,z_channels,curr_res,curr_res) print("Working with z of shape {} = {} dimensions.".format( self.z_shape, np.prod(self.z_shape))) # z to block_in self.conv_in = torch.nn.Conv2d(z_channels, block_in, kernel_size=3, stride=1, padding=1) # middle self.mid = nn.Module() self.mid.block_1 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) self.mid.attn_1 = AttnBlock(block_in) self.mid.block_2 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) # upsampling self.up = nn.ModuleList() for i_level in reversed(range(self.num_resolutions)): block = nn.ModuleList() attn = nn.ModuleList() block_out = ch*ch_mult[i_level] for i_block in range(self.num_res_blocks+1): block.append(ResnetBlock(in_channels=block_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out if curr_res in attn_resolutions: attn.append(AttnBlock(block_in)) up = nn.Module() up.block = block up.attn = attn if i_level != 0: up.upsample = Upsample(block_in, resamp_with_conv) curr_res = curr_res * 2 self.up.insert(0, up) # prepend to get consistent order # end self.norm_out = Normalize(block_in) self.conv_out = torch.nn.Conv2d(block_in, out_ch, kernel_size=3, stride=1, padding=1) def forward(self, z): #assert z.shape[1:] == self.z_shape[1:] self.last_z_shape = z.shape # timestep embedding temb = None # z to block_in h = self.conv_in(z) # middle h = self.mid.block_1(h, temb) h = self.mid.attn_1(h) h = self.mid.block_2(h, temb) # upsampling for i_level in reversed(range(self.num_resolutions)): for i_block in range(self.num_res_blocks+1): h = self.up[i_level].block[i_block](h, temb) if len(self.up[i_level].attn) > 0: h = self.up[i_level].attn[i_block](h) if i_level != 0: h = self.up[i_level].upsample(h) # end if self.give_pre_end: return h h = self.norm_out(h) h = nonlinearity(h) h = self.conv_out(h) return h class VUNet(nn.Module): def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, attn_resolutions, dropout=0.0, resamp_with_conv=True, in_channels, c_channels, resolution, z_channels, use_timestep=False, **ignore_kwargs): super().__init__() self.ch = ch self.temb_ch = self.ch*4 self.num_resolutions = len(ch_mult) self.num_res_blocks = num_res_blocks self.resolution = resolution self.use_timestep = use_timestep if self.use_timestep: # timestep embedding self.temb = nn.Module() self.temb.dense = nn.ModuleList([ torch.nn.Linear(self.ch, self.temb_ch), torch.nn.Linear(self.temb_ch, self.temb_ch), ]) # downsampling self.conv_in = torch.nn.Conv2d(c_channels, self.ch, kernel_size=3, stride=1, padding=1) curr_res = resolution in_ch_mult = (1,)+tuple(ch_mult) self.down = nn.ModuleList() for i_level in range(self.num_resolutions): block = nn.ModuleList() attn = nn.ModuleList() block_in = ch*in_ch_mult[i_level] block_out = ch*ch_mult[i_level] for i_block in range(self.num_res_blocks): block.append(ResnetBlock(in_channels=block_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out if curr_res in attn_resolutions: attn.append(AttnBlock(block_in)) down = nn.Module() down.block = block down.attn = attn if i_level != self.num_resolutions-1: down.downsample = Downsample(block_in, resamp_with_conv) curr_res = curr_res // 2 self.down.append(down) self.z_in = torch.nn.Conv2d(z_channels, block_in, kernel_size=1, stride=1, padding=0) # middle self.mid = nn.Module() self.mid.block_1 = ResnetBlock(in_channels=2*block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) self.mid.attn_1 = AttnBlock(block_in) self.mid.block_2 = ResnetBlock(in_channels=block_in, out_channels=block_in, temb_channels=self.temb_ch, dropout=dropout) # upsampling self.up = nn.ModuleList() for i_level in reversed(range(self.num_resolutions)): block = nn.ModuleList() attn = nn.ModuleList() block_out = ch*ch_mult[i_level] skip_in = ch*ch_mult[i_level] for i_block in range(self.num_res_blocks+1): if i_block == self.num_res_blocks: skip_in = ch*in_ch_mult[i_level] block.append(ResnetBlock(in_channels=block_in+skip_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out if curr_res in attn_resolutions: attn.append(AttnBlock(block_in)) up = nn.Module() up.block = block up.attn = attn if i_level != 0: up.upsample = Upsample(block_in, resamp_with_conv) curr_res = curr_res * 2 self.up.insert(0, up) # prepend to get consistent order # end self.norm_out = Normalize(block_in) self.conv_out = torch.nn.Conv2d(block_in, out_ch, kernel_size=3, stride=1, padding=1) def forward(self, x, z): #assert x.shape[2] == x.shape[3] == self.resolution if self.use_timestep: # timestep embedding assert t is not None temb = get_timestep_embedding(t, self.ch) temb = self.temb.dense[0](temb) temb = nonlinearity(temb) temb = self.temb.dense[1](temb) else: temb = None # downsampling hs = [self.conv_in(x)] for i_level in range(self.num_resolutions): for i_block in range(self.num_res_blocks): h = self.down[i_level].block[i_block](hs[-1], temb) if len(self.down[i_level].attn) > 0: h = self.down[i_level].attn[i_block](h) hs.append(h) if i_level != self.num_resolutions-1: hs.append(self.down[i_level].downsample(hs[-1])) # middle h = hs[-1] z = self.z_in(z) h = torch.cat((h,z),dim=1) h = self.mid.block_1(h, temb) h = self.mid.attn_1(h) h = self.mid.block_2(h, temb) # upsampling for i_level in reversed(range(self.num_resolutions)): for i_block in range(self.num_res_blocks+1): h = self.up[i_level].block[i_block]( torch.cat([h, hs.pop()], dim=1), temb) if len(self.up[i_level].attn) > 0: h = self.up[i_level].attn[i_block](h) if i_level != 0: h = self.up[i_level].upsample(h) # end h = self.norm_out(h) h = nonlinearity(h) h = self.conv_out(h) return h class SimpleDecoder(nn.Module): def __init__(self, in_channels, out_channels, *args, **kwargs): super().__init__() self.model = nn.ModuleList([nn.Conv2d(in_channels, in_channels, 1), ResnetBlock(in_channels=in_channels, out_channels=2 * in_channels, temb_channels=0, dropout=0.0), ResnetBlock(in_channels=2 * in_channels, out_channels=4 * in_channels, temb_channels=0, dropout=0.0), ResnetBlock(in_channels=4 * in_channels, out_channels=2 * in_channels, temb_channels=0, dropout=0.0), nn.Conv2d(2*in_channels, in_channels, 1), Upsample(in_channels, with_conv=True)]) # end self.norm_out = Normalize(in_channels) self.conv_out = torch.nn.Conv2d(in_channels, out_channels, kernel_size=3, stride=1, padding=1) def forward(self, x): for i, layer in enumerate(self.model): if i in [1,2,3]: x = layer(x, None) else: x = layer(x) h = self.norm_out(x) h = nonlinearity(h) x = self.conv_out(h) return x class UpsampleDecoder(nn.Module): def __init__(self, in_channels, out_channels, ch, num_res_blocks, resolution, ch_mult=(2,2), dropout=0.0): super().__init__() # upsampling self.temb_ch = 0 self.num_resolutions = len(ch_mult) self.num_res_blocks = num_res_blocks block_in = in_channels curr_res = resolution // 2 ** (self.num_resolutions - 1) self.res_blocks = nn.ModuleList() self.upsample_blocks = nn.ModuleList() for i_level in range(self.num_resolutions): res_block = [] block_out = ch * ch_mult[i_level] for i_block in range(self.num_res_blocks + 1): res_block.append(ResnetBlock(in_channels=block_in, out_channels=block_out, temb_channels=self.temb_ch, dropout=dropout)) block_in = block_out self.res_blocks.append(nn.ModuleList(res_block)) if i_level != self.num_resolutions - 1: self.upsample_blocks.append(Upsample(block_in, True)) curr_res = curr_res * 2 # end self.norm_out = Normalize(block_in) self.conv_out = torch.nn.Conv2d(block_in, out_channels, kernel_size=3, stride=1, padding=1) def forward(self, x): # upsampling h = x for k, i_level in enumerate(range(self.num_resolutions)): for i_block in range(self.num_res_blocks + 1): h = self.res_blocks[i_level][i_block](h, None) if i_level != self.num_resolutions - 1: h = self.upsample_blocks[k](h) h = self.norm_out(h) h = nonlinearity(h) h = self.conv_out(h) return h
30,221
37.895753
121
py
taming-transformers
taming-transformers-master/taming/modules/transformer/mingpt.py
""" taken from: https://github.com/karpathy/minGPT/ GPT model: - the initial stem consists of a combination of token encoding and a positional encoding - the meat of it is a uniform sequence of Transformer blocks - each Transformer is a sequential combination of a 1-hidden-layer MLP block and a self-attention block - all blocks feed into a central residual pathway similar to resnets - the final decoder is a linear projection into a vanilla Softmax classifier """ import math import logging import torch import torch.nn as nn from torch.nn import functional as F from transformers import top_k_top_p_filtering logger = logging.getLogger(__name__) class GPTConfig: """ base GPT config, params common to all GPT versions """ embd_pdrop = 0.1 resid_pdrop = 0.1 attn_pdrop = 0.1 def __init__(self, vocab_size, block_size, **kwargs): self.vocab_size = vocab_size self.block_size = block_size for k,v in kwargs.items(): setattr(self, k, v) class GPT1Config(GPTConfig): """ GPT-1 like network roughly 125M params """ n_layer = 12 n_head = 12 n_embd = 768 class CausalSelfAttention(nn.Module): """ A vanilla multi-head masked self-attention layer with a projection at the end. It is possible to use torch.nn.MultiheadAttention here but I am including an explicit implementation here to show that there is nothing too scary here. """ def __init__(self, config): super().__init__() assert config.n_embd % config.n_head == 0 # key, query, value projections for all heads self.key = nn.Linear(config.n_embd, config.n_embd) self.query = nn.Linear(config.n_embd, config.n_embd) self.value = nn.Linear(config.n_embd, config.n_embd) # regularization self.attn_drop = nn.Dropout(config.attn_pdrop) self.resid_drop = nn.Dropout(config.resid_pdrop) # output projection self.proj = nn.Linear(config.n_embd, config.n_embd) # causal mask to ensure that attention is only applied to the left in the input sequence mask = torch.tril(torch.ones(config.block_size, config.block_size)) if hasattr(config, "n_unmasked"): mask[:config.n_unmasked, :config.n_unmasked] = 1 self.register_buffer("mask", mask.view(1, 1, config.block_size, config.block_size)) self.n_head = config.n_head def forward(self, x, layer_past=None): B, T, C = x.size() # calculate query, key, values for all heads in batch and move head forward to be the batch dim k = self.key(x).view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs) q = self.query(x).view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs) v = self.value(x).view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs) present = torch.stack((k, v)) if layer_past is not None: past_key, past_value = layer_past k = torch.cat((past_key, k), dim=-2) v = torch.cat((past_value, v), dim=-2) # causal self-attention; Self-attend: (B, nh, T, hs) x (B, nh, hs, T) -> (B, nh, T, T) att = (q @ k.transpose(-2, -1)) * (1.0 / math.sqrt(k.size(-1))) if layer_past is None: att = att.masked_fill(self.mask[:,:,:T,:T] == 0, float('-inf')) att = F.softmax(att, dim=-1) att = self.attn_drop(att) y = att @ v # (B, nh, T, T) x (B, nh, T, hs) -> (B, nh, T, hs) y = y.transpose(1, 2).contiguous().view(B, T, C) # re-assemble all head outputs side by side # output projection y = self.resid_drop(self.proj(y)) return y, present # TODO: check that this does not break anything class Block(nn.Module): """ an unassuming Transformer block """ def __init__(self, config): super().__init__() self.ln1 = nn.LayerNorm(config.n_embd) self.ln2 = nn.LayerNorm(config.n_embd) self.attn = CausalSelfAttention(config) self.mlp = nn.Sequential( nn.Linear(config.n_embd, 4 * config.n_embd), nn.GELU(), # nice nn.Linear(4 * config.n_embd, config.n_embd), nn.Dropout(config.resid_pdrop), ) def forward(self, x, layer_past=None, return_present=False): # TODO: check that training still works if return_present: assert not self.training # layer past: tuple of length two with B, nh, T, hs attn, present = self.attn(self.ln1(x), layer_past=layer_past) x = x + attn x = x + self.mlp(self.ln2(x)) if layer_past is not None or return_present: return x, present return x class GPT(nn.Module): """ the full GPT language model, with a context size of block_size """ def __init__(self, vocab_size, block_size, n_layer=12, n_head=8, n_embd=256, embd_pdrop=0., resid_pdrop=0., attn_pdrop=0., n_unmasked=0): super().__init__() config = GPTConfig(vocab_size=vocab_size, block_size=block_size, embd_pdrop=embd_pdrop, resid_pdrop=resid_pdrop, attn_pdrop=attn_pdrop, n_layer=n_layer, n_head=n_head, n_embd=n_embd, n_unmasked=n_unmasked) # input embedding stem self.tok_emb = nn.Embedding(config.vocab_size, config.n_embd) self.pos_emb = nn.Parameter(torch.zeros(1, config.block_size, config.n_embd)) self.drop = nn.Dropout(config.embd_pdrop) # transformer self.blocks = nn.Sequential(*[Block(config) for _ in range(config.n_layer)]) # decoder head self.ln_f = nn.LayerNorm(config.n_embd) self.head = nn.Linear(config.n_embd, config.vocab_size, bias=False) self.block_size = config.block_size self.apply(self._init_weights) self.config = config logger.info("number of parameters: %e", sum(p.numel() for p in self.parameters())) def get_block_size(self): return self.block_size def _init_weights(self, module): if isinstance(module, (nn.Linear, nn.Embedding)): module.weight.data.normal_(mean=0.0, std=0.02) if isinstance(module, nn.Linear) and module.bias is not None: module.bias.data.zero_() elif isinstance(module, nn.LayerNorm): module.bias.data.zero_() module.weight.data.fill_(1.0) def forward(self, idx, embeddings=None, targets=None): # forward the GPT model token_embeddings = self.tok_emb(idx) # each index maps to a (learnable) vector if embeddings is not None: # prepend explicit embeddings token_embeddings = torch.cat((embeddings, token_embeddings), dim=1) t = token_embeddings.shape[1] assert t <= self.block_size, "Cannot forward, model block size is exhausted." position_embeddings = self.pos_emb[:, :t, :] # each position maps to a (learnable) vector x = self.drop(token_embeddings + position_embeddings) x = self.blocks(x) x = self.ln_f(x) logits = self.head(x) # if we are given some desired targets also calculate the loss loss = None if targets is not None: loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1)) return logits, loss def forward_with_past(self, idx, embeddings=None, targets=None, past=None, past_length=None): # inference only assert not self.training token_embeddings = self.tok_emb(idx) # each index maps to a (learnable) vector if embeddings is not None: # prepend explicit embeddings token_embeddings = torch.cat((embeddings, token_embeddings), dim=1) if past is not None: assert past_length is not None past = torch.cat(past, dim=-2) # n_layer, 2, b, nh, len_past, dim_head past_shape = list(past.shape) expected_shape = [self.config.n_layer, 2, idx.shape[0], self.config.n_head, past_length, self.config.n_embd//self.config.n_head] assert past_shape == expected_shape, f"{past_shape} =/= {expected_shape}" position_embeddings = self.pos_emb[:, past_length, :] # each position maps to a (learnable) vector else: position_embeddings = self.pos_emb[:, :token_embeddings.shape[1], :] x = self.drop(token_embeddings + position_embeddings) presents = [] # accumulate over layers for i, block in enumerate(self.blocks): x, present = block(x, layer_past=past[i, ...] if past is not None else None, return_present=True) presents.append(present) x = self.ln_f(x) logits = self.head(x) # if we are given some desired targets also calculate the loss loss = None if targets is not None: loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1)) return logits, loss, torch.stack(presents) # _, _, n_layer, 2, b, nh, 1, dim_head class DummyGPT(nn.Module): # for debugging def __init__(self, add_value=1): super().__init__() self.add_value = add_value def forward(self, idx): return idx + self.add_value, None class CodeGPT(nn.Module): """Takes in semi-embeddings""" def __init__(self, vocab_size, block_size, in_channels, n_layer=12, n_head=8, n_embd=256, embd_pdrop=0., resid_pdrop=0., attn_pdrop=0., n_unmasked=0): super().__init__() config = GPTConfig(vocab_size=vocab_size, block_size=block_size, embd_pdrop=embd_pdrop, resid_pdrop=resid_pdrop, attn_pdrop=attn_pdrop, n_layer=n_layer, n_head=n_head, n_embd=n_embd, n_unmasked=n_unmasked) # input embedding stem self.tok_emb = nn.Linear(in_channels, config.n_embd) self.pos_emb = nn.Parameter(torch.zeros(1, config.block_size, config.n_embd)) self.drop = nn.Dropout(config.embd_pdrop) # transformer self.blocks = nn.Sequential(*[Block(config) for _ in range(config.n_layer)]) # decoder head self.ln_f = nn.LayerNorm(config.n_embd) self.head = nn.Linear(config.n_embd, config.vocab_size, bias=False) self.block_size = config.block_size self.apply(self._init_weights) self.config = config logger.info("number of parameters: %e", sum(p.numel() for p in self.parameters())) def get_block_size(self): return self.block_size def _init_weights(self, module): if isinstance(module, (nn.Linear, nn.Embedding)): module.weight.data.normal_(mean=0.0, std=0.02) if isinstance(module, nn.Linear) and module.bias is not None: module.bias.data.zero_() elif isinstance(module, nn.LayerNorm): module.bias.data.zero_() module.weight.data.fill_(1.0) def forward(self, idx, embeddings=None, targets=None): # forward the GPT model token_embeddings = self.tok_emb(idx) # each index maps to a (learnable) vector if embeddings is not None: # prepend explicit embeddings token_embeddings = torch.cat((embeddings, token_embeddings), dim=1) t = token_embeddings.shape[1] assert t <= self.block_size, "Cannot forward, model block size is exhausted." position_embeddings = self.pos_emb[:, :t, :] # each position maps to a (learnable) vector x = self.drop(token_embeddings + position_embeddings) x = self.blocks(x) x = self.taming_cinln_f(x) logits = self.head(x) # if we are given some desired targets also calculate the loss loss = None if targets is not None: loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1)) return logits, loss #### sampling utils def top_k_logits(logits, k): v, ix = torch.topk(logits, k) out = logits.clone() out[out < v[:, [-1]]] = -float('Inf') return out @torch.no_grad() def sample(model, x, steps, temperature=1.0, sample=False, top_k=None): """ take a conditioning sequence of indices in x (of shape (b,t)) and predict the next token in the sequence, feeding the predictions back into the model each time. Clearly the sampling has quadratic complexity unlike an RNN that is only linear, and has a finite context window of block_size, unlike an RNN that has an infinite context window. """ block_size = model.get_block_size() model.eval() for k in range(steps): x_cond = x if x.size(1) <= block_size else x[:, -block_size:] # crop context if needed logits, _ = model(x_cond) # pluck the logits at the final step and scale by temperature logits = logits[:, -1, :] / temperature # optionally crop probabilities to only the top k options if top_k is not None: logits = top_k_logits(logits, top_k) # apply softmax to convert to probabilities probs = F.softmax(logits, dim=-1) # sample from the distribution or take the most likely if sample: ix = torch.multinomial(probs, num_samples=1) else: _, ix = torch.topk(probs, k=1, dim=-1) # append to the sequence and continue x = torch.cat((x, ix), dim=1) return x @torch.no_grad() def sample_with_past(x, model, steps, temperature=1., sample_logits=True, top_k=None, top_p=None, callback=None): # x is conditioning sample = x cond_len = x.shape[1] past = None for n in range(steps): if callback is not None: callback(n) logits, _, present = model.forward_with_past(x, past=past, past_length=(n+cond_len-1)) if past is None: past = [present] else: past.append(present) logits = logits[:, -1, :] / temperature if top_k is not None: logits = top_k_top_p_filtering(logits, top_k=top_k, top_p=top_p) probs = F.softmax(logits, dim=-1) if not sample_logits: _, x = torch.topk(probs, k=1, dim=-1) else: x = torch.multinomial(probs, num_samples=1) # append to the sequence and continue sample = torch.cat((sample, x), dim=1) del past sample = sample[:, cond_len:] # cut conditioning off return sample #### clustering utils class KMeans(nn.Module): def __init__(self, ncluster=512, nc=3, niter=10): super().__init__() self.ncluster = ncluster self.nc = nc self.niter = niter self.shape = (3,32,32) self.register_buffer("C", torch.zeros(self.ncluster,nc)) self.register_buffer('initialized', torch.tensor(0, dtype=torch.uint8)) def is_initialized(self): return self.initialized.item() == 1 @torch.no_grad() def initialize(self, x): N, D = x.shape assert D == self.nc, D c = x[torch.randperm(N)[:self.ncluster]] # init clusters at random for i in range(self.niter): # assign all pixels to the closest codebook element a = ((x[:, None, :] - c[None, :, :])**2).sum(-1).argmin(1) # move each codebook element to be the mean of the pixels that assigned to it c = torch.stack([x[a==k].mean(0) for k in range(self.ncluster)]) # re-assign any poorly positioned codebook elements nanix = torch.any(torch.isnan(c), dim=1) ndead = nanix.sum().item() print('done step %d/%d, re-initialized %d dead clusters' % (i+1, self.niter, ndead)) c[nanix] = x[torch.randperm(N)[:ndead]] # re-init dead clusters self.C.copy_(c) self.initialized.fill_(1) def forward(self, x, reverse=False, shape=None): if not reverse: # flatten bs,c,h,w = x.shape assert c == self.nc x = x.reshape(bs,c,h*w,1) C = self.C.permute(1,0) C = C.reshape(1,c,1,self.ncluster) a = ((x-C)**2).sum(1).argmin(-1) # bs, h*w indices return a else: # flatten bs, HW = x.shape """ c = self.C.reshape( 1, self.nc, 1, self.ncluster) c = c[bs*[0],:,:,:] c = c[:,:,HW*[0],:] x = x.reshape(bs, 1, HW, 1) x = x[:,3*[0],:,:] x = torch.gather(c, dim=3, index=x) """ x = self.C[x] x = x.permute(0,2,1) shape = shape if shape is not None else self.shape x = x.reshape(bs, *shape) return x
16,836
39.473558
140
py
taming-transformers
taming-transformers-master/taming/modules/transformer/permuter.py
import torch import torch.nn as nn import numpy as np class AbstractPermuter(nn.Module): def __init__(self, *args, **kwargs): super().__init__() def forward(self, x, reverse=False): raise NotImplementedError class Identity(AbstractPermuter): def __init__(self): super().__init__() def forward(self, x, reverse=False): return x class Subsample(AbstractPermuter): def __init__(self, H, W): super().__init__() C = 1 indices = np.arange(H*W).reshape(C,H,W) while min(H, W) > 1: indices = indices.reshape(C,H//2,2,W//2,2) indices = indices.transpose(0,2,4,1,3) indices = indices.reshape(C*4,H//2, W//2) H = H//2 W = W//2 C = C*4 assert H == W == 1 idx = torch.tensor(indices.ravel()) self.register_buffer('forward_shuffle_idx', nn.Parameter(idx, requires_grad=False)) self.register_buffer('backward_shuffle_idx', nn.Parameter(torch.argsort(idx), requires_grad=False)) def forward(self, x, reverse=False): if not reverse: return x[:, self.forward_shuffle_idx] else: return x[:, self.backward_shuffle_idx] def mortonify(i, j): """(i,j) index to linear morton code""" i = np.uint64(i) j = np.uint64(j) z = np.uint(0) for pos in range(32): z = (z | ((j & (np.uint64(1) << np.uint64(pos))) << np.uint64(pos)) | ((i & (np.uint64(1) << np.uint64(pos))) << np.uint64(pos+1)) ) return z class ZCurve(AbstractPermuter): def __init__(self, H, W): super().__init__() reverseidx = [np.int64(mortonify(i,j)) for i in range(H) for j in range(W)] idx = np.argsort(reverseidx) idx = torch.tensor(idx) reverseidx = torch.tensor(reverseidx) self.register_buffer('forward_shuffle_idx', idx) self.register_buffer('backward_shuffle_idx', reverseidx) def forward(self, x, reverse=False): if not reverse: return x[:, self.forward_shuffle_idx] else: return x[:, self.backward_shuffle_idx] class SpiralOut(AbstractPermuter): def __init__(self, H, W): super().__init__() assert H == W size = W indices = np.arange(size*size).reshape(size,size) i0 = size//2 j0 = size//2-1 i = i0 j = j0 idx = [indices[i0, j0]] step_mult = 0 for c in range(1, size//2+1): step_mult += 1 # steps left for k in range(step_mult): i = i - 1 j = j idx.append(indices[i, j]) # step down for k in range(step_mult): i = i j = j + 1 idx.append(indices[i, j]) step_mult += 1 if c < size//2: # step right for k in range(step_mult): i = i + 1 j = j idx.append(indices[i, j]) # step up for k in range(step_mult): i = i j = j - 1 idx.append(indices[i, j]) else: # end reached for k in range(step_mult-1): i = i + 1 idx.append(indices[i, j]) assert len(idx) == size*size idx = torch.tensor(idx) self.register_buffer('forward_shuffle_idx', idx) self.register_buffer('backward_shuffle_idx', torch.argsort(idx)) def forward(self, x, reverse=False): if not reverse: return x[:, self.forward_shuffle_idx] else: return x[:, self.backward_shuffle_idx] class SpiralIn(AbstractPermuter): def __init__(self, H, W): super().__init__() assert H == W size = W indices = np.arange(size*size).reshape(size,size) i0 = size//2 j0 = size//2-1 i = i0 j = j0 idx = [indices[i0, j0]] step_mult = 0 for c in range(1, size//2+1): step_mult += 1 # steps left for k in range(step_mult): i = i - 1 j = j idx.append(indices[i, j]) # step down for k in range(step_mult): i = i j = j + 1 idx.append(indices[i, j]) step_mult += 1 if c < size//2: # step right for k in range(step_mult): i = i + 1 j = j idx.append(indices[i, j]) # step up for k in range(step_mult): i = i j = j - 1 idx.append(indices[i, j]) else: # end reached for k in range(step_mult-1): i = i + 1 idx.append(indices[i, j]) assert len(idx) == size*size idx = idx[::-1] idx = torch.tensor(idx) self.register_buffer('forward_shuffle_idx', idx) self.register_buffer('backward_shuffle_idx', torch.argsort(idx)) def forward(self, x, reverse=False): if not reverse: return x[:, self.forward_shuffle_idx] else: return x[:, self.backward_shuffle_idx] class Random(nn.Module): def __init__(self, H, W): super().__init__() indices = np.random.RandomState(1).permutation(H*W) idx = torch.tensor(indices.ravel()) self.register_buffer('forward_shuffle_idx', idx) self.register_buffer('backward_shuffle_idx', torch.argsort(idx)) def forward(self, x, reverse=False): if not reverse: return x[:, self.forward_shuffle_idx] else: return x[:, self.backward_shuffle_idx] class AlternateParsing(AbstractPermuter): def __init__(self, H, W): super().__init__() indices = np.arange(W*H).reshape(H,W) for i in range(1, H, 2): indices[i, :] = indices[i, ::-1] idx = indices.flatten() assert len(idx) == H*W idx = torch.tensor(idx) self.register_buffer('forward_shuffle_idx', idx) self.register_buffer('backward_shuffle_idx', torch.argsort(idx)) def forward(self, x, reverse=False): if not reverse: return x[:, self.forward_shuffle_idx] else: return x[:, self.backward_shuffle_idx] if __name__ == "__main__": p0 = AlternateParsing(16, 16) print(p0.forward_shuffle_idx) print(p0.backward_shuffle_idx) x = torch.randint(0, 768, size=(11, 256)) y = p0(x) xre = p0(y, reverse=True) assert torch.equal(x, xre) p1 = SpiralOut(2, 2) print(p1.forward_shuffle_idx) print(p1.backward_shuffle_idx)
7,093
27.48996
83
py
taming-transformers
taming-transformers-master/taming/modules/losses/lpips.py
"""Stripped version of https://github.com/richzhang/PerceptualSimilarity/tree/master/models""" import torch import torch.nn as nn from torchvision import models from collections import namedtuple from taming.util import get_ckpt_path class LPIPS(nn.Module): # Learned perceptual metric def __init__(self, use_dropout=True): super().__init__() self.scaling_layer = ScalingLayer() self.chns = [64, 128, 256, 512, 512] # vg16 features self.net = vgg16(pretrained=True, requires_grad=False) self.lin0 = NetLinLayer(self.chns[0], use_dropout=use_dropout) self.lin1 = NetLinLayer(self.chns[1], use_dropout=use_dropout) self.lin2 = NetLinLayer(self.chns[2], use_dropout=use_dropout) self.lin3 = NetLinLayer(self.chns[3], use_dropout=use_dropout) self.lin4 = NetLinLayer(self.chns[4], use_dropout=use_dropout) self.load_from_pretrained() for param in self.parameters(): param.requires_grad = False def load_from_pretrained(self, name="vgg_lpips"): ckpt = get_ckpt_path(name, "taming/modules/autoencoder/lpips") self.load_state_dict(torch.load(ckpt, map_location=torch.device("cpu")), strict=False) print("loaded pretrained LPIPS loss from {}".format(ckpt)) @classmethod def from_pretrained(cls, name="vgg_lpips"): if name != "vgg_lpips": raise NotImplementedError model = cls() ckpt = get_ckpt_path(name) model.load_state_dict(torch.load(ckpt, map_location=torch.device("cpu")), strict=False) return model def forward(self, input, target): in0_input, in1_input = (self.scaling_layer(input), self.scaling_layer(target)) outs0, outs1 = self.net(in0_input), self.net(in1_input) feats0, feats1, diffs = {}, {}, {} lins = [self.lin0, self.lin1, self.lin2, self.lin3, self.lin4] for kk in range(len(self.chns)): feats0[kk], feats1[kk] = normalize_tensor(outs0[kk]), normalize_tensor(outs1[kk]) diffs[kk] = (feats0[kk] - feats1[kk]) ** 2 res = [spatial_average(lins[kk].model(diffs[kk]), keepdim=True) for kk in range(len(self.chns))] val = res[0] for l in range(1, len(self.chns)): val += res[l] return val class ScalingLayer(nn.Module): def __init__(self): super(ScalingLayer, self).__init__() self.register_buffer('shift', torch.Tensor([-.030, -.088, -.188])[None, :, None, None]) self.register_buffer('scale', torch.Tensor([.458, .448, .450])[None, :, None, None]) def forward(self, inp): return (inp - self.shift) / self.scale class NetLinLayer(nn.Module): """ A single linear layer which does a 1x1 conv """ def __init__(self, chn_in, chn_out=1, use_dropout=False): super(NetLinLayer, self).__init__() layers = [nn.Dropout(), ] if (use_dropout) else [] layers += [nn.Conv2d(chn_in, chn_out, 1, stride=1, padding=0, bias=False), ] self.model = nn.Sequential(*layers) class vgg16(torch.nn.Module): def __init__(self, requires_grad=False, pretrained=True): super(vgg16, self).__init__() vgg_pretrained_features = models.vgg16(pretrained=pretrained).features self.slice1 = torch.nn.Sequential() self.slice2 = torch.nn.Sequential() self.slice3 = torch.nn.Sequential() self.slice4 = torch.nn.Sequential() self.slice5 = torch.nn.Sequential() self.N_slices = 5 for x in range(4): self.slice1.add_module(str(x), vgg_pretrained_features[x]) for x in range(4, 9): self.slice2.add_module(str(x), vgg_pretrained_features[x]) for x in range(9, 16): self.slice3.add_module(str(x), vgg_pretrained_features[x]) for x in range(16, 23): self.slice4.add_module(str(x), vgg_pretrained_features[x]) for x in range(23, 30): self.slice5.add_module(str(x), vgg_pretrained_features[x]) if not requires_grad: for param in self.parameters(): param.requires_grad = False def forward(self, X): h = self.slice1(X) h_relu1_2 = h h = self.slice2(h) h_relu2_2 = h h = self.slice3(h) h_relu3_3 = h h = self.slice4(h) h_relu4_3 = h h = self.slice5(h) h_relu5_3 = h vgg_outputs = namedtuple("VggOutputs", ['relu1_2', 'relu2_2', 'relu3_3', 'relu4_3', 'relu5_3']) out = vgg_outputs(h_relu1_2, h_relu2_2, h_relu3_3, h_relu4_3, h_relu5_3) return out def normalize_tensor(x,eps=1e-10): norm_factor = torch.sqrt(torch.sum(x**2,dim=1,keepdim=True)) return x/(norm_factor+eps) def spatial_average(x, keepdim=True): return x.mean([2,3],keepdim=keepdim)
4,832
37.975806
104
py
taming-transformers
taming-transformers-master/taming/modules/losses/segmentation.py
import torch.nn as nn import torch.nn.functional as F class BCELoss(nn.Module): def forward(self, prediction, target): loss = F.binary_cross_entropy_with_logits(prediction,target) return loss, {} class BCELossWithQuant(nn.Module): def __init__(self, codebook_weight=1.): super().__init__() self.codebook_weight = codebook_weight def forward(self, qloss, target, prediction, split): bce_loss = F.binary_cross_entropy_with_logits(prediction,target) loss = bce_loss + self.codebook_weight*qloss return loss, {"{}/total_loss".format(split): loss.clone().detach().mean(), "{}/bce_loss".format(split): bce_loss.detach().mean(), "{}/quant_loss".format(split): qloss.detach().mean() }
816
34.521739
82
py
taming-transformers
taming-transformers-master/taming/modules/losses/vqperceptual.py
import torch import torch.nn as nn import torch.nn.functional as F from taming.modules.losses.lpips import LPIPS from taming.modules.discriminator.model import NLayerDiscriminator, weights_init class DummyLoss(nn.Module): def __init__(self): super().__init__() def adopt_weight(weight, global_step, threshold=0, value=0.): if global_step < threshold: weight = value return weight def hinge_d_loss(logits_real, logits_fake): loss_real = torch.mean(F.relu(1. - logits_real)) loss_fake = torch.mean(F.relu(1. + logits_fake)) d_loss = 0.5 * (loss_real + loss_fake) return d_loss def vanilla_d_loss(logits_real, logits_fake): d_loss = 0.5 * ( torch.mean(torch.nn.functional.softplus(-logits_real)) + torch.mean(torch.nn.functional.softplus(logits_fake))) return d_loss class VQLPIPSWithDiscriminator(nn.Module): def __init__(self, disc_start, codebook_weight=1.0, pixelloss_weight=1.0, disc_num_layers=3, disc_in_channels=3, disc_factor=1.0, disc_weight=1.0, perceptual_weight=1.0, use_actnorm=False, disc_conditional=False, disc_ndf=64, disc_loss="hinge"): super().__init__() assert disc_loss in ["hinge", "vanilla"] self.codebook_weight = codebook_weight self.pixel_weight = pixelloss_weight self.perceptual_loss = LPIPS().eval() self.perceptual_weight = perceptual_weight self.discriminator = NLayerDiscriminator(input_nc=disc_in_channels, n_layers=disc_num_layers, use_actnorm=use_actnorm, ndf=disc_ndf ).apply(weights_init) self.discriminator_iter_start = disc_start if disc_loss == "hinge": self.disc_loss = hinge_d_loss elif disc_loss == "vanilla": self.disc_loss = vanilla_d_loss else: raise ValueError(f"Unknown GAN loss '{disc_loss}'.") print(f"VQLPIPSWithDiscriminator running with {disc_loss} loss.") self.disc_factor = disc_factor self.discriminator_weight = disc_weight self.disc_conditional = disc_conditional def calculate_adaptive_weight(self, nll_loss, g_loss, last_layer=None): if last_layer is not None: nll_grads = torch.autograd.grad(nll_loss, last_layer, retain_graph=True)[0] g_grads = torch.autograd.grad(g_loss, last_layer, retain_graph=True)[0] else: nll_grads = torch.autograd.grad(nll_loss, self.last_layer[0], retain_graph=True)[0] g_grads = torch.autograd.grad(g_loss, self.last_layer[0], retain_graph=True)[0] d_weight = torch.norm(nll_grads) / (torch.norm(g_grads) + 1e-4) d_weight = torch.clamp(d_weight, 0.0, 1e4).detach() d_weight = d_weight * self.discriminator_weight return d_weight def forward(self, codebook_loss, inputs, reconstructions, optimizer_idx, global_step, last_layer=None, cond=None, split="train"): rec_loss = torch.abs(inputs.contiguous() - reconstructions.contiguous()) if self.perceptual_weight > 0: p_loss = self.perceptual_loss(inputs.contiguous(), reconstructions.contiguous()) rec_loss = rec_loss + self.perceptual_weight * p_loss else: p_loss = torch.tensor([0.0]) nll_loss = rec_loss #nll_loss = torch.sum(nll_loss) / nll_loss.shape[0] nll_loss = torch.mean(nll_loss) # now the GAN part if optimizer_idx == 0: # generator update if cond is None: assert not self.disc_conditional logits_fake = self.discriminator(reconstructions.contiguous()) else: assert self.disc_conditional logits_fake = self.discriminator(torch.cat((reconstructions.contiguous(), cond), dim=1)) g_loss = -torch.mean(logits_fake) try: d_weight = self.calculate_adaptive_weight(nll_loss, g_loss, last_layer=last_layer) except RuntimeError: assert not self.training d_weight = torch.tensor(0.0) disc_factor = adopt_weight(self.disc_factor, global_step, threshold=self.discriminator_iter_start) loss = nll_loss + d_weight * disc_factor * g_loss + self.codebook_weight * codebook_loss.mean() log = {"{}/total_loss".format(split): loss.clone().detach().mean(), "{}/quant_loss".format(split): codebook_loss.detach().mean(), "{}/nll_loss".format(split): nll_loss.detach().mean(), "{}/rec_loss".format(split): rec_loss.detach().mean(), "{}/p_loss".format(split): p_loss.detach().mean(), "{}/d_weight".format(split): d_weight.detach(), "{}/disc_factor".format(split): torch.tensor(disc_factor), "{}/g_loss".format(split): g_loss.detach().mean(), } return loss, log if optimizer_idx == 1: # second pass for discriminator update if cond is None: logits_real = self.discriminator(inputs.contiguous().detach()) logits_fake = self.discriminator(reconstructions.contiguous().detach()) else: logits_real = self.discriminator(torch.cat((inputs.contiguous().detach(), cond), dim=1)) logits_fake = self.discriminator(torch.cat((reconstructions.contiguous().detach(), cond), dim=1)) disc_factor = adopt_weight(self.disc_factor, global_step, threshold=self.discriminator_iter_start) d_loss = disc_factor * self.disc_loss(logits_real, logits_fake) log = {"{}/disc_loss".format(split): d_loss.clone().detach().mean(), "{}/logits_real".format(split): logits_real.detach().mean(), "{}/logits_fake".format(split): logits_fake.detach().mean() } return d_loss, log
6,179
44.109489
113
py
taming-transformers
taming-transformers-master/taming/modules/losses/__init__.py
from taming.modules.losses.vqperceptual import DummyLoss
58
18.666667
56
py
taming-transformers
taming-transformers-master/taming/models/dummy_cond_stage.py
from torch import Tensor class DummyCondStage: def __init__(self, conditional_key): self.conditional_key = conditional_key self.train = None def eval(self): return self @staticmethod def encode(c: Tensor): return c, None, (None, None, c) @staticmethod def decode(c: Tensor): return c @staticmethod def to_rgb(c: Tensor): return c
416
17.130435
46
py
taming-transformers
taming-transformers-master/taming/models/vqgan.py
import torch import torch.nn.functional as F import pytorch_lightning as pl from main import instantiate_from_config from taming.modules.diffusionmodules.model import Encoder, Decoder from taming.modules.vqvae.quantize import VectorQuantizer2 as VectorQuantizer from taming.modules.vqvae.quantize import GumbelQuantize from taming.modules.vqvae.quantize import EMAVectorQuantizer class VQModel(pl.LightningModule): def __init__(self, ddconfig, lossconfig, n_embed, embed_dim, ckpt_path=None, ignore_keys=[], image_key="image", colorize_nlabels=None, monitor=None, remap=None, sane_index_shape=False, # tell vector quantizer to return indices as bhw ): super().__init__() self.image_key = image_key self.encoder = Encoder(**ddconfig) self.decoder = Decoder(**ddconfig) self.loss = instantiate_from_config(lossconfig) self.quantize = VectorQuantizer(n_embed, embed_dim, beta=0.25, remap=remap, sane_index_shape=sane_index_shape) self.quant_conv = torch.nn.Conv2d(ddconfig["z_channels"], embed_dim, 1) self.post_quant_conv = torch.nn.Conv2d(embed_dim, ddconfig["z_channels"], 1) if ckpt_path is not None: self.init_from_ckpt(ckpt_path, ignore_keys=ignore_keys) self.image_key = image_key if colorize_nlabels is not None: assert type(colorize_nlabels)==int self.register_buffer("colorize", torch.randn(3, colorize_nlabels, 1, 1)) if monitor is not None: self.monitor = monitor def init_from_ckpt(self, path, ignore_keys=list()): sd = torch.load(path, map_location="cpu")["state_dict"] keys = list(sd.keys()) for k in keys: for ik in ignore_keys: if k.startswith(ik): print("Deleting key {} from state_dict.".format(k)) del sd[k] self.load_state_dict(sd, strict=False) print(f"Restored from {path}") def encode(self, x): h = self.encoder(x) h = self.quant_conv(h) quant, emb_loss, info = self.quantize(h) return quant, emb_loss, info def decode(self, quant): quant = self.post_quant_conv(quant) dec = self.decoder(quant) return dec def decode_code(self, code_b): quant_b = self.quantize.embed_code(code_b) dec = self.decode(quant_b) return dec def forward(self, input): quant, diff, _ = self.encode(input) dec = self.decode(quant) return dec, diff def get_input(self, batch, k): x = batch[k] if len(x.shape) == 3: x = x[..., None] x = x.permute(0, 3, 1, 2).to(memory_format=torch.contiguous_format) return x.float() def training_step(self, batch, batch_idx, optimizer_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x) if optimizer_idx == 0: # autoencode aeloss, log_dict_ae = self.loss(qloss, x, xrec, optimizer_idx, self.global_step, last_layer=self.get_last_layer(), split="train") self.log("train/aeloss", aeloss, prog_bar=True, logger=True, on_step=True, on_epoch=True) self.log_dict(log_dict_ae, prog_bar=False, logger=True, on_step=True, on_epoch=True) return aeloss if optimizer_idx == 1: # discriminator discloss, log_dict_disc = self.loss(qloss, x, xrec, optimizer_idx, self.global_step, last_layer=self.get_last_layer(), split="train") self.log("train/discloss", discloss, prog_bar=True, logger=True, on_step=True, on_epoch=True) self.log_dict(log_dict_disc, prog_bar=False, logger=True, on_step=True, on_epoch=True) return discloss def validation_step(self, batch, batch_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x) aeloss, log_dict_ae = self.loss(qloss, x, xrec, 0, self.global_step, last_layer=self.get_last_layer(), split="val") discloss, log_dict_disc = self.loss(qloss, x, xrec, 1, self.global_step, last_layer=self.get_last_layer(), split="val") rec_loss = log_dict_ae["val/rec_loss"] self.log("val/rec_loss", rec_loss, prog_bar=True, logger=True, on_step=True, on_epoch=True, sync_dist=True) self.log("val/aeloss", aeloss, prog_bar=True, logger=True, on_step=True, on_epoch=True, sync_dist=True) self.log_dict(log_dict_ae) self.log_dict(log_dict_disc) return self.log_dict def configure_optimizers(self): lr = self.learning_rate opt_ae = torch.optim.Adam(list(self.encoder.parameters())+ list(self.decoder.parameters())+ list(self.quantize.parameters())+ list(self.quant_conv.parameters())+ list(self.post_quant_conv.parameters()), lr=lr, betas=(0.5, 0.9)) opt_disc = torch.optim.Adam(self.loss.discriminator.parameters(), lr=lr, betas=(0.5, 0.9)) return [opt_ae, opt_disc], [] def get_last_layer(self): return self.decoder.conv_out.weight def log_images(self, batch, **kwargs): log = dict() x = self.get_input(batch, self.image_key) x = x.to(self.device) xrec, _ = self(x) if x.shape[1] > 3: # colorize with random projection assert xrec.shape[1] > 3 x = self.to_rgb(x) xrec = self.to_rgb(xrec) log["inputs"] = x log["reconstructions"] = xrec return log def to_rgb(self, x): assert self.image_key == "segmentation" if not hasattr(self, "colorize"): self.register_buffer("colorize", torch.randn(3, x.shape[1], 1, 1).to(x)) x = F.conv2d(x, weight=self.colorize) x = 2.*(x-x.min())/(x.max()-x.min()) - 1. return x class VQSegmentationModel(VQModel): def __init__(self, n_labels, *args, **kwargs): super().__init__(*args, **kwargs) self.register_buffer("colorize", torch.randn(3, n_labels, 1, 1)) def configure_optimizers(self): lr = self.learning_rate opt_ae = torch.optim.Adam(list(self.encoder.parameters())+ list(self.decoder.parameters())+ list(self.quantize.parameters())+ list(self.quant_conv.parameters())+ list(self.post_quant_conv.parameters()), lr=lr, betas=(0.5, 0.9)) return opt_ae def training_step(self, batch, batch_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x) aeloss, log_dict_ae = self.loss(qloss, x, xrec, split="train") self.log_dict(log_dict_ae, prog_bar=False, logger=True, on_step=True, on_epoch=True) return aeloss def validation_step(self, batch, batch_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x) aeloss, log_dict_ae = self.loss(qloss, x, xrec, split="val") self.log_dict(log_dict_ae, prog_bar=False, logger=True, on_step=True, on_epoch=True) total_loss = log_dict_ae["val/total_loss"] self.log("val/total_loss", total_loss, prog_bar=True, logger=True, on_step=True, on_epoch=True, sync_dist=True) return aeloss @torch.no_grad() def log_images(self, batch, **kwargs): log = dict() x = self.get_input(batch, self.image_key) x = x.to(self.device) xrec, _ = self(x) if x.shape[1] > 3: # colorize with random projection assert xrec.shape[1] > 3 # convert logits to indices xrec = torch.argmax(xrec, dim=1, keepdim=True) xrec = F.one_hot(xrec, num_classes=x.shape[1]) xrec = xrec.squeeze(1).permute(0, 3, 1, 2).float() x = self.to_rgb(x) xrec = self.to_rgb(xrec) log["inputs"] = x log["reconstructions"] = xrec return log class VQNoDiscModel(VQModel): def __init__(self, ddconfig, lossconfig, n_embed, embed_dim, ckpt_path=None, ignore_keys=[], image_key="image", colorize_nlabels=None ): super().__init__(ddconfig=ddconfig, lossconfig=lossconfig, n_embed=n_embed, embed_dim=embed_dim, ckpt_path=ckpt_path, ignore_keys=ignore_keys, image_key=image_key, colorize_nlabels=colorize_nlabels) def training_step(self, batch, batch_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x) # autoencode aeloss, log_dict_ae = self.loss(qloss, x, xrec, self.global_step, split="train") output = pl.TrainResult(minimize=aeloss) output.log("train/aeloss", aeloss, prog_bar=True, logger=True, on_step=True, on_epoch=True) output.log_dict(log_dict_ae, prog_bar=False, logger=True, on_step=True, on_epoch=True) return output def validation_step(self, batch, batch_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x) aeloss, log_dict_ae = self.loss(qloss, x, xrec, self.global_step, split="val") rec_loss = log_dict_ae["val/rec_loss"] output = pl.EvalResult(checkpoint_on=rec_loss) output.log("val/rec_loss", rec_loss, prog_bar=True, logger=True, on_step=True, on_epoch=True) output.log("val/aeloss", aeloss, prog_bar=True, logger=True, on_step=True, on_epoch=True) output.log_dict(log_dict_ae) return output def configure_optimizers(self): optimizer = torch.optim.Adam(list(self.encoder.parameters())+ list(self.decoder.parameters())+ list(self.quantize.parameters())+ list(self.quant_conv.parameters())+ list(self.post_quant_conv.parameters()), lr=self.learning_rate, betas=(0.5, 0.9)) return optimizer class GumbelVQ(VQModel): def __init__(self, ddconfig, lossconfig, n_embed, embed_dim, temperature_scheduler_config, ckpt_path=None, ignore_keys=[], image_key="image", colorize_nlabels=None, monitor=None, kl_weight=1e-8, remap=None, ): z_channels = ddconfig["z_channels"] super().__init__(ddconfig, lossconfig, n_embed, embed_dim, ckpt_path=None, ignore_keys=ignore_keys, image_key=image_key, colorize_nlabels=colorize_nlabels, monitor=monitor, ) self.loss.n_classes = n_embed self.vocab_size = n_embed self.quantize = GumbelQuantize(z_channels, embed_dim, n_embed=n_embed, kl_weight=kl_weight, temp_init=1.0, remap=remap) self.temperature_scheduler = instantiate_from_config(temperature_scheduler_config) # annealing of temp if ckpt_path is not None: self.init_from_ckpt(ckpt_path, ignore_keys=ignore_keys) def temperature_scheduling(self): self.quantize.temperature = self.temperature_scheduler(self.global_step) def encode_to_prequant(self, x): h = self.encoder(x) h = self.quant_conv(h) return h def decode_code(self, code_b): raise NotImplementedError def training_step(self, batch, batch_idx, optimizer_idx): self.temperature_scheduling() x = self.get_input(batch, self.image_key) xrec, qloss = self(x) if optimizer_idx == 0: # autoencode aeloss, log_dict_ae = self.loss(qloss, x, xrec, optimizer_idx, self.global_step, last_layer=self.get_last_layer(), split="train") self.log_dict(log_dict_ae, prog_bar=False, logger=True, on_step=True, on_epoch=True) self.log("temperature", self.quantize.temperature, prog_bar=False, logger=True, on_step=True, on_epoch=True) return aeloss if optimizer_idx == 1: # discriminator discloss, log_dict_disc = self.loss(qloss, x, xrec, optimizer_idx, self.global_step, last_layer=self.get_last_layer(), split="train") self.log_dict(log_dict_disc, prog_bar=False, logger=True, on_step=True, on_epoch=True) return discloss def validation_step(self, batch, batch_idx): x = self.get_input(batch, self.image_key) xrec, qloss = self(x, return_pred_indices=True) aeloss, log_dict_ae = self.loss(qloss, x, xrec, 0, self.global_step, last_layer=self.get_last_layer(), split="val") discloss, log_dict_disc = self.loss(qloss, x, xrec, 1, self.global_step, last_layer=self.get_last_layer(), split="val") rec_loss = log_dict_ae["val/rec_loss"] self.log("val/rec_loss", rec_loss, prog_bar=True, logger=True, on_step=False, on_epoch=True, sync_dist=True) self.log("val/aeloss", aeloss, prog_bar=True, logger=True, on_step=False, on_epoch=True, sync_dist=True) self.log_dict(log_dict_ae) self.log_dict(log_dict_disc) return self.log_dict def log_images(self, batch, **kwargs): log = dict() x = self.get_input(batch, self.image_key) x = x.to(self.device) # encode h = self.encoder(x) h = self.quant_conv(h) quant, _, _ = self.quantize(h) # decode x_rec = self.decode(quant) log["inputs"] = x log["reconstructions"] = x_rec return log class EMAVQ(VQModel): def __init__(self, ddconfig, lossconfig, n_embed, embed_dim, ckpt_path=None, ignore_keys=[], image_key="image", colorize_nlabels=None, monitor=None, remap=None, sane_index_shape=False, # tell vector quantizer to return indices as bhw ): super().__init__(ddconfig, lossconfig, n_embed, embed_dim, ckpt_path=None, ignore_keys=ignore_keys, image_key=image_key, colorize_nlabels=colorize_nlabels, monitor=monitor, ) self.quantize = EMAVectorQuantizer(n_embed=n_embed, embedding_dim=embed_dim, beta=0.25, remap=remap) def configure_optimizers(self): lr = self.learning_rate #Remove self.quantize from parameter list since it is updated via EMA opt_ae = torch.optim.Adam(list(self.encoder.parameters())+ list(self.decoder.parameters())+ list(self.quant_conv.parameters())+ list(self.post_quant_conv.parameters()), lr=lr, betas=(0.5, 0.9)) opt_disc = torch.optim.Adam(self.loss.discriminator.parameters(), lr=lr, betas=(0.5, 0.9)) return [opt_ae, opt_disc], []
16,760
40.487624
120
py
taming-transformers
taming-transformers-master/taming/models/cond_transformer.py
import os, math import torch import torch.nn.functional as F import pytorch_lightning as pl from main import instantiate_from_config from taming.modules.util import SOSProvider def disabled_train(self, mode=True): """Overwrite model.train with this function to make sure train/eval mode does not change anymore.""" return self class Net2NetTransformer(pl.LightningModule): def __init__(self, transformer_config, first_stage_config, cond_stage_config, permuter_config=None, ckpt_path=None, ignore_keys=[], first_stage_key="image", cond_stage_key="depth", downsample_cond_size=-1, pkeep=1.0, sos_token=0, unconditional=False, ): super().__init__() self.be_unconditional = unconditional self.sos_token = sos_token self.first_stage_key = first_stage_key self.cond_stage_key = cond_stage_key self.init_first_stage_from_ckpt(first_stage_config) self.init_cond_stage_from_ckpt(cond_stage_config) if permuter_config is None: permuter_config = {"target": "taming.modules.transformer.permuter.Identity"} self.permuter = instantiate_from_config(config=permuter_config) self.transformer = instantiate_from_config(config=transformer_config) if ckpt_path is not None: self.init_from_ckpt(ckpt_path, ignore_keys=ignore_keys) self.downsample_cond_size = downsample_cond_size self.pkeep = pkeep def init_from_ckpt(self, path, ignore_keys=list()): sd = torch.load(path, map_location="cpu")["state_dict"] for k in sd.keys(): for ik in ignore_keys: if k.startswith(ik): self.print("Deleting key {} from state_dict.".format(k)) del sd[k] self.load_state_dict(sd, strict=False) print(f"Restored from {path}") def init_first_stage_from_ckpt(self, config): model = instantiate_from_config(config) model = model.eval() model.train = disabled_train self.first_stage_model = model def init_cond_stage_from_ckpt(self, config): if config == "__is_first_stage__": print("Using first stage also as cond stage.") self.cond_stage_model = self.first_stage_model elif config == "__is_unconditional__" or self.be_unconditional: print(f"Using no cond stage. Assuming the training is intended to be unconditional. " f"Prepending {self.sos_token} as a sos token.") self.be_unconditional = True self.cond_stage_key = self.first_stage_key self.cond_stage_model = SOSProvider(self.sos_token) else: model = instantiate_from_config(config) model = model.eval() model.train = disabled_train self.cond_stage_model = model def forward(self, x, c): # one step to produce the logits _, z_indices = self.encode_to_z(x) _, c_indices = self.encode_to_c(c) if self.training and self.pkeep < 1.0: mask = torch.bernoulli(self.pkeep*torch.ones(z_indices.shape, device=z_indices.device)) mask = mask.round().to(dtype=torch.int64) r_indices = torch.randint_like(z_indices, self.transformer.config.vocab_size) a_indices = mask*z_indices+(1-mask)*r_indices else: a_indices = z_indices cz_indices = torch.cat((c_indices, a_indices), dim=1) # target includes all sequence elements (no need to handle first one # differently because we are conditioning) target = z_indices # make the prediction logits, _ = self.transformer(cz_indices[:, :-1]) # cut off conditioning outputs - output i corresponds to p(z_i | z_{<i}, c) logits = logits[:, c_indices.shape[1]-1:] return logits, target def top_k_logits(self, logits, k): v, ix = torch.topk(logits, k) out = logits.clone() out[out < v[..., [-1]]] = -float('Inf') return out @torch.no_grad() def sample(self, x, c, steps, temperature=1.0, sample=False, top_k=None, callback=lambda k: None): x = torch.cat((c,x),dim=1) block_size = self.transformer.get_block_size() assert not self.transformer.training if self.pkeep <= 0.0: # one pass suffices since input is pure noise anyway assert len(x.shape)==2 noise_shape = (x.shape[0], steps-1) #noise = torch.randint(self.transformer.config.vocab_size, noise_shape).to(x) noise = c.clone()[:,x.shape[1]-c.shape[1]:-1] x = torch.cat((x,noise),dim=1) logits, _ = self.transformer(x) # take all logits for now and scale by temp logits = logits / temperature # optionally crop probabilities to only the top k options if top_k is not None: logits = self.top_k_logits(logits, top_k) # apply softmax to convert to probabilities probs = F.softmax(logits, dim=-1) # sample from the distribution or take the most likely if sample: shape = probs.shape probs = probs.reshape(shape[0]*shape[1],shape[2]) ix = torch.multinomial(probs, num_samples=1) probs = probs.reshape(shape[0],shape[1],shape[2]) ix = ix.reshape(shape[0],shape[1]) else: _, ix = torch.topk(probs, k=1, dim=-1) # cut off conditioning x = ix[:, c.shape[1]-1:] else: for k in range(steps): callback(k) assert x.size(1) <= block_size # make sure model can see conditioning x_cond = x if x.size(1) <= block_size else x[:, -block_size:] # crop context if needed logits, _ = self.transformer(x_cond) # pluck the logits at the final step and scale by temperature logits = logits[:, -1, :] / temperature # optionally crop probabilities to only the top k options if top_k is not None: logits = self.top_k_logits(logits, top_k) # apply softmax to convert to probabilities probs = F.softmax(logits, dim=-1) # sample from the distribution or take the most likely if sample: ix = torch.multinomial(probs, num_samples=1) else: _, ix = torch.topk(probs, k=1, dim=-1) # append to the sequence and continue x = torch.cat((x, ix), dim=1) # cut off conditioning x = x[:, c.shape[1]:] return x @torch.no_grad() def encode_to_z(self, x): quant_z, _, info = self.first_stage_model.encode(x) indices = info[2].view(quant_z.shape[0], -1) indices = self.permuter(indices) return quant_z, indices @torch.no_grad() def encode_to_c(self, c): if self.downsample_cond_size > -1: c = F.interpolate(c, size=(self.downsample_cond_size, self.downsample_cond_size)) quant_c, _, [_,_,indices] = self.cond_stage_model.encode(c) if len(indices.shape) > 2: indices = indices.view(c.shape[0], -1) return quant_c, indices @torch.no_grad() def decode_to_img(self, index, zshape): index = self.permuter(index, reverse=True) bhwc = (zshape[0],zshape[2],zshape[3],zshape[1]) quant_z = self.first_stage_model.quantize.get_codebook_entry( index.reshape(-1), shape=bhwc) x = self.first_stage_model.decode(quant_z) return x @torch.no_grad() def log_images(self, batch, temperature=None, top_k=None, callback=None, lr_interface=False, **kwargs): log = dict() N = 4 if lr_interface: x, c = self.get_xc(batch, N, diffuse=False, upsample_factor=8) else: x, c = self.get_xc(batch, N) x = x.to(device=self.device) c = c.to(device=self.device) quant_z, z_indices = self.encode_to_z(x) quant_c, c_indices = self.encode_to_c(c) # create a "half"" sample z_start_indices = z_indices[:,:z_indices.shape[1]//2] index_sample = self.sample(z_start_indices, c_indices, steps=z_indices.shape[1]-z_start_indices.shape[1], temperature=temperature if temperature is not None else 1.0, sample=True, top_k=top_k if top_k is not None else 100, callback=callback if callback is not None else lambda k: None) x_sample = self.decode_to_img(index_sample, quant_z.shape) # sample z_start_indices = z_indices[:, :0] index_sample = self.sample(z_start_indices, c_indices, steps=z_indices.shape[1], temperature=temperature if temperature is not None else 1.0, sample=True, top_k=top_k if top_k is not None else 100, callback=callback if callback is not None else lambda k: None) x_sample_nopix = self.decode_to_img(index_sample, quant_z.shape) # det sample z_start_indices = z_indices[:, :0] index_sample = self.sample(z_start_indices, c_indices, steps=z_indices.shape[1], sample=False, callback=callback if callback is not None else lambda k: None) x_sample_det = self.decode_to_img(index_sample, quant_z.shape) # reconstruction x_rec = self.decode_to_img(z_indices, quant_z.shape) log["inputs"] = x log["reconstructions"] = x_rec if self.cond_stage_key in ["objects_bbox", "objects_center_points"]: figure_size = (x_rec.shape[2], x_rec.shape[3]) dataset = kwargs["pl_module"].trainer.datamodule.datasets["validation"] label_for_category_no = dataset.get_textual_label_for_category_no plotter = dataset.conditional_builders[self.cond_stage_key].plot log["conditioning"] = torch.zeros_like(log["reconstructions"]) for i in range(quant_c.shape[0]): log["conditioning"][i] = plotter(quant_c[i], label_for_category_no, figure_size) log["conditioning_rec"] = log["conditioning"] elif self.cond_stage_key != "image": cond_rec = self.cond_stage_model.decode(quant_c) if self.cond_stage_key == "segmentation": # get image from segmentation mask num_classes = cond_rec.shape[1] c = torch.argmax(c, dim=1, keepdim=True) c = F.one_hot(c, num_classes=num_classes) c = c.squeeze(1).permute(0, 3, 1, 2).float() c = self.cond_stage_model.to_rgb(c) cond_rec = torch.argmax(cond_rec, dim=1, keepdim=True) cond_rec = F.one_hot(cond_rec, num_classes=num_classes) cond_rec = cond_rec.squeeze(1).permute(0, 3, 1, 2).float() cond_rec = self.cond_stage_model.to_rgb(cond_rec) log["conditioning_rec"] = cond_rec log["conditioning"] = c log["samples_half"] = x_sample log["samples_nopix"] = x_sample_nopix log["samples_det"] = x_sample_det return log def get_input(self, key, batch): x = batch[key] if len(x.shape) == 3: x = x[..., None] if len(x.shape) == 4: x = x.permute(0, 3, 1, 2).to(memory_format=torch.contiguous_format) if x.dtype == torch.double: x = x.float() return x def get_xc(self, batch, N=None): x = self.get_input(self.first_stage_key, batch) c = self.get_input(self.cond_stage_key, batch) if N is not None: x = x[:N] c = c[:N] return x, c def shared_step(self, batch, batch_idx): x, c = self.get_xc(batch) logits, target = self(x, c) loss = F.cross_entropy(logits.reshape(-1, logits.size(-1)), target.reshape(-1)) return loss def training_step(self, batch, batch_idx): loss = self.shared_step(batch, batch_idx) self.log("train/loss", loss, prog_bar=True, logger=True, on_step=True, on_epoch=True) return loss def validation_step(self, batch, batch_idx): loss = self.shared_step(batch, batch_idx) self.log("val/loss", loss, prog_bar=True, logger=True, on_step=True, on_epoch=True) return loss def configure_optimizers(self): """ Following minGPT: This long function is unfortunately doing something very simple and is being very defensive: We are separating out all parameters of the model into two buckets: those that will experience weight decay for regularization and those that won't (biases, and layernorm/embedding weights). We are then returning the PyTorch optimizer object. """ # separate out all parameters to those that will and won't experience regularizing weight decay decay = set() no_decay = set() whitelist_weight_modules = (torch.nn.Linear, ) blacklist_weight_modules = (torch.nn.LayerNorm, torch.nn.Embedding) for mn, m in self.transformer.named_modules(): for pn, p in m.named_parameters(): fpn = '%s.%s' % (mn, pn) if mn else pn # full param name if pn.endswith('bias'): # all biases will not be decayed no_decay.add(fpn) elif pn.endswith('weight') and isinstance(m, whitelist_weight_modules): # weights of whitelist modules will be weight decayed decay.add(fpn) elif pn.endswith('weight') and isinstance(m, blacklist_weight_modules): # weights of blacklist modules will NOT be weight decayed no_decay.add(fpn) # special case the position embedding parameter in the root GPT module as not decayed no_decay.add('pos_emb') # validate that we considered every parameter param_dict = {pn: p for pn, p in self.transformer.named_parameters()} inter_params = decay & no_decay union_params = decay | no_decay assert len(inter_params) == 0, "parameters %s made it into both decay/no_decay sets!" % (str(inter_params), ) assert len(param_dict.keys() - union_params) == 0, "parameters %s were not separated into either decay/no_decay set!" \ % (str(param_dict.keys() - union_params), ) # create the pytorch optimizer object optim_groups = [ {"params": [param_dict[pn] for pn in sorted(list(decay))], "weight_decay": 0.01}, {"params": [param_dict[pn] for pn in sorted(list(no_decay))], "weight_decay": 0.0}, ] optimizer = torch.optim.AdamW(optim_groups, lr=self.learning_rate, betas=(0.9, 0.95)) return optimizer
15,648
43.331445
127
py
taming-transformers
taming-transformers-master/taming/data/annotated_objects_coco.py
import json from itertools import chain from pathlib import Path from typing import Iterable, Dict, List, Callable, Any from collections import defaultdict from tqdm import tqdm from taming.data.annotated_objects_dataset import AnnotatedObjectsDataset from taming.data.helper_types import Annotation, ImageDescription, Category COCO_PATH_STRUCTURE = { 'train': { 'top_level': '', 'instances_annotations': 'annotations/instances_train2017.json', 'stuff_annotations': 'annotations/stuff_train2017.json', 'files': 'train2017' }, 'validation': { 'top_level': '', 'instances_annotations': 'annotations/instances_val2017.json', 'stuff_annotations': 'annotations/stuff_val2017.json', 'files': 'val2017' } } def load_image_descriptions(description_json: List[Dict]) -> Dict[str, ImageDescription]: return { str(img['id']): ImageDescription( id=img['id'], license=img.get('license'), file_name=img['file_name'], coco_url=img['coco_url'], original_size=(img['width'], img['height']), date_captured=img.get('date_captured'), flickr_url=img.get('flickr_url') ) for img in description_json } def load_categories(category_json: Iterable) -> Dict[str, Category]: return {str(cat['id']): Category(id=str(cat['id']), super_category=cat['supercategory'], name=cat['name']) for cat in category_json if cat['name'] != 'other'} def load_annotations(annotations_json: List[Dict], image_descriptions: Dict[str, ImageDescription], category_no_for_id: Callable[[str], int], split: str) -> Dict[str, List[Annotation]]: annotations = defaultdict(list) total = sum(len(a) for a in annotations_json) for ann in tqdm(chain(*annotations_json), f'Loading {split} annotations', total=total): image_id = str(ann['image_id']) if image_id not in image_descriptions: raise ValueError(f'image_id [{image_id}] has no image description.') category_id = ann['category_id'] try: category_no = category_no_for_id(str(category_id)) except KeyError: continue width, height = image_descriptions[image_id].original_size bbox = (ann['bbox'][0] / width, ann['bbox'][1] / height, ann['bbox'][2] / width, ann['bbox'][3] / height) annotations[image_id].append( Annotation( id=ann['id'], area=bbox[2]*bbox[3], # use bbox area is_group_of=ann['iscrowd'], image_id=ann['image_id'], bbox=bbox, category_id=str(category_id), category_no=category_no ) ) return dict(annotations) class AnnotatedObjectsCoco(AnnotatedObjectsDataset): def __init__(self, use_things: bool = True, use_stuff: bool = True, **kwargs): """ @param data_path: is the path to the following folder structure: coco/ ├── annotations │ ├── instances_train2017.json │ ├── instances_val2017.json │ ├── stuff_train2017.json │ └── stuff_val2017.json ├── train2017 │ ├── 000000000009.jpg │ ├── 000000000025.jpg │ └── ... ├── val2017 │ ├── 000000000139.jpg │ ├── 000000000285.jpg │ └── ... @param: split: one of 'train' or 'validation' @param: desired image size (give square images) """ super().__init__(**kwargs) self.use_things = use_things self.use_stuff = use_stuff with open(self.paths['instances_annotations']) as f: inst_data_json = json.load(f) with open(self.paths['stuff_annotations']) as f: stuff_data_json = json.load(f) category_jsons = [] annotation_jsons = [] if self.use_things: category_jsons.append(inst_data_json['categories']) annotation_jsons.append(inst_data_json['annotations']) if self.use_stuff: category_jsons.append(stuff_data_json['categories']) annotation_jsons.append(stuff_data_json['annotations']) self.categories = load_categories(chain(*category_jsons)) self.filter_categories() self.setup_category_id_and_number() self.image_descriptions = load_image_descriptions(inst_data_json['images']) annotations = load_annotations(annotation_jsons, self.image_descriptions, self.get_category_number, self.split) self.annotations = self.filter_object_number(annotations, self.min_object_area, self.min_objects_per_image, self.max_objects_per_image) self.image_ids = list(self.annotations.keys()) self.clean_up_annotations_and_image_descriptions() def get_path_structure(self) -> Dict[str, str]: if self.split not in COCO_PATH_STRUCTURE: raise ValueError(f'Split [{self.split} does not exist for COCO data.]') return COCO_PATH_STRUCTURE[self.split] def get_image_path(self, image_id: str) -> Path: return self.paths['files'].joinpath(self.image_descriptions[str(image_id)].file_name) def get_image_description(self, image_id: str) -> Dict[str, Any]: # noinspection PyProtectedMember return self.image_descriptions[image_id]._asdict()
5,729
39.928571
119
py
taming-transformers
taming-transformers-master/taming/data/custom.py
import os import numpy as np import albumentations from torch.utils.data import Dataset from taming.data.base import ImagePaths, NumpyPaths, ConcatDatasetWithIndex class CustomBase(Dataset): def __init__(self, *args, **kwargs): super().__init__() self.data = None def __len__(self): return len(self.data) def __getitem__(self, i): example = self.data[i] return example class CustomTrain(CustomBase): def __init__(self, size, training_images_list_file): super().__init__() with open(training_images_list_file, "r") as f: paths = f.read().splitlines() self.data = ImagePaths(paths=paths, size=size, random_crop=False) class CustomTest(CustomBase): def __init__(self, size, test_images_list_file): super().__init__() with open(test_images_list_file, "r") as f: paths = f.read().splitlines() self.data = ImagePaths(paths=paths, size=size, random_crop=False)
998
24.615385
75
py
taming-transformers
taming-transformers-master/taming/data/base.py
import bisect import numpy as np import albumentations from PIL import Image from torch.utils.data import Dataset, ConcatDataset class ConcatDatasetWithIndex(ConcatDataset): """Modified from original pytorch code to return dataset idx""" def __getitem__(self, idx): if idx < 0: if -idx > len(self): raise ValueError("absolute value of index should not exceed dataset length") idx = len(self) + idx dataset_idx = bisect.bisect_right(self.cumulative_sizes, idx) if dataset_idx == 0: sample_idx = idx else: sample_idx = idx - self.cumulative_sizes[dataset_idx - 1] return self.datasets[dataset_idx][sample_idx], dataset_idx class ImagePaths(Dataset): def __init__(self, paths, size=None, random_crop=False, labels=None): self.size = size self.random_crop = random_crop self.labels = dict() if labels is None else labels self.labels["file_path_"] = paths self._length = len(paths) if self.size is not None and self.size > 0: self.rescaler = albumentations.SmallestMaxSize(max_size = self.size) if not self.random_crop: self.cropper = albumentations.CenterCrop(height=self.size,width=self.size) else: self.cropper = albumentations.RandomCrop(height=self.size,width=self.size) self.preprocessor = albumentations.Compose([self.rescaler, self.cropper]) else: self.preprocessor = lambda **kwargs: kwargs def __len__(self): return self._length def preprocess_image(self, image_path): image = Image.open(image_path) if not image.mode == "RGB": image = image.convert("RGB") image = np.array(image).astype(np.uint8) image = self.preprocessor(image=image)["image"] image = (image/127.5 - 1.0).astype(np.float32) return image def __getitem__(self, i): example = dict() example["image"] = self.preprocess_image(self.labels["file_path_"][i]) for k in self.labels: example[k] = self.labels[k][i] return example class NumpyPaths(ImagePaths): def preprocess_image(self, image_path): image = np.load(image_path).squeeze(0) # 3 x 1024 x 1024 image = np.transpose(image, (1,2,0)) image = Image.fromarray(image, mode="RGB") image = np.array(image).astype(np.uint8) image = self.preprocessor(image=image)["image"] image = (image/127.5 - 1.0).astype(np.float32) return image
2,609
35.760563
92
py
taming-transformers
taming-transformers-master/taming/data/helper_types.py
from typing import Dict, Tuple, Optional, NamedTuple, Union from PIL.Image import Image as pil_image from torch import Tensor try: from typing import Literal except ImportError: from typing_extensions import Literal Image = Union[Tensor, pil_image] BoundingBox = Tuple[float, float, float, float] # x0, y0, w, h CropMethodType = Literal['none', 'random', 'center', 'random-2d'] SplitType = Literal['train', 'validation', 'test'] class ImageDescription(NamedTuple): id: int file_name: str original_size: Tuple[int, int] # w, h url: Optional[str] = None license: Optional[int] = None coco_url: Optional[str] = None date_captured: Optional[str] = None flickr_url: Optional[str] = None flickr_id: Optional[str] = None coco_id: Optional[str] = None class Category(NamedTuple): id: str super_category: Optional[str] name: str class Annotation(NamedTuple): area: float image_id: str bbox: BoundingBox category_no: int category_id: str id: Optional[int] = None source: Optional[str] = None confidence: Optional[float] = None is_group_of: Optional[bool] = None is_truncated: Optional[bool] = None is_occluded: Optional[bool] = None is_depiction: Optional[bool] = None is_inside: Optional[bool] = None segmentation: Optional[Dict] = None
1,350
26.02
65
py
taming-transformers
taming-transformers-master/taming/data/open_images_helper.py
open_images_unify_categories_for_coco = { '/m/03bt1vf': '/m/01g317', '/m/04yx4': '/m/01g317', '/m/05r655': '/m/01g317', '/m/01bl7v': '/m/01g317', '/m/0cnyhnx': '/m/01xq0k1', '/m/01226z': '/m/018xm', '/m/05ctyq': '/m/018xm', '/m/058qzx': '/m/04ctx', '/m/06pcq': '/m/0l515', '/m/03m3pdh': '/m/02crq1', '/m/046dlr': '/m/01x3z', '/m/0h8mzrc': '/m/01x3z', } top_300_classes_plus_coco_compatibility = [ ('Man', 1060962), ('Clothing', 986610), ('Tree', 748162), ('Woman', 611896), ('Person', 610294), ('Human face', 442948), ('Girl', 175399), ('Building', 162147), ('Car', 159135), ('Plant', 155704), ('Human body', 137073), ('Flower', 133128), ('Window', 127485), ('Human arm', 118380), ('House', 114365), ('Wheel', 111684), ('Suit', 99054), ('Human hair', 98089), ('Human head', 92763), ('Chair', 88624), ('Boy', 79849), ('Table', 73699), ('Jeans', 57200), ('Tire', 55725), ('Skyscraper', 53321), ('Food', 52400), ('Footwear', 50335), ('Dress', 50236), ('Human leg', 47124), ('Toy', 46636), ('Tower', 45605), ('Boat', 43486), ('Land vehicle', 40541), ('Bicycle wheel', 34646), ('Palm tree', 33729), ('Fashion accessory', 32914), ('Glasses', 31940), ('Bicycle', 31409), ('Furniture', 30656), ('Sculpture', 29643), ('Bottle', 27558), ('Dog', 26980), ('Snack', 26796), ('Human hand', 26664), ('Bird', 25791), ('Book', 25415), ('Guitar', 24386), ('Jacket', 23998), ('Poster', 22192), ('Dessert', 21284), ('Baked goods', 20657), ('Drink', 19754), ('Flag', 18588), ('Houseplant', 18205), ('Tableware', 17613), ('Airplane', 17218), ('Door', 17195), ('Sports uniform', 17068), ('Shelf', 16865), ('Drum', 16612), ('Vehicle', 16542), ('Microphone', 15269), ('Street light', 14957), ('Cat', 14879), ('Fruit', 13684), ('Fast food', 13536), ('Animal', 12932), ('Vegetable', 12534), ('Train', 12358), ('Horse', 11948), ('Flowerpot', 11728), ('Motorcycle', 11621), ('Fish', 11517), ('Desk', 11405), ('Helmet', 10996), ('Truck', 10915), ('Bus', 10695), ('Hat', 10532), ('Auto part', 10488), ('Musical instrument', 10303), ('Sunglasses', 10207), ('Picture frame', 10096), ('Sports equipment', 10015), ('Shorts', 9999), ('Wine glass', 9632), ('Duck', 9242), ('Wine', 9032), ('Rose', 8781), ('Tie', 8693), ('Butterfly', 8436), ('Beer', 7978), ('Cabinetry', 7956), ('Laptop', 7907), ('Insect', 7497), ('Goggles', 7363), ('Shirt', 7098), ('Dairy Product', 7021), ('Marine invertebrates', 7014), ('Cattle', 7006), ('Trousers', 6903), ('Van', 6843), ('Billboard', 6777), ('Balloon', 6367), ('Human nose', 6103), ('Tent', 6073), ('Camera', 6014), ('Doll', 6002), ('Coat', 5951), ('Mobile phone', 5758), ('Swimwear', 5729), ('Strawberry', 5691), ('Stairs', 5643), ('Goose', 5599), ('Umbrella', 5536), ('Cake', 5508), ('Sun hat', 5475), ('Bench', 5310), ('Bookcase', 5163), ('Bee', 5140), ('Computer monitor', 5078), ('Hiking equipment', 4983), ('Office building', 4981), ('Coffee cup', 4748), ('Curtain', 4685), ('Plate', 4651), ('Box', 4621), ('Tomato', 4595), ('Coffee table', 4529), ('Office supplies', 4473), ('Maple', 4416), ('Muffin', 4365), ('Cocktail', 4234), ('Castle', 4197), ('Couch', 4134), ('Pumpkin', 3983), ('Computer keyboard', 3960), ('Human mouth', 3926), ('Christmas tree', 3893), ('Mushroom', 3883), ('Swimming pool', 3809), ('Pastry', 3799), ('Lavender (Plant)', 3769), ('Football helmet', 3732), ('Bread', 3648), ('Traffic sign', 3628), ('Common sunflower', 3597), ('Television', 3550), ('Bed', 3525), ('Cookie', 3485), ('Fountain', 3484), ('Paddle', 3447), ('Bicycle helmet', 3429), ('Porch', 3420), ('Deer', 3387), ('Fedora', 3339), ('Canoe', 3338), ('Carnivore', 3266), ('Bowl', 3202), ('Human eye', 3166), ('Ball', 3118), ('Pillow', 3077), ('Salad', 3061), ('Beetle', 3060), ('Orange', 3050), ('Drawer', 2958), ('Platter', 2937), ('Elephant', 2921), ('Seafood', 2921), ('Monkey', 2915), ('Countertop', 2879), ('Watercraft', 2831), ('Helicopter', 2805), ('Kitchen appliance', 2797), ('Personal flotation device', 2781), ('Swan', 2739), ('Lamp', 2711), ('Boot', 2695), ('Bronze sculpture', 2693), ('Chicken', 2677), ('Taxi', 2643), ('Juice', 2615), ('Cowboy hat', 2604), ('Apple', 2600), ('Tin can', 2590), ('Necklace', 2564), ('Ice cream', 2560), ('Human beard', 2539), ('Coin', 2536), ('Candle', 2515), ('Cart', 2512), ('High heels', 2441), ('Weapon', 2433), ('Handbag', 2406), ('Penguin', 2396), ('Rifle', 2352), ('Violin', 2336), ('Skull', 2304), ('Lantern', 2285), ('Scarf', 2269), ('Saucer', 2225), ('Sheep', 2215), ('Vase', 2189), ('Lily', 2180), ('Mug', 2154), ('Parrot', 2140), ('Human ear', 2137), ('Sandal', 2115), ('Lizard', 2100), ('Kitchen & dining room table', 2063), ('Spider', 1977), ('Coffee', 1974), ('Goat', 1926), ('Squirrel', 1922), ('Cello', 1913), ('Sushi', 1881), ('Tortoise', 1876), ('Pizza', 1870), ('Studio couch', 1864), ('Barrel', 1862), ('Cosmetics', 1841), ('Moths and butterflies', 1841), ('Convenience store', 1817), ('Watch', 1792), ('Home appliance', 1786), ('Harbor seal', 1780), ('Luggage and bags', 1756), ('Vehicle registration plate', 1754), ('Shrimp', 1751), ('Jellyfish', 1730), ('French fries', 1723), ('Egg (Food)', 1698), ('Football', 1697), ('Musical keyboard', 1683), ('Falcon', 1674), ('Candy', 1660), ('Medical equipment', 1654), ('Eagle', 1651), ('Dinosaur', 1634), ('Surfboard', 1630), ('Tank', 1628), ('Grape', 1624), ('Lion', 1624), ('Owl', 1622), ('Ski', 1613), ('Waste container', 1606), ('Frog', 1591), ('Sparrow', 1585), ('Rabbit', 1581), ('Pen', 1546), ('Sea lion', 1537), ('Spoon', 1521), ('Sink', 1512), ('Teddy bear', 1507), ('Bull', 1495), ('Sofa bed', 1490), ('Dragonfly', 1479), ('Brassiere', 1478), ('Chest of drawers', 1472), ('Aircraft', 1466), ('Human foot', 1463), ('Pig', 1455), ('Fork', 1454), ('Antelope', 1438), ('Tripod', 1427), ('Tool', 1424), ('Cheese', 1422), ('Lemon', 1397), ('Hamburger', 1393), ('Dolphin', 1390), ('Mirror', 1390), ('Marine mammal', 1387), ('Giraffe', 1385), ('Snake', 1368), ('Gondola', 1364), ('Wheelchair', 1360), ('Piano', 1358), ('Cupboard', 1348), ('Banana', 1345), ('Trumpet', 1335), ('Lighthouse', 1333), ('Invertebrate', 1317), ('Carrot', 1268), ('Sock', 1260), ('Tiger', 1241), ('Camel', 1224), ('Parachute', 1224), ('Bathroom accessory', 1223), ('Earrings', 1221), ('Headphones', 1218), ('Skirt', 1198), ('Skateboard', 1190), ('Sandwich', 1148), ('Saxophone', 1141), ('Goldfish', 1136), ('Stool', 1104), ('Traffic light', 1097), ('Shellfish', 1081), ('Backpack', 1079), ('Sea turtle', 1078), ('Cucumber', 1075), ('Tea', 1051), ('Toilet', 1047), ('Roller skates', 1040), ('Mule', 1039), ('Bust', 1031), ('Broccoli', 1030), ('Crab', 1020), ('Oyster', 1019), ('Cannon', 1012), ('Zebra', 1012), ('French horn', 1008), ('Grapefruit', 998), ('Whiteboard', 997), ('Zucchini', 997), ('Crocodile', 992), ('Clock', 960), ('Wall clock', 958), ('Doughnut', 869), ('Snail', 868), ('Baseball glove', 859), ('Panda', 830), ('Tennis racket', 830), ('Pear', 652), ('Bagel', 617), ('Oven', 616), ('Ladybug', 615), ('Shark', 615), ('Polar bear', 614), ('Ostrich', 609), ('Hot dog', 473), ('Microwave oven', 467), ('Fire hydrant', 20), ('Stop sign', 20), ('Parking meter', 20), ('Bear', 20), ('Flying disc', 20), ('Snowboard', 20), ('Tennis ball', 20), ('Kite', 20), ('Baseball bat', 20), ('Kitchen knife', 20), ('Knife', 20), ('Submarine sandwich', 20), ('Computer mouse', 20), ('Remote control', 20), ('Toaster', 20), ('Sink', 20), ('Refrigerator', 20), ('Alarm clock', 20), ('Wall clock', 20), ('Scissors', 20), ('Hair dryer', 20), ('Toothbrush', 20), ('Suitcase', 20) ]
8,930
22.502632
43
py
taming-transformers
taming-transformers-master/taming/data/ade20k.py
import os import numpy as np import cv2 import albumentations from PIL import Image from torch.utils.data import Dataset from taming.data.sflckr import SegmentationBase # for examples included in repo class Examples(SegmentationBase): def __init__(self, size=256, random_crop=False, interpolation="bicubic"): super().__init__(data_csv="data/ade20k_examples.txt", data_root="data/ade20k_images", segmentation_root="data/ade20k_segmentations", size=size, random_crop=random_crop, interpolation=interpolation, n_labels=151, shift_segmentation=False) # With semantic map and scene label class ADE20kBase(Dataset): def __init__(self, config=None, size=None, random_crop=False, interpolation="bicubic", crop_size=None): self.split = self.get_split() self.n_labels = 151 # unknown + 150 self.data_csv = {"train": "data/ade20k_train.txt", "validation": "data/ade20k_test.txt"}[self.split] self.data_root = "data/ade20k_root" with open(os.path.join(self.data_root, "sceneCategories.txt"), "r") as f: self.scene_categories = f.read().splitlines() self.scene_categories = dict(line.split() for line in self.scene_categories) with open(self.data_csv, "r") as f: self.image_paths = f.read().splitlines() self._length = len(self.image_paths) self.labels = { "relative_file_path_": [l for l in self.image_paths], "file_path_": [os.path.join(self.data_root, "images", l) for l in self.image_paths], "relative_segmentation_path_": [l.replace(".jpg", ".png") for l in self.image_paths], "segmentation_path_": [os.path.join(self.data_root, "annotations", l.replace(".jpg", ".png")) for l in self.image_paths], "scene_category": [self.scene_categories[l.split("/")[1].replace(".jpg", "")] for l in self.image_paths], } size = None if size is not None and size<=0 else size self.size = size if crop_size is None: self.crop_size = size if size is not None else None else: self.crop_size = crop_size if self.size is not None: self.interpolation = interpolation self.interpolation = { "nearest": cv2.INTER_NEAREST, "bilinear": cv2.INTER_LINEAR, "bicubic": cv2.INTER_CUBIC, "area": cv2.INTER_AREA, "lanczos": cv2.INTER_LANCZOS4}[self.interpolation] self.image_rescaler = albumentations.SmallestMaxSize(max_size=self.size, interpolation=self.interpolation) self.segmentation_rescaler = albumentations.SmallestMaxSize(max_size=self.size, interpolation=cv2.INTER_NEAREST) if crop_size is not None: self.center_crop = not random_crop if self.center_crop: self.cropper = albumentations.CenterCrop(height=self.crop_size, width=self.crop_size) else: self.cropper = albumentations.RandomCrop(height=self.crop_size, width=self.crop_size) self.preprocessor = self.cropper def __len__(self): return self._length def __getitem__(self, i): example = dict((k, self.labels[k][i]) for k in self.labels) image = Image.open(example["file_path_"]) if not image.mode == "RGB": image = image.convert("RGB") image = np.array(image).astype(np.uint8) if self.size is not None: image = self.image_rescaler(image=image)["image"] segmentation = Image.open(example["segmentation_path_"]) segmentation = np.array(segmentation).astype(np.uint8) if self.size is not None: segmentation = self.segmentation_rescaler(image=segmentation)["image"] if self.size is not None: processed = self.preprocessor(image=image, mask=segmentation) else: processed = {"image": image, "mask": segmentation} example["image"] = (processed["image"]/127.5 - 1.0).astype(np.float32) segmentation = processed["mask"] onehot = np.eye(self.n_labels)[segmentation] example["segmentation"] = onehot return example class ADE20kTrain(ADE20kBase): # default to random_crop=True def __init__(self, config=None, size=None, random_crop=True, interpolation="bicubic", crop_size=None): super().__init__(config=config, size=size, random_crop=random_crop, interpolation=interpolation, crop_size=crop_size) def get_split(self): return "train" class ADE20kValidation(ADE20kBase): def get_split(self): return "validation" if __name__ == "__main__": dset = ADE20kValidation() ex = dset[0] for k in ["image", "scene_category", "segmentation"]: print(type(ex[k])) try: print(ex[k].shape) except: print(ex[k])
5,378
42.032
107
py
taming-transformers
taming-transformers-master/taming/data/utils.py
import collections import os import tarfile import urllib import zipfile from pathlib import Path import numpy as np import torch from taming.data.helper_types import Annotation from torch._six import string_classes from torch.utils.data._utils.collate import np_str_obj_array_pattern, default_collate_err_msg_format from tqdm import tqdm def unpack(path): if path.endswith("tar.gz"): with tarfile.open(path, "r:gz") as tar: tar.extractall(path=os.path.split(path)[0]) elif path.endswith("tar"): with tarfile.open(path, "r:") as tar: tar.extractall(path=os.path.split(path)[0]) elif path.endswith("zip"): with zipfile.ZipFile(path, "r") as f: f.extractall(path=os.path.split(path)[0]) else: raise NotImplementedError( "Unknown file extension: {}".format(os.path.splitext(path)[1]) ) def reporthook(bar): """tqdm progress bar for downloads.""" def hook(b=1, bsize=1, tsize=None): if tsize is not None: bar.total = tsize bar.update(b * bsize - bar.n) return hook def get_root(name): base = "data/" root = os.path.join(base, name) os.makedirs(root, exist_ok=True) return root def is_prepared(root): return Path(root).joinpath(".ready").exists() def mark_prepared(root): Path(root).joinpath(".ready").touch() def prompt_download(file_, source, target_dir, content_dir=None): targetpath = os.path.join(target_dir, file_) while not os.path.exists(targetpath): if content_dir is not None and os.path.exists( os.path.join(target_dir, content_dir) ): break print( "Please download '{}' from '{}' to '{}'.".format(file_, source, targetpath) ) if content_dir is not None: print( "Or place its content into '{}'.".format( os.path.join(target_dir, content_dir) ) ) input("Press Enter when done...") return targetpath def download_url(file_, url, target_dir): targetpath = os.path.join(target_dir, file_) os.makedirs(target_dir, exist_ok=True) with tqdm( unit="B", unit_scale=True, unit_divisor=1024, miniters=1, desc=file_ ) as bar: urllib.request.urlretrieve(url, targetpath, reporthook=reporthook(bar)) return targetpath def download_urls(urls, target_dir): paths = dict() for fname, url in urls.items(): outpath = download_url(fname, url, target_dir) paths[fname] = outpath return paths def quadratic_crop(x, bbox, alpha=1.0): """bbox is xmin, ymin, xmax, ymax""" im_h, im_w = x.shape[:2] bbox = np.array(bbox, dtype=np.float32) bbox = np.clip(bbox, 0, max(im_h, im_w)) center = 0.5 * (bbox[0] + bbox[2]), 0.5 * (bbox[1] + bbox[3]) w = bbox[2] - bbox[0] h = bbox[3] - bbox[1] l = int(alpha * max(w, h)) l = max(l, 2) required_padding = -1 * min( center[0] - l, center[1] - l, im_w - (center[0] + l), im_h - (center[1] + l) ) required_padding = int(np.ceil(required_padding)) if required_padding > 0: padding = [ [required_padding, required_padding], [required_padding, required_padding], ] padding += [[0, 0]] * (len(x.shape) - 2) x = np.pad(x, padding, "reflect") center = center[0] + required_padding, center[1] + required_padding xmin = int(center[0] - l / 2) ymin = int(center[1] - l / 2) return np.array(x[ymin : ymin + l, xmin : xmin + l, ...]) def custom_collate(batch): r"""source: pytorch 1.9.0, only one modification to original code """ elem = batch[0] elem_type = type(elem) if isinstance(elem, torch.Tensor): out = None if torch.utils.data.get_worker_info() is not None: # If we're in a background process, concatenate directly into a # shared memory tensor to avoid an extra copy numel = sum([x.numel() for x in batch]) storage = elem.storage()._new_shared(numel) out = elem.new(storage) return torch.stack(batch, 0, out=out) elif elem_type.__module__ == 'numpy' and elem_type.__name__ != 'str_' \ and elem_type.__name__ != 'string_': if elem_type.__name__ == 'ndarray' or elem_type.__name__ == 'memmap': # array of string classes and object if np_str_obj_array_pattern.search(elem.dtype.str) is not None: raise TypeError(default_collate_err_msg_format.format(elem.dtype)) return custom_collate([torch.as_tensor(b) for b in batch]) elif elem.shape == (): # scalars return torch.as_tensor(batch) elif isinstance(elem, float): return torch.tensor(batch, dtype=torch.float64) elif isinstance(elem, int): return torch.tensor(batch) elif isinstance(elem, string_classes): return batch elif isinstance(elem, collections.abc.Mapping): return {key: custom_collate([d[key] for d in batch]) for key in elem} elif isinstance(elem, tuple) and hasattr(elem, '_fields'): # namedtuple return elem_type(*(custom_collate(samples) for samples in zip(*batch))) if isinstance(elem, collections.abc.Sequence) and isinstance(elem[0], Annotation): # added return batch # added elif isinstance(elem, collections.abc.Sequence): # check to make sure that the elements in batch have consistent size it = iter(batch) elem_size = len(next(it)) if not all(len(elem) == elem_size for elem in it): raise RuntimeError('each element in list of batch should be of equal size') transposed = zip(*batch) return [custom_collate(samples) for samples in transposed] raise TypeError(default_collate_err_msg_format.format(elem_type))
5,903
33.729412
100
py
taming-transformers
taming-transformers-master/taming/data/faceshq.py
import os import numpy as np import albumentations from torch.utils.data import Dataset from taming.data.base import ImagePaths, NumpyPaths, ConcatDatasetWithIndex class FacesBase(Dataset): def __init__(self, *args, **kwargs): super().__init__() self.data = None self.keys = None def __len__(self): return len(self.data) def __getitem__(self, i): example = self.data[i] ex = {} if self.keys is not None: for k in self.keys: ex[k] = example[k] else: ex = example return ex class CelebAHQTrain(FacesBase): def __init__(self, size, keys=None): super().__init__() root = "data/celebahq" with open("data/celebahqtrain.txt", "r") as f: relpaths = f.read().splitlines() paths = [os.path.join(root, relpath) for relpath in relpaths] self.data = NumpyPaths(paths=paths, size=size, random_crop=False) self.keys = keys class CelebAHQValidation(FacesBase): def __init__(self, size, keys=None): super().__init__() root = "data/celebahq" with open("data/celebahqvalidation.txt", "r") as f: relpaths = f.read().splitlines() paths = [os.path.join(root, relpath) for relpath in relpaths] self.data = NumpyPaths(paths=paths, size=size, random_crop=False) self.keys = keys class FFHQTrain(FacesBase): def __init__(self, size, keys=None): super().__init__() root = "data/ffhq" with open("data/ffhqtrain.txt", "r") as f: relpaths = f.read().splitlines() paths = [os.path.join(root, relpath) for relpath in relpaths] self.data = ImagePaths(paths=paths, size=size, random_crop=False) self.keys = keys class FFHQValidation(FacesBase): def __init__(self, size, keys=None): super().__init__() root = "data/ffhq" with open("data/ffhqvalidation.txt", "r") as f: relpaths = f.read().splitlines() paths = [os.path.join(root, relpath) for relpath in relpaths] self.data = ImagePaths(paths=paths, size=size, random_crop=False) self.keys = keys class FacesHQTrain(Dataset): # CelebAHQ [0] + FFHQ [1] def __init__(self, size, keys=None, crop_size=None, coord=False): d1 = CelebAHQTrain(size=size, keys=keys) d2 = FFHQTrain(size=size, keys=keys) self.data = ConcatDatasetWithIndex([d1, d2]) self.coord = coord if crop_size is not None: self.cropper = albumentations.RandomCrop(height=crop_size,width=crop_size) if self.coord: self.cropper = albumentations.Compose([self.cropper], additional_targets={"coord": "image"}) def __len__(self): return len(self.data) def __getitem__(self, i): ex, y = self.data[i] if hasattr(self, "cropper"): if not self.coord: out = self.cropper(image=ex["image"]) ex["image"] = out["image"] else: h,w,_ = ex["image"].shape coord = np.arange(h*w).reshape(h,w,1)/(h*w) out = self.cropper(image=ex["image"], coord=coord) ex["image"] = out["image"] ex["coord"] = out["coord"] ex["class"] = y return ex class FacesHQValidation(Dataset): # CelebAHQ [0] + FFHQ [1] def __init__(self, size, keys=None, crop_size=None, coord=False): d1 = CelebAHQValidation(size=size, keys=keys) d2 = FFHQValidation(size=size, keys=keys) self.data = ConcatDatasetWithIndex([d1, d2]) self.coord = coord if crop_size is not None: self.cropper = albumentations.CenterCrop(height=crop_size,width=crop_size) if self.coord: self.cropper = albumentations.Compose([self.cropper], additional_targets={"coord": "image"}) def __len__(self): return len(self.data) def __getitem__(self, i): ex, y = self.data[i] if hasattr(self, "cropper"): if not self.coord: out = self.cropper(image=ex["image"]) ex["image"] = out["image"] else: h,w,_ = ex["image"].shape coord = np.arange(h*w).reshape(h,w,1)/(h*w) out = self.cropper(image=ex["image"], coord=coord) ex["image"] = out["image"] ex["coord"] = out["coord"] ex["class"] = y return ex
4,640
33.377778
92
py
taming-transformers
taming-transformers-master/taming/data/annotated_objects_open_images.py
from collections import defaultdict from csv import DictReader, reader as TupleReader from pathlib import Path from typing import Dict, List, Any import warnings from taming.data.annotated_objects_dataset import AnnotatedObjectsDataset from taming.data.helper_types import Annotation, Category from tqdm import tqdm OPEN_IMAGES_STRUCTURE = { 'train': { 'top_level': '', 'class_descriptions': 'class-descriptions-boxable.csv', 'annotations': 'oidv6-train-annotations-bbox.csv', 'file_list': 'train-images-boxable.csv', 'files': 'train' }, 'validation': { 'top_level': '', 'class_descriptions': 'class-descriptions-boxable.csv', 'annotations': 'validation-annotations-bbox.csv', 'file_list': 'validation-images.csv', 'files': 'validation' }, 'test': { 'top_level': '', 'class_descriptions': 'class-descriptions-boxable.csv', 'annotations': 'test-annotations-bbox.csv', 'file_list': 'test-images.csv', 'files': 'test' } } def load_annotations(descriptor_path: Path, min_object_area: float, category_mapping: Dict[str, str], category_no_for_id: Dict[str, int]) -> Dict[str, List[Annotation]]: annotations: Dict[str, List[Annotation]] = defaultdict(list) with open(descriptor_path) as file: reader = DictReader(file) for i, row in tqdm(enumerate(reader), total=14620000, desc='Loading OpenImages annotations'): width = float(row['XMax']) - float(row['XMin']) height = float(row['YMax']) - float(row['YMin']) area = width * height category_id = row['LabelName'] if category_id in category_mapping: category_id = category_mapping[category_id] if area >= min_object_area and category_id in category_no_for_id: annotations[row['ImageID']].append( Annotation( id=i, image_id=row['ImageID'], source=row['Source'], category_id=category_id, category_no=category_no_for_id[category_id], confidence=float(row['Confidence']), bbox=(float(row['XMin']), float(row['YMin']), width, height), area=area, is_occluded=bool(int(row['IsOccluded'])), is_truncated=bool(int(row['IsTruncated'])), is_group_of=bool(int(row['IsGroupOf'])), is_depiction=bool(int(row['IsDepiction'])), is_inside=bool(int(row['IsInside'])) ) ) if 'train' in str(descriptor_path) and i < 14000000: warnings.warn(f'Running with subset of Open Images. Train dataset has length [{len(annotations)}].') return dict(annotations) def load_image_ids(csv_path: Path) -> List[str]: with open(csv_path) as file: reader = DictReader(file) return [row['image_name'] for row in reader] def load_categories(csv_path: Path) -> Dict[str, Category]: with open(csv_path) as file: reader = TupleReader(file) return {row[0]: Category(id=row[0], name=row[1], super_category=None) for row in reader} class AnnotatedObjectsOpenImages(AnnotatedObjectsDataset): def __init__(self, use_additional_parameters: bool, **kwargs): """ @param data_path: is the path to the following folder structure: open_images/ │ oidv6-train-annotations-bbox.csv ├── class-descriptions-boxable.csv ├── oidv6-train-annotations-bbox.csv ├── test │ ├── 000026e7ee790996.jpg │ ├── 000062a39995e348.jpg │ └── ... ├── test-annotations-bbox.csv ├── test-images.csv ├── train │ ├── 000002b66c9c498e.jpg │ ├── 000002b97e5471a0.jpg │ └── ... ├── train-images-boxable.csv ├── validation │ ├── 0001eeaf4aed83f9.jpg │ ├── 0004886b7d043cfd.jpg │ └── ... ├── validation-annotations-bbox.csv └── validation-images.csv @param: split: one of 'train', 'validation' or 'test' @param: desired image size (returns square images) """ super().__init__(**kwargs) self.use_additional_parameters = use_additional_parameters self.categories = load_categories(self.paths['class_descriptions']) self.filter_categories() self.setup_category_id_and_number() self.image_descriptions = {} annotations = load_annotations(self.paths['annotations'], self.min_object_area, self.category_mapping, self.category_number) self.annotations = self.filter_object_number(annotations, self.min_object_area, self.min_objects_per_image, self.max_objects_per_image) self.image_ids = list(self.annotations.keys()) self.clean_up_annotations_and_image_descriptions() def get_path_structure(self) -> Dict[str, str]: if self.split not in OPEN_IMAGES_STRUCTURE: raise ValueError(f'Split [{self.split} does not exist for Open Images data.]') return OPEN_IMAGES_STRUCTURE[self.split] def get_image_path(self, image_id: str) -> Path: return self.paths['files'].joinpath(f'{image_id:0>16}.jpg') def get_image_description(self, image_id: str) -> Dict[str, Any]: image_path = self.get_image_path(image_id) return {'file_path': str(image_path), 'file_name': image_path.name}
6,107
43.26087
115
py
taming-transformers
taming-transformers-master/taming/data/annotated_objects_dataset.py
from pathlib import Path from typing import Optional, List, Callable, Dict, Any, Union import warnings import PIL.Image as pil_image from torch import Tensor from torch.utils.data import Dataset from torchvision import transforms from taming.data.conditional_builder.objects_bbox import ObjectsBoundingBoxConditionalBuilder from taming.data.conditional_builder.objects_center_points import ObjectsCenterPointsConditionalBuilder from taming.data.conditional_builder.utils import load_object_from_string from taming.data.helper_types import BoundingBox, CropMethodType, Image, Annotation, SplitType from taming.data.image_transforms import CenterCropReturnCoordinates, RandomCrop1dReturnCoordinates, \ Random2dCropReturnCoordinates, RandomHorizontalFlipReturn, convert_pil_to_tensor class AnnotatedObjectsDataset(Dataset): def __init__(self, data_path: Union[str, Path], split: SplitType, keys: List[str], target_image_size: int, min_object_area: float, min_objects_per_image: int, max_objects_per_image: int, crop_method: CropMethodType, random_flip: bool, no_tokens: int, use_group_parameter: bool, encode_crop: bool, category_allow_list_target: str = "", category_mapping_target: str = "", no_object_classes: Optional[int] = None): self.data_path = data_path self.split = split self.keys = keys self.target_image_size = target_image_size self.min_object_area = min_object_area self.min_objects_per_image = min_objects_per_image self.max_objects_per_image = max_objects_per_image self.crop_method = crop_method self.random_flip = random_flip self.no_tokens = no_tokens self.use_group_parameter = use_group_parameter self.encode_crop = encode_crop self.annotations = None self.image_descriptions = None self.categories = None self.category_ids = None self.category_number = None self.image_ids = None self.transform_functions: List[Callable] = self.setup_transform(target_image_size, crop_method, random_flip) self.paths = self.build_paths(self.data_path) self._conditional_builders = None self.category_allow_list = None if category_allow_list_target: allow_list = load_object_from_string(category_allow_list_target) self.category_allow_list = {name for name, _ in allow_list} self.category_mapping = {} if category_mapping_target: self.category_mapping = load_object_from_string(category_mapping_target) self.no_object_classes = no_object_classes def build_paths(self, top_level: Union[str, Path]) -> Dict[str, Path]: top_level = Path(top_level) sub_paths = {name: top_level.joinpath(sub_path) for name, sub_path in self.get_path_structure().items()} for path in sub_paths.values(): if not path.exists(): raise FileNotFoundError(f'{type(self).__name__} data structure error: [{path}] does not exist.') return sub_paths @staticmethod def load_image_from_disk(path: Path) -> Image: return pil_image.open(path).convert('RGB') @staticmethod def setup_transform(target_image_size: int, crop_method: CropMethodType, random_flip: bool): transform_functions = [] if crop_method == 'none': transform_functions.append(transforms.Resize((target_image_size, target_image_size))) elif crop_method == 'center': transform_functions.extend([ transforms.Resize(target_image_size), CenterCropReturnCoordinates(target_image_size) ]) elif crop_method == 'random-1d': transform_functions.extend([ transforms.Resize(target_image_size), RandomCrop1dReturnCoordinates(target_image_size) ]) elif crop_method == 'random-2d': transform_functions.extend([ Random2dCropReturnCoordinates(target_image_size), transforms.Resize(target_image_size) ]) elif crop_method is None: return None else: raise ValueError(f'Received invalid crop method [{crop_method}].') if random_flip: transform_functions.append(RandomHorizontalFlipReturn()) transform_functions.append(transforms.Lambda(lambda x: x / 127.5 - 1.)) return transform_functions def image_transform(self, x: Tensor) -> (Optional[BoundingBox], Optional[bool], Tensor): crop_bbox = None flipped = None for t in self.transform_functions: if isinstance(t, (RandomCrop1dReturnCoordinates, CenterCropReturnCoordinates, Random2dCropReturnCoordinates)): crop_bbox, x = t(x) elif isinstance(t, RandomHorizontalFlipReturn): flipped, x = t(x) else: x = t(x) return crop_bbox, flipped, x @property def no_classes(self) -> int: return self.no_object_classes if self.no_object_classes else len(self.categories) @property def conditional_builders(self) -> ObjectsCenterPointsConditionalBuilder: # cannot set this up in init because no_classes is only known after loading data in init of superclass if self._conditional_builders is None: self._conditional_builders = { 'objects_center_points': ObjectsCenterPointsConditionalBuilder( self.no_classes, self.max_objects_per_image, self.no_tokens, self.encode_crop, self.use_group_parameter, getattr(self, 'use_additional_parameters', False) ), 'objects_bbox': ObjectsBoundingBoxConditionalBuilder( self.no_classes, self.max_objects_per_image, self.no_tokens, self.encode_crop, self.use_group_parameter, getattr(self, 'use_additional_parameters', False) ) } return self._conditional_builders def filter_categories(self) -> None: if self.category_allow_list: self.categories = {id_: cat for id_, cat in self.categories.items() if cat.name in self.category_allow_list} if self.category_mapping: self.categories = {id_: cat for id_, cat in self.categories.items() if cat.id not in self.category_mapping} def setup_category_id_and_number(self) -> None: self.category_ids = list(self.categories.keys()) self.category_ids.sort() if '/m/01s55n' in self.category_ids: self.category_ids.remove('/m/01s55n') self.category_ids.append('/m/01s55n') self.category_number = {category_id: i for i, category_id in enumerate(self.category_ids)} if self.category_allow_list is not None and self.category_mapping is None \ and len(self.category_ids) != len(self.category_allow_list): warnings.warn('Unexpected number of categories: Mismatch with category_allow_list. ' 'Make sure all names in category_allow_list exist.') def clean_up_annotations_and_image_descriptions(self) -> None: image_id_set = set(self.image_ids) self.annotations = {k: v for k, v in self.annotations.items() if k in image_id_set} self.image_descriptions = {k: v for k, v in self.image_descriptions.items() if k in image_id_set} @staticmethod def filter_object_number(all_annotations: Dict[str, List[Annotation]], min_object_area: float, min_objects_per_image: int, max_objects_per_image: int) -> Dict[str, List[Annotation]]: filtered = {} for image_id, annotations in all_annotations.items(): annotations_with_min_area = [a for a in annotations if a.area > min_object_area] if min_objects_per_image <= len(annotations_with_min_area) <= max_objects_per_image: filtered[image_id] = annotations_with_min_area return filtered def __len__(self): return len(self.image_ids) def __getitem__(self, n: int) -> Dict[str, Any]: image_id = self.get_image_id(n) sample = self.get_image_description(image_id) sample['annotations'] = self.get_annotation(image_id) if 'image' in self.keys: sample['image_path'] = str(self.get_image_path(image_id)) sample['image'] = self.load_image_from_disk(sample['image_path']) sample['image'] = convert_pil_to_tensor(sample['image']) sample['crop_bbox'], sample['flipped'], sample['image'] = self.image_transform(sample['image']) sample['image'] = sample['image'].permute(1, 2, 0) for conditional, builder in self.conditional_builders.items(): if conditional in self.keys: sample[conditional] = builder.build(sample['annotations'], sample['crop_bbox'], sample['flipped']) if self.keys: # only return specified keys sample = {key: sample[key] for key in self.keys} return sample def get_image_id(self, no: int) -> str: return self.image_ids[no] def get_annotation(self, image_id: str) -> str: return self.annotations[image_id] def get_textual_label_for_category_id(self, category_id: str) -> str: return self.categories[category_id].name def get_textual_label_for_category_no(self, category_no: int) -> str: return self.categories[self.get_category_id(category_no)].name def get_category_number(self, category_id: str) -> int: return self.category_number[category_id] def get_category_id(self, category_no: int) -> str: return self.category_ids[category_no] def get_image_description(self, image_id: str) -> Dict[str, Any]: raise NotImplementedError() def get_path_structure(self): raise NotImplementedError def get_image_path(self, image_id: str) -> Path: raise NotImplementedError
10,226
45.69863
122
py
taming-transformers
taming-transformers-master/taming/data/sflckr.py
import os import numpy as np import cv2 import albumentations from PIL import Image from torch.utils.data import Dataset class SegmentationBase(Dataset): def __init__(self, data_csv, data_root, segmentation_root, size=None, random_crop=False, interpolation="bicubic", n_labels=182, shift_segmentation=False, ): self.n_labels = n_labels self.shift_segmentation = shift_segmentation self.data_csv = data_csv self.data_root = data_root self.segmentation_root = segmentation_root with open(self.data_csv, "r") as f: self.image_paths = f.read().splitlines() self._length = len(self.image_paths) self.labels = { "relative_file_path_": [l for l in self.image_paths], "file_path_": [os.path.join(self.data_root, l) for l in self.image_paths], "segmentation_path_": [os.path.join(self.segmentation_root, l.replace(".jpg", ".png")) for l in self.image_paths] } size = None if size is not None and size<=0 else size self.size = size if self.size is not None: self.interpolation = interpolation self.interpolation = { "nearest": cv2.INTER_NEAREST, "bilinear": cv2.INTER_LINEAR, "bicubic": cv2.INTER_CUBIC, "area": cv2.INTER_AREA, "lanczos": cv2.INTER_LANCZOS4}[self.interpolation] self.image_rescaler = albumentations.SmallestMaxSize(max_size=self.size, interpolation=self.interpolation) self.segmentation_rescaler = albumentations.SmallestMaxSize(max_size=self.size, interpolation=cv2.INTER_NEAREST) self.center_crop = not random_crop if self.center_crop: self.cropper = albumentations.CenterCrop(height=self.size, width=self.size) else: self.cropper = albumentations.RandomCrop(height=self.size, width=self.size) self.preprocessor = self.cropper def __len__(self): return self._length def __getitem__(self, i): example = dict((k, self.labels[k][i]) for k in self.labels) image = Image.open(example["file_path_"]) if not image.mode == "RGB": image = image.convert("RGB") image = np.array(image).astype(np.uint8) if self.size is not None: image = self.image_rescaler(image=image)["image"] segmentation = Image.open(example["segmentation_path_"]) assert segmentation.mode == "L", segmentation.mode segmentation = np.array(segmentation).astype(np.uint8) if self.shift_segmentation: # used to support segmentations containing unlabeled==255 label segmentation = segmentation+1 if self.size is not None: segmentation = self.segmentation_rescaler(image=segmentation)["image"] if self.size is not None: processed = self.preprocessor(image=image, mask=segmentation ) else: processed = {"image": image, "mask": segmentation } example["image"] = (processed["image"]/127.5 - 1.0).astype(np.float32) segmentation = processed["mask"] onehot = np.eye(self.n_labels)[segmentation] example["segmentation"] = onehot return example class Examples(SegmentationBase): def __init__(self, size=None, random_crop=False, interpolation="bicubic"): super().__init__(data_csv="data/sflckr_examples.txt", data_root="data/sflckr_images", segmentation_root="data/sflckr_segmentations", size=size, random_crop=random_crop, interpolation=interpolation)
4,097
43.543478
104
py
taming-transformers
taming-transformers-master/taming/data/imagenet.py
import os, tarfile, glob, shutil import yaml import numpy as np from tqdm import tqdm from PIL import Image import albumentations from omegaconf import OmegaConf from torch.utils.data import Dataset from taming.data.base import ImagePaths from taming.util import download, retrieve import taming.data.utils as bdu def give_synsets_from_indices(indices, path_to_yaml="data/imagenet_idx_to_synset.yaml"): synsets = [] with open(path_to_yaml) as f: di2s = yaml.load(f) for idx in indices: synsets.append(str(di2s[idx])) print("Using {} different synsets for construction of Restriced Imagenet.".format(len(synsets))) return synsets def str_to_indices(string): """Expects a string in the format '32-123, 256, 280-321'""" assert not string.endswith(","), "provided string '{}' ends with a comma, pls remove it".format(string) subs = string.split(",") indices = [] for sub in subs: subsubs = sub.split("-") assert len(subsubs) > 0 if len(subsubs) == 1: indices.append(int(subsubs[0])) else: rang = [j for j in range(int(subsubs[0]), int(subsubs[1]))] indices.extend(rang) return sorted(indices) class ImageNetBase(Dataset): def __init__(self, config=None): self.config = config or OmegaConf.create() if not type(self.config)==dict: self.config = OmegaConf.to_container(self.config) self._prepare() self._prepare_synset_to_human() self._prepare_idx_to_synset() self._load() def __len__(self): return len(self.data) def __getitem__(self, i): return self.data[i] def _prepare(self): raise NotImplementedError() def _filter_relpaths(self, relpaths): ignore = set([ "n06596364_9591.JPEG", ]) relpaths = [rpath for rpath in relpaths if not rpath.split("/")[-1] in ignore] if "sub_indices" in self.config: indices = str_to_indices(self.config["sub_indices"]) synsets = give_synsets_from_indices(indices, path_to_yaml=self.idx2syn) # returns a list of strings files = [] for rpath in relpaths: syn = rpath.split("/")[0] if syn in synsets: files.append(rpath) return files else: return relpaths def _prepare_synset_to_human(self): SIZE = 2655750 URL = "https://heibox.uni-heidelberg.de/f/9f28e956cd304264bb82/?dl=1" self.human_dict = os.path.join(self.root, "synset_human.txt") if (not os.path.exists(self.human_dict) or not os.path.getsize(self.human_dict)==SIZE): download(URL, self.human_dict) def _prepare_idx_to_synset(self): URL = "https://heibox.uni-heidelberg.de/f/d835d5b6ceda4d3aa910/?dl=1" self.idx2syn = os.path.join(self.root, "index_synset.yaml") if (not os.path.exists(self.idx2syn)): download(URL, self.idx2syn) def _load(self): with open(self.txt_filelist, "r") as f: self.relpaths = f.read().splitlines() l1 = len(self.relpaths) self.relpaths = self._filter_relpaths(self.relpaths) print("Removed {} files from filelist during filtering.".format(l1 - len(self.relpaths))) self.synsets = [p.split("/")[0] for p in self.relpaths] self.abspaths = [os.path.join(self.datadir, p) for p in self.relpaths] unique_synsets = np.unique(self.synsets) class_dict = dict((synset, i) for i, synset in enumerate(unique_synsets)) self.class_labels = [class_dict[s] for s in self.synsets] with open(self.human_dict, "r") as f: human_dict = f.read().splitlines() human_dict = dict(line.split(maxsplit=1) for line in human_dict) self.human_labels = [human_dict[s] for s in self.synsets] labels = { "relpath": np.array(self.relpaths), "synsets": np.array(self.synsets), "class_label": np.array(self.class_labels), "human_label": np.array(self.human_labels), } self.data = ImagePaths(self.abspaths, labels=labels, size=retrieve(self.config, "size", default=0), random_crop=self.random_crop) class ImageNetTrain(ImageNetBase): NAME = "ILSVRC2012_train" URL = "http://www.image-net.org/challenges/LSVRC/2012/" AT_HASH = "a306397ccf9c2ead27155983c254227c0fd938e2" FILES = [ "ILSVRC2012_img_train.tar", ] SIZES = [ 147897477120, ] def _prepare(self): self.random_crop = retrieve(self.config, "ImageNetTrain/random_crop", default=True) cachedir = os.environ.get("XDG_CACHE_HOME", os.path.expanduser("~/.cache")) self.root = os.path.join(cachedir, "autoencoders/data", self.NAME) self.datadir = os.path.join(self.root, "data") self.txt_filelist = os.path.join(self.root, "filelist.txt") self.expected_length = 1281167 if not bdu.is_prepared(self.root): # prep print("Preparing dataset {} in {}".format(self.NAME, self.root)) datadir = self.datadir if not os.path.exists(datadir): path = os.path.join(self.root, self.FILES[0]) if not os.path.exists(path) or not os.path.getsize(path)==self.SIZES[0]: import academictorrents as at atpath = at.get(self.AT_HASH, datastore=self.root) assert atpath == path print("Extracting {} to {}".format(path, datadir)) os.makedirs(datadir, exist_ok=True) with tarfile.open(path, "r:") as tar: tar.extractall(path=datadir) print("Extracting sub-tars.") subpaths = sorted(glob.glob(os.path.join(datadir, "*.tar"))) for subpath in tqdm(subpaths): subdir = subpath[:-len(".tar")] os.makedirs(subdir, exist_ok=True) with tarfile.open(subpath, "r:") as tar: tar.extractall(path=subdir) filelist = glob.glob(os.path.join(datadir, "**", "*.JPEG")) filelist = [os.path.relpath(p, start=datadir) for p in filelist] filelist = sorted(filelist) filelist = "\n".join(filelist)+"\n" with open(self.txt_filelist, "w") as f: f.write(filelist) bdu.mark_prepared(self.root) class ImageNetValidation(ImageNetBase): NAME = "ILSVRC2012_validation" URL = "http://www.image-net.org/challenges/LSVRC/2012/" AT_HASH = "5d6d0df7ed81efd49ca99ea4737e0ae5e3a5f2e5" VS_URL = "https://heibox.uni-heidelberg.de/f/3e0f6e9c624e45f2bd73/?dl=1" FILES = [ "ILSVRC2012_img_val.tar", "validation_synset.txt", ] SIZES = [ 6744924160, 1950000, ] def _prepare(self): self.random_crop = retrieve(self.config, "ImageNetValidation/random_crop", default=False) cachedir = os.environ.get("XDG_CACHE_HOME", os.path.expanduser("~/.cache")) self.root = os.path.join(cachedir, "autoencoders/data", self.NAME) self.datadir = os.path.join(self.root, "data") self.txt_filelist = os.path.join(self.root, "filelist.txt") self.expected_length = 50000 if not bdu.is_prepared(self.root): # prep print("Preparing dataset {} in {}".format(self.NAME, self.root)) datadir = self.datadir if not os.path.exists(datadir): path = os.path.join(self.root, self.FILES[0]) if not os.path.exists(path) or not os.path.getsize(path)==self.SIZES[0]: import academictorrents as at atpath = at.get(self.AT_HASH, datastore=self.root) assert atpath == path print("Extracting {} to {}".format(path, datadir)) os.makedirs(datadir, exist_ok=True) with tarfile.open(path, "r:") as tar: tar.extractall(path=datadir) vspath = os.path.join(self.root, self.FILES[1]) if not os.path.exists(vspath) or not os.path.getsize(vspath)==self.SIZES[1]: download(self.VS_URL, vspath) with open(vspath, "r") as f: synset_dict = f.read().splitlines() synset_dict = dict(line.split() for line in synset_dict) print("Reorganizing into synset folders") synsets = np.unique(list(synset_dict.values())) for s in synsets: os.makedirs(os.path.join(datadir, s), exist_ok=True) for k, v in synset_dict.items(): src = os.path.join(datadir, k) dst = os.path.join(datadir, v) shutil.move(src, dst) filelist = glob.glob(os.path.join(datadir, "**", "*.JPEG")) filelist = [os.path.relpath(p, start=datadir) for p in filelist] filelist = sorted(filelist) filelist = "\n".join(filelist)+"\n" with open(self.txt_filelist, "w") as f: f.write(filelist) bdu.mark_prepared(self.root) def get_preprocessor(size=None, random_crop=False, additional_targets=None, crop_size=None): if size is not None and size > 0: transforms = list() rescaler = albumentations.SmallestMaxSize(max_size = size) transforms.append(rescaler) if not random_crop: cropper = albumentations.CenterCrop(height=size,width=size) transforms.append(cropper) else: cropper = albumentations.RandomCrop(height=size,width=size) transforms.append(cropper) flipper = albumentations.HorizontalFlip() transforms.append(flipper) preprocessor = albumentations.Compose(transforms, additional_targets=additional_targets) elif crop_size is not None and crop_size > 0: if not random_crop: cropper = albumentations.CenterCrop(height=crop_size,width=crop_size) else: cropper = albumentations.RandomCrop(height=crop_size,width=crop_size) transforms = [cropper] preprocessor = albumentations.Compose(transforms, additional_targets=additional_targets) else: preprocessor = lambda **kwargs: kwargs return preprocessor def rgba_to_depth(x): assert x.dtype == np.uint8 assert len(x.shape) == 3 and x.shape[2] == 4 y = x.copy() y.dtype = np.float32 y = y.reshape(x.shape[:2]) return np.ascontiguousarray(y) class BaseWithDepth(Dataset): DEFAULT_DEPTH_ROOT="data/imagenet_depth" def __init__(self, config=None, size=None, random_crop=False, crop_size=None, root=None): self.config = config self.base_dset = self.get_base_dset() self.preprocessor = get_preprocessor( size=size, crop_size=crop_size, random_crop=random_crop, additional_targets={"depth": "image"}) self.crop_size = crop_size if self.crop_size is not None: self.rescaler = albumentations.Compose( [albumentations.SmallestMaxSize(max_size = self.crop_size)], additional_targets={"depth": "image"}) if root is not None: self.DEFAULT_DEPTH_ROOT = root def __len__(self): return len(self.base_dset) def preprocess_depth(self, path): rgba = np.array(Image.open(path)) depth = rgba_to_depth(rgba) depth = (depth - depth.min())/max(1e-8, depth.max()-depth.min()) depth = 2.0*depth-1.0 return depth def __getitem__(self, i): e = self.base_dset[i] e["depth"] = self.preprocess_depth(self.get_depth_path(e)) # up if necessary h,w,c = e["image"].shape if self.crop_size and min(h,w) < self.crop_size: # have to upscale to be able to crop - this just uses bilinear out = self.rescaler(image=e["image"], depth=e["depth"]) e["image"] = out["image"] e["depth"] = out["depth"] transformed = self.preprocessor(image=e["image"], depth=e["depth"]) e["image"] = transformed["image"] e["depth"] = transformed["depth"] return e class ImageNetTrainWithDepth(BaseWithDepth): # default to random_crop=True def __init__(self, random_crop=True, sub_indices=None, **kwargs): self.sub_indices = sub_indices super().__init__(random_crop=random_crop, **kwargs) def get_base_dset(self): if self.sub_indices is None: return ImageNetTrain() else: return ImageNetTrain({"sub_indices": self.sub_indices}) def get_depth_path(self, e): fid = os.path.splitext(e["relpath"])[0]+".png" fid = os.path.join(self.DEFAULT_DEPTH_ROOT, "train", fid) return fid class ImageNetValidationWithDepth(BaseWithDepth): def __init__(self, sub_indices=None, **kwargs): self.sub_indices = sub_indices super().__init__(**kwargs) def get_base_dset(self): if self.sub_indices is None: return ImageNetValidation() else: return ImageNetValidation({"sub_indices": self.sub_indices}) def get_depth_path(self, e): fid = os.path.splitext(e["relpath"])[0]+".png" fid = os.path.join(self.DEFAULT_DEPTH_ROOT, "val", fid) return fid class RINTrainWithDepth(ImageNetTrainWithDepth): def __init__(self, config=None, size=None, random_crop=True, crop_size=None): sub_indices = "30-32, 33-37, 151-268, 281-285, 80-100, 365-382, 389-397, 118-121, 300-319" super().__init__(config=config, size=size, random_crop=random_crop, sub_indices=sub_indices, crop_size=crop_size) class RINValidationWithDepth(ImageNetValidationWithDepth): def __init__(self, config=None, size=None, random_crop=False, crop_size=None): sub_indices = "30-32, 33-37, 151-268, 281-285, 80-100, 365-382, 389-397, 118-121, 300-319" super().__init__(config=config, size=size, random_crop=random_crop, sub_indices=sub_indices, crop_size=crop_size) class DRINExamples(Dataset): def __init__(self): self.preprocessor = get_preprocessor(size=256, additional_targets={"depth": "image"}) with open("data/drin_examples.txt", "r") as f: relpaths = f.read().splitlines() self.image_paths = [os.path.join("data/drin_images", relpath) for relpath in relpaths] self.depth_paths = [os.path.join("data/drin_depth", relpath.replace(".JPEG", ".png")) for relpath in relpaths] def __len__(self): return len(self.image_paths) def preprocess_image(self, image_path): image = Image.open(image_path) if not image.mode == "RGB": image = image.convert("RGB") image = np.array(image).astype(np.uint8) image = self.preprocessor(image=image)["image"] image = (image/127.5 - 1.0).astype(np.float32) return image def preprocess_depth(self, path): rgba = np.array(Image.open(path)) depth = rgba_to_depth(rgba) depth = (depth - depth.min())/max(1e-8, depth.max()-depth.min()) depth = 2.0*depth-1.0 return depth def __getitem__(self, i): e = dict() e["image"] = self.preprocess_image(self.image_paths[i]) e["depth"] = self.preprocess_depth(self.depth_paths[i]) transformed = self.preprocessor(image=e["image"], depth=e["depth"]) e["image"] = transformed["image"] e["depth"] = transformed["depth"] return e def imscale(x, factor, keepshapes=False, keepmode="bicubic"): if factor is None or factor==1: return x dtype = x.dtype assert dtype in [np.float32, np.float64] assert x.min() >= -1 assert x.max() <= 1 keepmode = {"nearest": Image.NEAREST, "bilinear": Image.BILINEAR, "bicubic": Image.BICUBIC}[keepmode] lr = (x+1.0)*127.5 lr = lr.clip(0,255).astype(np.uint8) lr = Image.fromarray(lr) h, w, _ = x.shape nh = h//factor nw = w//factor assert nh > 0 and nw > 0, (nh, nw) lr = lr.resize((nw,nh), Image.BICUBIC) if keepshapes: lr = lr.resize((w,h), keepmode) lr = np.array(lr)/127.5-1.0 lr = lr.astype(dtype) return lr class ImageNetScale(Dataset): def __init__(self, size=None, crop_size=None, random_crop=False, up_factor=None, hr_factor=None, keep_mode="bicubic"): self.base = self.get_base() self.size = size self.crop_size = crop_size if crop_size is not None else self.size self.random_crop = random_crop self.up_factor = up_factor self.hr_factor = hr_factor self.keep_mode = keep_mode transforms = list() if self.size is not None and self.size > 0: rescaler = albumentations.SmallestMaxSize(max_size = self.size) self.rescaler = rescaler transforms.append(rescaler) if self.crop_size is not None and self.crop_size > 0: if len(transforms) == 0: self.rescaler = albumentations.SmallestMaxSize(max_size = self.crop_size) if not self.random_crop: cropper = albumentations.CenterCrop(height=self.crop_size,width=self.crop_size) else: cropper = albumentations.RandomCrop(height=self.crop_size,width=self.crop_size) transforms.append(cropper) if len(transforms) > 0: if self.up_factor is not None: additional_targets = {"lr": "image"} else: additional_targets = None self.preprocessor = albumentations.Compose(transforms, additional_targets=additional_targets) else: self.preprocessor = lambda **kwargs: kwargs def __len__(self): return len(self.base) def __getitem__(self, i): example = self.base[i] image = example["image"] # adjust resolution image = imscale(image, self.hr_factor, keepshapes=False) h,w,c = image.shape if self.crop_size and min(h,w) < self.crop_size: # have to upscale to be able to crop - this just uses bilinear image = self.rescaler(image=image)["image"] if self.up_factor is None: image = self.preprocessor(image=image)["image"] example["image"] = image else: lr = imscale(image, self.up_factor, keepshapes=True, keepmode=self.keep_mode) out = self.preprocessor(image=image, lr=lr) example["image"] = out["image"] example["lr"] = out["lr"] return example class ImageNetScaleTrain(ImageNetScale): def __init__(self, random_crop=True, **kwargs): super().__init__(random_crop=random_crop, **kwargs) def get_base(self): return ImageNetTrain() class ImageNetScaleValidation(ImageNetScale): def get_base(self): return ImageNetValidation() from skimage.feature import canny from skimage.color import rgb2gray class ImageNetEdges(ImageNetScale): def __init__(self, up_factor=1, **kwargs): super().__init__(up_factor=1, **kwargs) def __getitem__(self, i): example = self.base[i] image = example["image"] h,w,c = image.shape if self.crop_size and min(h,w) < self.crop_size: # have to upscale to be able to crop - this just uses bilinear image = self.rescaler(image=image)["image"] lr = canny(rgb2gray(image), sigma=2) lr = lr.astype(np.float32) lr = lr[:,:,None][:,:,[0,0,0]] out = self.preprocessor(image=image, lr=lr) example["image"] = out["image"] example["lr"] = out["lr"] return example class ImageNetEdgesTrain(ImageNetEdges): def __init__(self, random_crop=True, **kwargs): super().__init__(random_crop=random_crop, **kwargs) def get_base(self): return ImageNetTrain() class ImageNetEdgesValidation(ImageNetEdges): def get_base(self): return ImageNetValidation()
20,815
36.237925
112
py
taming-transformers
taming-transformers-master/taming/data/coco.py
import os import json import albumentations import numpy as np from PIL import Image from tqdm import tqdm from torch.utils.data import Dataset from taming.data.sflckr import SegmentationBase # for examples included in repo class Examples(SegmentationBase): def __init__(self, size=256, random_crop=False, interpolation="bicubic"): super().__init__(data_csv="data/coco_examples.txt", data_root="data/coco_images", segmentation_root="data/coco_segmentations", size=size, random_crop=random_crop, interpolation=interpolation, n_labels=183, shift_segmentation=True) class CocoBase(Dataset): """needed for (image, caption, segmentation) pairs""" def __init__(self, size=None, dataroot="", datajson="", onehot_segmentation=False, use_stuffthing=False, crop_size=None, force_no_crop=False, given_files=None): self.split = self.get_split() self.size = size if crop_size is None: self.crop_size = size else: self.crop_size = crop_size self.onehot = onehot_segmentation # return segmentation as rgb or one hot self.stuffthing = use_stuffthing # include thing in segmentation if self.onehot and not self.stuffthing: raise NotImplemented("One hot mode is only supported for the " "stuffthings version because labels are stored " "a bit different.") data_json = datajson with open(data_json) as json_file: self.json_data = json.load(json_file) self.img_id_to_captions = dict() self.img_id_to_filepath = dict() self.img_id_to_segmentation_filepath = dict() assert data_json.split("/")[-1] in ["captions_train2017.json", "captions_val2017.json"] if self.stuffthing: self.segmentation_prefix = ( "data/cocostuffthings/val2017" if data_json.endswith("captions_val2017.json") else "data/cocostuffthings/train2017") else: self.segmentation_prefix = ( "data/coco/annotations/stuff_val2017_pixelmaps" if data_json.endswith("captions_val2017.json") else "data/coco/annotations/stuff_train2017_pixelmaps") imagedirs = self.json_data["images"] self.labels = {"image_ids": list()} for imgdir in tqdm(imagedirs, desc="ImgToPath"): self.img_id_to_filepath[imgdir["id"]] = os.path.join(dataroot, imgdir["file_name"]) self.img_id_to_captions[imgdir["id"]] = list() pngfilename = imgdir["file_name"].replace("jpg", "png") self.img_id_to_segmentation_filepath[imgdir["id"]] = os.path.join( self.segmentation_prefix, pngfilename) if given_files is not None: if pngfilename in given_files: self.labels["image_ids"].append(imgdir["id"]) else: self.labels["image_ids"].append(imgdir["id"]) capdirs = self.json_data["annotations"] for capdir in tqdm(capdirs, desc="ImgToCaptions"): # there are in average 5 captions per image self.img_id_to_captions[capdir["image_id"]].append(np.array([capdir["caption"]])) self.rescaler = albumentations.SmallestMaxSize(max_size=self.size) if self.split=="validation": self.cropper = albumentations.CenterCrop(height=self.crop_size, width=self.crop_size) else: self.cropper = albumentations.RandomCrop(height=self.crop_size, width=self.crop_size) self.preprocessor = albumentations.Compose( [self.rescaler, self.cropper], additional_targets={"segmentation": "image"}) if force_no_crop: self.rescaler = albumentations.Resize(height=self.size, width=self.size) self.preprocessor = albumentations.Compose( [self.rescaler], additional_targets={"segmentation": "image"}) def __len__(self): return len(self.labels["image_ids"]) def preprocess_image(self, image_path, segmentation_path): image = Image.open(image_path) if not image.mode == "RGB": image = image.convert("RGB") image = np.array(image).astype(np.uint8) segmentation = Image.open(segmentation_path) if not self.onehot and not segmentation.mode == "RGB": segmentation = segmentation.convert("RGB") segmentation = np.array(segmentation).astype(np.uint8) if self.onehot: assert self.stuffthing # stored in caffe format: unlabeled==255. stuff and thing from # 0-181. to be compatible with the labels in # https://github.com/nightrome/cocostuff/blob/master/labels.txt # we shift stuffthing one to the right and put unlabeled in zero # as long as segmentation is uint8 shifting to right handles the # latter too assert segmentation.dtype == np.uint8 segmentation = segmentation + 1 processed = self.preprocessor(image=image, segmentation=segmentation) image, segmentation = processed["image"], processed["segmentation"] image = (image / 127.5 - 1.0).astype(np.float32) if self.onehot: assert segmentation.dtype == np.uint8 # make it one hot n_labels = 183 flatseg = np.ravel(segmentation) onehot = np.zeros((flatseg.size, n_labels), dtype=np.bool) onehot[np.arange(flatseg.size), flatseg] = True onehot = onehot.reshape(segmentation.shape + (n_labels,)).astype(int) segmentation = onehot else: segmentation = (segmentation / 127.5 - 1.0).astype(np.float32) return image, segmentation def __getitem__(self, i): img_path = self.img_id_to_filepath[self.labels["image_ids"][i]] seg_path = self.img_id_to_segmentation_filepath[self.labels["image_ids"][i]] image, segmentation = self.preprocess_image(img_path, seg_path) captions = self.img_id_to_captions[self.labels["image_ids"][i]] # randomly draw one of all available captions per image caption = captions[np.random.randint(0, len(captions))] example = {"image": image, "caption": [str(caption[0])], "segmentation": segmentation, "img_path": img_path, "seg_path": seg_path, "filename_": img_path.split(os.sep)[-1] } return example class CocoImagesAndCaptionsTrain(CocoBase): """returns a pair of (image, caption)""" def __init__(self, size, onehot_segmentation=False, use_stuffthing=False, crop_size=None, force_no_crop=False): super().__init__(size=size, dataroot="data/coco/train2017", datajson="data/coco/annotations/captions_train2017.json", onehot_segmentation=onehot_segmentation, use_stuffthing=use_stuffthing, crop_size=crop_size, force_no_crop=force_no_crop) def get_split(self): return "train" class CocoImagesAndCaptionsValidation(CocoBase): """returns a pair of (image, caption)""" def __init__(self, size, onehot_segmentation=False, use_stuffthing=False, crop_size=None, force_no_crop=False, given_files=None): super().__init__(size=size, dataroot="data/coco/val2017", datajson="data/coco/annotations/captions_val2017.json", onehot_segmentation=onehot_segmentation, use_stuffthing=use_stuffthing, crop_size=crop_size, force_no_crop=force_no_crop, given_files=given_files) def get_split(self): return "validation"
8,121
44.887006
115
py
taming-transformers
taming-transformers-master/taming/data/image_transforms.py
import random import warnings from typing import Union import torch from torch import Tensor from torchvision.transforms import RandomCrop, functional as F, CenterCrop, RandomHorizontalFlip, PILToTensor from torchvision.transforms.functional import _get_image_size as get_image_size from taming.data.helper_types import BoundingBox, Image pil_to_tensor = PILToTensor() def convert_pil_to_tensor(image: Image) -> Tensor: with warnings.catch_warnings(): # to filter PyTorch UserWarning as described here: https://github.com/pytorch/vision/issues/2194 warnings.simplefilter("ignore") return pil_to_tensor(image) class RandomCrop1dReturnCoordinates(RandomCrop): def forward(self, img: Image) -> (BoundingBox, Image): """ Additionally to cropping, returns the relative coordinates of the crop bounding box. Args: img (PIL Image or Tensor): Image to be cropped. Returns: Bounding box: x0, y0, w, h PIL Image or Tensor: Cropped image. Based on: torchvision.transforms.RandomCrop, torchvision 1.7.0 """ if self.padding is not None: img = F.pad(img, self.padding, self.fill, self.padding_mode) width, height = get_image_size(img) # pad the width if needed if self.pad_if_needed and width < self.size[1]: padding = [self.size[1] - width, 0] img = F.pad(img, padding, self.fill, self.padding_mode) # pad the height if needed if self.pad_if_needed and height < self.size[0]: padding = [0, self.size[0] - height] img = F.pad(img, padding, self.fill, self.padding_mode) i, j, h, w = self.get_params(img, self.size) bbox = (j / width, i / height, w / width, h / height) # x0, y0, w, h return bbox, F.crop(img, i, j, h, w) class Random2dCropReturnCoordinates(torch.nn.Module): """ Additionally to cropping, returns the relative coordinates of the crop bounding box. Args: img (PIL Image or Tensor): Image to be cropped. Returns: Bounding box: x0, y0, w, h PIL Image or Tensor: Cropped image. Based on: torchvision.transforms.RandomCrop, torchvision 1.7.0 """ def __init__(self, min_size: int): super().__init__() self.min_size = min_size def forward(self, img: Image) -> (BoundingBox, Image): width, height = get_image_size(img) max_size = min(width, height) if max_size <= self.min_size: size = max_size else: size = random.randint(self.min_size, max_size) top = random.randint(0, height - size) left = random.randint(0, width - size) bbox = left / width, top / height, size / width, size / height return bbox, F.crop(img, top, left, size, size) class CenterCropReturnCoordinates(CenterCrop): @staticmethod def get_bbox_of_center_crop(width: int, height: int) -> BoundingBox: if width > height: w = height / width h = 1.0 x0 = 0.5 - w / 2 y0 = 0. else: w = 1.0 h = width / height x0 = 0. y0 = 0.5 - h / 2 return x0, y0, w, h def forward(self, img: Union[Image, Tensor]) -> (BoundingBox, Union[Image, Tensor]): """ Additionally to cropping, returns the relative coordinates of the crop bounding box. Args: img (PIL Image or Tensor): Image to be cropped. Returns: Bounding box: x0, y0, w, h PIL Image or Tensor: Cropped image. Based on: torchvision.transforms.RandomHorizontalFlip (version 1.7.0) """ width, height = get_image_size(img) return self.get_bbox_of_center_crop(width, height), F.center_crop(img, self.size) class RandomHorizontalFlipReturn(RandomHorizontalFlip): def forward(self, img: Image) -> (bool, Image): """ Additionally to flipping, returns a boolean whether it was flipped or not. Args: img (PIL Image or Tensor): Image to be flipped. Returns: flipped: whether the image was flipped or not PIL Image or Tensor: Randomly flipped image. Based on: torchvision.transforms.RandomHorizontalFlip (version 1.7.0) """ if torch.rand(1) < self.p: return True, F.hflip(img) return False, img
4,511
32.924812
109
py
taming-transformers
taming-transformers-master/taming/data/conditional_builder/objects_center_points.py
import math import random import warnings from itertools import cycle from typing import List, Optional, Tuple, Callable from PIL import Image as pil_image, ImageDraw as pil_img_draw, ImageFont from more_itertools.recipes import grouper from taming.data.conditional_builder.utils import COLOR_PALETTE, WHITE, GRAY_75, BLACK, FULL_CROP, filter_annotations, \ additional_parameters_string, horizontally_flip_bbox, pad_list, get_circle_size, get_plot_font_size, \ absolute_bbox, rescale_annotations from taming.data.helper_types import BoundingBox, Annotation from taming.data.image_transforms import convert_pil_to_tensor from torch import LongTensor, Tensor class ObjectsCenterPointsConditionalBuilder: def __init__(self, no_object_classes: int, no_max_objects: int, no_tokens: int, encode_crop: bool, use_group_parameter: bool, use_additional_parameters: bool): self.no_object_classes = no_object_classes self.no_max_objects = no_max_objects self.no_tokens = no_tokens self.encode_crop = encode_crop self.no_sections = int(math.sqrt(self.no_tokens)) self.use_group_parameter = use_group_parameter self.use_additional_parameters = use_additional_parameters @property def none(self) -> int: return self.no_tokens - 1 @property def object_descriptor_length(self) -> int: return 2 @property def embedding_dim(self) -> int: extra_length = 2 if self.encode_crop else 0 return self.no_max_objects * self.object_descriptor_length + extra_length def tokenize_coordinates(self, x: float, y: float) -> int: """ Express 2d coordinates with one number. Example: assume self.no_tokens = 16, then no_sections = 4: 0 0 0 0 0 0 # 0 0 0 0 0 0 0 0 x Then the # position corresponds to token 6, the x position to token 15. @param x: float in [0, 1] @param y: float in [0, 1] @return: discrete tokenized coordinate """ x_discrete = int(round(x * (self.no_sections - 1))) y_discrete = int(round(y * (self.no_sections - 1))) return y_discrete * self.no_sections + x_discrete def coordinates_from_token(self, token: int) -> (float, float): x = token % self.no_sections y = token // self.no_sections return x / (self.no_sections - 1), y / (self.no_sections - 1) def bbox_from_token_pair(self, token1: int, token2: int) -> BoundingBox: x0, y0 = self.coordinates_from_token(token1) x1, y1 = self.coordinates_from_token(token2) return x0, y0, x1 - x0, y1 - y0 def token_pair_from_bbox(self, bbox: BoundingBox) -> Tuple[int, int]: return self.tokenize_coordinates(bbox[0], bbox[1]), \ self.tokenize_coordinates(bbox[0] + bbox[2], bbox[1] + bbox[3]) def inverse_build(self, conditional: LongTensor) \ -> Tuple[List[Tuple[int, Tuple[float, float]]], Optional[BoundingBox]]: conditional_list = conditional.tolist() crop_coordinates = None if self.encode_crop: crop_coordinates = self.bbox_from_token_pair(conditional_list[-2], conditional_list[-1]) conditional_list = conditional_list[:-2] table_of_content = grouper(conditional_list, self.object_descriptor_length) assert conditional.shape[0] == self.embedding_dim return [ (object_tuple[0], self.coordinates_from_token(object_tuple[1])) for object_tuple in table_of_content if object_tuple[0] != self.none ], crop_coordinates def plot(self, conditional: LongTensor, label_for_category_no: Callable[[int], str], figure_size: Tuple[int, int], line_width: int = 3, font_size: Optional[int] = None) -> Tensor: plot = pil_image.new('RGB', figure_size, WHITE) draw = pil_img_draw.Draw(plot) circle_size = get_circle_size(figure_size) font = ImageFont.truetype('/usr/share/fonts/truetype/lato/Lato-Regular.ttf', size=get_plot_font_size(font_size, figure_size)) width, height = plot.size description, crop_coordinates = self.inverse_build(conditional) for (representation, (x, y)), color in zip(description, cycle(COLOR_PALETTE)): x_abs, y_abs = x * width, y * height ann = self.representation_to_annotation(representation) label = label_for_category_no(ann.category_no) + ' ' + additional_parameters_string(ann) ellipse_bbox = [x_abs - circle_size, y_abs - circle_size, x_abs + circle_size, y_abs + circle_size] draw.ellipse(ellipse_bbox, fill=color, width=0) draw.text((x_abs, y_abs), label, anchor='md', fill=BLACK, font=font) if crop_coordinates is not None: draw.rectangle(absolute_bbox(crop_coordinates, width, height), outline=GRAY_75, width=line_width) return convert_pil_to_tensor(plot) / 127.5 - 1. def object_representation(self, annotation: Annotation) -> int: modifier = 0 if self.use_group_parameter: modifier |= 1 * (annotation.is_group_of is True) if self.use_additional_parameters: modifier |= 2 * (annotation.is_occluded is True) modifier |= 4 * (annotation.is_depiction is True) modifier |= 8 * (annotation.is_inside is True) return annotation.category_no + self.no_object_classes * modifier def representation_to_annotation(self, representation: int) -> Annotation: category_no = representation % self.no_object_classes modifier = representation // self.no_object_classes # noinspection PyTypeChecker return Annotation( area=None, image_id=None, bbox=None, category_id=None, id=None, source=None, confidence=None, category_no=category_no, is_group_of=bool((modifier & 1) * self.use_group_parameter), is_occluded=bool((modifier & 2) * self.use_additional_parameters), is_depiction=bool((modifier & 4) * self.use_additional_parameters), is_inside=bool((modifier & 8) * self.use_additional_parameters) ) def _crop_encoder(self, crop_coordinates: BoundingBox) -> List[int]: return list(self.token_pair_from_bbox(crop_coordinates)) def _make_object_descriptors(self, annotations: List[Annotation]) -> List[Tuple[int, ...]]: object_tuples = [ (self.object_representation(a), self.tokenize_coordinates(a.bbox[0] + a.bbox[2] / 2, a.bbox[1] + a.bbox[3] / 2)) for a in annotations ] empty_tuple = (self.none, self.none) object_tuples = pad_list(object_tuples, empty_tuple, self.no_max_objects) return object_tuples def build(self, annotations: List, crop_coordinates: Optional[BoundingBox] = None, horizontal_flip: bool = False) \ -> LongTensor: if len(annotations) == 0: warnings.warn('Did not receive any annotations.') if len(annotations) > self.no_max_objects: warnings.warn('Received more annotations than allowed.') annotations = annotations[:self.no_max_objects] if not crop_coordinates: crop_coordinates = FULL_CROP random.shuffle(annotations) annotations = filter_annotations(annotations, crop_coordinates) if self.encode_crop: annotations = rescale_annotations(annotations, FULL_CROP, horizontal_flip) if horizontal_flip: crop_coordinates = horizontally_flip_bbox(crop_coordinates) extra = self._crop_encoder(crop_coordinates) else: annotations = rescale_annotations(annotations, crop_coordinates, horizontal_flip) extra = [] object_tuples = self._make_object_descriptors(annotations) flattened = [token for tuple_ in object_tuples for token in tuple_] + extra assert len(flattened) == self.embedding_dim assert all(0 <= value < self.no_tokens for value in flattened) return LongTensor(flattened)
8,165
47.319527
120
py
taming-transformers
taming-transformers-master/taming/data/conditional_builder/objects_bbox.py
from itertools import cycle from typing import List, Tuple, Callable, Optional from PIL import Image as pil_image, ImageDraw as pil_img_draw, ImageFont from more_itertools.recipes import grouper from taming.data.image_transforms import convert_pil_to_tensor from torch import LongTensor, Tensor from taming.data.helper_types import BoundingBox, Annotation from taming.data.conditional_builder.objects_center_points import ObjectsCenterPointsConditionalBuilder from taming.data.conditional_builder.utils import COLOR_PALETTE, WHITE, GRAY_75, BLACK, additional_parameters_string, \ pad_list, get_plot_font_size, absolute_bbox class ObjectsBoundingBoxConditionalBuilder(ObjectsCenterPointsConditionalBuilder): @property def object_descriptor_length(self) -> int: return 3 def _make_object_descriptors(self, annotations: List[Annotation]) -> List[Tuple[int, ...]]: object_triples = [ (self.object_representation(ann), *self.token_pair_from_bbox(ann.bbox)) for ann in annotations ] empty_triple = (self.none, self.none, self.none) object_triples = pad_list(object_triples, empty_triple, self.no_max_objects) return object_triples def inverse_build(self, conditional: LongTensor) -> Tuple[List[Tuple[int, BoundingBox]], Optional[BoundingBox]]: conditional_list = conditional.tolist() crop_coordinates = None if self.encode_crop: crop_coordinates = self.bbox_from_token_pair(conditional_list[-2], conditional_list[-1]) conditional_list = conditional_list[:-2] object_triples = grouper(conditional_list, 3) assert conditional.shape[0] == self.embedding_dim return [ (object_triple[0], self.bbox_from_token_pair(object_triple[1], object_triple[2])) for object_triple in object_triples if object_triple[0] != self.none ], crop_coordinates def plot(self, conditional: LongTensor, label_for_category_no: Callable[[int], str], figure_size: Tuple[int, int], line_width: int = 3, font_size: Optional[int] = None) -> Tensor: plot = pil_image.new('RGB', figure_size, WHITE) draw = pil_img_draw.Draw(plot) font = ImageFont.truetype( "/usr/share/fonts/truetype/lato/Lato-Regular.ttf", size=get_plot_font_size(font_size, figure_size) ) width, height = plot.size description, crop_coordinates = self.inverse_build(conditional) for (representation, bbox), color in zip(description, cycle(COLOR_PALETTE)): annotation = self.representation_to_annotation(representation) class_label = label_for_category_no(annotation.category_no) + ' ' + additional_parameters_string(annotation) bbox = absolute_bbox(bbox, width, height) draw.rectangle(bbox, outline=color, width=line_width) draw.text((bbox[0] + line_width, bbox[1] + line_width), class_label, anchor='la', fill=BLACK, font=font) if crop_coordinates is not None: draw.rectangle(absolute_bbox(crop_coordinates, width, height), outline=GRAY_75, width=line_width) return convert_pil_to_tensor(plot) / 127.5 - 1.
3,223
51.852459
120
py
taming-transformers
taming-transformers-master/taming/data/conditional_builder/utils.py
import importlib from typing import List, Any, Tuple, Optional from taming.data.helper_types import BoundingBox, Annotation # source: seaborn, color palette tab10 COLOR_PALETTE = [(30, 118, 179), (255, 126, 13), (43, 159, 43), (213, 38, 39), (147, 102, 188), (139, 85, 74), (226, 118, 193), (126, 126, 126), (187, 188, 33), (22, 189, 206)] BLACK = (0, 0, 0) GRAY_75 = (63, 63, 63) GRAY_50 = (127, 127, 127) GRAY_25 = (191, 191, 191) WHITE = (255, 255, 255) FULL_CROP = (0., 0., 1., 1.) def intersection_area(rectangle1: BoundingBox, rectangle2: BoundingBox) -> float: """ Give intersection area of two rectangles. @param rectangle1: (x0, y0, w, h) of first rectangle @param rectangle2: (x0, y0, w, h) of second rectangle """ rectangle1 = rectangle1[0], rectangle1[1], rectangle1[0] + rectangle1[2], rectangle1[1] + rectangle1[3] rectangle2 = rectangle2[0], rectangle2[1], rectangle2[0] + rectangle2[2], rectangle2[1] + rectangle2[3] x_overlap = max(0., min(rectangle1[2], rectangle2[2]) - max(rectangle1[0], rectangle2[0])) y_overlap = max(0., min(rectangle1[3], rectangle2[3]) - max(rectangle1[1], rectangle2[1])) return x_overlap * y_overlap def horizontally_flip_bbox(bbox: BoundingBox) -> BoundingBox: return 1 - (bbox[0] + bbox[2]), bbox[1], bbox[2], bbox[3] def absolute_bbox(relative_bbox: BoundingBox, width: int, height: int) -> Tuple[int, int, int, int]: bbox = relative_bbox bbox = bbox[0] * width, bbox[1] * height, (bbox[0] + bbox[2]) * width, (bbox[1] + bbox[3]) * height return int(bbox[0]), int(bbox[1]), int(bbox[2]), int(bbox[3]) def pad_list(list_: List, pad_element: Any, pad_to_length: int) -> List: return list_ + [pad_element for _ in range(pad_to_length - len(list_))] def rescale_annotations(annotations: List[Annotation], crop_coordinates: BoundingBox, flip: bool) -> \ List[Annotation]: def clamp(x: float): return max(min(x, 1.), 0.) def rescale_bbox(bbox: BoundingBox) -> BoundingBox: x0 = clamp((bbox[0] - crop_coordinates[0]) / crop_coordinates[2]) y0 = clamp((bbox[1] - crop_coordinates[1]) / crop_coordinates[3]) w = min(bbox[2] / crop_coordinates[2], 1 - x0) h = min(bbox[3] / crop_coordinates[3], 1 - y0) if flip: x0 = 1 - (x0 + w) return x0, y0, w, h return [a._replace(bbox=rescale_bbox(a.bbox)) for a in annotations] def filter_annotations(annotations: List[Annotation], crop_coordinates: BoundingBox) -> List: return [a for a in annotations if intersection_area(a.bbox, crop_coordinates) > 0.0] def additional_parameters_string(annotation: Annotation, short: bool = True) -> str: sl = slice(1) if short else slice(None) string = '' if not (annotation.is_group_of or annotation.is_occluded or annotation.is_depiction or annotation.is_inside): return string if annotation.is_group_of: string += 'group'[sl] + ',' if annotation.is_occluded: string += 'occluded'[sl] + ',' if annotation.is_depiction: string += 'depiction'[sl] + ',' if annotation.is_inside: string += 'inside'[sl] return '(' + string.strip(",") + ')' def get_plot_font_size(font_size: Optional[int], figure_size: Tuple[int, int]) -> int: if font_size is None: font_size = 10 if max(figure_size) >= 256: font_size = 12 if max(figure_size) >= 512: font_size = 15 return font_size def get_circle_size(figure_size: Tuple[int, int]) -> int: circle_size = 2 if max(figure_size) >= 256: circle_size = 3 if max(figure_size) >= 512: circle_size = 4 return circle_size def load_object_from_string(object_string: str) -> Any: """ Source: https://stackoverflow.com/a/10773699 """ module_name, class_name = object_string.rsplit(".", 1) return getattr(importlib.import_module(module_name), class_name)
3,963
36.396226
113
py
Fanoos
Fanoos-master/fanoos.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from handleDisplayOfCopyrightDuringInteractions import *; # run the # display and any iteraction relavent to showing the copyright terms... from config import randomSeedForNumpy, randomSeedForPython3LibRandom; # just setting the random seed here so # it can be used everywhere else.... from UI.fanoosFrontend import FanoosFrontend; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface; import sys; from UI import captureTerminalOutput; ObjDatabaseValueTracker.set_sessionUUID(); objDatabaseInterface.open(); objDatabaseInterface.executeScriptFile("databaseInterface/makeTables.sql"); terminalIOWritterObject = captureTerminalOutput.thingToWriteTo(); stdoutWrapper = captureTerminalOutput.wrapperForCopy(sys.stdout, terminalIOWritterObject, "stdout"); sys.stdout = stdoutWrapper; stderrWrapper = captureTerminalOutput.wrapperForCopy(sys.stderr, terminalIOWritterObject, "stderr"); sys.stderr = stderrWrapper; stdinWrapper = captureTerminalOutput.wrapperForCopy(sys.stdin, terminalIOWritterObject, "stdin"); sys.stdin = stdinWrapper; my_cmd = FanoosFrontend(stdin=stdinWrapper, stdout=stdoutWrapper); my_cmd.cmdloop(); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ commandToEndSessionInDatabase = \ "UPDATE sessionInfo SET dateAndTimeFinished = CURRENT_TIMESTAMP WHERE sessionUUID = '" + str(ObjDatabaseValueTracker.get_sessionUUID()) +"';"; objDatabaseInterface.exec(commandToEndSessionInDatabase); sys.stdout = stdoutWrapper.thingToWrap; sys.stderr = stderrWrapper.thingToWrap; objDatabaseInterface.commit(); objDatabaseInterface.close();
5,548
73.986486
2,785
py
Fanoos
Fanoos-master/run_drawBoxes.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from collectingResults.drawBoxes import frontEnd_drawBox; frontEnd_drawBox("tmp/resultsPresentedInPaper_CLA_boxes_0ab2cbdb-5a9e-4359-8b96-f935017311fc.bin") """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """
4,101
92.227273
2,781
py
Fanoos
Fanoos-master/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/handleDisplayOfCopyrightDuringInteractions.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); def handleDisplayOfCopyrightDuringInteractions(): fh = open("COPYRIGHT.txt", "r"); contentOfCopyRight = fh.read(); fh.close(); print(""" V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V Copyright notice ------------------------------------------------------------------------------- Note: a copy of the license terms can be found in the file LICENSE.txt distributed with this code =============================================================================== """ + \ contentOfCopyRight + \ """ ^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ """, flush=True); handleDisplayOfCopyrightDuringInteractions();
1,845
30.288136
165
py
Fanoos
Fanoos-master/config/debugFlags.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # v_print=0; # stands for verbrosity of printing. # v_print=0 - print nothing. # v_print > 0 - print various debug information. assert(isinstance(v_print, int)); assert(v_print >= 0); # If <fileName> is not listed below, then get_v_print_ForThisFile(<fileName>) # returns the default verbrosity level, v_print . If <fileName> is in the # below dictionary, then get_v_print_ForThisFile(<fileName>) returns # dictMappingFileToDebugLevel[<fileName>] as the verbrosity level. # Note that this framework could easily be extended to have matches use # regular expressions as oppossed to strict matches in a dictionary. # For now, however, this first our purposes. # # Currently, we list some examples in the below dictionary for illustration # purposes. dictMappingFileToDebugLevel = {\ "fanoos.py" : 2, \ "domainsAndConditions/classesDefiningQuestions.py" : 1, \ "CEGARLikeAnalysis/CEGARLikeAnalysisMain.py" : 1, \ "utils/distributionStatics.py" : 2 \ }; def get_v_print_ForThisFile(filePath): pathToThisConfigFile = __file__; relativePathToThisConfigFile = "config/debugFlags.py"; assert(pathToThisConfigFile.endswith(relativePathToThisConfigFile)); directoryStartPath = pathToThisConfigFile[:-len(relativePathToThisConfigFile)]; relativeFilePath = filePath; if(filePath.startswith(directoryStartPath)): relativeFilePath = filePath[len(directoryStartPath):]; return dictMappingFileToDebugLevel.get(relativeFilePath, v_print);
2,550
34.929577
165
py
Fanoos
Fanoos-master/config/randomSeedSetting.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import time; # The value upperboundExclusiveOnRandomKeys is used to keep the # random seeds in the range accepted by the libraries # envoked. For instance, the version of numpy used # requires the "seed must be between 0 and 2**32 - 1". The # "exclusive" part of the variable name indicates that the # random seeds should be strictly below it. upperboundExclusiveOnRandomKeys = 2**32; randomSeedForNumpy=int( (time.time() - 1578000000) * 100) % upperboundExclusiveOnRandomKeys; randomSeedForPython3LibRandom=int( (time.time() - 1578000000) * 100) % upperboundExclusiveOnRandomKeys; assert(isinstance(randomSeedForNumpy, int)); assert(randomSeedForNumpy >= 0); assert(randomSeedForNumpy < upperboundExclusiveOnRandomKeys); assert(isinstance(randomSeedForPython3LibRandom, int)); assert(randomSeedForPython3LibRandom >= 0); assert(randomSeedForPython3LibRandom < upperboundExclusiveOnRandomKeys); import numpy; numpy.random.seed(randomSeedForNumpy); import random; random.seed(randomSeedForPython3LibRandom); def functToRun(): # Printing the below lists helps ensure that, across runs, indeed the # random seed is changing... print(str(numpy.random.randint(0, 10, 10))); """ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg""" print(str([random.randint(0,10) for x in range(0,10)])); print("Random seeds set." ,flush=True); functToRun();
5,165
69.767123
2,791
py
Fanoos
Fanoos-master/config/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # from config.randomSeedSetting import \ randomSeedForNumpy, randomSeedForPython3LibRandom; # so that other code # may access and record these values as meta-data, if necessary. import config.defaultValues as defaultValues; import config.debugFlags as debugFlags;
1,341
32.55
165
py
Fanoos
Fanoos-master/config/defaultValues.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # # We may add more thorough descriptions of some parameters at a later # point, as well as possibly merging some of these parameters into one. # Note that the the state-operator model of Fanoos allows operators to alter # most of these parameters (outside of determining file names and database # timeouts,etc.), with the alterations stored in states. limitOnNumberOfTimesToMerge=0; #None: loop until convergence; 0: no of boxes merging ; <integer n> : at most n cycles of examining all boxes then joining some precisionForMerging=3; # higher->more precise->match more exactly numberOfSamplesToTry=5; exponentialComponentNeededWhenIncreasingAbstractionLevel=0.1; # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg formConditionToSatisfy_statistical_numberOfSamples=50; userInterface_maximumNumberOfLinesToPrintAtOneTime=15; databaseName="houseForDatabase/KB.db"; # KB is short for "Knowledge Base" numberOfStatisticalSamplesToTakeIn_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = 10; numberOfStatisticalSamplesToTakeIn_numberOfStatisticalSamplesToTakeIn_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = 10; responceDelayTimeForUnexpectedInputes=3;# this value is in seconds databaseWriteTimeoutLimit= 100; # this is in seconds.
5,051
94.320755
2,783
py
Fanoos
Fanoos-master/trainedNetworks/invertedDoublePendulumBulletEnv_v0/convertInvertedPendulumNetworkToAlternateFormat.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; import pickle; def getTrainedModelWeightAndBiases(nameOfFileToLoad): fh = open(nameOfFileToLoad, "rb"); A = pickle.load(fh); fh.close(); return A; pathToFile ="./networkLayers.pickle"; tempDict = getTrainedModelWeightAndBiases(pathToFile); def getParam(name): return np.array(tempDict["model/pi" + name + ":0"], dtype=float); def getParam_valueFunct(name): return np.array(tempDict["model/vf" + name + ":0"], dtype=float); modelToSave = {"weights" : [0,0,0], "biases" : [0,0,0]}; modelToSave["weights"][0] = getParam('_fc0/w'); modelToSave["weights"][1] = getParam('_fc1/w'); modelToSave["weights"][2] = getParam('/w'); modelToSave["biases"][0] = getParam('_fc0/b').flatten(); modelToSave["biases"][1] = getParam('_fc1/b').flatten(); modelToSave["biases"][2] = getParam('/b').flatten(); modelToSave["v_weights"] = [0,0,0]; modelToSave["v_biases"] = [0,0,0]; modelToSave["v_weights"][0] = getParam_valueFunct('_fc0/w'); modelToSave["v_weights"][1] = getParam_valueFunct('_fc1/w'); modelToSave["v_weights"][2] = getParam_valueFunct('/w'); modelToSave["v_biases"][0] = getParam_valueFunct('_fc0/b').flatten(); modelToSave["v_biases"][1] = getParam_valueFunct('_fc1/b').flatten(); modelToSave["v_biases"][2] = getParam_valueFunct('/b').flatten(); codeForClippingFunction = """ def properlyTransformAction(thisAction): requires(isinstance(thisAction, np.ndarray)); requires(len(thisAction.shape) == 1); requires(thisAction.shape[0] == 2); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg lb = np.array([ -1 , -float("inf")]); ub = np.array([ 1, float("inf")]); scaled_action = np.clip(thisAction, lb, ub); ensures(isinstance(scaled_action, np.ndarray)); ensures(len(scaled_action.shape) == 1); ensures(scaled_action.shape[0] == 2); ensures(np.all(lb <= scaled_action)); ensures(np.all(scaled_action <= ub)); return scaled_action; """; codeForPushingValuesThroughNetwork = """ def pushBoxThrough(self, thisBoxInitial): requires(getDimensionOfBox(thisBoxInitial) == 7); def normalize_obs(obs: np.ndarray) -> np.ndarray: ''' See: https://stable-baselines.readthedocs.io/en/master/_modules/stable_baselines/common/vec_env/vec_normalize.html#VecNormalize Normalize observations using this VecNormalize's observations statistics. Calling this method does not update statistics. ''' obs_rms_mean = np.array([float.fromhex(x) for x in \ ['-0x1.b77efabdeab5cp-2', '-0x1.b59fd13b14562p-4', '-0x1.c4bcd6925511fp-2', \ '0x1.fc28fe890e974p-1', '-0x1.b05cdd0a57ae1p-6', '-0x1.3f597d80568ffp-4', \ '0x1.f8352e45950f2p-1', '0x1.caea4c56e0357p-6', '0x1.7e22dba621535p-3'] ]); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg obs_rms_var = np.array([float.fromhex(x) for x in \ ['0x1.544b06860465cp-3', '0x1.ed35aa7e5a76bp-3', '0x1.4cf8dfe3c1d18p-3', \ '0x1.8d9cec886a9a1p-11', '0x1.b9fd4ce62ce65p-7', '0x1.8ddbf1a6a94adp-2', \ '0x1.a2e34153dfe07p-8', '0x1.79556f17af25ep-6', '0x1.00a9a20606534p+0'] ]); epsilon = float.fromhex('0x1.5798ee2308c3ap-27'); clip_obs = float.fromhex('0x1.4000000000000p+3'); obs = np.clip((obs - obs_rms_mean) / np.sqrt(obs_rms_var + epsilon), -clip_obs, clip_obs) return obs; def normalize_obs_box(thisBox): requires(isProperBox(thisBox)); A = normalize_obs(thisBox[:,0]); B = normalize_obs(thisBox[:,1]); boxToReturn = getBox(A, B); ensures(isProperBox(boxToReturn)); return boxToReturn; def helper_pushBoxThrough_noClipping(weights, biases, thisBox): A1 = computeBoxOutputForLayer( thisBox, weights[0], biases[0], activationFunctionName ="tanh" ); B2 = computeBoxOutputForLayer( A1, weights[1], biases[1], activationFunctionName ="tanh" ); C3 = computeBoxOutputForLayer( B2, weights[2], biases[2], activationFunctionName ="linear" ); return C3; def maximumCos(bound): # bound = np.array([start, end]) piPortions = tuple(np.floor(bound / np.pi)); assert(piPortions[1] >= piPortions[0]); if(piPortions[0] == piPortions[1]): return np.max(np.cos(bound)); if(piPortions[0] % 2 == 0): assert(piPortions[1] - piPortions[0] >= 1); if(piPortions[1] == piPortions[0] + 1): return np.max(np.cos(bound)); else: assert(piPortions[1] >= 2 + piPortions[0]); return 1; assert(piPortions[0] % 2 == 1); assert(piPortions[1] > piPortions[0]); return 1; def minimumCos(bound): # bound = np.array([start, end]) piPortions = tuple(np.floor(bound / np.pi)); assert(piPortions[1] >= piPortions[0]); if(piPortions[0] == piPortions[1]): return np.min(np.cos(bound)); if(piPortions[0] % 2 == 1): if(piPortions[1] == piPortions[0] + 1): return np.min(np.cos(bound)); else: assert(piPortions[1] >= 2 + piPortions[0]); return -1; assert(piPortions[0] % 2 == 0); assert(piPortions[1] > piPortions[0]); return -1; def pushBoundThroughCos(bound): lb = minimumCos(bound); ub = maximumCos(bound); return np.array([lb, ub]); def pushBoundThroughSin(bound): return pushBoundThroughCos(bound - (np.pi / 2)); weights = self.thisModel["weights"]; biases = self.thisModel["biases"]; v_weights = self.thisModel["v_weights"]; v_biases = self.thisModel["v_biases"]; # Below, we have thisBoxInitial[2,:] + thisBoxInitial[1,:] becuase: # thisBoxInitial[2,:] is the DELTA between the base of the robot and # x-value for the CENTER of pole2 (the code in rl-baselines-zoo # refers to it as the end of pole2, but the bullet simulator actually # get the position from the center of the pole - this behaviour # is actually explicitly considered in the code defining the inverted # doulbe-pendulum environment - see gym_pendulum_envs.py). # thisBoxInitial[1,:] is the x-value for the base of the robot # # thus, there sum gives the x-value of the pole2 "end" in respect to the # global reference frame, which is what the policy has been trained to expect. thisBox = np.array([\ thisBoxInitial[0,:], \ thisBoxInitial[1,:], \ thisBoxInitial[2,:] + thisBoxInitial[0,:], \ pushBoundThroughCos(thisBoxInitial[3,:]), \ pushBoundThroughSin(thisBoxInitial[3,:]), \ thisBoxInitial[4,:], \ pushBoundThroughCos(thisBoxInitial[5,:]), \ pushBoundThroughSin(thisBoxInitial[5,:]), \ thisBoxInitial[6,:] \ ]); thisBox = normalize_obs_box(thisBox); unclippedAction = helper_pushBoxThrough_noClipping(weights, biases, thisBox) unclippedValue = helper_pushBoxThrough_noClipping(v_weights, v_biases, thisBox) clippedAndTransformedC3 = self.clipAndScaleResultBox(getJointBox([unclippedAction, unclippedValue])); return clippedAndTransformedC3; """;""" ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg""" modelToSave["clippingFunctionCode"] = codeForClippingFunction; modelToSave["functionToPushBoxThroughModel"] = codeForPushingValuesThroughNetwork; nameOfFileToLoad = "./networkLayers_putIntoProperFormat.pickle"; fh = open(nameOfFileToLoad, "wb"); pickle.dump(modelToSave, fh); fh.close();
17,040
73.414847
2,792
py
Fanoos
Fanoos-master/trainedNetworks/modelForTesting/formModelForTesting_twoDimInput_threeDimOutput_identityFunctionAndAddition.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; from sklearn.linear_model import LinearRegression; from scipy.special import comb; from sklearn.datasets import fetch_openml; def requires(booleanStatement): assert(booleanStatement); return; def getDefaultTargetAndNonTargetData(valueNames, targetNames): requires(isinstance(targetNames, list)); requires(all([isinstance(x, str) for x in targetNames])); requires(set(valueNames).issuperset(targetNames)); requires(len(targetNames) == len(set(targetNames))); setOfTargetNames = set(targetNames); indicesOfNewTargets = np.zeros(len(valueNames), dtype=bool); for thisIndex in range(0, len(valueNames)): if(valueNames[thisIndex] in setOfTargetNames): indicesOfNewTargets[thisIndex] = 1; indicesOfNonTargets = ~indicesOfNewTargets; return {"indicesOfNewTargets" : indicesOfNewTargets, \ "indicesOfNonTargets": indicesOfNonTargets}; def normalizeThisData(thisData, indicesOfTheTrainSet): B = thisData - np.min(thisData[indicesOfTheTrainSet, :], axis=0); B = B / np.max(B, axis=0); return B; def tryDatasetWithPolynomialLinear(indicesOfNewTargets, indicesOfNonTargets): requires(isinstance(indicesOfNewTargets, np.ndarray)); requires(isinstance(indicesOfNonTargets, np.ndarray)); requires(not np.any(indicesOfNewTargets & indicesOfNonTargets)); requires(indicesOfNewTargets.dtype == np.dtype('bool')); requires(indicesOfNonTargets.dtype == np.dtype('bool')); newFeatureNames = ["in_x", "in_y"] ; newTargetData_featureNames = ["out_u", "out_v", "out_w"]; newBaseData_featureNames = newFeatureNames; """ note: the layout of coefficients is: np.array( [ [<coefficients for first output variable>], [<coefficients for second output variable>], . . . [<coefficients for nth output variable>], ]); example: input box: [[0. 1.] [0. 1.]] output: [[ 3.5 6.105 ] [-100.2 -98.99999]] """ resultsToReturn = \ {"newFeatureNames" : newFeatureNames ,\ "coefficients" : np.array([[1.0, 0.0], [0.0, 1.0], [1.0, 1.0]]) , \ "intercept" : np.array([0.0, 0.0, -1.0]) , \ "namesOfTargetValues" : newTargetData_featureNames, \ "orderOfNonFeaturizedObservationNames" : newBaseData_featureNames \ }; return resultsToReturn; import pickle; valueNames = ["in_x", "in_y", "out_u", "out_v", "out_w"]; targetNames =["out_u", "out_v", "out_w"] A = getDefaultTargetAndNonTargetData(valueNames, targetNames); trainedModelInfoDicts = tryDatasetWithPolynomialLinear(\ A["indicesOfNewTargets"], \ A["indicesOfNonTargets"] ); # TODO: see the todos at the end of the code for preparing the one # dimensional identity function... fh = open("./modelForTesting_twoDimInput_threeDimOutput_identityFunctionAndAddition.pickle", "wb"); pickle.dump(trainedModelInfoDicts, fh); fh.close();
4,182
29.311594
165
py
Fanoos
Fanoos-master/trainedNetworks/modelForTesting/formModelForTesting_oneDimInput_oneDimOutput_identityFunction.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; from sklearn.linear_model import LinearRegression; from scipy.special import comb; from sklearn.datasets import fetch_openml; def requires(booleanStatement): assert(booleanStatement); return; def getDefaultTargetAndNonTargetData(valueNames, targetNames): requires(isinstance(targetNames, list)); requires(all([isinstance(x, str) for x in targetNames])); requires(set(valueNames).issuperset(targetNames)); requires(len(targetNames) == len(set(targetNames))); setOfTargetNames = set(targetNames); indicesOfNewTargets = np.zeros(len(valueNames), dtype=bool); for thisIndex in range(0, len(valueNames)): if(valueNames[thisIndex] in setOfTargetNames): indicesOfNewTargets[thisIndex] = 1; indicesOfNonTargets = ~indicesOfNewTargets; return {"indicesOfNewTargets" : indicesOfNewTargets, \ "indicesOfNonTargets": indicesOfNonTargets}; def normalizeThisData(thisData, indicesOfTheTrainSet): B = thisData - np.min(thisData[indicesOfTheTrainSet, :], axis=0); B = B / np.max(B, axis=0); return B; def tryDatasetWithPolynomialLinear(indicesOfNewTargets, indicesOfNonTargets): requires(isinstance(indicesOfNewTargets, np.ndarray)); requires(isinstance(indicesOfNonTargets, np.ndarray)); requires(not np.any(indicesOfNewTargets & indicesOfNonTargets)); requires(indicesOfNewTargets.dtype == np.dtype('bool')); requires(indicesOfNonTargets.dtype == np.dtype('bool')); newFeatureNames = ["in_x"] ; newTargetData_featureNames = ["out_y"]; newBaseData_featureNames = newFeatureNames; resultsToReturn = \ {"newFeatureNames" : newFeatureNames ,\ "coefficients" : np.array([[1.00]]) , \ "intercept" : np.array([[0.0]]) , \ "namesOfTargetValues" : newTargetData_featureNames, \ "orderOfNonFeaturizedObservationNames" : newBaseData_featureNames \ }; return resultsToReturn; import pickle; valueNames = ["in_x", "out_y"]; targetNames =['out_y'] A = getDefaultTargetAndNonTargetData(valueNames, targetNames); trainedModelInfoDicts = tryDatasetWithPolynomialLinear(\ A["indicesOfNewTargets"], \ A["indicesOfNonTargets"] ); # TODO: in general, for models stored or prepared, for the sake of record-keeping # and future error checking, put in the data trained and the code used to train it... # TODO: include UUIDs with the trained model material.... fh = open("./modelForTesting_oneDimInput_oneDimOutput_identityFunction.pickle", "wb"); pickle.dump(trainedModelInfoDicts, fh); fh.close();
3,776
31.282051
165
py
Fanoos
Fanoos-master/trainedNetworks/cpuPolynomialRegressionModel/polynomialRegressionTrial.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; from sklearn.linear_model import LinearRegression; from sklearn.datasets import fetch_california_housing; from sklearn.preprocessing import PolynomialFeatures; from scipy.special import comb; from sklearn.datasets import fetch_openml; def requires(booleanStatement): assert(booleanStatement); return; def getDefaultTargetAndNonTargetData(thisDataset, targetNames): requires(isinstance(targetNames, list)); requires(all([isinstance(x, str) for x in targetNames])); requires(set(thisDataset.feature_names).issuperset(targetNames)); requires(len(targetNames) == len(set(targetNames))); setOfTargetNames = set(targetNames); featureNames = thisDataset.feature_names; indicesOfNewTargets = np.zeros(thisDataset.data.shape[1], dtype=bool); for thisIndex in range(0, thisDataset.data.shape[1]): if(featureNames[thisIndex] in setOfTargetNames): indicesOfNewTargets[thisIndex] = 1; indicesOfNonTargets = ~indicesOfNewTargets; return {"indicesOfNewTargets" : indicesOfNewTargets, \ "indicesOfNonTargets": indicesOfNonTargets}; def normalizeThisData(thisData, indicesOfTheTrainSet): B = thisData - np.min(thisData[indicesOfTheTrainSet, :], axis=0); # .reshape((thisData.shape[0], 1)) B = B / np.max(B, axis=0); return B; def featurizeData(thisData, indicesOfTheTrainSet, maximumDegree, normalizePriorToFeaturing=True): # NOTE: We disable the intercept term since the regression already generates an intercept... featurizer = PolynomialFeatures(maximumDegree, include_bias=False); numberOfInstances = thisData.shape[0]; indicesOfTheTestSet = sorted(list(set(range(0, numberOfInstances)).difference(indicesOfTheTrainSet))) trainingSet = thisData[indicesOfTheTrainSet, :]; featureNormalizedData = None; if(normalizePriorToFeaturing): B = thisData - \ np.min(trainingSet, axis=0); B = B / np.max(B[indicesOfTheTrainSet, :], axis=0); featureNormalizedData = B; else: featureNormalizedData = thisData; assert(type(featureNormalizedData) != type(None)); # Notice we normalize BEFORE taking the polynomial transforms, not after - this # helps prevent blow-up. newFeaturizedData = featurizer.fit_transform(featureNormalizedData); return newFeaturizedData, featurizer; def tryDatasetWithPolynomialLinear(thisDataset, indicesOfNewTargets, indicesOfNonTargets, includeOriginalTarget=True, normalizeFeatures=True): requires(isinstance(indicesOfNewTargets, np.ndarray)); requires(isinstance(indicesOfNonTargets, np.ndarray)); requires(not np.any(indicesOfNewTargets & indicesOfNonTargets)); requires(indicesOfNewTargets.dtype == np.dtype('bool')); requires(indicesOfNonTargets.dtype == np.dtype('bool')); numberOfInstances = thisDataset.data.shape[0]; newBaseData = thisDataset.data[:, indicesOfNonTargets]; newBaseData_featureNames = np.array(thisDataset.feature_names)[indicesOfNonTargets]; print("newBaseData_featureNames:" + str(newBaseData_featureNames)); newTargetData = np.zeros((numberOfInstances, sum(indicesOfNewTargets) + (1 if includeOriginalTarget else 0))); newTargetData[:, :(sum(indicesOfNewTargets))] = thisDataset.data[:,indicesOfNewTargets]; if(includeOriginalTarget): newTargetData[:, sum(indicesOfNewTargets)] = thisDataset.target; newTargetData_featureNames = np.array(thisDataset.feature_names + (["usr"] if includeOriginalTarget else []) )[\ (indicesOfNewTargets if (not includeOriginalTarget) else (list(indicesOfNewTargets) + [includeOriginalTarget]))]; print("newTargetData_featureNames:" + str(newTargetData_featureNames)); numberOfTrainingInstances=int( numberOfInstances * 0.90); indicesOfTheTrainSet = np.random.choice(range(0, numberOfInstances), numberOfTrainingInstances, replace=False); indicesOfTheTestSet = sorted(list(set(range(0, numberOfInstances)).difference(indicesOfTheTrainSet))) newTargetData = normalizeThisData(newTargetData, indicesOfTheTrainSet); resultsToReturn = {}; resultsToReturn["experimentSetup"] = {\ "indicesOfNewTargets" : indicesOfNewTargets, \ "indicesOfNonTargets" : indicesOfNonTargets, \ "includeOriginalTarget" : includeOriginalTarget, \ "newBaseData" : newBaseData, \ "newBaseData_featureNames" : newBaseData_featureNames, \ "newTargetData" : newTargetData, \ "newTargetData_featureNames" : newTargetData_featureNames, \ "indicesOfTheTrainSet" : indicesOfTheTrainSet, \ "indicesOfTheTestSet" : indicesOfTheTestSet \ }; for maximumDegree in [1, 2, 3, 4]: print("v`V~V~~V~V~V~V~V~V~V~V~V~V~V~~V~V~~V~VV~~V~V~~V~V"); print("maximumDegree:" + str(maximumDegree)); print("================================================"); newFeaturizedData, featurizer = featurizeData(newBaseData, \ indicesOfTheTrainSet, maximumDegree, normalizePriorToFeaturing=normalizeFeatures) expectedNumberOfNewFeatures = comb( ((newBaseData.shape[1] + 1) - 1) + maximumDegree, maximumDegree) - 1; """ The below should hold for the following reasons: by pirates-and-gold argument, we have newBaseData.shape[1] pirates. To distribute maximumDegree pieces of gold among them, we have comb( (newBaseData.shape[1] - 1) + maximumDegree, maximumDegree) many ways. However, we want not just monomials where the sum of the degrees of the variables is maximumDegree, but also monomials where the sum of the degrees is less than than. So we include an additional bin/pirate that repressents "throwing away" a degree (equivalently, raising the number 1 to a power). That gives that the number of combinations should be comb( ((newBaseData.shape[1] + 1) - 1) + maximumDegree, maximumDegree) We subtract the final 1 (the one outside comb) after all of this because we exclude the term resulting from all the variables having power-zero. This is because we have the regression handle the intercept term - having a constant column in the features would be repetative. """ assert(expectedNumberOfNewFeatures == newFeaturizedData.shape[1]); newFeaturizedData = normalizeThisData(newFeaturizedData, indicesOfTheTrainSet); reg = LinearRegression(\ fit_intercept=True, \ copy_X=True \ ).fit(\ newFeaturizedData[indicesOfTheTrainSet, :], \ newTargetData[indicesOfTheTrainSet, :] \ ); # Notice that we did not pre-center the data and we excluded the # constant term from the featurization . print("newFeaturizedData.shape:" + str(newFeaturizedData.shape)); scoreOnTestSet = reg.score(newFeaturizedData[indicesOfTheTestSet, :], \ newTargetData[indicesOfTheTestSet, :]); print("scoreOnTestSet:" + str(scoreOnTestSet), flush=True); scoreOnTrainSet = reg.score(newFeaturizedData[indicesOfTheTrainSet, :], \ newTargetData[indicesOfTheTrainSet, :]); print("scoreOnTrainSet:" + str(scoreOnTrainSet), flush=True); print("^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^"); assert( maximumDegree not in resultsToReturn); resultsToReturn[maximumDegree] = \ {"newFeatureNames" : featurizer.get_feature_names(newBaseData_featureNames) ,\ "coefficients" : reg.coef_ , \ "intercept" : reg.intercept_ , \ "scoreOnTestSet" : scoreOnTestSet , \ "scoreOnTrainSet" : scoreOnTrainSet, \ "maximumDegree" : maximumDegree, \ "namesOfTargetValues" : newTargetData_featureNames, \ "orderOfNonFeaturizedObservationNames" : newBaseData_featureNames, \ "maxValuesNonFeaturizedObservations" : np.max(newBaseData, axis=0), \ "minValuesNonFeaturizedObservations" : np.min(newBaseData, axis=0), \ "" : """ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg""" \ }; return resultsToReturn; import pickle; thisDataset = fetch_openml(data_id=562); targetNames =['lwrite', 'swrite']; A = getDefaultTargetAndNonTargetData(thisDataset, targetNames) A["indicesOfNonTargets"] = np.array([False for x in range(0, 12)]); A["indicesOfNonTargets"][0] = True; A["indicesOfNonTargets"][3] = True; A["indicesOfNonTargets"][2] = True; A["indicesOfNonTargets"][10] = True; A["indicesOfNonTargets"][11] = True; pathToFile = "./trainedPolynomialModelInfo.pickle"; for includeOriginalTarget in [False, True]: """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ print("\nV~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V"); print("includeOriginalTarget:" + str(includeOriginalTarget)); print("=============================================================="); trainedModelInfoDicts = tryDatasetWithPolynomialLinear(\ thisDataset, \ A["indicesOfNewTargets"], \ A["indicesOfNonTargets"], normalizeFeatures=True, \ includeOriginalTarget=includeOriginalTarget); if(includeOriginalTarget): print("Saving Models for includeOriginalTarget:" + str(includeOriginalTarget)); # TODO: put in the data trained and the code used to train it... # TODO: include UUIDs with the trained model stuff.... fh = open(pathToFile, "wb"); pickle.dump(trainedModelInfoDicts, fh); fh.close(); print("^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^\n\n\n"); #~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Below code added later (originally in a seperate, small script) to simply the # format of the interface to the learned model. #============================================================================== fh = open(pathToFile, "rb"); A = pickle.load(fh); fh.close(); assert(isinstance(A, dict)); fh = open(pathToFile, "wb"); pickle.dump(A[3], fh); fh.close(); #_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^
16,913
66.386454
2,810
py
Fanoos
Fanoos-master/UI/fanoosFrontend.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from utils.getGitCommitHash import getGitCommitHash; import uuid; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox, boxSize; import re; import inspect; import z3; from boxesAndBoxOperations.codeForGettingSamplesBetweenBoxes import getSampleVectorsToCheckAgainst, getBoxCenter; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass,\ Condition_TheBoxItself, MetaCondition_Conjunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation; from boxesAndBoxOperations.splitBox import splitBox; from domainsAndConditions.classesDefiningQuestions import QuestionBaseClass, QuestionClass_What_Do_You_Do_When, \ QuestionClass_When_Do_You, QuestionClass_What_Are_The_Circumstances_In_Which, QuestionClass_What_Do_You_Ussually_Do_When, \ QuestionClass_When_Do_You_Ussually, QuestionClass_What_Are_The_Usual_Circumstances_In_Which; from boxesAndBoxOperations.mergeBoxes import mergeBoxes; from propagateBoxThroughLearnedSystem.classesToPropogateBoxThroughModels import ModelBoxProgatorManager, \ PropogatorForPolyFeatLinearModel , PropogatorForNueralNet ; import config; from UI.cycleToRespondToUserQuestion import respondToUserQuestion as externalCall_respondToUserQuestion; import cmd; import readline; import traceback; import os; import time as timePackageToUseForSleep; from UI.commandLineAutocompleter import readInLineAllowingForPathCompletion; from UI.genericUIFunctions import promptToSelectFromList; from domainsAndConditions.domainAndConditionsForCircleFollowing import DomainForCircleFollowing; from domainsAndConditions.domainAndConditionsForInvertedDoublePendulum import DomainForInvertedDoublePendulum; from domainsAndConditions.domainAndConditionsForCPUUse import DomainForCPUUse; from domainsAndConditions.domainAndConditionsFor_modelForTesting_oneDimInput_oneDimOutput import \ DomainFor_modelForTesting_oneDimInput_oneDimOutput; from domainsAndConditions.domainAndConditionsFor_modelForTesting_twoDimInput_threeDimOutput import \ DomainFor_modelForTesting_twoDimInput_threeDimOutput; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface; def getAvailableDomains(): listOfDomains = [DomainForCircleFollowing, DomainForInvertedDoublePendulum, DomainForCPUUse, DomainFor_modelForTesting_oneDimInput_oneDimOutput, DomainFor_modelForTesting_twoDimInput_threeDimOutput]; return listOfDomains; def getAndParseUserInput(domainInformation, questionType, text, dictMappingConditionTokenToCondition): requires(isinstance(domainInformation, BaseClassDomainInformation)); # MetaCondition_Conjunction if(len(text) == 0): raise Exception("Improperly formed question. There are not conditions specified."); tokenForOr = "or"; tokenForAnd = "and("; tokenForEndOfAnd = ")"; stringOfUserInput = text.replace("\n", ""); #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # a few hacks on the input that are useful in our case to make typing input in a little # less fragile. #=================================================== stringOfUserInput = stringOfUserInput.replace(",", " "); stringOfUserInput = stringOfUserInput.replace(tokenForAnd, tokenForAnd + " "); stringOfUserInput = stringOfUserInput.replace(tokenForEndOfAnd, " " + tokenForEndOfAnd); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ if(len(stringOfUserInput) == 0): raise Exception("Empty input by user."); if(stringOfUserInput[-1] in {".", "?"}): stringOfUserInput = stringOfUserInput[:-1]; textSplitOnSpaces = [x for x in re.split(" +", stringOfUserInput) if len(x) > 0]; listOfConditionsSpecified = []; inConjunctCondition = False; listOfElementsToConjunct = []; addedTokenLastIndex = False; for thisString in textSplitOnSpaces: if(thisString == tokenForAnd): if(addedTokenLastIndex): raise Exception("Misformated expression: question not in DNF form."); inConjunctCondition = True; addedTokenLastIndex = True; continue; if(thisString == tokenForEndOfAnd): inConjunctCondition = False; addedTokenLastIndex = True; if(len(listOfElementsToConjunct) == 0): raise Exception("Empty list of elements to conjunct."); listOfConditionsSpecified.append( \ MetaCondition_Conjunction(listOfElementsToConjunct) ); listOfElementsToConjunct = []; continue; if(inConjunctCondition): addedTokenLastIndex = True; if(thisString not in dictMappingConditionTokenToCondition.keys()): raise Exception("Unrecognized string in conjunct: " + str(thisString)); listOfElementsToConjunct.append(dictMappingConditionTokenToCondition[thisString]); continue; assert(not inConjunctCondition); if(addedTokenLastIndex): if(thisString != tokenForOr): raise Exception("Misformated expression: missing disjunctive connector between conditions." + \ " Remember, questions, after the question indicator token, must be in DNF form."); addedTokenLastIndex = False; continue; assert(not addedTokenLastIndex); if(thisString not in dictMappingConditionTokenToCondition.keys()): raise Exception("Unrecognized condition: " + str(thisString)); listOfConditionsSpecified.append(dictMappingConditionTokenToCondition[thisString]); addedTokenLastIndex = True; if(inConjunctCondition): raise Exception("Unclosed conjunct condition."); if(not addedTokenLastIndex): raise Exception("Misformated expression: question not in DNF form. Ended with open or statement."); assert(len(listOfElementsToConjunct) == 0); thisQuestion = questionType( listOfConditionsSpecified, domainInformation); ensures(isinstance(thisQuestion , QuestionBaseClass)); ensures(len(listOfConditionsSpecified) > 0); ensures(all([isinstance(x, CharacterizationConditionsBaseClass) for x in listOfConditionsSpecified])); return thisQuestion; class FanoosFrontend(cmd.Cmd): prompt = '(Fanoos) '; @staticmethod def _standardWait(): timeForSleep=config.defaultValues.responceDelayTimeForUnexpectedInputes; print("Sleeping " + str(timeForSleep) + " seconds before responding....", flush=True); timePackageToUseForSleep.sleep(timeForSleep); return; def default(*x, **kwargs): __class__._standardWait(); cmd.Cmd.default(*x, **kwargs); return; def emptyline(*x, **kwargs): __class__._standardWait(); print("Ignoring Empty Line.", flush=True); return; def tempFuct(self, questionType, *arg): try: parsedUserQuestion = getAndParseUserInput(self.domainInformation, questionType, arg[0], self.dictMappingConditionTokenToCondition); self.history.append((parsedUserQuestion, [])); # This object is modified by reference by the events in # externalCall_respondToUserQuestion ObjDatabaseValueTracker.set_questionInstanceUUID(uuidToUse=parsedUserQuestion.getID()); commandToStartQuestionInDatabase_1 = \ "INSERT INTO session_questionInstance_relation (sessionUUID, questionInstanceUUID) VALUES ('" + \ str(ObjDatabaseValueTracker.get_sessionUUID()) + "' , '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "')"; objDatabaseInterface.exec(commandToStartQuestionInDatabase_1); # below, the value of database column questionInstanceUUID is set by default.... commandToStartQuestionInDatabase_2 = \ "INSERT INTO questionInstanceInfo (questionInstanceUUID, questionInstanceType) VALUES ('" + \ str(ObjDatabaseValueTracker.get_questionInstanceUUID()) +"', '" + \ str(parsedUserQuestion.__class__.__name__) + "');"; objDatabaseInterface.exec(commandToStartQuestionInDatabase_2); objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ "UPDATE questionInstanceInfo SET questionInstanceContentTextUncleaned = ? WHERE questionInstanceUUID = '" + \ str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';", \ [arg[0]]); objDatabaseInterface.commit(); externalCall_respondToUserQuestion(\ self.domainInformation, self.loadedLearnedModel, \ parsedUserQuestion, self.history); # record the end of the question..... commandToEndQuestionInDatabase = \ "UPDATE questionInstanceInfo SET dateAndTimeFinished = CURRENT_TIMESTAMP WHERE questionInstanceUUID = '" + \ str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';"; objDatabaseInterface.exec(commandToEndQuestionInDatabase); objDatabaseInterface.commit(); except: errorMessageIndented = " " + traceback.format_exc().replace("\n", "\n "); sys.stderr.write(errorMessageIndented); sys.stderr.flush(); timePackageToUseForSleep.sleep(3); return; def preloop(self): availableDomains = getAvailableDomains(); userResponce = promptToSelectFromList([x.getName() for x in availableDomains], "the domain to use"); integerDomainSelection = userResponce[1]; s = z3.Solver(); self.domainInformation = availableDomains[integerDomainSelection](s); print("Enter path to the neural-net weights to use. Spaces in the path name will be ignored.", flush=True); thisLineMissingNewLine = readInLineAllowingForPathCompletion(); pathToWeights = os.path.realpath(thisLineMissingNewLine.replace(" ", "")); assert(isinstance(pathToWeights, str)); assert("\n" not in pathToWeights); if(not os.path.isfile(pathToWeights)): raise Exception("Path not found: " + str(pathToWeights)); # userUUID is NULL for now, and the time started is set by default. The end-time will be set later.... randomSeeds = {"randomSeedForNumpy" : config.randomSeedForNumpy, "randomSeedForPython3LibRandom": config.randomSeedForPython3LibRandom}; commandToStartSessionInDatabase = \ "INSERT INTO sessionInfo ( sessionUUID, domainUUID , pathToSystemAnalyzed, gitCommitHash, randomSeeds ) VALUES " + \ "('" + str(ObjDatabaseValueTracker.get_sessionUUID()) + "' , '" + \ str(self.domainInformation.getUUID()) + "' , '" + \ str(pathToWeights) + "', '" + \ str(getGitCommitHash(str(uuid.uuid4()))) + "', ?);"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(commandToStartSessionInDatabase, [randomSeeds]);""" ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ self.loadedLearnedModel = None; dictMappingModelTypeDescriptionToClass = {\ "Nueral Net: input>32 hidden->[fully connected]->32 hidden->linear output unit->output clipped" : PropogatorForNueralNet, \ "Linear Regression with Degree-3 Polynomial Features" : PropogatorForPolyFeatLinearModel, \ "Models Described in README.txt for sanity checking" : PropogatorForPolyFeatLinearModel \ }; userChoice = promptToSelectFromList(list(dictMappingModelTypeDescriptionToClass.keys()), "type of learned model."); self.loadedLearnedModel = dictMappingModelTypeDescriptionToClass[userChoice[0]](pathToWeights); assert(isinstance(self.loadedLearnedModel , ModelBoxProgatorManager)); self.history = []; def functorFor_do_methods(thisQuestionType): return ( lambda *args: self.tempFuct(self.dictMappingLeadingTokenToQuestionClass[thisQuestionType], args[1] )); def functorFor_complete_methods(thisQuestionType): return (lambda s, text, line, start_index, end_index : \ self.dictMappingLeadingTokenToQuestionClass[thisQuestionType].getUseableConditions(\ self.domainInformation, self.dictMappingLeadingTokenToQuestionClass[thisQuestionType], conditionNameStartsWith=text) ); for thisQuestionType in self.dictMappingLeadingTokenToQuestionClass: setattr(FanoosFrontend, ("do_" + thisQuestionType), classmethod( functorFor_do_methods(thisQuestionType) )); setattr(FanoosFrontend, ("complete_" + thisQuestionType), classmethod(functorFor_complete_methods(thisQuestionType)) ); return; @property def dictMappingConditionTokenToCondition(self): tempDict = dict(); baseConditionsForDomain = self.domainInformation.getBaseConditions(); for thisCond in baseConditionsForDomain: thisKey = str(thisCond).replace(" ", "_").lower(); tempDict[thisKey] = thisCond; return tempDict dictMappingLeadingTokenToQuestionClass = {\ "what_do_you_do_when" : QuestionClass_What_Do_You_Do_When, \ "when_do_you" : QuestionClass_When_Do_You , \ "what_are_the_circumstances_in_which" : QuestionClass_What_Are_The_Circumstances_In_Which, \ "what_do_you_ussually_do_when" : QuestionClass_What_Do_You_Ussually_Do_When, \ "when_do_you_ussually" : QuestionClass_When_Do_You_Ussually , \ "what_are_the_usual_circumstances_in_which" : QuestionClass_What_Are_The_Usual_Circumstances_In_Which \ }; @property def intro(self): return ("""\ V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V Type in your question. Questions are expected to follow this pattern: <question type> <condition> [or <condition>]* ? Where: question type is one of: """ + ("\n ".join(list(self.dictMappingLeadingTokenToQuestionClass.keys()))) + \ """ condition is one of the following or a conjunction of them (i.e., "and(condition1, condition2, condition3, ....)" ): """ + ("\n ".join(list(self.dictMappingConditionTokenToCondition.keys()))) + \ "\n\nTo simply exit, type in the command exit" + \ "\n\n^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^"); def do_exit(self, arg): return True;
18,586
54.98494
2,902
py
Fanoos
Fanoos-master/UI/genericUIFunctions.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; import sys; import config; import time as timePackageToUseForSleep; def displayForUser(thisState, dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered, useMoreToLimitOutput=True): requires(isinstance(useMoreToLimitOutput, bool)); listOfLines = []; def properPrintHandling(thisString): requires(isinstance(thisString, str)); if(useMoreToLimitOutput): listOfLines.append(thisString); else: print(thisString, flush=True); return properPrintHandling("===================================================================="); properPrintHandling("\nSum of Analyzed Box Volumes: " + \ str(dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered["totalVolumeOfBoxesInList"])); # In theory, this #could just be read from the # database using the database value tracker to get the most recent UUIDs and # read the box volume statistic for the most recent state. This will be done in later pushes. TODO: further # consider the points made in this comment this and implement if deemed best . properPrintHandling("\nDescription:"); descriptionSortedByValues = [\ (dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[x.getID()]["uniqueVolumeCovered"], \ dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[x.getID()]["volumeCovered"], \ str(x) ) \ for x in thisState.getDescription()]; descriptionSortedByValues.sort(reverse=True); for thisElem in descriptionSortedByValues: properPrintHandling(str(thisElem)); if(useMoreToLimitOutput): myLinuxStyleMoreCommand(listOfLines); return; from config.defaultValues import userInterface_maximumNumberOfLinesToPrintAtOneTime; def myLinuxStyleMoreCommand(linesToPrint): requires(isinstance(linesToPrint, list)); requires(all([isinstance(x, str) for x in linesToPrint])); numberOfLinesToPrint = userInterface_maximumNumberOfLinesToPrintAtOneTime; sys.stdout.write(str(min(numberOfLinesToPrint,len(linesToPrint))) + " of " + str(len(linesToPrint)) + " lines to print shown."); if(numberOfLinesToPrint < len(linesToPrint)): sys.stdout.write(" Press enter to show more. Hit ctrl+C or enter letter q to break. Hit a to list all."); sys.stdout.flush(); index = 0; for thisLine in linesToPrint: if(index >= numberOfLinesToPrint): thisLine = sys.stdin.readline(); if(thisLine.lower() == "q\n"): break; elif(thisLine.lower() == "a\n"): numberOfLinesToPrint = float("inf"); else: """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ print(""); sys.stdout.write(linesToPrint[index]); sys.stdout.flush(); assert(index + 1 > index); # weak overflow check... index = index + 1; print("\n"); return; import re; def promptToSelectFromList(listOfOptions, descriptionOfSelection): requires(isinstance(listOfOptions, list)); requires(len(listOfOptions) > 0); requires(all([isinstance(x, str) for x in listOfOptions])); requires(isinstance(descriptionOfSelection, str)); requires(len(descriptionOfSelection) > 0); print("Enter the integer for " + descriptionOfSelection + ". Options are as follows:", flush = True); for index in range(0, len(listOfOptions)): print(str(index) + " : " + listOfOptions[index]); while(True): thisLine = sys.stdin.readline(); if(re.match("^ *[0-9]+ *\n$", thisLine)): break; print("Unrecognized input: \"" + thisLine + "\"", flush=True); print("Try again.", flush=True); timeForSleep=config.defaultValues.responceDelayTimeForUnexpectedInputes; print("Sleeping " + str(timeForSleep) + " seconds before responding....", flush=True); timePackageToUseForSleep.sleep(timeForSleep); integerDomainSelection = int(thisLine[:-1]); assert(integerDomainSelection >= 0); assert(integerDomainSelection < len(listOfOptions)); return (listOfOptions[integerDomainSelection], integerDomainSelection);
8,130
57.496403
2,793
py
Fanoos
Fanoos-master/UI/captureTerminalOutput.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import sys; import inspect; from utils.contracts import *; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface; class wrapperForCopy(): def _wrapperFunct_capture(self, nameOfThis, properFunction): def tempFunct(*x, **kwargs): tempResult = properFunction(*x, **kwargs); self.thingToWriteTo.write(self.label, (nameOfThis, x, kwargs, tempResult)); # This is going to lose input if an error occurs running the function... return tempResult; return tempFunct; def __init__(self, thingToWrap, thingToWriteTo, label, debug=False): self.thingToWrap = thingToWrap; self.thingToWriteTo = thingToWriteTo; self.label = label; self.thingsNotToCapture = {'__class__', '__delattr__', '__dict__', '__dir__', '__doc__', '__eq__', '__format__', '__ge__', '__getattribute__', '__gt__', '__hash__', '__init__', '__init_subclass__', '__le__', '__lt__', '__module__', '__ne__', '__new__', '__reduce__', '__reduce_ex__', '__repr__', '__setattr__', '__sizeof__', '__str__', '__subclasshook__', '__weakref__'}; for thisThing in inspect.getmembers(thingToWrap): if((thisThing[0] not in self.thingsNotToCapture) and \ ("function" in type(thisThing[1]).__name__)): if(debug): print("captured:" + str(thisThing[0]), flush=True); setattr(self, thisThing[0], self._wrapperFunct_capture(thisThing[0], thisThing[1])); elif(thisThing[0] not in self.thingsNotToCapture): setattr(self, thisThing[0], thisThing[1]); else: if(debug): print("not captured:" + str(thisThing[0]), flush=True); return; import pickle; class thingToWriteTo(): def __init__(self): self.orderIndex = 0; return; def write(self, label, thisThing): requires(isinstance(label, str)); # TODO: safety checks on the label.... byteObj = pickle.dumps(thisThing); commandToExecute = \ "INSERT INTO terminalOutput ( sessionUUID, orderIndex , channelName , valueOnChannel ) VALUES " + \ "('" + str(ObjDatabaseValueTracker.get_sessionUUID()) + "' , " + \ str(self.orderIndex) + " , '" + \ label + "', ?);"; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ objDatabaseInterface.interfaceBleed_insertValuesForBlob( commandToExecute, [byteObj]); assert(self.orderIndex + 1 > self.orderIndex); # weak overflow check, though strictly speaking not # needed in python..... self.orderIndex = self.orderIndex + 1; objDatabaseInterface.commit(); return;
6,706
67.438776
2,819
py
Fanoos
Fanoos-master/UI/commandLineAutocompleter.py
import os import re import readline # This code largely copied from https://stackoverflow.com/questions/5637124/tab-completion-in-pythons-raw-input RE_SPACE = re.compile('.*\s+$', re.M) class Completer(object): def _listdir(self, root): "List directory 'root' appending the path separator to subdirs." res = [] for name in os.listdir(root): path = os.path.join(root, name) if os.path.isdir(path): name += os.sep res.append(name) return res def _complete_path(self, path=None): "Perform completion of filesystem path." if not path: return self._listdir('.') dirname, rest = os.path.split(path) tmp = dirname if dirname else '.' res = [os.path.join(dirname, p) for p in self._listdir(tmp) if p.startswith(rest)] # more than one match, or single match which does not exist (typo) if len(res) > 1 or not os.path.exists(path): return res # resolved to a single directory, so return list of files below it if os.path.isdir(path): return [os.path.join(path, p) for p in self._listdir(path)] # exact file match terminates this completion return [path + ' '] def complete_extra(self, args): "Completions for the 'extra' command." if not args: return self._complete_path('.') # treat the last arg as a path and complete it return self._complete_path(args[-1]) def complete(self, text, state): "Generic readline completion entry point." buffer = readline.get_line_buffer() line = readline.get_line_buffer().split() # account for last argument ending in a space if RE_SPACE.match(buffer): line.append('') # resolve command to the implementation function if(True): impl = getattr(self, 'complete_extra') args = line[0:] if args: return (impl(args) + [None])[state] return [cmd + ' '][state] results = [c + ' ' for c in COMMANDS if c.startswith(cmd)] + [None] return results[state] def readInLineAllowingForPathCompletion(): comp = Completer() # we want to treat '/' as part of a word, so override the delimiters readline.set_completer_delims(' \t\n;') readline.parse_and_bind("tab: complete") readline.set_completer(comp.complete) return input(); #tried just "input" here to get the copying behavior to work... """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """
5,353
72.342466
2,781
py
Fanoos
Fanoos-master/UI/cycleToRespondToUserQuestion.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from statesAndOperatorsAndSelection.descriptionState import FirstState_DescriptionState, DescriptionState; from statesAndOperatorsAndSelection.chooseOperatorToApply import chooseOperatorToApply ; from statesAndOperatorsAndSelection.descriptionOperator import DescriptionOperator, Operator_StartOperator; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface; from UI.genericUIFunctions import displayForUser; from domainsAndConditions.classesDefiningQuestions import QuestionBaseClass ; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; from propagateBoxThroughLearnedSystem.classesToPropogateBoxThroughModels import ModelBoxProgatorManager; def executeDatabaseCommandList(commandsToExecute): for thisCommand in commandsToExecute: objDatabaseInterface.exec(thisCommand); objDatabaseInterface.commit(); return; def recordStateGeneral(stateToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(stateToRecord ,DescriptionState)); # The OR IGNORE part below is to avoid volating unique constraints on the table. # recall that the history travel operator allows revisiting an old state executeDatabaseCommandList([ "INSERT INTO questionInstance_QAState_relation (questionInstanceUUID, answerIndex, QAStateUUID) VALUES ('" + \ str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "', " + \ str(indexIntoQA) + ", '" + \ str(stateToRecord.getID()) + "');" , "INSERT OR IGNORE INTO QAStateInfo(QAStateUUID) VALUES ('" + str(stateToRecord.getID())+ "');", ]); return; def recordStateShown(stateToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(stateToRecord ,DescriptionState)); executeDatabaseCommandList([ "UPDATE questionInstance_QAState_relation SET dateAndTimeAnswerShown = CURRENT_TIMESTAMP WHERE " + \ " answerIndex = " + str(indexIntoQA) + " and " + \ " questionInstanceUUID = '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';" ]); return; def recordOperatorGeneral(operatorToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(operatorToRecord ,DescriptionOperator)); executeDatabaseCommandList([ "INSERT INTO questionInstance_QAOperator_relation (questionInstanceUUID, startingAnswerIndex, QAOperatorUUID)" +\ "VALUES ('" + \ str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "', " + \ str(indexIntoQA) + ", '" + \ str(operatorToRecord.getID()) + "');" ]); return; def recordOperatorComputationStarted(operatorToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(operatorToRecord ,DescriptionOperator)); executeDatabaseCommandList([ "UPDATE questionInstance_QAOperator_relation SET dateAndTimeComputationStarted = CURRENT_TIMESTAMP WHERE " + \ " startingAnswerIndex = " + str(indexIntoQA) + " and " + \ " questionInstanceUUID = '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';" ]); return; def recordOperatorComputationFinished(operatorToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(operatorToRecord ,DescriptionOperator)); executeDatabaseCommandList([ "UPDATE questionInstance_QAOperator_relation SET dateAndTimeComputationFinished = CURRENT_TIMESTAMP WHERE " + \ " startingAnswerIndex = " + str(indexIntoQA) + " and " + \ " questionInstanceUUID = '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';" ]); return; from statesAndOperatorsAndSelection.automaticOperatorSelection.operationSelectionManagers import \ Manual_SelectorManager, OriginalMethod_SelectorManager; def respondToUserQuestion(domainInformation, loadedLearnedModel, parsedUserQuestion, historyFromOtherTrials): requires(isinstance(domainInformation, BaseClassDomainInformation)); requires(isinstance(parsedUserQuestion, QuestionBaseClass )); requires(isinstance(historyFromOtherTrials, list)); requires(len(historyFromOtherTrials) > 0); requires(isinstance(historyFromOtherTrials[-1], tuple)); requires(len(historyFromOtherTrials[-1]) == 2); requires(isinstance(historyFromOtherTrials[-1][1], list)); requires(len(historyFromOtherTrials[-1][1]) == 0); objectForHistory = historyFromOtherTrials[-1][1]; startState = FirstState_DescriptionState(); startOperator = Operator_StartOperator(); indexIntoQA = 0; objectForHistory.append( (startState, startOperator) ); recordStateGeneral(startState, indexIntoQA); inputUniverseBox = domainInformation.getInputSpaceUniverseBox(); startState.recordInDatabase(parsedUserQuestion.variablesBoxesProducedMayBeOver, inputUniverseBox); # The start and end times for the states may be NULL. As such, since the initial state is NOT shown to the user, # we record NULL for its start and end time.... recordOperatorGeneral(startOperator, indexIntoQA); # Operator: start computation recordOperatorComputationStarted(startOperator, indexIntoQA); thisState = \ startOperator.apply(parsedUserQuestion, domainInformation, loadedLearnedModel, startState, historyFromOtherTrials, indexIntoQA); dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered = thisState.getSideInformation("dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered"); recordOperatorComputationFinished(startOperator, indexIntoQA); manualSelectionManager = Manual_SelectorManager(domainInformation, loadedLearnedModel); autoSelectionManager = OriginalMethod_SelectorManager(domainInformation, loadedLearnedModel); while(True): assert(indexIntoQA + 1 > indexIntoQA); # Weak overflow check. Strictly speaking, this is not necessary for python since it has infinite # precision integers, but it is good to check.... indexIntoQA = indexIntoQA + 1; recordStateGeneral(thisState, indexIntoQA); thisState.recordInDatabase(parsedUserQuestion.variablesBoxesProducedMayBeOver, inputUniverseBox); recordStateShown(thisState, indexIntoQA) # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg displayForUser(thisState, dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered); # NOTE: the time that the user responds to the responce shown, as well as what their responce is, # is recorded in the file chooseOperatorToApply.py .... operatorChosen = chooseOperatorToApply(parsedUserQuestion, domainInformation, loadedLearnedModel, \ thisState, historyFromOtherTrials, indexIntoQA,\ manualSelectionManager, autoSelectionManager ); objectForHistory.append( (thisState, operatorChosen) ); if(operatorChosen == None): print("Breaking abstraction refinement, presumably due to user command.", flush=True); break; requires(isinstance(operatorChosen, DescriptionOperator)); recordOperatorGeneral(operatorChosen, indexIntoQA); recordOperatorComputationStarted(operatorChosen, indexIntoQA) thisState = \ operatorChosen.apply(parsedUserQuestion, domainInformation, loadedLearnedModel, thisState, historyFromOtherTrials, indexIntoQA); recordOperatorComputationFinished(operatorChosen, indexIntoQA) dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered = thisState.getSideInformation("dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered"); manualSelectionManager.cleanUp(); autoSelectionManager.cleanUp(); return;
11,975
59.180905
2,832
py
Fanoos
Fanoos-master/UI/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/boxesAndBoxOperations/codeForGettingSamplesBetweenBoxes.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import uuid; import re; import struct; from boxesAndBoxOperations.readAndWriteBoxes import *; from boxesAndBoxOperations.splitBox import splitBox; import inspect; import config; def boxSize(thisBox): """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ requires(isProperBox(thisBox)); return np.product(thisBox[:,1] - thisBox[:,0]); def getBoxCenter(thisBox): requires(isProperBox(thisBox)); # below, we add first because we are more concerned about precision loss / under flow than overflow. return (thisBox[:,0].reshape((getDimensionOfBox(thisBox), 1)) + thisBox[:,1].reshape((getDimensionOfBox(thisBox), 1))) * 0.5; def getRandomVectorBetweenScaledExtensionBoxes(thisBox, minInfinityNormPorportionOfDistanceFromBox, maxInfinityNormPorportionOfDistanceFromBox): requires(isProperBox(thisBox)); requires(boxSize(thisBox) > 0.0); # Does not work with flat boxes.... requires(isinstance(minInfinityNormPorportionOfDistanceFromBox ,float)); requires(minInfinityNormPorportionOfDistanceFromBox >= 0); requires(isinstance(maxInfinityNormPorportionOfDistanceFromBox ,float)); requires(maxInfinityNormPorportionOfDistanceFromBox >= 0); requires(maxInfinityNormPorportionOfDistanceFromBox >= minInfinityNormPorportionOfDistanceFromBox); center = getBoxCenter(thisBox); scalarToPushOut = np.random.rand() * (maxInfinityNormPorportionOfDistanceFromBox - minInfinityNormPorportionOfDistanceFromBox) + \ minInfinityNormPorportionOfDistanceFromBox; assert(scalarToPushOut >= minInfinityNormPorportionOfDistanceFromBox); # assuming precision limits don't break this assert(scalarToPushOut <= maxInfinityNormPorportionOfDistanceFromBox); parametersForShiftAboutCenter = 1 - (2 * np.random.rand(getDimensionOfBox(thisBox), 1)); assert(np.all(parametersForShiftAboutCenter >= -1)); assert(np.all(parametersForShiftAboutCenter <= 1)); parametersForShiftAboutCenter = parametersForShiftAboutCenter * \ (scalarToPushOut / np.max(np.abs(parametersForShiftAboutCenter))); vectorToReturn = (0.5 * (thisBox[:,1] - thisBox[:,0]).reshape((getDimensionOfBox(thisBox), 1)) * parametersForShiftAboutCenter) + center; return vectorToReturn; def getSampleVectorsToCheckAgainst(thisBox, minInfinityNormPorportionOfDistanceFromBox, \ maxInfinityNormPorportionOfDistanceFromBox, numberOfSamplesToGet): # use both getRandomVectorOutsideOfBox and, later, a procedural approach to make sure the space is covered.... requires(isinstance(minInfinityNormPorportionOfDistanceFromBox ,float)); requires(minInfinityNormPorportionOfDistanceFromBox >= 0); requires(isinstance(maxInfinityNormPorportionOfDistanceFromBox ,float)); requires(maxInfinityNormPorportionOfDistanceFromBox >= 0); requires(maxInfinityNormPorportionOfDistanceFromBox >= minInfinityNormPorportionOfDistanceFromBox); requires(isinstance(numberOfSamplesToGet, int)); requires(numberOfSamplesToGet > 0); samplesToReturn = []; while(len(samplesToReturn) < numberOfSamplesToGet): thisSample = \ getRandomVectorBetweenScaledExtensionBoxes(thisBox, minInfinityNormPorportionOfDistanceFromBox, \ maxInfinityNormPorportionOfDistanceFromBox); samplesToReturn.append(thisSample); ensures(len(samplesToReturn) == numberOfSamplesToGet); # TODO check that the dimension of the samples returned matches the dimension of the box. return samplesToReturn;
7,588
63.313559
2,785
py
Fanoos
Fanoos-master/boxesAndBoxOperations/mergeBoxes.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; from boxesAndBoxOperations.getBox import *; import config; def boxSize(thisBox): requires(isProperBox(thisBox)); return np.product(thisBox[:,1] - thisBox[:,0]); def boxesCanMergeWithNoGap(boxA, boxB): requires(isProperBox(boxA)); requires(isProperBox(boxB)); # if two boxes in n-dimensional space share 2 *(n-1) + 1 = 2n - 1 bounding coordinates (not necessarly serving # as the same role - ie, [a, b] and [b,c] are considered to share the "bounding coordinate" c) (out of the 2n they have), # then they share a hyper-face and can merge. To see this, notice that the bounds only differ on one dimension, and # since there can only be one bounding-value different between the boxes, the boxes must touch on this dimension. # Thus, we can just take the smallest bound that contains the elements along the dimension where the boxes are different # then keep the rest of the dimensions the same, and we found a tight box holding them. # # For example: # boxA : [[x1, x2], # [y1, y2], # [z1, z2]] # boxB : [[x1, x2], # [y1, y2'], # [z1, z2]] # Since it must be that y1< y2 and y1 < y2' we see that one of the boxes actually must be contained in the other. # Another example: # boxA : [[x1, x2], # [y1, y2], # [z1, z2]] # boxB : [[x1, x2], # [y2, y3], # [z1, z2]] # The union of these two boxes is exactly the box: # [[x1, x2], # [y1, y3], # [z1, z2]] # # Because of how the code calling this works - only checking corners - # below would only ever trigger if the boxes share a face - in which # case it is actually a less efficient version of the check that comes # after it... As such, I am commenting it out for now, but leaving in this # deadcode as minimal documentation of this fact and to show the idea # was evaluated then hit such a snag. indicesWhereDifferent = np.where(~np.isclose(boxA, boxB)); assert(len(indicesWhereDifferent) == 2); # first coordinate: np.array with # row-values of indices different, second coordinate: np.array with # column-values of indices different if(len(indicesWhereDifferent[0]) <= 1): return True; # This is a degenerate case and not the main intent of this function... variablesDifferent = indicesWhereDifferent[0]; numberOfVariablesDifferent = len(set(variablesDifferent)); if(numberOfVariablesDifferent > 1): return False; variableDifferent = variablesDifferent[0]; # By the previous conditions, we know # that the list variablesDifferent must contain all the same element and # be non-empty - otherwise the code would have already returned. if(np.isclose(boxB[variableDifferent, 1], boxA[variableDifferent, 0]) or \ np.isclose(boxA[variableDifferent, 1], boxB[variableDifferent, 0]) ): return True; return False; def getAllRoughCornerPoints(thisBox, precision): # this function largely based on the function getInitialAbstraction_boxesBySign # from CEGARLikeAnalysis/CEGARLikeAnalysisMain.py requires(isProperBox(thisBox)); numberOfCorners = 2 ** (thisBox.shape[0]); listToReturn = []; for thisCornerIndex in range(0, numberOfCorners): thisCorner = []; binaryRepresentationOfIndex = np.binary_repr(thisCornerIndex, width=getDimensionOfBox(thisBox)); assert(isinstance(binaryRepresentationOfIndex, str)); assert(all([(x in {"1", "0"}) for x in binaryRepresentationOfIndex])); for thisVariableIndex in range(0, getDimensionOfBox(thisBox)): valueToRecord = np.round(thisBox[thisVariableIndex, int(binaryRepresentationOfIndex[thisVariableIndex])], precision); thisCorner.append(float(valueToRecord)); # if I did not convert it to a float, it would be recorded # as a larger numpy-value, taking up more space.... listToReturn.append(pickle.dumps(thisCorner)); return listToReturn; def getSetupDicts(listOfBoxes, precision=3): dictMappingIndexToBox = dict(); dictMappingCornerToIDsOfBoxesWithThem = dict(); dictMappingIDsOfBoxesToTheirCorners= dict(); for thisBox in listOfBoxes: indexIDOfThisBox = len(dictMappingIndexToBox); sizeOfThisBox = boxSize(thisBox); dictMappingIndexToBox[indexIDOfThisBox] = thisBox; cornersOfThisBox = getAllRoughCornerPoints(thisBox, precision); cornersOfThisBox = list(set(cornersOfThisBox )); # making sure the corners we consider are unique, # even if the box is flat.... dictMappingIDsOfBoxesToTheirCorners[indexIDOfThisBox] = cornersOfThisBox; for thisCorner in cornersOfThisBox: if(thisCorner not in dictMappingCornerToIDsOfBoxesWithThem): dictMappingCornerToIDsOfBoxesWithThem[thisCorner] = []; dictMappingCornerToIDsOfBoxesWithThem[thisCorner].append((sizeOfThisBox, indexIDOfThisBox)); for thisKey in dictMappingCornerToIDsOfBoxesWithThem: # this list of boxes that have a certain corner are sorted in reverse order of size # so that we can try to aggressively try and greadily merge boxes... dictMappingCornerToIDsOfBoxesWithThem[thisKey] = \ sorted(dictMappingCornerToIDsOfBoxesWithThem[thisKey], reverse=True); return {"dictMappingIndexToBox" : dictMappingIndexToBox, \ "dictMappingCornerToIDsOfBoxesWithThem" : dictMappingCornerToIDsOfBoxesWithThem, \ "dictMappingIDsOfBoxesToTheirCorners" : dictMappingIDsOfBoxesToTheirCorners}; def addBox(dictMappingIndexToBox, dictMappingCornerToIDsOfBoxesWithThem, \ dictMappingIDsOfBoxesToTheirCorners, boxToAdd, precision=3): indexIDOfThisBox = max(dictMappingIndexToBox) ; assert(indexIDOfThisBox + 1 > indexIDOfThisBox); # weak overflow check... indexIDOfThisBox = indexIDOfThisBox + 1; sizeOfThisBox = boxSize(boxToAdd); dictMappingIndexToBox[indexIDOfThisBox] = boxToAdd; cornersOfThisBox = getAllRoughCornerPoints(boxToAdd, precision); cornersOfThisBox = list(set(cornersOfThisBox )); # making sure the corners we consider are unique, # even if the box is flat.... dictMappingIDsOfBoxesToTheirCorners[indexIDOfThisBox] = cornersOfThisBox; for thisCorner in cornersOfThisBox: if(thisCorner not in dictMappingCornerToIDsOfBoxesWithThem): dictMappingCornerToIDsOfBoxesWithThem[thisCorner] = []; dictMappingCornerToIDsOfBoxesWithThem[thisCorner].append((sizeOfThisBox, indexIDOfThisBox)); # this list of boxes that have a certain corner are sorted in reverse order (i.e., descending order) # of size so that we can try to aggressively try and greadily merge boxes... dictMappingCornerToIDsOfBoxesWithThem[thisCorner] = \ sorted(dictMappingCornerToIDsOfBoxesWithThem[thisCorner], reverse=True); return indexIDOfThisBox; def removeBox(dictMappingIndexToBox, dictMappingCornerToIDsOfBoxesWithThem, \ dictMappingIDsOfBoxesToTheirCorners, IDNumberOfBoxToRemove, precision=3): cornersToSearchAndRemoveBoxFrom = dictMappingIDsOfBoxesToTheirCorners[IDNumberOfBoxToRemove]; # removing record of this box dictMappingIndexToBox.pop(IDNumberOfBoxToRemove); dictMappingIDsOfBoxesToTheirCorners.pop(IDNumberOfBoxToRemove); for thisCorner in cornersToSearchAndRemoveBoxFrom: index = 0; while(index < len(dictMappingCornerToIDsOfBoxesWithThem[thisCorner])): if(dictMappingCornerToIDsOfBoxesWithThem[thisCorner][index][1] == IDNumberOfBoxToRemove): dictMappingCornerToIDsOfBoxesWithThem[thisCorner].pop(index); if(len(dictMappingCornerToIDsOfBoxesWithThem[thisCorner]) == 0): # no box has this corner dictMappingCornerToIDsOfBoxesWithThem.pop(thisCorner); break; # recall that we inserted boxes into the dict ensuring that the # the coordinates we got for each box are unique, even if the box is flat... assert(index < index + 1); # weak overflow check... index = index + 1; # below assert checks that the sorting order still remains. Note we must use .get for # the dict-access in the range since we might have gotten rid of the key thisCorner . assert(\ all([ (dictMappingCornerToIDsOfBoxesWithThem[thisCorner][index][0] >= \ dictMappingCornerToIDsOfBoxesWithThem[thisCorner][index+1][0]) \ for index in range(0, len(dictMappingCornerToIDsOfBoxesWithThem.get(thisCorner, [])) -1)])); return; import sys; from utils.contracts import *; def mergeBoxesInTargetSet(dictMappingIndexToBox, dictMappingCornerToIDsOfBoxesWithThem, \ dictMappingIDsOfBoxesToTheirCorners, cornerToExamine, precision=3): listOfIDsOfBoxesToMerge = dictMappingCornerToIDsOfBoxesWithThem.get(cornerToExamine, []); # NOTE: this means that # if boxes overlap but DO NOT share a corner, THEN THIS CODE WILL NEVER MERGE IT. thisFirstBoxIndex = 0; while(thisFirstBoxIndex < len(listOfIDsOfBoxesToMerge)): thisSecondBoxIndex = thisFirstBoxIndex + 1; while(thisSecondBoxIndex < len(listOfIDsOfBoxesToMerge)): thisFirstBoxID = listOfIDsOfBoxesToMerge[thisFirstBoxIndex][1]; thisSecondBoxID = listOfIDsOfBoxesToMerge[thisSecondBoxIndex][1]; if(boxesCanMergeWithNoGap(dictMappingIndexToBox[thisFirstBoxID], \ dictMappingIndexToBox[thisSecondBoxID])): newBox = getContainingBox([dictMappingIndexToBox[thisFirstBoxID], \ dictMappingIndexToBox[thisSecondBoxID]]); # NOTE: We MUST add in a box prior to removing others to prevent # the mechanisms in add-box from (1) repeating an ID (this actually # should not impact correctness, but I suppose it is nice... actually, # that is silly), (2) make sure that we do not work over an # empty sequence in the list for that corner (... i.e., violate # an invariant... that sounds more like a bug in the code than # an issue with doing operations in a certain order).... addBox(dictMappingIndexToBox, dictMappingCornerToIDsOfBoxesWithThem, \ dictMappingIDsOfBoxesToTheirCorners, newBox, precision=precision); for thisIDToRemove in [thisFirstBoxID, thisSecondBoxID]: removeBox(dictMappingIndexToBox, \ dictMappingCornerToIDsOfBoxesWithThem, dictMappingIDsOfBoxesToTheirCorners,\ thisIDToRemove, precision=precision); if(cornerToExamine not in dictMappingCornerToIDsOfBoxesWithThem): return; # There should be a more efficient way of encorporating new boxes # instead of doing the below. listOfIDsOfBoxesToMerge = dictMappingCornerToIDsOfBoxesWithThem[cornerToExamine]; thisFirstBoxIndex = 0; thisSecondBoxIndex = 1; continue; assert( thisSecondBoxIndex + 1 > thisSecondBoxIndex);# weak overflow check. thisSecondBoxIndex = thisSecondBoxIndex + 1; assert(thisFirstBoxIndex + 1 > thisFirstBoxIndex); # weak overflow check. thisFirstBoxIndex = thisFirstBoxIndex + 1 return; def mergeBoxes(listOfInitialBoxes, precision=3, maxNumberOfIterations=None): requires(isinstance(maxNumberOfIterations, type(None)) or isinstance(maxNumberOfIterations, int)); requires(isinstance(maxNumberOfIterations, type(None)) or (maxNumberOfIterations > 0)); initialDictionary = getSetupDicts(listOfInitialBoxes, precision=precision); dictMappingIndexToBox = initialDictionary["dictMappingIndexToBox"]; dictMappingCornerToIDsOfBoxesWithThem = \ initialDictionary["dictMappingCornerToIDsOfBoxesWithThem"]; dictMappingIDsOfBoxesToTheirCorners = \ initialDictionary["dictMappingIDsOfBoxesToTheirCorners"]; lengthOfPreviousNumberOfBoxes = None; listOfCorners = list(dictMappingCornerToIDsOfBoxesWithThem.keys()); iterationNumber = 0; while(lengthOfPreviousNumberOfBoxes == None or \ lengthOfPreviousNumberOfBoxes != len(dictMappingIndexToBox)): if(maxNumberOfIterations != None and (iterationNumber > maxNumberOfIterations)): break; sys.stdout.flush(); assert(iterationNumber < iterationNumber + 1); # weak overflow check iterationNumber = iterationNumber + 1; lengthOfPreviousNumberOfBoxes = len(dictMappingIndexToBox); randomOrderingOfCorners = np.random.permutation(range(0, len(dictMappingCornerToIDsOfBoxesWithThem))); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg # Randomly- # or trying to randonly- address corners so to avoid constantly # trying to remerge the corners in the same order, which would help form # pathological examples. for thisIndex in randomOrderingOfCorners: thisCorner = listOfCorners[thisIndex]; mergeBoxesInTargetSet(dictMappingIndexToBox, \ dictMappingCornerToIDsOfBoxesWithThem, dictMappingIDsOfBoxesToTheirCorners, \ thisCorner, precision=precision); return {"dictMappingIndexToBox" : dictMappingIndexToBox, \ "dictMappingCornerToIDsOfBoxesWithThem" : dictMappingCornerToIDsOfBoxesWithThem, \ "dictMappingIDsOfBoxesToTheirCorners" : dictMappingIDsOfBoxesToTheirCorners}; from boxesAndBoxOperations.readAndWriteBoxes import *; def mergeBoxesOnRealData(): #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~VV~V~V~V~V # Code copied from analysisOfResults_<redacted>.py #============================================================================== def getBoxesFromFile(): fhDict = {\ "positionAndTheta_notTE" : "CLA_statesWhereNotTE_016410bf-c229-4655-85bf-351c2642d421.bin", \ "positionAndTheta_TE" : "CLA_statesWhereTrueEverywhere_016410bf-c229-4655-85bf-351c2642d421.bin", \ "postionAndThetaAndDerivatives_notTE" : "CLA_statesWhereNotTE_29393cea-4a98-4c63-bf00-4035ebe2aee6.bin", \ "postionAndThetaAndDerivatives_TE" : "CLA_statesWhereTrueEverywhere_29393cea-4a98-4c63-bf00-4035ebe2aee6.bin" \ }; fhDict = {\ "CLA_statesWhereTrueEverywhere_ed43a878-714b-4ff9-be38-2d02dbad9b2c" : \ "CLA_statesWhereTrueEverywhere_ed43a878-714b-4ff9-be38-2d02dbad9b2c.bin"}; boxDict = dict(); for thisKey in fhDict: fhDict[thisKey] = open("resultsCEGARLikeAnalysis/" + fhDict[thisKey], "rb"); boxDict[thisKey] = readBoxes(fhDict[thisKey]); fhDict[thisKey].close(); return boxDict; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ temp = getBoxesFromFile(); mergeBoxes(temp["CLA_statesWhereTrueEverywhere_ed43a878-714b-4ff9-be38-2d02dbad9b2c"]); #"positionAndTheta_notTE"]) return; def generateBoxDivisionsForTesting(numberOfDivisionsAlongAxis, dimensionOfBoxes): requires(isinstance(numberOfDivisionsAlongAxis, int)); requires(numberOfDivisionsAlongAxis > 0); axisIntervals = [[x, x + 1] for x in range(0, numberOfDivisionsAlongAxis)]; boxesInPreviousInteration = [[]]; dimensionIndex = 0; while(dimensionIndex < dimensionOfBoxes): boxesInThisIteration = []; for thisPreviousBox in boxesInPreviousInteration: for thisAddedAxis in axisIntervals: boxesInThisIteration.append(thisPreviousBox + [thisAddedAxis]); boxesInPreviousInteration = boxesInThisIteration; assert(dimensionIndex < dimensionIndex + 1); # weak overflow check... dimensionIndex = dimensionIndex + 1; for thisIndex in range(0, len(boxesInThisIteration)): boxesInThisIteration[thisIndex] = np.array(boxesInThisIteration[thisIndex]); ensures(len(boxesInThisIteration) == numberOfDivisionsAlongAxis ** dimensionOfBoxes); ensures(all([isProperBox(x) for x in boxesInThisIteration])); return boxesInThisIteration; def generateMutlidimensionalBlockStairsForTesting(numberOfGroups, dimension): requires(isinstance(numberOfGroups, int)); requires(numberOfGroups > 0); requires(isinstance(dimension, int)); requires(dimension >= 2); startBox = getRandomBox(dimension); # Note that in-and-of themselves, the four boxes - boxA, boxB, boxC, boxD - # should all be able to merge into a single box, [0,2]x[0,2] # TODO: pick the two dimensions that form the basis of the four boxes merged randomly... boxA = startBox.copy(); boxA[0, :] = np.array([0,1]); boxA[1, :] = np.array([0,1]); boxB = startBox.copy(); boxB[0, :] = np.array([1,2]); boxB[1, :] = np.array([0,1]); boxC = startBox.copy(); boxC[0, :] = np.array([0,1]); boxC[1, :] = np.array([1,2]); boxD = startBox.copy(); boxD[0, :] = np.array([1,2]); boxD[1, :] = np.array([1,2]); boxAAndBAndCAndDMergedTogether = startBox.copy(); boxAAndBAndCAndDMergedTogether[1, :] = np.array([0,2]); boxAAndBAndCAndDMergedTogether[0, :] = np.array([0,2]); listToMerge = []; whatMergedResultShouldBe = []; def copyAndMoveOnDiagonalFourBasicBoxes(displacementOfDiagonal): requires(isinstance(displacementOfDiagonal, float) or isinstance(displacementOfDiagonal, int)); thisCopy = [boxA.copy(), boxB.copy(), boxC.copy(), boxD.copy()]; for thisIndex in range(0, len(thisCopy)): thisCopy[thisIndex][:2,:] = thisCopy[thisIndex][:2,:] + displacementOfDiagonal; return thisCopy; for thisIndex in range(0, numberOfGroups): displacement = 2 * thisIndex; # why times 2? Becuase the side-lengths of # boxAAndBAndCAndDMergedTogether are that long. listToMerge= listToMerge + copyAndMoveOnDiagonalFourBasicBoxes(displacement); whatMergedResultShouldBe.append(boxAAndBAndCAndDMergedTogether.copy()); whatMergedResultShouldBe[-1][:2, :] = whatMergedResultShouldBe[-1][:2, :] + displacement; return {"listToMerge" : listToMerge, "whatMergedResultShouldBe" : whatMergedResultShouldBe}; from boxesAndBoxOperations.getBox import *; def mergeBoxes_quadraticTime_usefulForOutputSpaceBoxes_mergeBoxesThatContainOneAnother(thisListOfBoxes): indicesToKeep = []; indicesToContinueCheckingAgainst = set(range(0, len(thisListOfBoxes))); for thisStartIndex in range(0, len(thisListOfBoxes)): keep = True; for thisEndIndex in indicesToContinueCheckingAgainst: if(thisEndIndex == thisStartIndex): continue; if(boxAContainsBoxB(thisListOfBoxes[thisEndIndex], thisListOfBoxes[thisStartIndex])): keep = False; break; if(keep): indicesToKeep.append(thisStartIndex); else: indicesToContinueCheckingAgainst.remove(thisStartIndex); assert(set(indicesToKeep) == indicesToContinueCheckingAgainst); # this should apply # by the end of the process.... return [thisListOfBoxes[thisIndex] for thisIndex in indicesToKeep];
23,515
54.857482
2,906
py
Fanoos
Fanoos-master/boxesAndBoxOperations/readAndWriteBoxes.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; from boxesAndBoxOperations.getBox import getBox, getRandomBox, isProperBox, getDimensionOfBox; import struct; import numpy as np; def writeBox(fileHandle, thisBox, metaData): requires(isProperBox(thisBox)); requires(isinstance(metaData, list)); requires(len(metaData) == 2); requires(isinstance(metaData[0], int)); requires(isinstance(metaData[1], int)); requires(metaData[0] >= 0); requires(metaData[1] >= 0); requires(metaData[0] <= 255); requires(metaData[1] <= 255); # Below we write out the metadata for thisMetaDataInt in metaData: fileHandle.write(struct.pack( "B", thisMetaDataInt)); numberOfBytesThatAFloatHasOnThisSystem = 4; # makes assumptions # about the system running this code, unless the library # in question standardized it.... numberOfDimensions = getDimensionOfBox(thisBox); boundIndex = 0; while(boundIndex < 2): variableIndex = 0; while(variableIndex < numberOfDimensions): valueToWriteOut = thisBox[variableIndex, boundIndex]; assert(isinstance(valueToWriteOut, float)); numberOfBytesWritten = fileHandle.write(struct.pack( "f", valueToWriteOut)); assert(numberOfBytesWritten == numberOfBytesThatAFloatHasOnThisSystem); assert(variableIndex < variableIndex + 1); # weak overflow check... should not # be necessary in python, but always good practice to include. variableIndex = variableIndex + 1; assert(boundIndex in {0, 1}); boundIndex = boundIndex + 1 return; def readBoxFromFile(fhToRead, dimensionOfBox, numberOfBytesThatAFloatHasOnThisSystem): requires(isinstance(dimensionOfBox, int)); requires(dimensionOfBox > 0); boxToReturn = np.zeros((dimensionOfBox, 2)); valueToUnpack = [fhToRead.read(1), fhToRead.read(1)]; if(len(valueToUnpack[0]) == 0): return None; metaData = [struct.unpack( "B", valueToUnpack[x])[0] for x in [0,1]]; # somewhat hacky way to write this..... boundIndex = 0; # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg while(boundIndex < 2): variableIndex = 0; while(variableIndex < dimensionOfBox): valueToUnpack = fhToRead.read(numberOfBytesThatAFloatHasOnThisSystem); if(len(valueToUnpack) == 0): assert(boundIndex == 0 and variableIndex == 0); # in other words, not just # part of the way through reading a box. return None; unpackedValue = struct.unpack( "f", valueToUnpack)[0]; assert(isinstance(unpackedValue, float)); assert(not np.isnan(unpackedValue)); boxToReturn[variableIndex , boundIndex] = unpackedValue; assert(variableIndex < variableIndex + 1); # weak overflow check... should not # be necessary in python, but always good practice to include. variableIndex = variableIndex + 1; assert(boundIndex in {0, 1}); boundIndex = boundIndex + 1 ensures(isProperBox(boxToReturn)); return (boxToReturn, metaData); def readBoxes(fhToRead): numberOfBytesThatAStructIntHasOnThisSystem = 4; numberOfBytesThatAFloatHasOnThisSystem = 4; # makes assumptions # about the system running this code, unless the library # in question standardized it.... contentOfFile = fhToRead.read(numberOfBytesThatAStructIntHasOnThisSystem); dimensionOfBoxes = struct.unpack("i", contentOfFile)[0]; assert(isinstance(dimensionOfBoxes, int)); assert(dimensionOfBoxes > 0); numberOfFloatsNeededToDefineABox = 2 * dimensionOfBoxes; # one set of dimensionOfBoxes # many floats to define the lower bound, one set of dimensionOfBoxes to define the # upper bound. numberOfBytesPerBox = numberOfBytesThatAFloatHasOnThisSystem * numberOfFloatsNeededToDefineABox; listOfBoxesToReturn = []; boxIndex = 0; while(True): thisPotentialBox = readBoxFromFile(fhToRead,dimensionOfBoxes,\ numberOfBytesThatAFloatHasOnThisSystem); if(not isinstance(thisPotentialBox, type(None))): listOfBoxesToReturn.append(thisPotentialBox); else: break; assert(boxIndex < boxIndex + 1); # weak overflow check... should not # be necessary in python, but always good practice to include. boxIndex = boxIndex + 1; assert(len(listOfBoxesToReturn) == boxIndex); # assert(boxIndex == numberOfBoxes); return listOfBoxesToReturn;
8,517
52.2375
2,803
py
Fanoos
Fanoos-master/boxesAndBoxOperations/splitBox.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import numpy as np; # The scalingFactors option below is useful for times when the box under # consideration is composed of axis that represent different units of have # different sensitivites. For example, if a box is (number of inches to the moon) X (radians of orientation of travel) # , it may be allowable to have the first coordinate vary by +/-4 without # effecting anything , while if the second coordinate can change by that much, what # it represents can before completely altered. That is an example of both a difference # in scale and a difference in type leading to scaling to be desireable. Another example # that is simply a difference in scaling is (number of inches to reach the moon) x (number of inches surgery is from major artery) # # When scalingFactors as a nan value in a position, the resulting # box in not split along the axis that has that nan value... def splitBox(thisBox, stringSpecifyingStyleOfSplit, scalingFactors=None): requires(isProperBox(thisBox)); requires(isinstance(stringSpecifyingStyleOfSplit, str)); requires((isinstance(scalingFactors, type(None))) or (isinstance(scalingFactors, np.ndarray))); requires((isinstance(scalingFactors, type(None))) or (scalingFactors.shape==tuple([getDimensionOfBox(thisBox)]))); if(isinstance(scalingFactors, type(None))): raise Exception("While scalingFactors = None is supported in splitBox currently, use of " + \ "None as the scalling axis has been disabled elsewhere in the code. Thus, this is a " +\ "sign that the caller is in error - which is why this function (splitBox) is raising an " + \ "exception to alert the user to something that should not be occuring if the code is in " + \ "sync on the latest revisions."); dictMappingStyleOfSplitToFunction = {\ "halvingAllAxis" : splitBox_halvingAllAxis ,\ "halfLongestAxis" : splitBox_halfLongestAxis, \ "randomSplittingLongestAxis" : splitBox_randomSplittingLongestAxis, \ "randomNumberOfUniformSplits_old" : splitBox_randomNumberOfUniformSplits_old, \ "randomNumberOfUniformSplits" : splitBox_randomNumberOfUniformSplits \ }; if(stringSpecifyingStyleOfSplit not in dictMappingStyleOfSplitToFunction.keys()): raise Exception("splitBox: stringSpecifyingStyleOfSplit not in dictMappingStyleOfSplitToFunction.keys()"); boxToSplit = thisBox; if(not isinstance(scalingFactors, type(None))): boxToSplit = thisBox / scalingFactors.reshape((getDimensionOfBox(thisBox), 1)); boxToSplit[np.isnan(boxToSplit)] = 0; rawReturn = dictMappingStyleOfSplitToFunction[stringSpecifyingStyleOfSplit](boxToSplit); valueToReturn = []; if(not isinstance(scalingFactors, type(None))): for thisTempBox in rawReturn: valueToReturn.append(thisTempBox * scalingFactors.reshape((getDimensionOfBox(thisBox), 1))); valueToReturn[-1][np.isnan(scalingFactors)] = \ thisBox[np.isnan(scalingFactors)] else: valueToReturn = rawReturn; # The value valueToReturn is allowed to be None when the splitting # procedure values to produce a new box - that is, the point is a # fixed point for the splitting procedure. Letting the calling # function now this by returning None explicitly can be helpful. ensures((valueToReturn == None) or (isinstance(valueToReturn, list))); ensures((valueToReturn == None) or (len(valueToReturn) >= 2)); ensures((valueToReturn == None) or all([isProperBox(x) for x in valueToReturn])); ensures((valueToReturn == None) or np.all(np.isclose(getContainingBox(valueToReturn), thisBox))); # Below basically says that if valueToReturn is not-None and # the user did not specify a scaling factor, than equality should hold.... this is probably # a skectchy guarentee to even try to make considering we want to allow # more advanced splitting, but for now it looks like it will hold, and it is nice to know. ensures((valueToReturn == None) or (isinstance(scalingFactors, type(None)) or np.all(np.isclose(getContainingBox(valueToReturn), thisBox)))); return valueToReturn; def splitBox_halvingAllAxis(thisBox): raise Exception("Not Yet Implemented"); def splitBox_halfLongestAxis(thisBox): requires(isProperBox(thisBox)); if(np.all(thisBox[:, 0] == thisBox[:, 1])): # The box is completely flat, nothing can be split... return None; indexOfAxisToSplitOn = np.argmax(thisBox[:, 1] - thisBox[:,0]); middleOfAxisValue = np.mean(thisBox[indexOfAxisToSplitOn, :]); boxA = thisBox.copy(); boxB = thisBox.copy(); boxA[indexOfAxisToSplitOn, :] = np.array([thisBox[indexOfAxisToSplitOn, 0], middleOfAxisValue]); boxB[indexOfAxisToSplitOn, :] = np.array([middleOfAxisValue, thisBox[indexOfAxisToSplitOn, 1]]); return [boxA, boxB]; def splitBox_randomSplittingLongestAxis(thisBox): requires(isProperBox(thisBox)); if(np.all(thisBox[:, 0] == thisBox[:, 1])): # The box is completely flat, nothing can be split... return None; indexOfAxisToSplitOn = np.argmax(thisBox[:, 1] - thisBox[:,0]); # middleOfAxisValue = np.mean(thisBox[indexOfAxisToSplitOn, :]); numberOfRandomValuesToUse = np.random.randint(2,5); newCutOffs = \ np.sort(np.random.rand(numberOfRandomValuesToUse) * (thisBox[indexOfAxisToSplitOn,1] - thisBox[indexOfAxisToSplitOn,0]) + thisBox[indexOfAxisToSplitOn,0]); newCutoffs = [thisBox[indexOfAxisToSplitOn, 0]] + list(newCutOffs) + [thisBox[indexOfAxisToSplitOn, 1]]; boxesToReturn = []; assert(len(newCutoffs) > 1); for thisIndex in range(0, len(newCutoffs) - 1): newBox = thisBox.copy(); newBox[indexOfAxisToSplitOn, :] = np.array([newCutoffs[thisIndex], newCutoffs[thisIndex + 1]]); boxesToReturn.append(newBox); return boxesToReturn; def splitBox_randomNumberOfUniformSplits_old(thisBox): requires(isProperBox(thisBox)); if(np.all(thisBox[:, 0] == thisBox[:, 1])): # The box is completely flat, nothing can be split... return None; indexOfAxisToSplitOn = np.argmax(thisBox[:, 1] - thisBox[:,0]); # middleOfAxisValue = np.mean(thisBox[indexOfAxisToSplitOn, :]); numberOfRandomValuesToUse = np.random.randint(0,11); if(numberOfRandomValuesToUse < 4): numberOfRandomValuesToUse = 2; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ else: numberOfRandomValuesToUse = int(numberOfRandomValuesToUse / 2); newCutOffs = \ [ ((float(x) / numberOfRandomValuesToUse) * (thisBox[indexOfAxisToSplitOn,1] - thisBox[indexOfAxisToSplitOn,0]) + thisBox[indexOfAxisToSplitOn,0]) \ for x in range(1, numberOfRandomValuesToUse)]; newCutoffs = [thisBox[indexOfAxisToSplitOn, 0]] + list(newCutOffs) + [thisBox[indexOfAxisToSplitOn, 1]]; boxesToReturn = []; assert(len(newCutoffs) > 1); for thisIndex in range(0, len(newCutoffs) - 1): newBox = thisBox.copy(); newBox[indexOfAxisToSplitOn, :] = np.array([newCutoffs[thisIndex], newCutoffs[thisIndex + 1]]); boxesToReturn.append(newBox); return boxesToReturn; def splitBox_randomNumberOfUniformSplits(thisBox): requires(isProperBox(thisBox)); if(np.all(thisBox[:, 0] == thisBox[:, 1])): # The box is completely flat, nothing can be split... return None; indexOfAxisToSplitOn = np.argmax(thisBox[:, 1] - thisBox[:,0]); # 4/5 probability of numberOfRandomValuesToUse == 2, 1/5 probability of numberOfRandomValuesToUse == 3 numberOfRandomValuesToUse = int(np.random.randint(0,5) >= 4) + 2; newCutOffs = \ [ ((float(x) / numberOfRandomValuesToUse) * (thisBox[indexOfAxisToSplitOn,1] - thisBox[indexOfAxisToSplitOn,0]) + thisBox[indexOfAxisToSplitOn,0]) \ for x in range(1, numberOfRandomValuesToUse)]; newCutoffs = [thisBox[indexOfAxisToSplitOn, 0]] + list(newCutOffs) + [thisBox[indexOfAxisToSplitOn, 1]]; boxesToReturn = []; assert(len(newCutoffs) > 1); for thisIndex in range(0, len(newCutoffs) - 1): newBox = thisBox.copy(); newBox[indexOfAxisToSplitOn, :] = np.array([newCutoffs[thisIndex], newCutoffs[thisIndex + 1]]); boxesToReturn.append(newBox); return boxesToReturn;
12,391
56.637209
2,829
py
Fanoos
Fanoos-master/boxesAndBoxOperations/CEGARFileWrittingManager.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import uuid; from utils.getGitCommitHash import gitCommitHashWhenThisCodeStartedRunning; from utils.getStringTimeNow import *; import re; import struct; import config; def recordMetaDataForStart(thisUUID, dictMappingFileTypeToFileHandleToUse): requires(isinstance(thisUUID, str)); requires(re.match("^[0-9a-f\-]+$", thisUUID) != None); requires(isinstance(dictMappingFileTypeToFileHandleToUse, dict)); requires(list(dictMappingFileTypeToFileHandleToUse.keys()) == \ ['boxes', 'metaData']); dictMappingFileTypeToFileHandleToUse["metaData"].write("UUID:" + thisUUID + "\n"); dictMappingFileTypeToFileHandleToUse["metaData"].write("gitCommitHashWhenThisCodeStartedRunning:" + gitCommitHashWhenThisCodeStartedRunning + "\n"); dictMappingFileTypeToFileHandleToUse["metaData"].write("time started running:" + getStringTimeNow() + "\n"); dictMappingFileTypeToFileHandleToUse["metaData"].flush(); return; def recordMetaDataForFinish(dictMappingFileTypeToFileHandleToUse): requires(isinstance(dictMappingFileTypeToFileHandleToUse, dict)); requires(list(dictMappingFileTypeToFileHandleToUse.keys()) == \ ['boxes', 'metaData']); dictMappingFileTypeToFileHandleToUse["metaData"].write("time finished running:" + getStringTimeNow() + "\n"); dictMappingFileTypeToFileHandleToUse["metaData"].flush(); return; def getFilesToSaveResultsIn(thisUUID, universeBox): requires(isinstance(thisUUID, str)); requires(re.match("^[0-9a-f\-]+$", thisUUID) != None); requires(isProperBox(universeBox)); #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~VV~~V~V~V~VV~V~V~V~V~VV~V~V~V~V~V~V~V~V # Setting up files to write to #------------------------------------------------------------------------- # Below, CLA stands for CEGAR-Like Analysis basePath = "./tmp/CLA_"; dictMappingFileTypeToFileHandleToUse = {\ "boxes" : "",\ "metaData" : "" \ }; for thisKey in dictMappingFileTypeToFileHandleToUse: dictMappingFileTypeToFileHandleToUse[thisKey] = \ basePath + thisKey + "_" + thisUUID; if(thisKey == "boxes"): dictMappingFileTypeToFileHandleToUse[thisKey] = dictMappingFileTypeToFileHandleToUse[thisKey] + ".bin"; fh = open(dictMappingFileTypeToFileHandleToUse[thisKey], "wb"); else: fh = open(dictMappingFileTypeToFileHandleToUse[thisKey], "w"); dictMappingFileTypeToFileHandleToUse[thisKey] = fh; # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~VV~~V~V~V~VV~V~V~V~V~VV~V~V~V~V~V~V~V~V # Writting important meta-data to set up files, etc. #------------------------------------------------------------------------- recordMetaDataForStart(thisUUID, dictMappingFileTypeToFileHandleToUse); dictMappingFileTypeToFileHandleToUse["metaData"].write("universeBox:" + str(universeBox).replace("\n", " ") + "\n"); dictMappingFileTypeToFileHandleToUse["boxes"].write(struct.pack("i", getDimensionOfBox(universeBox))); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ for thisFh in dictMappingFileTypeToFileHandleToUse.values(): thisFh.flush(); ensures(isinstance(dictMappingFileTypeToFileHandleToUse, dict)); ensures(list(dictMappingFileTypeToFileHandleToUse.keys()) == \ ['boxes', 'metaData']); return dictMappingFileTypeToFileHandleToUse; def closeFilesToSaveResultsIn(dictMappingFileTypeToFileHandleToUse): requires(isinstance(dictMappingFileTypeToFileHandleToUse, dict)); requires(list(dictMappingFileTypeToFileHandleToUse.keys()) == \ ['boxes', 'metaData']); recordMetaDataForFinish(dictMappingFileTypeToFileHandleToUse); for thisFH in dictMappingFileTypeToFileHandleToUse.values(): thisFH.flush(); thisFH.close(); return; from boxesAndBoxOperations.readAndWriteBoxes import writeBox; class CEGARFileWrittingManager(): def __init__(self, universeBox): requires(isProperBox(universeBox)); self.uuid = str(uuid.uuid4()); self.universeBox = universeBox; self.dictMappingFileTypeToFileHandleToUse = getFilesToSaveResultsIn(self.uuid, self.universeBox); self.writeBufferForBoxes = []; self.boxBufferSizeLimit = 1000; return; def closeFilesToSaveResultsIn(self): self._flushBoxBuffer(); closeFilesToSaveResultsIn(self.dictMappingFileTypeToFileHandleToUse); return; def _flushBoxBuffer(self): for thisBoxAndMetaData in self.writeBufferForBoxes: writeBox(thisBoxAndMetaData[0], thisBoxAndMetaData[1], thisBoxAndMetaData[2]); for thisKey in self.dictMappingFileTypeToFileHandleToUse: self.dictMappingFileTypeToFileHandleToUse[thisKey].flush(); self.writeBufferForBoxes = []; return; def writeBox(self, thisBox, thisBoxMetaData): requires(isProperBox(thisBox)); requires(getDimensionOfBox(thisBox) == getDimensionOfBox(self.universeBox)); requires(isinstance(thisBoxMetaData, list)); requires(len(thisBoxMetaData) == 2); requires(isinstance(thisBoxMetaData[0], int)); requires(isinstance(thisBoxMetaData[1], int)); requires(thisBoxMetaData[0] >= 0); requires(thisBoxMetaData[1] >= 0); requires(thisBoxMetaData[0] <= 255); requires(thisBoxMetaData[1] <= 255); self.writeBufferForBoxes.append((self.dictMappingFileTypeToFileHandleToUse["boxes"], thisBox, thisBoxMetaData)); if(len(self.writeBufferForBoxes) > self.boxBufferSizeLimit): self._flushBoxBuffer(); return; def writeMetadata(self, tagValue, thisLine): requires(isinstance(tagValue, str)); requires(len(tagValue) > 0); requires("\n" not in tagValue); requires(":" not in tagValue); requires(isinstance(thisLine, str)); self.dictMappingFileTypeToFileHandleToUse["metaData"].write(\ tagValue + ":" + thisLine.replace("\n", "<NEWLINE>") + "\n\n"); self.dictMappingFileTypeToFileHandleToUse["metaData"].flush(); return;
10,336
53.405263
2,843
py
Fanoos
Fanoos-master/boxesAndBoxOperations/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/boxesAndBoxOperations/getBox.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; import numpy as np; # A Box: # A 2d numpy array # 0th coordinate - the variable to consider # 1st coordinate - the minimum (coordinate zero) or maximum (coordinate one) value. def isProperBox(thisProposedBox): if(not(isinstance(thisProposedBox, np.ndarray))): return False; if(len(thisProposedBox.shape) != 2): return False; if(thisProposedBox.shape[1] != 2): return False; if(thisProposedBox.shape[0] <= 0): return False; if(np.any(thisProposedBox[:,0] > thisProposedBox[:,1])): return False; return True; def getBox(minimums, maximums): requires(isinstance(minimums, np.ndarray)); requires(isinstance(maximums, np.ndarray)); requires(len(maximums.shape) == 1); requires(maximums.shape[0] > 0); requires(maximums.shape == maximums.shape); requires(np.all(maximums <= maximums)); # TODO: consider making it np.float64 instead of just float below.. thisBox = np.array([minimums, maximums], dtype=float).transpose(); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ ensures(isProperBox(thisBox)); return thisBox; def getDimensionOfBox(thisBox): requires(isProperBox(thisBox)); valToReturn = thisBox.shape[0]; ensures(isinstance(valToReturn, int)); ensures(valToReturn > 0); return valToReturn ; def getJointBox(listOfBoxesToFindJiontOf): # For example of what this does, consider a box over variables [x,y] # and a box over [w,z]. The "jiont box" defines a box over [x,y,w,z]; # WE MUST GUARENTEE THE ORDER OF INPUT IS PRESERVED IN THE OUTPUT so # that we know which coordinate coresponds to which variable. requires(isinstance(listOfBoxesToFindJiontOf, list)); requires(len(listOfBoxesToFindJiontOf) >= 1); requires(np.all(isProperBox(x) for x in listOfBoxesToFindJiontOf)); thisJointBox = np.concatenate(listOfBoxesToFindJiontOf, axis=0) ensures(isProperBox(thisJointBox)); ensures(getDimensionOfBox(thisJointBox) == \ sum([getDimensionOfBox(thisBox) for thisBox in listOfBoxesToFindJiontOf])); return thisJointBox; def getContainingBox(listOfBoxesToFindContainingBoxOf): requires(isinstance(listOfBoxesToFindContainingBoxOf, list)); requires(len(listOfBoxesToFindContainingBoxOf) >= 1); requires(np.all(isProperBox(x) for x in listOfBoxesToFindContainingBoxOf)); # Below basically says that all boxes must be over the same # dimension (if not, then certaintly the boxes cannot be over the same # set of variables, but since we do not store variable names, we # cannot check that directly) requires(len(set([getDimensionOfBox(thisBox) for thisBox in listOfBoxesToFindContainingBoxOf])) == 1); maxValues = np.max([x[:,1] for x in listOfBoxesToFindContainingBoxOf], axis=0); minValues = np.min([x[:,0] for x in listOfBoxesToFindContainingBoxOf], axis=0); thisBox = getBox(minValues, maxValues); ensures(isProperBox(thisBox)); # By the requires, we know that the below is box safe (accessing # index zero of listOfBoxesToFindContainingBoxOf) and means that # thisBox has the same dimension as each member of listOfBoxesToFindContainingBoxOf ; ensures(getDimensionOfBox(thisBox) == getDimensionOfBox(listOfBoxesToFindContainingBoxOf[0])); return thisBox; def getRandomBox(numberOfVariables): requires(isinstance(numberOfVariables, int)); requires(numberOfVariables > 0); thisBox = np.random.rand(numberOfVariables, 2); thisBox[:,0] = -thisBox[:,0]; ensures(isProperBox(thisBox)); return thisBox; def boxSize(thisBox): requires(isProperBox(thisBox)); return np.product(thisBox[:,1] - thisBox[:,0]); def boxContainsVector(thisBox, vector): requires(isProperBox(thisBox)); requires(isinstance(vector, np.ndarray)); requires(len(vector.shape) == 1); if(getDimensionOfBox(thisBox) != vector.shape[0]): return False; # TODO: consider making the condition in the conditional-guard a requirement. # or raising an exception here. return (np.all(thisBox[:, 0] <= vector) and np.all(thisBox[:, 1] >= vector)); def boxAContainsBoxB(thisBoxA, thisBoxB): requires(isProperBox(thisBoxA)); requires(isProperBox(thisBoxB)); if(getDimensionOfBox(thisBoxA) != getDimensionOfBox(thisBoxB)): return False; # TODO: consider making the condition in the conditional-guard a requirement. # or raising an exception here. return (np.all(thisBoxA[:, 0] <= thisBoxB[:, 0]) and np.all(thisBoxA[:, 1] >= thisBoxB[:, 1])); def getRandomVectorInBox(thisBox): convexityParameters = np.random.rand(getDimensionOfBox(thisBox)); thisVector = thisBox[:,0] * convexityParameters + (thisBox[:,1] * (1 - convexityParameters)); ensures(boxContainsVector(thisBox, thisVector)); return thisVector;
8,786
52.254545
2,860
py
Fanoos
Fanoos-master/collectingResults/drawBoxes.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import config; import numpy as np import matplotlib.pyplot as plt from matplotlib.collections import PatchCollection from matplotlib.patches import Rectangle from boxesAndBoxOperations.getBox import *; from utils.contracts import *; class drawBoxes(): def __init__(self, axis1ToProjectOnto, axis2ToProjectOnto, universeBox): requires(isProperBox(universeBox)); requires(isinstance(axis1ToProjectOnto, int)); requires(isinstance(axis2ToProjectOnto, int)); requires(axis1ToProjectOnto >= 0); requires(axis2ToProjectOnto >= 0); requires(axis1ToProjectOnto < getDimensionOfBox(universeBox)); requires(axis2ToProjectOnto < getDimensionOfBox(universeBox)); self.axis1ToProjectOnto = axis1ToProjectOnto; self.axis2ToProjectOnto = axis2ToProjectOnto; self.universeBox = universeBox; self.relavantSubVolumeOfUniverseBox = self.volumeOfBoxExcludingChosenAxis(self.universeBox); return; def convertBoxToPatch(self, thisBox): requires(isProperBox(thisBox)); xVals = thisBox[self.axis1ToProjectOnto, :]; assert(xVals[0] <= xVals[1]); yVals = thisBox[self.axis2ToProjectOnto, :]; assert(yVals[0] <= yVals[1]); thisAlphaValue = self.volumeOfBoxExcludingChosenAxis(thisBox) / self.relavantSubVolumeOfUniverseBox; assert(thisAlphaValue >= 0.0); assert(thisAlphaValue <= 1.0); return Rectangle((xVals[0], yVals[0]), xVals[1] - xVals[0], yVals[1] - yVals[0], alpha=thisAlphaValue, ec=(0,0,0,1), linewidth=1.5,facecolor="g"); @staticmethod def getPlotLimits(thisAxisIndex, theseBoxes): requires(hasattr(theseBoxes, '__iter__')); # checking it is iterable... requires(all([isProperBox(x) for x in theseBoxes])); requires(thisAxisIndex >= 0); requires(thisAxisIndex < getDimensionOfBox(theseBoxes[0])); minVal = np.min([thisBox[thisAxisIndex, 0] for thisBox in theseBoxes]); maxVal = np.max([thisBox[thisAxisIndex, 1] for thisBox in theseBoxes]); return (minVal, maxVal); def volumeOfBoxExcludingChosenAxis(self, thisBox): denominator = np.prod(np.diff(thisBox[[self.axis1ToProjectOnto, self.axis2ToProjectOnto], :], axis=1)); if(np.isclose(denominator, 0.0)): return 0.0; return np.prod(np.diff(thisBox, axis=1)) / np.prod(np.diff(thisBox[[self.axis1ToProjectOnto, self.axis2ToProjectOnto], :], axis=1)); def setAxisLimits(self, ax, theseBoxes): requires(hasattr(theseBoxes, '__iter__')); # checking it is iterable... requires(all([isProperBox(x) for x in theseBoxes])); firstAxisLimits = self.getPlotLimits(self.axis1ToProjectOnto, theseBoxes); twoDotFivePercentOfRangeOfFirstAxis = 0.025 * (firstAxisLimits[1] - firstAxisLimits[0]); secondAxisLimits = self.getPlotLimits(self.axis2ToProjectOnto, theseBoxes); twoDotFivePercentOfRangeOfSecondAxis = 0.025 * (secondAxisLimits[1] - secondAxisLimits[0]); ax.set_xlim(firstAxisLimits[0] - twoDotFivePercentOfRangeOfFirstAxis, \ firstAxisLimits[1] + twoDotFivePercentOfRangeOfFirstAxis); ax.set_ylim(secondAxisLimits[0] - twoDotFivePercentOfRangeOfSecondAxis, \ secondAxisLimits[1] + twoDotFivePercentOfRangeOfSecondAxis); return; def drawBoxesOnAxis(self, ax, fig, theseBoxes, boxColors): requires(hasattr(theseBoxes, '__iter__')); # checking it is iterable... requires(all([isProperBox(x) for x in theseBoxes])); requires(hasattr(boxColors, '__iter__')); # checking it is iterable... requires(len(theseBoxes) == len(boxColors)); # probably not safe to check given the # weaker requirements above.... requires(all([isinstance(x, str) for x in boxColors])); self.miniminumVolBox = np.min([self.volumeOfBoxExcludingChosenAxis(x) for x in theseBoxes]); boxesConvertedToPltRectangles = []; for thisBox in theseBoxes: boxesConvertedToPltRectangles.append(self.convertBoxToPatch(thisBox)); [ax.add_artist(x) for x in boxesConvertedToPltRectangles]; self.setAxisLimits(ax, theseBoxes); leftPadSize = (ax.get_xlim()[1] - ax.get_xlim()[0]) * 0.07; hieght = ax.get_ylim()[1] - ax.get_ylim()[0]; leftX = ax.get_xlim()[0] - leftPadSize; alphaScale = [ Rectangle((leftX, ax.get_ylim()[0] + x * hieght),\ leftPadSize, hieght * 0.2, alpha=(x * 1.25), ec=(0,0,0,1), linewidth=1.5,facecolor="g") for x in \ [0.0, 0.20, 0.4, 0.6, 0.8] ]; [ax.add_artist(x) for x in alphaScale]; [ax.text( (ax.get_xlim()[0] - leftPadSize * 0.66), (ax.get_ylim()[0] + (x + 0.1) * hieght), str((x * 1.25)) , rotation=90, fontsize=20) for x in \ [0.0, 0.20, 0.4, 0.6, 0.8] ]; ax.set_xlim(ax.get_xlim()[0] - leftPadSize, ax.get_xlim()[1]); return; from boxesAndBoxOperations.readAndWriteBoxes import readBoxes; from CEGARLikeAnalysis import labelsForBoxes ; from CEGARLikeAnalysis.labelsForBoxes import *; def helper_frontEnd_drawBox(boxes): requires(all([isProperBox(x) for x in boxes])); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ # Below is a bit of a hack to allow for handling both the input and the output # space boxes in a semi-sensible manner without having to shove the input-space # box as a whole through the model we are trying to analyze. There is nothing wrong # with that, it just would take more work and be a tangent to what we are trying to # do at the moment. universeBox = getBox(np.min(boxes, axis =0)[:, 0], np.max(boxes, axis=0)[:, 1]); print("apparent bounds on boxes:" + str(universeBox)); colors = ['g' for index in boxes]; maxIndex = getDimensionOfBox(universeBox); for firstIndex in range(0, maxIndex): for secondIndex in range(firstIndex + 1, maxIndex): if((firstIndex, secondIndex) != (0, 3)): continue; fig, ax = plt.subplots(1); plt.tick_params(labelsize=20) plt.xlabel('lread', fontsize=40); plt.ylabel('freemem', fontsize=40); DB = drawBoxes(firstIndex,secondIndex, universeBox); DB.drawBoxesOnAxis(ax, fig, boxes, colors); plt.show() return; def frontEnd_drawBox(fileName): requires(isinstance(fileName, str)); requires(len(fileName) > 0); tempFH = open(fileName, "rb"); boxesToPlot = [x[0] for x in readBoxes(tempFH) if ((x[1][1] & labelsForBoxes.LOWESTLEVEL_FALSESOMEWHEREANDEXHAUSTEDLOOKING) > 0)]; tempFH.close(); helper_frontEnd_drawBox(boxesToPlot); return;
10,698
55.909574
2,781
py
Fanoos
Fanoos-master/descriptionGeneration/removePredicatesImpliedByOthers.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox, boxSize, getRandomVectorInBox; import z3; from boxesAndBoxOperations.codeForGettingSamplesBetweenBoxes import getSampleVectorsToCheckAgainst, getBoxCenter; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass,\ Condition_TheBoxItself, MetaCondition_Conjunction; import config; def removePredicatesImpliedByOthers(coveringDescriptionsFiltered, \ dictMappingConditionToBoxesItIsConsistentWith, listOfBoxes, \ listMappingAxisIndexToVariableInQuestion, dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered_initial): z3Solver = coveringDescriptionsFiltered[0].z3Solver; boxItselfConditionIDs = frozenset([x.getID() for x in coveringDescriptionsFiltered if \ isinstance(x, Condition_TheBoxItself)]); conjunctionConditionIDs = frozenset([x.getID() for x in coveringDescriptionsFiltered if \ isinstance(x, MetaCondition_Conjunction)]); assert(boxItselfConditionIDs.isdisjoint(conjunctionConditionIDs)); otherPredicateIDs = frozenset([x.getID() for x in coveringDescriptionsFiltered if \ not ((x.getID() in boxItselfConditionIDs) or (x.getID() in conjunctionConditionIDs))]); assert(otherPredicateIDs.isdisjoint(boxItselfConditionIDs)); assert(conjunctionConditionIDs.isdisjoint(otherPredicateIDs)); orderToConsiderElements = [\ sorted( list(x), \ key=(lambda x: dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered_initial[x][ "uniqueVolumeCovered"] ) \ ) for x in [\ boxItselfConditionIDs, conjunctionConditionIDs, otherPredicateIDs \ ] \ ]; newDescription = [x.getID() for x in coveringDescriptionsFiltered]; perminentSetOfBoxesToCheckOver = set([]); for thisListToConsider in orderToConsiderElements: for thisPredID in thisListToConsider: restOfPreds = [x for x in coveringDescriptionsFiltered if \ ((x.getID() in newDescription) and (x.getID() != thisPredID))]; assert(len(restOfPreds) == len(newDescription) -1 ); boxesCoveredByThisPred = dictMappingConditionToBoxesItIsConsistentWith[thisPredID]; setOfBoxesToCheckOver = perminentSetOfBoxesToCheckOver.union(boxesCoveredByThisPred); removeThisPred = True; for thisBoxIndex in setOfBoxesToCheckOver : verdict = checkIfPredicateRepetativeForThisBox(listOfBoxes[thisBoxIndex], restOfPreds, z3Solver, \ listMappingAxisIndexToVariableInQuestion); if( not verdict ): removeThisPred = False; break; if(removeThisPred): assert(thisPredID in newDescription); newDescription.remove(thisPredID); assert(thisPredID not in newDescription); perminentSetOfBoxesToCheckOver = setOfBoxesToCheckOver; assert(boxesCoveredByThisPred.issubset(perminentSetOfBoxesToCheckOver)); coveringDescriptionsFiltered = [x for x in coveringDescriptionsFiltered if (x.getID() in newDescription)]; return coveringDescriptionsFiltered; def _helper_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition_convertBoxToFormulaConstraints(listMappingAxisIndexToVariableInQuestion, thisBox): requires(isProperBox(thisBox)); requires(getDimensionOfBox(thisBox) == len(listMappingAxisIndexToVariableInQuestion)); F = z3.And([ \ z3.And( float(thisBox[index, 0]) <= listMappingAxisIndexToVariableInQuestion[index], \ listMappingAxisIndexToVariableInQuestion[index] <= float(thisBox[index, 1]) \ ) \ for index in range(0, len(listMappingAxisIndexToVariableInQuestion)) ]); return F; def checkIfPredicateRepetativeForThisBox(thisBox, restOfConditions, z3Solver, listMappingAxisIndexToVariableInQuestion): # TODO: split the two sections below into two functions... #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # for efficiency, some probabilistic checks to see if some quick random sampling # shows that the disjunction of the conditions fail to cover the whole box... #=========================================================================== numberOfSamples=\ config.defaultValues.numberOfStatisticalSamplesToTakeIn_numberOfStatisticalSamplesToTakeIn_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition for thisSampleIndex in range(0, numberOfSamples): randomVector = getRandomVectorInBox(thisBox).reshape(getDimensionOfBox(thisBox), 1); noConditionsHold = True; for thisCondition in restOfConditions: if(thisCondition.pythonFormatEvaluation(randomVector)): noConditionsHold = False; break; if(noConditionsHold): return False; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Formal check in the case that probabilistic sampling wasn't able to disprove # that the disjunction of the conditions to prove the statement... #=========================================================================== z3Solver.reset(); # this might be the expensive.... TODO: check # disjunctive normal form - each element in the list is a clause which we or-together.... formulaToCheck = \ (\ z3.ForAll( listMappingAxisIndexToVariableInQuestion , \ z3.Implies(\ _helper_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition_convertBoxToFormulaConstraints(\ listMappingAxisIndexToVariableInQuestion, thisBox), \ z3.Or([x.z3FormattedCondition for x in restOfConditions]) \ ) \ ) \ ); z3Solver.add(formulaToCheck); verdict = (z3Solver.check() == z3.z3.sat); z3Solver.reset(); return verdict; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^
7,447
43.86747
165
py
Fanoos
Fanoos-master/descriptionGeneration/generateDescription.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox, boxSize; from statesAndOperatorsAndSelection.descriptionState import DescriptionState ; import z3; from boxesAndBoxOperations.codeForGettingSamplesBetweenBoxes import getSampleVectorsToCheckAgainst, getBoxCenter; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass,\ Condition_TheBoxItself, MetaCondition_Conjunction; from boxesAndBoxOperations.splitBox import splitBox; from descriptionGeneration.draftCodeForMulitVariantConditionLearning import getApproximateMultivariateSetCover; from boxesAndBoxOperations.mergeBoxes import mergeBoxes, \ mergeBoxes_quadraticTime_usefulForOutputSpaceBoxes_mergeBoxesThatContainOneAnother ; import config; from descriptionGeneration.removePredicatesImpliedByOthers import removePredicatesImpliedByOthers ; def getConsistentConditions(thisBox, listOfConditions, thisState): requires(isinstance(listOfConditions, list)); requires(all([isinstance(x, CharacterizationConditionsBaseClass) for x in listOfConditions])); requires(isProperBox(thisBox)); requires(isinstance(thisState, DescriptionState)); numberOfSamplesToTry = thisState.readParameter("numberOfSamplesToTry"); assert(isinstance(numberOfSamplesToTry, int)); assert(numberOfSamplesToTry >= 0); indicesOfCandidateConditions = list(range(0, len(listOfConditions))); samplesToTry = getSampleVectorsToCheckAgainst(thisBox, 0.0, 1.0, numberOfSamplesToTry); indicesOfCandidateConditionsAfterFeasibilityCheck = []; for thisConditionIndex in indicesOfCandidateConditions: success = True; for thisSample in samplesToTry: verdict = listOfConditions[thisConditionIndex].pythonFormatEvaluation(thisSample); if(not verdict): success = False; break; if(success): indicesOfCandidateConditionsAfterFeasibilityCheck.append(thisConditionIndex); listOfConsistentCondtions = []; for thisConditionIndex in indicesOfCandidateConditionsAfterFeasibilityCheck: if(listOfConditions[thisConditionIndex].allMembersOfBoxSatisfyCondition(thisBox)): listOfConsistentCondtions.append(thisConditionIndex); return listOfConsistentCondtions; def getMostSpecificCondition(thisBox, listOfConditions, thisState): """ Given a list of predicates that are consistent with the box provided, randomly samples AROUND (not inside) the box provided to determine which of the predicates is most specific. In particular, we sample vectors at increasing distances from the box (based on l_{infinity} norm from the closest box side) and whichever predicates become false first (i.e., at a certain "raduis" of sampling, a subset of the predicates is no longer consistent with the data sampled) are considered more specifically consistent with the box in question. Bear in mind that this is a randomized approach, so results may vary. """ requires(isinstance(listOfConditions, list)); requires(all([isinstance(x, CharacterizationConditionsBaseClass) for x in listOfConditions])); # Below is a sanity check- we want all the members of listOfConditions to be true over every element in # thisBox, so they should at least be true for the vector at the center of the box.... requires(all([ x.pythonFormatEvaluation(getBoxCenter(thisBox)) for x in listOfConditions])); requires(isProperBox(thisBox)); requires(isinstance(thisState, DescriptionState)); numberOfSamplesToTry = thisState.readParameter("numberOfSamplesToTry"); assert(isinstance(numberOfSamplesToTry, int)); assert(numberOfSamplesToTry >= 0); exponentialComponent = thisState.readParameter("exponentialComponent"); assert(isinstance(exponentialComponent, float)); assert(exponentialComponent >= 0.0); numberOfDimensionToCover = getDimensionOfBox(thisBox); if(len(listOfConditions) == 0): return None; dimensionsCovered = set(); minAndMaxInfinityNormPorportionOfDistanceFromBox = [\ [1.0, 1.01], \ [1.01,1.05], \ [1.05, 1.10], \ [1.10, 1.20], \ [1.20, 1.40], \ [1.40, 1.80], \ [1.80, 2.60] ]; for thisRowIndex in range(0, len(minAndMaxInfinityNormPorportionOfDistanceFromBox)): for columnIndex in [0, 1]: thisValue = minAndMaxInfinityNormPorportionOfDistanceFromBox[thisRowIndex][columnIndex]; minAndMaxInfinityNormPorportionOfDistanceFromBox[thisRowIndex][columnIndex] = \ thisValue * np.exp(exponentialComponent * thisValue); setOfIndicesOfCandidateMostSpecificValues = set(); success=False; for thisMinAndMaxPorpDistance in minAndMaxInfinityNormPorportionOfDistanceFromBox: assert(thisMinAndMaxPorpDistance[0] >= 1.0); assert(thisMinAndMaxPorpDistance[0] < thisMinAndMaxPorpDistance[1]); samples = getSampleVectorsToCheckAgainst(thisBox, \ thisMinAndMaxPorpDistance[0], thisMinAndMaxPorpDistance[1], numberOfSamplesToTry); temp_newVariablesCovered = set(); for thisIndex in range(0, len(listOfConditions)): if(thisIndex in setOfIndicesOfCandidateMostSpecificValues): continue; if(set(listOfConditions[thisIndex].relaventVariables()).issubset(dimensionsCovered)): continue; for thisSample in samples: if( (not listOfConditions[thisIndex].pythonFormatEvaluation(thisSample)) ): # and \ setOfIndicesOfCandidateMostSpecificValues.add(thisIndex); temp_newVariablesCovered.update(listOfConditions[thisIndex].relaventVariables()); # notice that we append in the new variables covered only AFTER we are done evaluating for candidate # conditions based on thisMinAndMaxPorpDistance. That way, conditions covering the same variables # but in the same partition are both drawn in, as desired. dimensionsCovered.update(temp_newVariablesCovered) if(len(dimensionsCovered) == numberOfDimensionToCover): assert(len(setOfIndicesOfCandidateMostSpecificValues) > 0); return setOfIndicesOfCandidateMostSpecificValues; assert(len(dimensionsCovered) < numberOfDimensionToCover); if(len(setOfIndicesOfCandidateMostSpecificValues) != 0): assert(len(dimensionsCovered) > 0); return setOfIndicesOfCandidateMostSpecificValues; assert(setOfIndicesOfCandidateMostSpecificValues == set()); return None; def handleNewInstancesOf_MetaCondition_Conjunction(coveringDescriptions, dictMappingConditionToBoxesItIsConsistentWith): """ updates dictMappingConditionToBoxesItIsConsistentWith to list the boxes that newly-introduced conjuncts are consistent with. """ for thisConditionIndex in range(0, len(coveringDescriptions)): if(coveringDescriptions[thisConditionIndex].getID() in dictMappingConditionToBoxesItIsConsistentWith): continue; # Not 100% this is correct to do.... if(isinstance(coveringDescriptions[thisConditionIndex], MetaCondition_Conjunction)): idsOfConjunctsInCondition = coveringDescriptions[thisConditionIndex].getID(); assert(isinstance(idsOfConjunctsInCondition, frozenset)); assert(len(idsOfConjunctsInCondition) > 0); assert( idsOfConjunctsInCondition not in dictMappingConditionToBoxesItIsConsistentWith); setOfBoxesThatConditionIsConsistentWith = None; for thisID in idsOfConjunctsInCondition: boxesConsistentWithThisID = dictMappingConditionToBoxesItIsConsistentWith[thisID].copy(); # CRITICAL TO USE .copy() HERE TO AVOID ACCIDENTALLY MODIFYING THE INFORMATION FOR THE ORIGINAL CONDTIONS USING THE intersection_update BELOW - NOTE THAT THE ORIGINAL CONIDITIONS CAN BE AND LIKLEY ARE IN USE ELSEWHERE.... assert(isinstance(boxesConsistentWithThisID, set)); assert(len(boxesConsistentWithThisID) > 0); # otherwise this condition would not be in the minimal covering..... if(setOfBoxesThatConditionIsConsistentWith == None): setOfBoxesThatConditionIsConsistentWith = boxesConsistentWithThisID; else: setOfBoxesThatConditionIsConsistentWith.intersection_update(boxesConsistentWithThisID); assert(isinstance(setOfBoxesThatConditionIsConsistentWith, set)); assert(len(setOfBoxesThatConditionIsConsistentWith) > 0); assert(isinstance(setOfBoxesThatConditionIsConsistentWith, set)); assert(len(setOfBoxesThatConditionIsConsistentWith) > 0); assert(idsOfConjunctsInCondition not in dictMappingConditionToBoxesItIsConsistentWith); dictMappingConditionToBoxesItIsConsistentWith[idsOfConjunctsInCondition] = setOfBoxesThatConditionIsConsistentWith; assert(idsOfConjunctsInCondition in dictMappingConditionToBoxesItIsConsistentWith); return; def handleNewInstancesOf_BoxItself(coveringDescriptionsFiltered, listOfConditions_after, listMappingAxisIndexToVariableInQuestion, dictMappingConditionToBoxesItIsConsistentWith): """ Takes the description provided (coveringDescriptionsFiltered), and extracts occurances of Condition_TheBoxItself. This function then produces a new description by removing the boxes from the original description and inserting new instances of Condition_TheBoxItself that result from merging the boxes extracted from the original description. """ requires(isinstance(coveringDescriptionsFiltered, list)); requires(all([isinstance(x, CharacterizationConditionsBaseClass) for x in coveringDescriptionsFiltered])); requires({x.getID() for x in coveringDescriptionsFiltered if not isinstance(x, MetaCondition_Conjunction)}.issubset({x.getID() for x in listOfConditions_after})); localMappingFromBoxIDToIndex = dict(); localMappingFromBoxIDToBox = dict(); for thisCondition in coveringDescriptionsFiltered: if( isinstance(thisCondition, Condition_TheBoxItself) ): # Notice that we do the copy below so that, when pop is done after, it does not # effect the original set in dictMappingConditionToBoxesItIsConsistentWith... indexes = dictMappingConditionToBoxesItIsConsistentWith[thisCondition.getID()].copy(); assert(len(indexes) == 1); assert(isinstance(indexes, set)); localMappingFromBoxIDToIndex[thisCondition.getID()] = indexes.pop(); localMappingFromBoxIDToBox[thisCondition.getID()] = thisCondition.personalBox; assert(len(indexes) == 0); listOfCandidateBoxes = [thisCondition.personalBox for thisCondition in coveringDescriptionsFiltered \ if isinstance(thisCondition, Condition_TheBoxItself)]; listOfConditions_after = [thisCondition for thisCondition in listOfConditions_after if not isinstance(thisCondition, Condition_TheBoxItself)]; coveringDescriptionsFiltered = [thisCondition for thisCondition in coveringDescriptionsFiltered \ if not isinstance(thisCondition, Condition_TheBoxItself)]; temp = mergeBoxes(listOfCandidateBoxes, precision=5, maxNumberOfIterations=None); newMergedBoxes = list(temp["dictMappingIndexToBox"].values()); newMergedBoxes = mergeBoxes_quadraticTime_usefulForOutputSpaceBoxes_mergeBoxesThatContainOneAnother(newMergedBoxes); temp = mergeBoxes(listOfCandidateBoxes, precision=5, maxNumberOfIterations=None); newConditionsToAddIn = [\ Condition_TheBoxItself(listOfConditions_after[0].z3Solver, thisBox, listMappingAxisIndexToVariableInQuestion) \ for thisBox in newMergedBoxes]; # note that we have NOT set .listMappingPositionToVariableName for these new variables... for thisCondition in newConditionsToAddIn: listOfConditions_after.append(thisCondition); coveringDescriptionsFiltered.append(thisCondition); # Below is a bit of a hacky and slow way to do things - # ideally we would keep track of this information when we did the original # merges - but it is a reasonable start subsequent to any extensive rewritting... dictMappingConditionToBoxesItIsConsistentWith[thisCondition.getID()] = \ set([localMappingFromBoxIDToIndex[x] for x in localMappingFromBoxIDToIndex.keys() if thisCondition.allMembersOfBoxSatisfyCondition(localMappingFromBoxIDToBox[x]) ]); assert({x.getID() for x in coveringDescriptionsFiltered if not isinstance(x, MetaCondition_Conjunction)}.issubset({x.getID() for x in listOfConditions_after})); return (coveringDescriptionsFiltered, listOfConditions_after, dictMappingConditionToBoxesItIsConsistentWith); def getVolumesCoveredInformation(listOfBoxes, coveringDescriptionsFiltered, dictMappingConditionToBoxesItIsConsistentWith): requires( {x.getID() for x in coveringDescriptionsFiltered}.issubset(set(dictMappingConditionToBoxesItIsConsistentWith.keys()))); #V~V~V~V~V~V~VV~V~VV~V~V~VV~V~V~VV~V~V~V~V~V~V~V~VV~V~VV~~V~V~V~V # Volumes covered information #---------------------------------------------------------------- # Each member of coveringDescriptionsFiltered should uniquely cover at # least on axis of at least one box, or else it would not have been selected # in the approximal set covering... #================================================================ listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt = [set() for x in range(0, len(listOfBoxes))]; for thisConditionIndex in range(0, len(coveringDescriptionsFiltered)): thisConditionID = coveringDescriptionsFiltered[thisConditionIndex].getID(); # this works as intended even when dealing with # instances of MetaCondition_Conjunction indicesOfBoxesCoveredByCondition = dictMappingConditionToBoxesItIsConsistentWith.get(thisConditionID, set()); assert(thisConditionID in dictMappingConditionToBoxesItIsConsistentWith); assert( len(indicesOfBoxesCoveredByCondition) > 0); for thisBoxIndex in indicesOfBoxesCoveredByCondition: listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt[thisBoxIndex].update([thisConditionID]); assert(thisConditionID in listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt[thisBoxIndex]); dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered = \ {x.getID() : {"volumeCovered" : 0.0, "uniqueVolumeCovered" : 0.0} for x in coveringDescriptionsFiltered}; totalVolumeOfBoxesInList = 0.0; for thisBoxIndex in range(0, len(listOfBoxes)): volumeOfThisBox = boxSize(listOfBoxes[thisBoxIndex]); assert((totalVolumeOfBoxesInList + volumeOfThisBox > totalVolumeOfBoxesInList) or np.isclose(volumeOfThisBox, 0.0)); totalVolumeOfBoxesInList = totalVolumeOfBoxesInList + volumeOfThisBox; # Notice that since we do not ensure the boxes in the # list are disjoint, this sum is NOT necessarly the volume of the union of the boxes. However, since we do some merging of # the boxes prior to trying to describe them, it should be reasonable close or have a reasonable correspondance. It is not # overly important, it just provides a reasonable scaling - see the section below titled "normalizing results". conditionsCoveringBox = listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt[thisBoxIndex]; # NOTE: conditionsCoveringBox may be an empty set if the box was covered by an instance of Condition_TheBoxItself for thisCoveringBoxID in conditionsCoveringBox: assert("," not in thisCoveringBoxID); dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[thisCoveringBoxID]["volumeCovered"] = \ dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[thisCoveringBoxID]["volumeCovered"] + volumeOfThisBox; if(len(conditionsCoveringBox) == 1): # TODO: improve this for the multi-dimensional cases (i.e., multi boxes, but each covering non-subset collections of variables) dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[thisCoveringBoxID]["uniqueVolumeCovered"] = \ dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[thisCoveringBoxID]["uniqueVolumeCovered"] + volumeOfThisBox; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Normalizing results #----------------------------------------------------------------- #================================================================= assert(totalVolumeOfBoxesInList >= 0.0); if(totalVolumeOfBoxesInList == 0.0): totalVolumeOfBoxesInList = 1.0; for thisKey in dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered: for thisSubKey in ["volumeCovered", "uniqueVolumeCovered"]: dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[thisKey][thisSubKey] = \ dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered[thisKey][thisSubKey] / totalVolumeOfBoxesInList; dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered["totalVolumeOfBoxesInList"] = totalVolumeOfBoxesInList; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ return dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered; def getInitialListOfConditionsConsistentWithBoxes(\ listOfBoxes, listOfConditions, listMappingAxisIndexToVariableInQuestion, thisState): requires(isinstance(thisState, DescriptionState)); requires(isinstance(listOfConditions, list)); requires(all([isinstance(x, CharacterizationConditionsBaseClass) for x in listOfConditions])); requires(isinstance(listOfBoxes, list)); requires(len(listOfBoxes) > 0); requires(all([isProperBox(thisBox) for thisBox in listOfBoxes])); requires(isinstance(listMappingAxisIndexToVariableInQuestion, list)); requires(all([isinstance(x, z3.z3.ArithRef) for x in listMappingAxisIndexToVariableInQuestion])); requires(len(listMappingAxisIndexToVariableInQuestion) == getDimensionOfBox(listOfBoxes[0])); listOfConditions_after = listOfConditions.copy(); # see the condition where "if(len(consistentConditions) == 0):" # below for why this was necessary... dictMappingConditionToBoxesItIsConsistentWith = dict(); listOfSetsCoveringBox = []; boxIndex = 0; for thisBox in listOfBoxes: consistentConditions = [listOfConditions[x] for x in \ getConsistentConditions(thisBox, listOfConditions, thisState) ]; # below chunk is important for removing redundant (in respect to logical implication) predicates # at the end. for thisCondition in consistentConditions: thisConditionID = thisCondition.getID(); if(thisConditionID not in dictMappingConditionToBoxesItIsConsistentWith): dictMappingConditionToBoxesItIsConsistentWith[thisConditionID] = set(); dictMappingConditionToBoxesItIsConsistentWith[thisConditionID].add(boxIndex); assert(boxIndex + 1 > boxIndex); # weak overflow check.... boxIndex = boxIndex + 1; subIndicesOfMostSpecificConsistentConditions = getMostSpecificCondition(\ thisBox, \ consistentConditions, \ thisState); if( (len(consistentConditions) == 0) or ( subIndicesOfMostSpecificConsistentConditions == None ) ): setOfConditionsCoveringThisBox = set(); produceGreaterAbstraction = thisState.readParameter("produceGreaterAbstraction"); assert(isinstance(produceGreaterAbstraction, bool)); if( (not produceGreaterAbstraction) or (len(consistentConditions) == 0) ): thisConditionForBox = Condition_TheBoxItself(\ listOfConditions[0].z3Solver,thisBox, listMappingAxisIndexToVariableInQuestion); listOfConditions_after.append(thisConditionForBox); # In the line below, we have boxIndex -1 because boxIndex was INCREMENTED prior to reaching # here - the boxIndex corresponding to thisBox is thus (boxIndex -1) dictMappingConditionToBoxesItIsConsistentWith[thisConditionForBox.getID()] = set([boxIndex -1]); assert(boxIndex -1 >= 0); assert(boxIndex -1 < boxIndex); # though technically not possible in python integers, checking # integer arithmetic behavior.... though, given the assert prior, underflow should be impossible... setOfConditionsCoveringThisBox.update([thisConditionForBox.getID()]); else: setOfConditionsCoveringThisBox.update({x.getID() for x in consistentConditions}); listOfSetsCoveringBox.append(setOfConditionsCoveringThisBox); else: mostSpecificConsistentConditions = [consistentConditions[x] for x in \ subIndicesOfMostSpecificConsistentConditions]; listOfSetsCoveringBox.append({x.getID() for x in mostSpecificConsistentConditions}); return (listOfConditions_after, dictMappingConditionToBoxesItIsConsistentWith, listOfSetsCoveringBox); import time; def generateDescription(listOfBoxes, listOfConditions, listMappingAxisIndexToVariableInQuestion, thisState): requires(isinstance(listOfConditions, list)); requires(all([isinstance(x, CharacterizationConditionsBaseClass) for x in listOfConditions])); requires(isinstance(listOfBoxes, list)); requires(all([isProperBox(thisBox) for thisBox in listOfBoxes])); requires(isinstance(listMappingAxisIndexToVariableInQuestion, list)); requires(all([isinstance(x, z3.z3.ArithRef) for x in listMappingAxisIndexToVariableInQuestion])); requires( len(listOfBoxes) == 0 or \ len(listMappingAxisIndexToVariableInQuestion) == getDimensionOfBox(listOfBoxes[0])); requires(isinstance(thisState, DescriptionState)); if(len(listOfBoxes) == 0): raise NotImplementedError("There is No Situation where the State of Affairs Asked-About Occurs "+ \ "(in terms of mechanics, this means no boxes where found containing any elements "+\ "that are applicable to your question)."); assert(len(listOfBoxes) > 0); (listOfConditions_after, dictMappingConditionToBoxesItIsConsistentWith, listOfSetsCoveringBox) = \ getInitialListOfConditionsConsistentWithBoxes(\ listOfBoxes, listOfConditions, listMappingAxisIndexToVariableInQuestion, thisState); # Below two variables are for use in the set coverings. dimensionOfSpace = getDimensionOfBox(listOfBoxes[0]); dictMappingBElementsToConditions = {x.getID() : x for x in listOfConditions_after}; for thisElem in dictMappingBElementsToConditions: if(isinstance(thisElem, frozenset)): # occurs in the case of an instance of MetaCondition_Conjunction for thisSubElem in thisElem: dictMappingBElementsToConditions[thisSubElem] = [x for x in listOfConditions_after if x.getID() == thisSubElem][0]; coveringDescriptionsInitial = getApproximateMultivariateSetCover(listOfSetsCoveringBox, dimensionOfSpace, dictMappingBElementsToConditions); assert(isinstance(coveringDescriptionsInitial, list)); coveringDescriptionsFiltered = set(); # coveringDescriptionsInitial.copy(); coveringDescriptionsInitial = list(coveringDescriptionsInitial); listOfConditions_after = listOfConditions_after + list(coveringDescriptionsInitial); handleNewInstancesOf_MetaCondition_Conjunction(coveringDescriptionsInitial, dictMappingConditionToBoxesItIsConsistentWith) listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt = [set() for x in range(0, len(listOfBoxes))]; dictMappingBElementsToConditions = dict(); forCheckPurposes_setOfBoxesCovered = set(); for thisConditionIndex in range(0, len(coveringDescriptionsInitial)): thisConditionID = coveringDescriptionsInitial[thisConditionIndex].getID(); dictMappingBElementsToConditions[thisConditionID] = coveringDescriptionsInitial[thisConditionIndex]; indicesOfBoxesCoveredByCondition = dictMappingConditionToBoxesItIsConsistentWith[thisConditionID]; assert(len(indicesOfBoxesCoveredByCondition) > 0); # otherwise it should not be in the minimal set covering # description... assert(isinstance(indicesOfBoxesCoveredByCondition, set)); forCheckPurposes_setOfBoxesCovered.update(indicesOfBoxesCoveredByCondition); assert(forCheckPurposes_setOfBoxesCovered.issuperset(indicesOfBoxesCoveredByCondition)); for thisBoxIndex in indicesOfBoxesCoveredByCondition: listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt[thisBoxIndex].update([thisConditionID]); assert(thisConditionID in listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt[thisBoxIndex]); assert(forCheckPurposes_setOfBoxesCovered == {x for x in range(0, len(listOfBoxes))}); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg for thisElem in listOfConditions: dictMappingBElementsToConditions[thisElem.getID()] = thisElem; coveringDescriptionsFiltered = list(getApproximateMultivariateSetCover(\ listMappingFromBoxIndexToSetOfChosenConditionsCoveringIt, dimensionOfSpace, dictMappingBElementsToConditions)); listOfConditions_after = listOfConditions_after + list(coveringDescriptionsFiltered); handleNewInstancesOf_MetaCondition_Conjunction(coveringDescriptionsFiltered, dictMappingConditionToBoxesItIsConsistentWith) dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered_initial = \ getVolumesCoveredInformation(listOfBoxes, coveringDescriptionsFiltered, \ dictMappingConditionToBoxesItIsConsistentWith); coveringDescriptionsFiltered = removePredicatesImpliedByOthers(\ coveringDescriptionsFiltered, dictMappingConditionToBoxesItIsConsistentWith, \ listOfBoxes, listMappingAxisIndexToVariableInQuestion, dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered_initial); (coveringDescriptionsFiltered, listOfConditions_after, dictMappingConditionToBoxesItIsConsistentWith) = \ handleNewInstancesOf_BoxItself( \ coveringDescriptionsFiltered, listOfConditions_after, listMappingAxisIndexToVariableInQuestion, dictMappingConditionToBoxesItIsConsistentWith); dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered = \ getVolumesCoveredInformation(listOfBoxes, coveringDescriptionsFiltered, dictMappingConditionToBoxesItIsConsistentWith); assert(set(listOfConditions_after).issuperset(listOfConditions)); ensures({x.getID() for x in coveringDescriptionsFiltered if not isinstance(x, MetaCondition_Conjunction)}.issubset({x.getID() for x in listOfConditions_after})); ensures({x.getID() for x in coveringDescriptionsFiltered}.union(set(["totalVolumeOfBoxesInList"])) == \ set(dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered.keys())); return {"description" : coveringDescriptionsFiltered, \ "listOfConditions_after" : listOfConditions_after ,\ "dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered" : dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered} ;
31,769
63.836735
2,874
py
Fanoos
Fanoos-master/descriptionGeneration/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/descriptionGeneration/draftCodeForMulitVariantConditionLearning.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import uuid; import re; import struct; from boxesAndBoxOperations.codeForGettingSamplesBetweenBoxes import getSampleVectorsToCheckAgainst, getBoxCenter; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass,\ Condition_TheBoxItself, MetaCondition_Conjunction; import config; # A : the list index - boxes... # B : the elements in the sets - conditions... def getApproximateMultivariateSetCover(listOfSets, dimensionOfSpace, dictMappingBElementsToConditions): requires(isinstance(listOfSets, list)); requires(all([isinstance(x, set) for x in listOfSets])); requires(isinstance(dimensionOfSpace, int)); requires(dimensionOfSpace > 0); requires(isinstance(dictMappingBElementsToConditions, dict)); requires(all([listOfSets[x].issubset(list(dictMappingBElementsToConditions.keys())) \ for x in range(0, len(listOfSets))])); # recomputing list(dictMappingBElementsToConditions.keys()) here is expensive, # and doing the subset check for each element is expensive.... requires(all([isinstance(x, CharacterizationConditionsBaseClass) \ for x in dictMappingBElementsToConditions.values()])); # NOTE: at the moment, dimensionOfSpace is not used, but I can see it potentially being useful. So, for now, # assuming it is of minimum effort to provide, we leave it in as a required parameter in case it ultimately # proves to be useful for efficiency or decision making later. # This implements the greedy approximate solution to the # set-cover problem, where the universe to cover is the # indices of the list (i.e., list(range(0, len(listOfSets))) ) # https://en.wikipedia.org/wiki/Set_cover_problem # The phrase "UsedToCover" in the below variable name is important: these are not # the boxes that the condition happens to cover, but the boxes that were present when it # became the clear best decision to include this box in the results. forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover = dict(); dictMappingAToCoveringB = dict(); dictMappingBElementsToAElementsTheyCover = dict(); for thisIndex in range(0, len(listOfSets)): assert(thisIndex not in dictMappingAToCoveringB); dictMappingAToCoveringB[thisIndex] = listOfSets[thisIndex]; for thisBElement in listOfSets[thisIndex]: if(thisBElement not in dictMappingBElementsToAElementsTheyCover): dictMappingBElementsToAElementsTheyCover[thisBElement] = set(); assert(thisIndex not in dictMappingBElementsToAElementsTheyCover[thisBElement]); dictMappingBElementsToAElementsTheyCover[thisBElement].add(thisIndex); assert(thisIndex in dictMappingBElementsToAElementsTheyCover[thisBElement]); assert(thisIndex in dictMappingAToCoveringB); dictMappingAElementToVariablesCoveredForEachSet = {x : set() for x in dictMappingAToCoveringB}; while(len(dictMappingAToCoveringB) > 0): maxCoveringBElement = None; AElementsCovered = set(); for thisBElement in dictMappingBElementsToAElementsTheyCover: if(\ (len(dictMappingBElementsToAElementsTheyCover[thisBElement]) > len(AElementsCovered)) or \ (len(dictMappingBElementsToAElementsTheyCover[thisBElement]) == len(AElementsCovered) and np.random.rand() > 0.5) \ ): maxCoveringBElement = thisBElement; AElementsCovered = dictMappingBElementsToAElementsTheyCover[thisBElement].copy(); if(maxCoveringBElement not in forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover): forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover[maxCoveringBElement] = set(); forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover[maxCoveringBElement].update(AElementsCovered); assert(maxCoveringBElement in forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover); variablesCoveredByTheMaxElement = dictMappingBElementsToConditions[maxCoveringBElement].relaventVariables(); assert(isinstance(variablesCoveredByTheMaxElement, frozenset)); assert(len(variablesCoveredByTheMaxElement) > 0); for thisAElement in AElementsCovered: assert(thisAElement in dictMappingAToCoveringB); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg dictMappingAElementToVariablesCoveredForEachSet[thisAElement].update(variablesCoveredByTheMaxElement); assert(dictMappingAElementToVariablesCoveredForEachSet[thisAElement].issuperset(variablesCoveredByTheMaxElement)); BElementsCoverThisAElement = dictMappingAToCoveringB[thisAElement].copy(); assert(len(BElementsCoverThisAElement) > 0); assert(isinstance(BElementsCoverThisAElement, set)); for thisBElement in BElementsCoverThisAElement: if(not dictMappingBElementsToConditions[thisBElement].relaventVariables().issubset(dictMappingAElementToVariablesCoveredForEachSet[thisAElement])): continue; # make a dictionary mapping each box to the set of variables for it that have been covered.... assert(thisAElement in dictMappingBElementsToAElementsTheyCover[thisBElement]); dictMappingBElementsToAElementsTheyCover[thisBElement].remove(thisAElement); assert(thisAElement not in dictMappingBElementsToAElementsTheyCover[thisBElement]); assert(thisBElement in dictMappingAToCoveringB[thisAElement]); dictMappingAToCoveringB[thisAElement].remove(thisBElement); assert(thisBElement not in dictMappingAToCoveringB[thisAElement]); if(len(dictMappingAToCoveringB[thisAElement]) == 0): BElementsCoverThisAElement = dictMappingAToCoveringB.pop(thisAElement); assert(BElementsCoverThisAElement == set()); # The below note is the same as the one done for forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover, # done for exactly the same reason... # The phrase "UsedToCover" in the below variable name is important: these are not # the boxes that the condition happens to cover, but the boxes that were present when it # became the clear best decision to include this box in the results. forFinalResult_dictMappingAElementToBElementsThatWereUsedToCoverIt = {x : set() for x in range(0, len(listOfSets))} ; for thisBElement in forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover: for thisAElement in forFinalResult_dictMappingBElementToAElementTheyWereUsedToCover[thisBElement]: forFinalResult_dictMappingAElementToBElementsThatWereUsedToCoverIt[thisAElement].add(thisBElement); setOfSetsOfBElementsToCouple = {frozenset(x) for x in forFinalResult_dictMappingAElementToBElementsThatWereUsedToCoverIt.values()}; assert(all(len(x) > 0 for x in setOfSetsOfBElementsToCouple)); finalListOfElementsToReturn = []; for thisSet in setOfSetsOfBElementsToCouple: assert(len(thisSet) > 0); tempA = set(); for thisCondID in thisSet: if(isinstance(thisCondID, frozenset)): tempA.update(thisCondID); else: assert(isinstance(thisCondID, str)); tempA.add(thisCondID); listOfConditionsWithIDsInThisSet = [ dictMappingBElementsToConditions[x] for x in tempA ]; if(len(thisSet) == 1): finalListOfElementsToReturn.append( \ listOfConditionsWithIDsInThisSet[0] ); else: assert(len(thisSet) > 1); finalListOfElementsToReturn.append( \ MetaCondition_Conjunction(listOfConditionsWithIDsInThisSet)); return finalListOfElementsToReturn;
11,996
64.201087
2,844
py
Fanoos
Fanoos-master/domainsAndConditions/classesDefiningQuestions.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from propagateBoxThroughLearnedSystem.classesToPropogateBoxThroughModels import ModelBoxProgatorManager; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox, getRandomVectorInBox, getRandomVectorInBox, boxSize; import uuid; from utils.getGitCommitHash import gitCommitHashWhenThisCodeStartedRunning; from utils.getStringTimeNow import *; import re; import struct; from boxesAndBoxOperations.readAndWriteBoxes import readBoxes; from boxesAndBoxOperations.splitBox import splitBox; from CEGARLikeAnalysis.CEGARLikeAnalysisMain import analysis ; from boxesAndBoxOperations.mergeBoxes import mergeBoxes, mergeBoxes_quadraticTime_usefulForOutputSpaceBoxes_mergeBoxesThatContainOneAnother; import inspect; from CEGARLikeAnalysis import labelsForBoxes; from boxesAndBoxOperations.codeForGettingSamplesBetweenBoxes import getSampleVectorsToCheckAgainst, getBoxCenter; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass,\ Condition_TheBoxItself, MetaCondition_Conjunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation; import z3; import config; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface, executeDatabaseCommandList; from utils.distributionStatics import distributionStatics; from statesAndOperatorsAndSelection.descriptionState import DescriptionState ; class QuestionBaseClass(): def setID(self): self.uuid = str(uuid.uuid4()); return def getID(self): return self.uuid;# .copy(); def __init__(self, conditionsToBeConsistentWith): requires(isinstance(conditionsToBeConsistentWith, list)); requires(len(conditionsToBeConsistentWith) > 0); requires(all([isinstance(x, CharacterizationConditionsBaseClass) for x in conditionsToBeConsistentWith])); self.conditionsToBeConsistentWith = conditionsToBeConsistentWith; self.setID(); return; def getCopyOfConditionsToBeConsistentWith(self): return self.conditionsToBeConsistentWith.copy(); def formConditionToSatisfy(self, numberOfSamples): raise NotImplementedError(); # child classes must override. ############################################################################### ############################################################################### ############################################################################### import traceback; import os; import time as timePackageToUseForSleep; class Question_DomainOfVariablesInResponce(QuestionBaseClass): @staticmethod def setVariablesInformation(domainInfo): raise NotImplementedError(); @staticmethod def getUseableConditions(domainInfo, classType, conditionNameStartsWith=""): requires(issubclass(classType, Question_DomainOfVariablesInResponce)); requires(isinstance(conditionNameStartsWith, str)); requires(isinstance(domainInfo, BaseClassDomainInformation)); (variablesConditionMayInclude, variablesBoxesProducedMayBeOver) = \ classType.setVariablesInformation(domainInfo); def convertConditionName(thisCondition): return str(thisCondition).replace(" ", "_").lower(); temp= [convertConditionName(thisCondition) for thisCondition in domainInfo.getBaseConditions() if \ ( thisCondition.relaventVariables().issubset(variablesConditionMayInclude) and \ convertConditionName(thisCondition).startswith(conditionNameStartsWith) ) ]; return temp; def helper_getRelaventInputBoxes_get_functionToStatisfy(self): functionToStatisfy_initial = self.formConditionToSatisfy(); return self.setupProperFeedToFunctionforThisQuestionType(functionToStatisfy_initial); def setupProperFeedToFunctionforThisQuestionType(self, thisFunct): raise NotImplementedError(); def getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition(self): raise NotImplementedError(); def helper_getBoxesToDescribe_get_boxesToDescribePriorToMerging(self, thisInstanceOfModelBoxProgatorManager, \ inputDomainBoxes): raise NotImplementedError(); def __init__(self, conditionsToBeConsistentWith, domainInformation): requires(isinstance(domainInformation, BaseClassDomainInformation)); QuestionBaseClass.__init__(self, conditionsToBeConsistentWith); self.domainInfo = domainInformation; self.variablesConditionMayInclude = None; self.variablesBoxesProducedMayBeOver = None; (self.variablesConditionMayInclude, self.variablesBoxesProducedMayBeOver) = \ self.setVariablesInformation(self.domainInfo); assert(isinstance(self.variablesConditionMayInclude, list)); assert(isinstance(self.variablesBoxesProducedMayBeOver, list)); assert(len(self.variablesConditionMayInclude) > 0); assert(len(self.variablesBoxesProducedMayBeOver) > 0); if(not all([thisCondition.relaventVariables().issubset(self.variablesConditionMayInclude) \ for thisCondition in self.conditionsToBeConsistentWith])): raise Exception("Conditions provided involve variables that are not restricted to the domain of interest."); ensures(isinstance(self.variablesConditionMayInclude, list)); ensures(isinstance(self.variablesBoxesProducedMayBeOver, list)); ensures(len(self.variablesConditionMayInclude) > 0); ensures(len(self.variablesBoxesProducedMayBeOver) > 0); ensures(all([isinstance(x, z3.z3.ArithRef) for x in self.variablesConditionMayInclude])); ensures(all([isinstance(x, z3.z3.ArithRef) for x in self.variablesBoxesProducedMayBeOver])); ensures(set(self.variablesConditionMayInclude + self.variablesBoxesProducedMayBeOver).issubset(\ self.domainInfo.inputSpaceVariables() + self.domainInfo.outputSpaceVariables() ) ); return def getRelaventInputBoxes(self, thisInstanceOfModelBoxProgatorManager, thisState): requires(isinstance(thisState, DescriptionState)); functionToStatisfy = self.helper_getRelaventInputBoxes_get_functionToStatisfy(); functionToCheckWhetherNoPointsInTheBoxStatisfyCondition = self.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition(); epsilonForBoxSize = thisState.readParameter("floatValueForBoxDivisionCutoff"); splitOnlyOnRelaventVariables = thisState.readParameter("splitOnlyOnRelaventVariables"); assert(isinstance(epsilonForBoxSize, float)); assert(epsilonForBoxSize > 0); assert(isinstance(splitOnlyOnRelaventVariables, bool)); axisToSplitOn= list( range(0, getDimensionOfBox(self.domainInfo.getInputSpaceUniverseBox())) ); if(splitOnlyOnRelaventVariables): setOfRelaventVariables = set(); for thisCondition in self.conditionsToBeConsistentWith: setOfRelaventVariables.update(thisCondition.relaventVariables()); tempAxisToSplitOn=[\ thisIndex for thisIndex in axisToSplitOn \ if (self.domainInfo.inputSpaceVariables()[thisIndex] in setOfRelaventVariables)]; if(len(tempAxisToSplitOn) > 0): #This would fail to happen, for instanc, when # self.conditionsToBeConsistentWith specify conditions over the output space # as oppossed to the input space. axisToSplitOn=tempAxisToSplitOn; assert(len(axisToSplitOn) > 0); # The below assert is most easily understood from converting (p or q) to (if (not p) then q). # That is, if axisToSplitOn does not include the whole input space, it is because # splitOnlyOnRelaventVariables has been enabled. Note that this does NOT say the converse, # since even if splitOnlyOnRelaventVariables is True, axisToSplitOn might not reduce in # size for a number of reasons. assert( (len(axisToSplitOn) == len(self.domainInfo.inputSpaceVariables())) or \ splitOnlyOnRelaventVariables); assert( len(axisToSplitOn) <= len(self.domainInfo.inputSpaceVariables()) ); assert( set(axisToSplitOn).issubset(range(0, len(self.domainInfo.inputSpaceVariables())) )); # Below checks that the elements of axisToSplitOn are unique assert( len(set(axisToSplitOn)) == len(axisToSplitOn) ); axisScaling = self.domainInfo.getInputSpaceUniverseBox()[:, 1] - self.domainInfo.getInputSpaceUniverseBox()[:, 0]; assert(all(axisScaling >= 0.0)); indicesWhereAxisNotFlat = np.where(axisScaling[axisToSplitOn] > 0)[-1]; assert(np.all(axisScaling[axisToSplitOn][indicesWhereAxisNotFlat] > 0)); def functionToDetermineWhenToGiveUpOnBox_axisSmallAfterScalingByUniverseSize(thisBox): return np.max( (thisBox[axisToSplitOn,:][indicesWhereAxisNotFlat,1] - thisBox[axisToSplitOn,:][indicesWhereAxisNotFlat,0]) /\ axisScaling[axisToSplitOn][indicesWhereAxisNotFlat]) <= epsilonForBoxSize; thisCEGARFileWrittingManagerInstance = analysis(self.domainInfo.getInputSpaceUniverseBox(), thisInstanceOfModelBoxProgatorManager, \ functionToStatisfy, functionToDetermineWhenToGiveUpOnBox_axisSmallAfterScalingByUniverseSize, \ limitSplittingToAxisWithIndicesInThisList=axisToSplitOn, \ functionToCheckWhetherNoPointsInTheBoxStatisfyCondition=functionToCheckWhetherNoPointsInTheBoxStatisfyCondition); fileNameHoldingDesiredBoxes = \ thisCEGARFileWrittingManagerInstance.dictMappingFileTypeToFileHandleToUse["boxes"].name; tempFH = open(fileNameHoldingDesiredBoxes, "rb"); boxesToReturn = [x[0] for x in readBoxes(tempFH) if ((x[1][1] & labelsForBoxes.LOWESTLEVEL_FALSESOMEWHEREANDEXHAUSTEDLOOKING) > 0)]; #<---------------- NOTICE THE NOT (i.e., the check for falsity).... tempFH.close(); # Note that the second disjunct in the below depends on the specifics # of the splitting strategy and stop-criteria we use, and as such may # need to be adjusted if we employ different functions later. ensures(splitOnlyOnRelaventVariables or \ all([ np.all((x[:,1] - x[:,0]) <= (epsilonForBoxSize * axisScaling)) \ for x in boxesToReturn ]) \ ); ensures(isinstance(boxesToReturn, list)); ensures(all([isProperBox(x) for x in boxesToReturn])); return boxesToReturn; @staticmethod def recordBoxStats(listOfBoxes, phaseLabel): requires(isinstance(phaseLabel, str)); requires(len(phaseLabel) > 0); requires(isinstance(listOfBoxes, list)); requires(all([isProperBox(x) for x in listOfBoxes])); labelForThisPhaseInQuestionAnswering = "Question_DomainOfVariablesInResponce:" + phaseLabel; #V~~VV~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Below is largely copied from descriptionState.py # TODO: add this sort of code as a funtion in the utils to call upon.... #========================================================================= generalSummaryFunctionsAndLabelsForThem = [\ ( (lambda A : np.prod(np.diff(A, axis=1)) ), "bvolume"), \ ( (lambda A : np.min(np.diff(A, axis=1)) ), "bminSideLength"), \ ( (lambda A : np.max(np.diff(A, axis=1)) ), "bmaxSideLength"), \ ( (lambda A : np.sum(np.diff(A, axis=1)) ), "bsumSideLengths"), \ ]; for thisFunctAndLabel in generalSummaryFunctionsAndLabelsForThem: theseValues = [ thisFunctAndLabel[0](x) for x in listOfBoxes]; resultValue = distributionStatics(theseValues); specificLabel = labelForThisPhaseInQuestionAnswering + ":" + thisFunctAndLabel[1]; for thisKey in resultValue: commandToExecute = \ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, fieldValue) VALUES ('" + \ ObjDatabaseValueTracker.get_QAStateUUID_mostRecentBeingComputed() + "', '" + (specificLabel + ":" + thisKey) + "', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [resultValue[thisKey]] ); objDatabaseInterface.commit(); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ return; def getBoxesToDescribe(self, thisInstanceOfModelBoxProgatorManager, thisState ): requires(isinstance(thisState, DescriptionState)); inputDomainBoxes = self.getRelaventInputBoxes(\ thisInstanceOfModelBoxProgatorManager, thisState ); assert(isinstance(inputDomainBoxes, list)); listMappingBoxIndexToVariable = self.variablesBoxesProducedMayBeOver; if(len(inputDomainBoxes) == 0): return ([], listMappingBoxIndexToVariable, []); assert(len(inputDomainBoxes) > 0); boxesToDescribePriorToMerging = self.helper_getBoxesToDescribe_get_boxesToDescribePriorToMerging(\ thisInstanceOfModelBoxProgatorManager, inputDomainBoxes); assert(all([ (getDimensionOfBox(thisBox) == \ len(self.variablesBoxesProducedMayBeOver) ) \ for thisBox in boxesToDescribePriorToMerging])); boxesToReturn = None; limitOnNumberOfTimesToMerge = thisState.readParameter("limitOnNumberOfTimesToMerge"); assert(isinstance(limitOnNumberOfTimesToMerge, int) or (type(limitOnNumberOfTimesToMerge) == type(None))); assert((limitOnNumberOfTimesToMerge == None) or (limitOnNumberOfTimesToMerge >= 0)); if((limitOnNumberOfTimesToMerge == None) or (limitOnNumberOfTimesToMerge > 0)): middleLabelForBoxStatsRecording = "getBoxesToDescribe:" self.recordBoxStats(boxesToDescribePriorToMerging, \ middleLabelForBoxStatsRecording + "boxesOfInterestPriorToMerging"); precisionForMerging = thisState.readParameter("precisionForMerging"); temp = mergeBoxes(boxesToDescribePriorToMerging, precision=precisionForMerging, \ maxNumberOfIterations=limitOnNumberOfTimesToMerge); boxesToReturn = list(temp["dictMappingIndexToBox"].values()); boxesToReturn = mergeBoxes_quadraticTime_usefulForOutputSpaceBoxes_mergeBoxesThatContainOneAnother(boxesToReturn); self.recordBoxStats(boxesToReturn, \ middleLabelForBoxStatsRecording + "boxesOfInterestAfterMerging"); else: temp = boxesToDescribePriorToMerging; boxesToReturn = temp; assert(boxesToReturn is not None); # It is important to have this function return listMappingBoxIndexToVariable, not # only to possibly save labels and provide some centeralization, but also to ensure # that the variable list in question is always returned in the same order... ensures(isinstance(listMappingBoxIndexToVariable, list)); ensures(all([isinstance(x, z3.z3.ArithRef) for x in listMappingBoxIndexToVariable])); ensures(all([ (getDimensionOfBox(thisBox) == \ len(listMappingBoxIndexToVariable) ) \ for thisBox in boxesToReturn])); return (boxesToReturn, listMappingBoxIndexToVariable, inputDomainBoxes); class Question_InputDomain(Question_DomainOfVariablesInResponce): @staticmethod def setVariablesInformation(domainInfo): variablesConditionMayInclude = domainInfo.outputSpaceVariables(); variablesBoxesProducedMayBeOver = domainInfo.inputSpaceVariables(); return (variablesConditionMayInclude, variablesBoxesProducedMayBeOver); def setupProperFeedToFunctionforThisQuestionType(self, thisFunctIntial): # The CEGAR-like analysis uses functions of form that take (inputBox, outputBox) as # input ... thisFunctFinal = (lambda inputBox, outputBox : thisFunctIntial(outputBox)); return thisFunctFinal; def helper_getBoxesToDescribe_get_boxesToDescribePriorToMerging(self, thisInstanceOfModelBoxProgatorManager, \ inputDomainBoxes): boxesToDescribePriorToMerging = inputDomainBoxes; return boxesToDescribePriorToMerging; class Question_OutputDomain(Question_DomainOfVariablesInResponce): @staticmethod def setVariablesInformation(domainInfo): variablesConditionMayInclude = domainInfo.inputSpaceVariables(); variablesBoxesProducedMayBeOver = domainInfo.outputSpaceVariables(); return (variablesConditionMayInclude, variablesBoxesProducedMayBeOver); def setupProperFeedToFunctionforThisQuestionType(self, thisFunctIntial): # The CEGAR-like analysis uses functions of form that take (inputBox, outputBox) as # input ... thisFunctFinal = (lambda inputBox, outputBox : thisFunctIntial(inputBox)); return thisFunctFinal; def helper_getBoxesToDescribe_get_boxesToDescribePriorToMerging(self, thisInstanceOfModelBoxProgatorManager, \ inputDomainBoxes): boxesToDescribePriorToMerging = [\ thisInstanceOfModelBoxProgatorManager.pushBoxThrough(thisInputBox) \ for thisInputBox in inputDomainBoxes ]; return boxesToDescribePriorToMerging; class Question_JointInputAndOutputDomain(Question_DomainOfVariablesInResponce): @staticmethod def setVariablesInformation(domainInfo): variablesConditionMayInclude = domainInfo.inputSpaceVariables() + domainInfo.outputSpaceVariables(); variablesBoxesProducedMayBeOver = domainInfo.inputSpaceVariables() + domainInfo.outputSpaceVariables(); return (variablesConditionMayInclude, variablesBoxesProducedMayBeOver); def setupProperFeedToFunctionforThisQuestionType(self, thisFunctIntial): thisFunctFinal = (lambda inputBox, outputBox : thisFunctIntial(getJointBox([inputBox, outputBox]))); return thisFunctFinal; def helper_getBoxesToDescribe_get_boxesToDescribePriorToMerging(self, thisInstanceOfModelBoxProgatorManager, \ inputDomainBoxes): correspondingOutputBoxes = [\ thisInstanceOfModelBoxProgatorManager.pushBoxThrough(thisInputBox) \ for thisInputBox in inputDomainBoxes ]; assert(len(correspondingOutputBoxes) == len(inputDomainBoxes)); assert(all([ (getDimensionOfBox(thisBox) == len(self.domainInfo.outputSpaceVariables())) \ for thisBox in correspondingOutputBoxes])); boxesToDescribePriorToMerging = [\ getJointBox(list(x)) for x in zip(inputDomainBoxes, correspondingOutputBoxes)]; return boxesToDescribePriorToMerging; ############################################################################### ############################################################################### ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg ############################################################################### class Question_WayConditionIsChecked(QuestionBaseClass): pass; class Question_FormalUniversalQuant(Question_WayConditionIsChecked): def _helper_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition_convertBoxToFormulaConstraints(self, thisBox): requires(isProperBox(thisBox)); requires(getDimensionOfBox(thisBox) == len(self.variablesConditionMayInclude)); F = z3.And([ \ z3.And( float(thisBox[index, 0]) <= self.variablesConditionMayInclude[index], \ self.variablesConditionMayInclude[index] <= float(thisBox[index, 1]) \ ) \ for index in range(0, len(self.variablesConditionMayInclude)) ]); return F; def getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition(self): def functToCheck(thisBox): #V~V~V~V~V~VV~V~V~V~V~VV~V~V~V~V~V~V~V~V~V~V~V~V~VV~V~V~V~V~VV~V # Statistical sampling to help speed of the forall-checking. If we # find one counter-example to the forall-claim, then there is no # reason to do the expensive z3 check... #=============================================================== for thisSample in range(0, config.defaultValues.numberOfStatisticalSamplesToTakeIn_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition): randomVector = getRandomVectorInBox(thisBox).reshape(getDimensionOfBox(thisBox), 1); if(not any([thisCondition.pythonFormatEvaluation(randomVector) for thisCondition in self.conditionsToBeConsistentWith]) ): return False; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ z3Solver = self.conditionsToBeConsistentWith[0].z3Solver; z3Solver.reset(); # this might be the expensive.... I have to check # disjunctive normal form - each element in the list is a clause which we or-together.... formulaToCheck = \ (\ z3.ForAll( self.variablesConditionMayInclude , \ z3.Implies(\ self._helper_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition_convertBoxToFormulaConstraints(thisBox), \ z3.Or([x.z3FormattedCondition for x in self.conditionsToBeConsistentWith]) \ ) \ ) \ ); z3Solver.add(formulaToCheck); verdict = (z3Solver.check() == z3.z3.sat); z3Solver.reset(); # this might be the expensive.... I have to check return verdict; # NOTE: <<<<<<<<<<<<<<<<<<<<<<<<< notice I AM NOT NEGATING functToCheck <<<<<<<<<<<< return self.setupProperFeedToFunctionforThisQuestionType(functToCheck); def formConditionToSatisfy_hackyCopyFrom_Question_Probabilistic(self): # This function returns FALSE when it found a random sample that statisfies the DNF in question, # and returns TRUE when it DOES NOT numberOfSamples=\ config.defaultValues.numberOfStatisticalSamplesToTakeIn_numberOfStatisticalSamplesToTakeIn_getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition def statisticalCheckForallConditionFails(thisBox, thisCondition): # Statistically checking that thisCondition FAILS EVERYWHERE in #<<<<<<<<<<<<<<<<<<<<<<<<<<< NOTICE IT TRIES TO CHECK FOR FAILURE # the box. for thisSampleIndex in range(0, numberOfSamples): randomVector = getRandomVectorInBox(thisBox).reshape(getDimensionOfBox(thisBox), 1); if(thisCondition.pythonFormatEvaluation(randomVector)): return False; return True; return (lambda thisBox : all([\ statisticalCheckForallConditionFails(thisBox, thisCondition) \ for thisCondition in self.conditionsToBeConsistentWith])); # Below, numberOfSamples is not used def formConditionToSatisfy(self, numberOfSamples=config.defaultValues.formConditionToSatisfy_statistical_numberOfSamples): # disjunctive normal form - each element in the list is a clause which we or-together.... mainCondition = \ (lambda thisBox : any([x.existsMemberOfBoxSatifyingCondition(thisBox) for x in self.conditionsToBeConsistentWith])); def functToGive(thisBox): # In the case where random sampling could not find a satisfying assigment, we still need # to do a formal check that no such assignment exists. If, however, random sampling # found it is possible to satisfy, no formal checking is needed. As we do elsewhere, the # goal of this function is TO RETURN FALSE WHEN AN ASSIGNEMNT IS TRUE, AND TRUE WHEN IT # DOES NOT. if(self.formConditionToSatisfy_hackyCopyFrom_Question_Probabilistic()(thisBox)): return (not mainCondition(thisBox)); #<<<<<<<<<<<<<<<<<<<<<<<<< NOTICE THE NOT else: return False; return functToGive; class Question_Probabilistic(Question_WayConditionIsChecked): def getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition(self): return None; def formConditionToSatisfy(self, \ numberOfSamples=config.defaultValues.formConditionToSatisfy_statistical_numberOfSamples): requires(isinstance(numberOfSamples, int) or (numberOfSamples == None)); requires( ( numberOfSamples == None) or (numberOfSamples > 0)); # We have to explicitly pass in numberOfSamples below since it is potentially # modified in the function, and thus a variable local to this local-function. def statisticalCheckForallConditionFails(thisBox, thisCondition, numberOfSamples): # Statistically checking that thisCondition FAILS EVERYWHERE in #<<<<<<<<<<<<<<<<<<<<<<<<<<< NOTICE IT TRIES TO CHECK FOR FAILURE # the box. if(numberOfSamples == None): numberOfSamples = (2 * getDimensionOfBox(thisBox)) + 1; for thisSampleIndex in range(0, numberOfSamples): randomVector = getRandomVectorInBox(thisBox).reshape(getDimensionOfBox(thisBox), 1); if(thisCondition.pythonFormatEvaluation(randomVector)): return False; # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg return True; return (lambda thisBox : all([\ statisticalCheckForallConditionFails(thisBox, thisCondition, numberOfSamples) \ for thisCondition in self.conditionsToBeConsistentWith])); ############################################################################### ############################################################################### ############################################################################### class QuestionClass_What_Do_You_Do_When(Question_OutputDomain, Question_FormalUniversalQuant): getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = Question_FormalUniversalQuant.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition; pass; class QuestionClass_When_Do_You(Question_InputDomain, Question_FormalUniversalQuant): getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = Question_FormalUniversalQuant.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition; pass; class QuestionClass_What_Are_The_Circumstances_In_Which(Question_JointInputAndOutputDomain, Question_FormalUniversalQuant): getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = Question_FormalUniversalQuant.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition; pass; class QuestionClass_What_Do_You_Ussually_Do_When(Question_OutputDomain, Question_Probabilistic): getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = Question_Probabilistic.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition; pass; class QuestionClass_When_Do_You_Ussually(Question_InputDomain, Question_Probabilistic): getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = Question_Probabilistic.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition; pass; class QuestionClass_What_Are_The_Usual_Circumstances_In_Which(Question_JointInputAndOutputDomain, Question_Probabilistic): getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition = Question_Probabilistic.getFunctionToCheckWhetherNoPointsInTheBoxStatisfyCondition; pass;
34,638
57.610829
2,861
py
Fanoos
Fanoos-master/domainsAndConditions/domainAndConditionsFor_modelForTesting_twoDimInput_threeDimOutput.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; from utils.contracts import *; from boxesAndBoxOperations.getBox import getDimensionOfBox; import z3; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass, CharacterizationCondition_FromPythonFunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; from domainsAndConditions.utilsForDefiningPredicates import *; class DomainFor_modelForTesting_twoDimInput_threeDimOutput(BaseClassDomainInformation): def __init__(self, z3SolverInstance): requires(isinstance(z3SolverInstance, z3.z3.Solver)); self.initializedConditions = None; self.initialize_baseConditions(z3SolverInstance); assert(self.initializedConditions != None); return; @staticmethod def getUUID(): return "862fb8c4-fee4-4181-a79f-9127cd8b2b64"; @staticmethod def getInputSpaceUniverseBox(): orderOfVariables = __class__.inputSpaceVariables(); dictMappingVariableToBound = {\ "in_x" : [-15.0, 15.0], \ "in_y" : [-10.0, 10.0] \ }; thisUniverseBox = __class__._helper_getInputSpaceUniverseBox(\ orderOfVariables, dictMappingVariableToBound); ensures(getDimensionOfBox(thisUniverseBox) == len(DomainFor_modelForTesting_twoDimInput_threeDimOutput.inputSpaceVariables())); return thisUniverseBox; @staticmethod def inputSpaceVariables(): return [\ z3.Real(x) for x in ["in_x", "in_y"] ]; @staticmethod def outputSpaceVariables(): return [z3.Real(x) for x in ["out_u", "out_v", "out_w"]]; @staticmethod def getName(): return "Domain For modelForTesting_twoDimInput_threeDimOutput"; def initialize_baseConditions(self, z3SolverInstance): dictMappingPredicateStringNameToUUID = \ { "IN_X_NEG20DOT0TONEG19DOT0" : "a4433189-76b4-40ee-b9ca-3cc9a37287e0" , \ "IN_Y_NEG20DOT0TONEG19DOT0" : "af2466bf-d1f8-4038-9925-9bbfe361d2bf" , \ "IN_X_NEG20DOT0TONEG19DOT5" : "00fa6c5b-297d-4ac3-98cd-77f3e1adb219" , \ "IN_Y_NEG20DOT0TONEG19DOT5" : "ada870ed-9146-4c1c-871f-79e7b3cc874a" , \ "IN_X_NEG19DOT5TONEG19DOT0" : "27db3e7c-2325-4d57-bd14-195801daa63e" , \ "IN_Y_NEG19DOT5TONEG19DOT0" : "8fb0f9e5-a8e1-4d9a-b8a8-d137f6b71752" , \ "IN_X_NEG19DOT0TONEG18DOT0" : "cc1adca9-7447-4533-8a67-23c48bfe308a" , \ "IN_Y_NEG19DOT0TONEG18DOT0" : "4baa96b3-d607-4349-a0d2-a3c3049f8e58" , \ "IN_X_NEG19DOT0TONEG18DOT5" : "439f46f3-1069-4113-8ffe-189290726769" , \ "IN_Y_NEG19DOT0TONEG18DOT5" : "6b654a18-e769-4e33-907a-8d82ee17ffa7" , \ "IN_X_NEG18DOT5TONEG18DOT0" : "968532b2-229b-4bbe-a081-229869a84b88" , \ "IN_Y_NEG18DOT5TONEG18DOT0" : "fadef770-7076-4caa-87cd-8c98a678f8c2" , \ "IN_X_NEG18DOT0TONEG17DOT0" : "587249c7-08e4-4c96-bfcc-48814c846835" , \ "IN_Y_NEG18DOT0TONEG17DOT0" : "853a5e21-7938-427d-b7de-6b57829a2b77" , \ "IN_X_NEG18DOT0TONEG17DOT5" : "29980b88-66f3-4a02-92b7-ac31cd4996c6" , \ "IN_Y_NEG18DOT0TONEG17DOT5" : "c76acb90-aca2-49b6-a6ec-52b9c7ba160e" , \ "IN_X_NEG17DOT5TONEG17DOT0" : "5618eeed-2ac3-4493-b41d-b9e7a5f7514a" , \ "IN_Y_NEG17DOT5TONEG17DOT0" : "a2882b0c-3e14-4d05-bb66-01a5a21a87c0" , \ "IN_X_NEG17DOT0TONEG16DOT0" : "64d06e42-7b30-420d-ae68-56ee9752e1a1" , \ "IN_Y_NEG17DOT0TONEG16DOT0" : "0ff594d9-7b0f-47df-bd51-2b33dbe962ec" , \ "IN_X_NEG17DOT0TONEG16DOT5" : "ace4d588-dd75-4ea5-af22-c1ba593e7f61" , \ "IN_Y_NEG17DOT0TONEG16DOT5" : "e77d0f66-7d91-4ff7-baa6-799678ffe5c8" , \ "IN_X_NEG16DOT5TONEG16DOT0" : "8540b40f-7c06-4d0d-bccb-01f8ab531e3b" , \ "IN_Y_NEG16DOT5TONEG16DOT0" : "f1c9493c-9909-4d79-94e3-ad836a2402c3" , \ "IN_X_NEG16DOT0TONEG15DOT0" : "25940dfb-139a-4818-9506-e5b6ccbefbf8" , \ "IN_Y_NEG16DOT0TONEG15DOT0" : "f52f964c-6806-48dc-8120-b0dbe6dd5e1b" , \ "IN_X_NEG16DOT0TONEG15DOT5" : "69104c3c-d41d-4a3a-b2d0-007835efc458" , \ "IN_Y_NEG16DOT0TONEG15DOT5" : "43edc681-1ce2-4a66-99bc-f30645215daa" , \ "IN_X_NEG15DOT5TONEG15DOT0" : "45f2dafc-1aa2-463a-88bc-ee966121e715" , \ "IN_Y_NEG15DOT5TONEG15DOT0" : "71c14806-e483-4cb1-99a3-75d626d15300" , \ "IN_X_NEG15DOT0TONEG14DOT0" : "626a60a3-0861-4d2c-9927-841ba0ac532b" , \ "IN_Y_NEG15DOT0TONEG14DOT0" : "db645a1d-5432-42fe-8767-f542cde05416" , \ "IN_X_NEG15DOT0TONEG14DOT5" : "c1abdd83-3ff1-4e50-a56b-d58184e5cd8a" , \ "IN_Y_NEG15DOT0TONEG14DOT5" : "0565882b-39c1-427d-9767-8499c7b515a0" , \ "IN_X_NEG14DOT5TONEG14DOT0" : "c601caa0-319d-4f7f-b1e1-13bccbd24f41" , \ "IN_Y_NEG14DOT5TONEG14DOT0" : "cbd840b3-e2d0-4c2b-876a-cc64ba1cc427" , \ "IN_X_NEG14DOT0TONEG13DOT0" : "51300a19-6787-4a41-bc79-7691d961cf70" , \ "IN_Y_NEG14DOT0TONEG13DOT0" : "b047ac7a-b6bd-4b09-b81b-d1f8a9dcfab9" , \ "IN_X_NEG14DOT0TONEG13DOT5" : "b4d2ec8c-710f-4592-a869-80dab2d1414f" , \ "IN_Y_NEG14DOT0TONEG13DOT5" : "1ee85eae-0993-4e97-b752-e463f376efe9" , \ "IN_X_NEG13DOT5TONEG13DOT0" : "78a666f7-804a-4ce3-923b-a14045920887" , \ "IN_Y_NEG13DOT5TONEG13DOT0" : "40731390-2894-48fd-b0d1-cd7f80f10d18" , \ "IN_X_NEG13DOT0TONEG12DOT0" : "2c3b3591-85d9-45c0-baf5-1909cefacd5a" , \ "IN_Y_NEG13DOT0TONEG12DOT0" : "d36c9989-9814-4f8b-980a-50bd4c8f238d" , \ "IN_X_NEG13DOT0TONEG12DOT5" : "69c53f8c-0d10-4778-bed0-cec7bcfb2063" , \ "IN_Y_NEG13DOT0TONEG12DOT5" : "03dad021-26ed-4179-96b7-523f9dc912e8" , \ "IN_X_NEG12DOT5TONEG12DOT0" : "2ccfd41f-2a05-448c-a6a2-142da6637b39" , \ "IN_Y_NEG12DOT5TONEG12DOT0" : "d317fcc9-ae63-4161-b880-a7bd67375e73" , \ "IN_X_NEG12DOT0TONEG11DOT0" : "51479b7d-f175-40d7-be9e-f16137771e31" , \ "IN_Y_NEG12DOT0TONEG11DOT0" : "f1eee679-5a5f-4a9c-bf63-16d91caec1aa" , \ "IN_X_NEG12DOT0TONEG11DOT5" : "3b91a11e-c4a4-40d6-b140-a5b501d070a6" , \ "IN_Y_NEG12DOT0TONEG11DOT5" : "6974f47e-3252-4a2b-9f63-d7ae9b8c8217" , \ "IN_X_NEG11DOT5TONEG11DOT0" : "be66388d-96a7-4716-9def-e3e0af1f1752" , \ "IN_Y_NEG11DOT5TONEG11DOT0" : "16e137f2-4ba1-49a6-bffb-19dc4dcc6d81" , \ "IN_X_NEG11DOT0TONEG10DOT0" : "39a9f3e7-4231-4df6-93e1-64eb7b7f6341" , \ "IN_Y_NEG11DOT0TONEG10DOT0" : "67af9c29-ab5e-442b-9a64-d797feda8468" , \ "IN_X_NEG11DOT0TONEG10DOT5" : "625c48f3-9f57-4a86-bca4-59ec473b3f8e" , \ "IN_Y_NEG11DOT0TONEG10DOT5" : "b0f237bc-58f3-459b-8366-2656f47d3cc4" , \ "IN_X_NEG10DOT5TONEG10DOT0" : "4b8522a3-c832-468e-a916-8b11fc624794" , \ "IN_Y_NEG10DOT5TONEG10DOT0" : "2b7dcb52-9072-4dab-b418-c0232ec392f1" , \ "IN_X_NEG10DOT0TONEG9DOT0" : "4d62bd4d-a6d9-4f82-9a6c-2ba109458ca6" , \ "IN_Y_NEG10DOT0TONEG9DOT0" : "8ff7ed78-4fdf-45b6-a3f2-cabd94b96f94" , \ "IN_X_NEG10DOT0TONEG9DOT5" : "4b1e2b67-6eff-42cf-ab9a-d01bd5423b5e" , \ "IN_Y_NEG10DOT0TONEG9DOT5" : "48134cc8-fe5a-4606-a09b-a927ef70ebe3" , \ "IN_X_NEG9DOT5TONEG9DOT0" : "770095c5-cb48-4f56-93ba-724eb4c0a7a0" , \ "IN_Y_NEG9DOT5TONEG9DOT0" : "047e9fdb-1264-46e5-81d5-c1e7091056f1" , \ "IN_X_NEG9DOT0TONEG8DOT0" : "8d7c5b32-8d27-4648-b6b5-e4f2efd27f7f" , \ "IN_Y_NEG9DOT0TONEG8DOT0" : "da5d8383-804c-42f4-96a0-eeea550ee60b" , \ "IN_X_NEG9DOT0TONEG8DOT5" : "40638a98-e293-4e75-a6b3-a246623edc72" , \ "IN_Y_NEG9DOT0TONEG8DOT5" : "8e3181f8-1f2d-4641-96b6-182617994179" , \ "IN_X_NEG8DOT5TONEG8DOT0" : "395a242b-22ff-4d59-95dc-0ad22709b9fb" , \ "IN_Y_NEG8DOT5TONEG8DOT0" : "9c2be67e-0795-48c3-924d-e8d926301ffb" , \ "IN_X_NEG8DOT0TONEG7DOT0" : "84a1bd60-113f-49ea-a754-ef57b0f3ce43" , \ "IN_Y_NEG8DOT0TONEG7DOT0" : "4836315b-1edb-469b-94ab-3e8a03b5c730" , \ "IN_X_NEG8DOT0TONEG7DOT5" : "383a743b-7298-4896-87fc-8ceafc4d9a40" , \ "IN_Y_NEG8DOT0TONEG7DOT5" : "c25f7a16-b487-4469-a3bd-3e718dcc13a3" , \ "IN_X_NEG7DOT5TONEG7DOT0" : "ceb05436-82d1-4cf8-97f5-79d3c1540733" , \ "IN_Y_NEG7DOT5TONEG7DOT0" : "a0b257a9-0780-4417-837b-8f1203191bf2" , \ "IN_X_NEG7DOT0TONEG6DOT0" : "ce82425e-5534-4234-a0ba-756d84bdb139" , \ "IN_Y_NEG7DOT0TONEG6DOT0" : "226a157c-a16e-47b9-b7d6-6dc84d58b00f" , \ "IN_X_NEG7DOT0TONEG6DOT5" : "3adfa009-6f21-4e6f-b5cc-585226624db6" , \ "IN_Y_NEG7DOT0TONEG6DOT5" : "9cc45874-6aa3-4b5f-a4b6-b9847c0199d6" , \ "IN_X_NEG6DOT5TONEG6DOT0" : "532b4414-5489-4826-9274-40c2af334581" , \ "IN_Y_NEG6DOT5TONEG6DOT0" : "4b6a8963-ee0c-476c-bf49-fd44037dc249" , \ "IN_X_NEG6DOT0TONEG5DOT0" : "9ff58550-049f-4868-9f57-7fc0fcd5e52e" , \ "IN_Y_NEG6DOT0TONEG5DOT0" : "618a76c3-8bae-4e3b-9059-337840da93b5" , \ "IN_X_NEG6DOT0TONEG5DOT5" : "6aebb372-f16c-4397-8f9d-e5d9100c53ee" , \ "IN_Y_NEG6DOT0TONEG5DOT5" : "f61ccd75-76c4-4a68-8240-78936a100940" , \ "IN_X_NEG5DOT5TONEG5DOT0" : "b7b4cbdd-d288-4ae9-82bb-b85acd8e01bd" , \ "IN_Y_NEG5DOT5TONEG5DOT0" : "92a4f6ca-85cc-48ea-b34f-2860e87a8aa3" , \ "IN_X_NEG5DOT0TONEG4DOT0" : "7ba75dc7-aa8a-4077-a876-f28d82346d8a" , \ "IN_Y_NEG5DOT0TONEG4DOT0" : "4ab1118e-e2f2-4fd3-8d0b-a668596d550e" , \ "IN_X_NEG5DOT0TONEG4DOT5" : "92c3bbf0-fac3-4d14-9e7c-0345329fb813" , \ "IN_Y_NEG5DOT0TONEG4DOT5" : "7aa4bd98-db81-4701-8a33-086fe6cc912d" , \ "IN_X_NEG4DOT5TONEG4DOT0" : "d8a547d6-01ab-404e-bb63-343e0a7fd9b9" , \ "IN_Y_NEG4DOT5TONEG4DOT0" : "2b539615-fe2b-4b11-8ea4-233cb4ba14b7" , \ "IN_X_NEG4DOT0TONEG3DOT0" : "5faf30a6-2399-46d6-bc2f-c12743a37442" , \ "IN_Y_NEG4DOT0TONEG3DOT0" : "b81341ff-a584-46d8-9df8-783c7c2768f3" , \ "IN_X_NEG4DOT0TONEG3DOT5" : "a79603c6-ea18-4e39-984f-6d49ecd39557" , \ "IN_Y_NEG4DOT0TONEG3DOT5" : "5b3183d5-7301-4e86-a0c5-1ed46e24e22d" , \ "IN_X_NEG3DOT5TONEG3DOT0" : "bee6978c-d3d8-4e9a-9736-3f1ecf43e782" , \ "IN_Y_NEG3DOT5TONEG3DOT0" : "06583dc8-c512-4c78-be9b-271d784886b3" , \ "IN_X_NEG3DOT0TONEG2DOT0" : "1a11629b-cacf-4482-b531-7938991dc1ca" , \ "IN_Y_NEG3DOT0TONEG2DOT0" : "3ff6efc1-2040-4f03-bef7-de4ac0172a22" , \ "IN_X_NEG3DOT0TONEG2DOT5" : "b50b4ad7-da01-4984-a5dc-6188d44a439a" , \ "IN_Y_NEG3DOT0TONEG2DOT5" : "e4d05686-0da9-40b3-a1dd-4b8b899c20a1" , \ "IN_X_NEG2DOT5TONEG2DOT0" : "e42efd59-cb86-4006-8727-f9387216de01" , \ "IN_Y_NEG2DOT5TONEG2DOT0" : "01753372-fcb9-44d5-b43e-fffd5a95e863" , \ "IN_X_NEG2DOT0TONEG1DOT0" : "0fcee40c-128d-4f26-a411-63044702a2a6" , \ "IN_Y_NEG2DOT0TONEG1DOT0" : "a81363db-2b5f-4e26-9129-5cec5d7ac394" , \ "IN_X_NEG2DOT0TONEG1DOT5" : "4f3949fd-5c94-4b55-aeac-09d04a4f9e01" , \ "IN_Y_NEG2DOT0TONEG1DOT5" : "b02bd089-f2c0-44f7-bba2-33eff483e523" , \ "IN_X_NEG1DOT5TONEG1DOT0" : "3e8d1c9d-4db4-4a5b-b802-c45c336ad36e" , \ "IN_Y_NEG1DOT5TONEG1DOT0" : "4c3f8d47-c56d-467f-9404-03da78b8179c" , \ "IN_X_NEG1DOT0TO0DOT0" : "1bb3e692-595c-4dfa-8f81-f8f2b08737fe" , \ "IN_Y_NEG1DOT0TO0DOT0" : "a52a4f32-82b2-4159-9c13-e0e315b82afb" , \ "IN_X_NEG1DOT0TONEG0DOT5" : "528b9ed9-0653-41fc-bdf8-005f8ada4612" , \ "IN_Y_NEG1DOT0TONEG0DOT5" : "ce11302c-a9d7-47ae-a08e-e79606ee9a3e" , \ "IN_X_NEG0DOT5TO0DOT0" : "fc631ce8-50ff-4144-bcdc-503da3087b83" , \ "IN_Y_NEG0DOT5TO0DOT0" : "85b4d189-2cf9-41a7-b5a9-44cc63e238d0" , \ "IN_X_0DOT0TO1DOT0" : "5e7c692d-ed60-4bf2-939a-8ee88a8e11f0" , \ "IN_Y_0DOT0TO1DOT0" : "d82fd6e7-00ea-4c35-8ca5-65dcc3fb1323" , \ "IN_X_0DOT0TO0DOT5" : "80e2ac72-f9a3-40fa-9233-34b6b9d5129f" , \ "IN_Y_0DOT0TO0DOT5" : "8ebeb4ad-a3e0-4d51-8e28-ec0d9b39f4a1" , \ "IN_X_0DOT5TO1DOT0" : "cd2bae95-fa25-4585-bd8b-5c4e447895e3" , \ "IN_Y_0DOT5TO1DOT0" : "cb4bf7bc-de85-41cd-8cf6-0aaaa7d8089d" , \ "IN_X_1DOT0TO2DOT0" : "101ba361-d53b-4d97-b551-d4b34c1b157a" , \ "IN_Y_1DOT0TO2DOT0" : "752f195f-55a8-4217-b312-156942f8991c" , \ "IN_X_1DOT0TO1DOT5" : "84b8bd46-4458-45cc-aa36-b0b1d8ca957e" , \ "IN_Y_1DOT0TO1DOT5" : "82ba291a-b3fe-44f3-a351-aca4701b3284" , \ "IN_X_1DOT5TO2DOT0" : "329307b5-60ef-4220-ae17-d58d0a634746" , \ "IN_Y_1DOT5TO2DOT0" : "342cc253-b267-414e-8f05-2ff955cfd809" , \ "IN_X_2DOT0TO3DOT0" : "bae8d113-3126-4d62-a7a5-91e63a4bcc9c" , \ "IN_Y_2DOT0TO3DOT0" : "4b02ecef-8e81-4531-8fc2-e50b2128e222" , \ "IN_X_2DOT0TO2DOT5" : "1b33c149-7a56-48b2-b135-af394bb81360" , \ "IN_Y_2DOT0TO2DOT5" : "92ddb9ac-93bc-4e6d-b680-96f42134c110" , \ "IN_X_2DOT5TO3DOT0" : "178084f0-3a6b-49d2-8946-531d84cba704" , \ "IN_Y_2DOT5TO3DOT0" : "7011e195-2d5e-4872-9973-514b7f1a5bcf" , \ "IN_X_3DOT0TO4DOT0" : "68f24b75-751b-4d1b-8c24-eef3105f1326" , \ "IN_Y_3DOT0TO4DOT0" : "289705c4-1bf5-4140-9ab1-cb6e2b67e44f" , \ "IN_X_3DOT0TO3DOT5" : "f8ec99b4-1682-4521-b797-442633940930" , \ "IN_Y_3DOT0TO3DOT5" : "6cfbc0c6-cad1-484a-9173-1e3275a2cc0b" , \ "IN_X_3DOT5TO4DOT0" : "7aa418e6-bc01-4af8-9309-01393148e694" , \ "IN_Y_3DOT5TO4DOT0" : "564ab9f0-3b40-4285-ab99-23a4a98a9310" , \ "IN_X_4DOT0TO5DOT0" : "33b6d03b-a9e2-4fbc-9252-4197df1bd46b" , \ "IN_Y_4DOT0TO5DOT0" : "788bee6a-4599-43c9-93ab-b20bfe592a92" , \ "IN_X_4DOT0TO4DOT5" : "d71f7753-c5ba-4b38-b3d8-4668d5dc23e0" , \ "IN_Y_4DOT0TO4DOT5" : "3c6b0c26-635a-44f8-8b0b-c4cda931104e" , \ "IN_X_4DOT5TO5DOT0" : "6fe5df4a-0c10-466e-80d4-431deff596a4" , \ "IN_Y_4DOT5TO5DOT0" : "d072c736-5c6e-48e2-bebb-d95009621f94" , \ "IN_X_5DOT0TO6DOT0" : "4163d4af-1555-4508-b1e4-0dc471550090" , \ "IN_Y_5DOT0TO6DOT0" : "f845a53b-dd7e-4dbf-a902-787f47f8510a" , \ "IN_X_5DOT0TO5DOT5" : "dee6c8cd-3a34-4ebd-b36e-4d16cbd30b7d" , \ "IN_Y_5DOT0TO5DOT5" : "4981a768-4a3d-490d-93e8-b688663ebf4c" , \ "IN_X_5DOT5TO6DOT0" : "c8950432-f64e-48c7-a3b9-09f076d8cd4a" , \ "IN_Y_5DOT5TO6DOT0" : "00c69e7a-07d6-4780-baba-58a40d7e0617" , \ "IN_X_6DOT0TO7DOT0" : "3eeaabb3-0187-4c94-88d2-e08df5667046" , \ "IN_Y_6DOT0TO7DOT0" : "7513a1bd-dd23-4674-8fbc-e0fdfebb7c9f" , \ "IN_X_6DOT0TO6DOT5" : "0cf7d9ba-8ce7-4bd9-9d9b-319cc3752c48" , \ "IN_Y_6DOT0TO6DOT5" : "dc9851cf-f5d6-4bb6-9b9b-0d42078f34a1" , \ "IN_X_6DOT5TO7DOT0" : "176d7020-63da-4639-a322-644f2dd4bcce" , \ "IN_Y_6DOT5TO7DOT0" : "2f930d1c-5f50-4830-aa5e-7393dd8bfe1b" , \ "IN_X_7DOT0TO8DOT0" : "98af9e31-4e34-4e34-acb5-600c8e30af51" , \ "IN_Y_7DOT0TO8DOT0" : "dbec776a-4404-40e1-8725-10e840aaabe1" , \ "IN_X_7DOT0TO7DOT5" : "3573bed3-51d9-489d-ba17-a1615ea49d86" , \ "IN_Y_7DOT0TO7DOT5" : "9462d943-92ea-49d2-9143-45f43fb8c8e5" , \ "IN_X_7DOT5TO8DOT0" : "cc077f35-3585-462f-8e55-b6f9f702c1a4" , \ "IN_Y_7DOT5TO8DOT0" : "accab2ce-14e7-495f-b89e-a8b504b27b24" , \ "IN_X_8DOT0TO9DOT0" : "be42244e-a669-40c6-9b2a-bb5670793a47" , \ "IN_Y_8DOT0TO9DOT0" : "040af208-5557-4179-83c8-0d95e177806d" , \ "IN_X_8DOT0TO8DOT5" : "b4b5f321-5947-4a21-9384-f124619521d0" , \ "IN_Y_8DOT0TO8DOT5" : "21273ece-77ae-4090-bd9f-5d4b4ab9fbff" , \ "IN_X_8DOT5TO9DOT0" : "108710ca-1e87-45ff-878f-0aa0445936ac" , \ "IN_Y_8DOT5TO9DOT0" : "15c66172-c735-47e3-9e75-4012c6ad62ca" , \ "IN_X_9DOT0TO10DOT0" : "8db2dcbb-36bb-4362-a833-0e0b8ffe8e1c" , \ "IN_Y_9DOT0TO10DOT0" : "acec202b-f003-4fa3-9e25-574ea85d1727" , \ "IN_X_9DOT0TO9DOT5" : "be84f09a-0043-41b0-b5cd-0ff6a26ec3f8" , \ "IN_Y_9DOT0TO9DOT5" : "0477ca4f-ec8d-4dc4-9b6e-7ded67c48f33" , \ "IN_X_9DOT5TO10DOT0" : "da99305c-9fae-46c8-adc6-c8da9a968521" , \ "IN_Y_9DOT5TO10DOT0" : "ec8af7ea-9203-4039-b239-ab102c986d0c" , \ "IN_X_10DOT0TO11DOT0" : "b7a71df0-5802-4a76-892c-b18df73ab8a7" , \ "IN_Y_10DOT0TO11DOT0" : "013cc284-9815-42d6-8f9f-b7d9ad1aeffa" , \ "IN_X_10DOT0TO10DOT5" : "137ff452-1f00-4d1d-90ac-6a6c95da329a" , \ "IN_Y_10DOT0TO10DOT5" : "6a7e3c62-e10f-40f8-b0b8-bda5ec3ab977" , \ "IN_X_10DOT5TO11DOT0" : "d0ac2499-f368-4552-af2f-c95efc0a2a8c" , \ "IN_Y_10DOT5TO11DOT0" : "0933204f-1b6d-4222-b10e-1035cd1168af" , \ "IN_X_11DOT0TO12DOT0" : "9d9fb80e-6e68-40b4-941e-07655eff7793" , \ "IN_Y_11DOT0TO12DOT0" : "5383099e-f2fd-4cae-9e74-9877e2cff74b" , \ "IN_X_11DOT0TO11DOT5" : "c0c94728-c407-4c1d-bda9-e529781157c6" , \ "IN_Y_11DOT0TO11DOT5" : "7c0b0531-62b3-4e42-bf21-9ffbee857c19" , \ "IN_X_11DOT5TO12DOT0" : "4e731d8c-4af1-41c0-b646-35274a27e0a7" , \ "IN_Y_11DOT5TO12DOT0" : "4b670337-5fb6-4cc7-8cf1-9b13c874107b" , \ "IN_X_12DOT0TO13DOT0" : "4354b8bd-175d-47c2-bcb1-6430099915ad" , \ "IN_Y_12DOT0TO13DOT0" : "e2c2ef43-e96c-48b0-ba05-9a7c80c478e3" , \ "IN_X_12DOT0TO12DOT5" : "0477c0b5-d26e-4ae3-9e1f-d5ea5a687a59" , \ "IN_Y_12DOT0TO12DOT5" : "5fbeee8e-66f3-487d-843c-570f66fd8ce7" , \ "IN_X_12DOT5TO13DOT0" : "8d65ed79-824d-4f4a-9509-d6af07894159" , \ "IN_Y_12DOT5TO13DOT0" : "6af9fec5-346d-41c7-9d23-bb79b43addc6" , \ "IN_X_13DOT0TO14DOT0" : "7b8bf3c2-3795-4353-96eb-e8a4dcc9240d" , \ "IN_Y_13DOT0TO14DOT0" : "587c0470-2f6d-412b-b74c-b06acd291e98" , \ "IN_X_13DOT0TO13DOT5" : "1811a603-a153-4187-8c7c-1cb8157dae64" , \ "IN_Y_13DOT0TO13DOT5" : "374d040f-dce0-4710-99a1-5f2f3ebd0fbe" , \ "IN_X_13DOT5TO14DOT0" : "2dbdc1e8-7097-4a7a-99cd-357beb04fb94" , \ "IN_Y_13DOT5TO14DOT0" : "5d094076-1438-4ffa-b37d-e70b924a1a89" , \ "IN_X_14DOT0TO15DOT0" : "41c99974-a930-439e-9af9-1c4a874ecd0b" , \ "IN_Y_14DOT0TO15DOT0" : "56df92f8-9a5e-4938-b321-4340909c7857" , \ "IN_X_14DOT0TO14DOT5" : "bb7122e1-fd3d-4de3-bbd1-e8c22727dc4d" , \ "IN_Y_14DOT0TO14DOT5" : "cb669d00-ed49-4392-82b1-eb7415537f62" , \ "IN_X_14DOT5TO15DOT0" : "7929c7eb-96ad-4a3b-a9c0-94307f6010c2" , \ "IN_Y_14DOT5TO15DOT0" : "9eae5ee9-8521-48bf-9e84-bbe35e9de6b4" , \ "IN_X_15DOT0TO16DOT0" : "d5aed402-3c2f-483e-840a-e387224e66af" , \ "IN_Y_15DOT0TO16DOT0" : "2ef013d6-af60-4a3d-aab6-ef2762717167" , \ "IN_X_15DOT0TO15DOT5" : "9148b03d-0580-49cf-916f-4b53f7d0080a" , \ "IN_Y_15DOT0TO15DOT5" : "c4b8ff1e-96f8-4e86-bd1f-a3ccaceb203d" , \ "IN_X_15DOT5TO16DOT0" : "49c70262-2575-406e-97b1-281a8f89c7d4" , \ "IN_Y_15DOT5TO16DOT0" : "54febe2b-ea0d-4bd5-a1d1-0683fac97221" , \ "IN_X_16DOT0TO17DOT0" : "a56c472d-16b6-49ec-aabd-aabce909248c" , \ "IN_Y_16DOT0TO17DOT0" : "ccb746ba-2def-41c3-951d-28850a6a00ad" , \ "IN_X_16DOT0TO16DOT5" : "c4e59fc3-0d75-4d73-8933-a29b505a5822" , \ "IN_Y_16DOT0TO16DOT5" : "2adc6b8e-af50-4816-aee1-7e18318fe509" , \ "IN_X_16DOT5TO17DOT0" : "a80cb4b0-2fbb-45be-a5a7-44e2fa50c18f" , \ "IN_Y_16DOT5TO17DOT0" : "a0f6c3c3-2aae-4661-b94f-25549fee7bde" , \ "IN_X_17DOT0TO18DOT0" : "2fdd6618-0b7a-417d-a3e8-51a23d654e84" , \ "IN_Y_17DOT0TO18DOT0" : "4f256ce6-89fe-4d9a-8fbd-def4eca91a9b" , \ "IN_X_17DOT0TO17DOT5" : "b5234917-1341-40bf-b435-01f5f2bc4d49" , \ "IN_Y_17DOT0TO17DOT5" : "ad7c8e5a-3f28-4fe5-8912-7f5d99a43ef1" , \ "IN_X_17DOT5TO18DOT0" : "1356f838-2af7-4a71-97c9-2faf64a27d3b" , \ "IN_Y_17DOT5TO18DOT0" : "7481ea0c-8a38-4f63-a57a-d7ab78582ac7" , \ "IN_X_18DOT0TO19DOT0" : "e4a691bf-3dba-43b1-876b-02f320b76174" , \ "IN_Y_18DOT0TO19DOT0" : "0e4fa640-ec9a-49c8-8406-9d62c054adc2" , \ "IN_X_18DOT0TO18DOT5" : "d3362e51-8b6b-4821-81e3-d79f5358bd0c" , \ "IN_Y_18DOT0TO18DOT5" : "a0288b89-deae-4985-a222-d245347e4425" , \ "IN_X_18DOT5TO19DOT0" : "cc3110ce-4e51-4d3c-a727-7890492a591f" , \ "IN_Y_18DOT5TO19DOT0" : "d81cbdb0-b895-47fc-b00c-ed1804ecb3d5" , \ "IN_X_19DOT0TO20DOT0" : "13f564ca-de0a-47ae-bf9a-eef973cbea35" , \ "IN_Y_19DOT0TO20DOT0" : "10bd1380-bb72-4e72-9879-9947150cedee" , \ "IN_X_19DOT0TO19DOT5" : "a936a781-dca0-497a-8c1a-8f941de73e68" , \ "IN_Y_19DOT0TO19DOT5" : "64837bca-80bf-4032-a24f-1c8f73752a47" , \ "IN_X_19DOT5TO20DOT0" : "52b574b4-b3f6-4bfa-9511-655c0598aa85" , \ "IN_Y_19DOT5TO20DOT0" : "9a14e009-09ec-408b-ae67-4d1cba135d28" , \ "OUT_U_NEG25DOT0TONEG24DOT0" : "36aec72b-242d-42e7-9a95-b4be4214ad32" , \ "OUT_V_NEG25DOT0TONEG24DOT0" : "d047b7cc-7d72-4637-a808-41b35cdeda48" , \ "OUT_W_NEG25DOT0TONEG24DOT0" : "f2ba7e85-7089-4364-a5a6-55412aae1684" , \ "OUT_U_NEG25DOT0TONEG24DOT5" : "7103ed36-4bb2-4d02-85b6-7d2a20ecba11" , \ "OUT_V_NEG25DOT0TONEG24DOT5" : "ba81511a-bbf8-44fa-b00c-2a418f83a709" , \ "OUT_W_NEG25DOT0TONEG24DOT5" : "6a27fc7e-f3a9-4402-a54d-5dd5fe912c6f" , \ "OUT_U_NEG24DOT5TONEG24DOT0" : "00de3d55-1268-4496-946c-43423d3733bd" , \ "OUT_V_NEG24DOT5TONEG24DOT0" : "30b9472c-5913-4754-b863-908a14bce311" , \ "OUT_W_NEG24DOT5TONEG24DOT0" : "3ddbda5e-6476-4024-af5a-a31a66fdc9b0" , \ "OUT_U_NEG24DOT0TONEG23DOT0" : "07963d8c-45e9-4be1-ade7-90ddfc9731cd" , \ "OUT_V_NEG24DOT0TONEG23DOT0" : "e3b96ac9-b454-430a-a28d-409c90c491a7" , \ "OUT_W_NEG24DOT0TONEG23DOT0" : "c258b08c-382b-46f2-92a6-964893ba19ef" , \ "OUT_U_NEG24DOT0TONEG23DOT5" : "1f9424e8-f35b-41e2-a412-a51b32b30f81" , \ "OUT_V_NEG24DOT0TONEG23DOT5" : "0613b0c1-b459-4d28-b72f-c30429d6867d" , \ "OUT_W_NEG24DOT0TONEG23DOT5" : "982af104-b96e-437b-b580-2a8d2b2ce2d2" , \ "OUT_U_NEG23DOT5TONEG23DOT0" : "b0c09ddb-e399-45f5-be5d-e235a2f5cee7" , \ "OUT_V_NEG23DOT5TONEG23DOT0" : "df7c7744-f491-485f-82a1-8c0ba1a01457" , \ "OUT_W_NEG23DOT5TONEG23DOT0" : "ad1f2272-a6d3-477f-bbf4-d7a16b6b21b9" , \ "OUT_U_NEG23DOT0TONEG22DOT0" : "33d6e154-d40a-4c88-8949-83b93afc21d5" , \ "OUT_V_NEG23DOT0TONEG22DOT0" : "e5da1cf3-e977-4261-af63-c54d2d855e9c" , \ "OUT_W_NEG23DOT0TONEG22DOT0" : "5371a203-13c6-4ba3-bea9-3a4c99a057f5" , \ "OUT_U_NEG23DOT0TONEG22DOT5" : "1205b423-669f-4878-be73-886d09f7351f" , \ "OUT_V_NEG23DOT0TONEG22DOT5" : "4513144d-4d5a-4e35-aeaa-a0616d96a58a" , \ "OUT_W_NEG23DOT0TONEG22DOT5" : "488b91c1-b5f6-4f61-9691-87a62be6f838" , \ "OUT_U_NEG22DOT5TONEG22DOT0" : "00ecfac4-7989-436d-8cbd-b0f425699dfe" , \ "OUT_V_NEG22DOT5TONEG22DOT0" : "2fa22855-e1c6-412d-94c5-45d135d192f0" , \ "OUT_W_NEG22DOT5TONEG22DOT0" : "bfdf149d-4705-4783-84db-ded323dc73d0" , \ "OUT_U_NEG22DOT0TONEG21DOT0" : "298aab04-bbb3-4f14-8016-88251a60c48b" , \ "OUT_V_NEG22DOT0TONEG21DOT0" : "a8e3d378-aaa1-4d98-aefb-21bb8c16388b" , \ "OUT_W_NEG22DOT0TONEG21DOT0" : "cdec1eba-fd32-4c14-8785-5c442ec8d3dc" , \ "OUT_U_NEG22DOT0TONEG21DOT5" : "c91fffef-bbe9-4ae2-a122-4be44557d78e" , \ "OUT_V_NEG22DOT0TONEG21DOT5" : "ec3faeb6-3a5c-4dd8-9a5f-4bd832066685" , \ "OUT_W_NEG22DOT0TONEG21DOT5" : "9ae377de-95f0-4f57-8030-e2d7e7cc8ca8" , \ "OUT_U_NEG21DOT5TONEG21DOT0" : "d11647b0-40e3-45f5-b38b-1fca9f5ff860" , \ "OUT_V_NEG21DOT5TONEG21DOT0" : "63f7c161-a6f6-4144-a4c8-85ccad0c6b98" , \ "OUT_W_NEG21DOT5TONEG21DOT0" : "7b6f6a05-7c11-494d-9e13-90743995bc04" , \ "OUT_U_NEG21DOT0TONEG20DOT0" : "c92b796f-2b4d-4bb3-b707-d04dc1764c33" , \ "OUT_V_NEG21DOT0TONEG20DOT0" : "d1a56b79-028d-4400-9c94-8ea07c2d6b5d" , \ "OUT_W_NEG21DOT0TONEG20DOT0" : "c1612ed7-77ac-4366-94ed-2551d9e4cb81" , \ "OUT_U_NEG21DOT0TONEG20DOT5" : "89253478-92d6-4425-adbd-538eb7986354" , \ "OUT_V_NEG21DOT0TONEG20DOT5" : "1ae06644-5b79-4b52-8ffc-78ba545f8bb2" , \ "OUT_W_NEG21DOT0TONEG20DOT5" : "72ac71f5-ae86-4a2e-b6aa-ee6b47f5a36d" , \ "OUT_U_NEG20DOT5TONEG20DOT0" : "88f23d6d-39df-4d1a-afaf-6a4150b96f76" , \ "OUT_V_NEG20DOT5TONEG20DOT0" : "3d9a4a34-0ab0-4da4-86fc-51766c2010f2" , \ "OUT_W_NEG20DOT5TONEG20DOT0" : "8efd2da1-1c5f-412b-a403-879d82007166" , \ "OUT_U_NEG20DOT0TONEG19DOT0" : "246a9bca-1872-4e70-9ca6-85ed4667c1fa" , \ "OUT_V_NEG20DOT0TONEG19DOT0" : "3830270e-b33a-4b84-aeb4-6f974d769dac" , \ "OUT_W_NEG20DOT0TONEG19DOT0" : "67e7da95-fb40-4bc9-af3b-fbfcf52feaeb" , \ "OUT_U_NEG20DOT0TONEG19DOT5" : "c1c0ed2c-20f3-41bc-94d8-7bd79a5dde05" , \ "OUT_V_NEG20DOT0TONEG19DOT5" : "34d9d356-88f9-4731-9980-63160ebfcd28" , \ "OUT_W_NEG20DOT0TONEG19DOT5" : "fc5c6a08-3b4c-4505-8c8e-7d5860de6b88" , \ "OUT_U_NEG19DOT5TONEG19DOT0" : "96cc7c7d-27b7-49e6-9b0a-efac9eb425f5" , \ "OUT_V_NEG19DOT5TONEG19DOT0" : "45a29a44-4f28-4f4f-8a5b-54928364cd91" , \ "OUT_W_NEG19DOT5TONEG19DOT0" : "42ae4b85-e78c-4d49-b47d-7fab475a10b9" , \ "OUT_U_NEG19DOT0TONEG18DOT0" : "db61d5a0-92c4-46d9-9ea2-588571cfa3f8" , \ "OUT_V_NEG19DOT0TONEG18DOT0" : "68f6da40-a393-42f6-9d23-d2fd512eaf75" , \ "OUT_W_NEG19DOT0TONEG18DOT0" : "d8a87da0-2009-44a8-951e-e941e81e5952" , \ "OUT_U_NEG19DOT0TONEG18DOT5" : "868d6ffa-196d-4b95-9c6b-9915037c2a79" , \ "OUT_V_NEG19DOT0TONEG18DOT5" : "f6f31575-d3d4-481c-a3e8-0d5654f7951e" , \ "OUT_W_NEG19DOT0TONEG18DOT5" : "9518a350-a115-49c8-a956-cb4b88efcdaf" , \ "OUT_U_NEG18DOT5TONEG18DOT0" : "fee9939a-f502-4d6e-9f25-a4e8b86f2b6f" , \ "OUT_V_NEG18DOT5TONEG18DOT0" : "90b8356b-e1a9-4624-8521-de5922918191" , \ "OUT_W_NEG18DOT5TONEG18DOT0" : "a6c7fc39-6556-4a24-af83-08c489747485" , \ "OUT_U_NEG18DOT0TONEG17DOT0" : "1f08d63d-b2dc-46ee-a1e6-759ea78635cc" , \ "OUT_V_NEG18DOT0TONEG17DOT0" : "c9cdc6c4-9176-45ea-9eda-dca020c6eac1" , \ "OUT_W_NEG18DOT0TONEG17DOT0" : "06fd9f20-8562-49ea-90b9-1840eae3cf4b" , \ "OUT_U_NEG18DOT0TONEG17DOT5" : "b83ee8e3-65ad-4c8e-9804-1a2b414fd2e9" , \ "OUT_V_NEG18DOT0TONEG17DOT5" : "a65addd3-4070-4d31-a918-d45b5833e94f" , \ "OUT_W_NEG18DOT0TONEG17DOT5" : "5ba4aa72-da6c-4667-ae8d-5e4f8a009efd" , \ "OUT_U_NEG17DOT5TONEG17DOT0" : "5eed0347-24ec-4c94-abd7-d622dbe56a1f" , \ "OUT_V_NEG17DOT5TONEG17DOT0" : "72f7d856-ce6e-4c3a-928c-2e86abd7b0c8" , \ "OUT_W_NEG17DOT5TONEG17DOT0" : "c57b8910-a49f-4b15-9630-8dd9ead0bcf9" , \ "OUT_U_NEG17DOT0TONEG16DOT0" : "e80e4e1f-9053-43e1-974a-c942b65b3e31" , \ "OUT_V_NEG17DOT0TONEG16DOT0" : "42e98f7c-e6df-4dd8-b514-0eeab46e277d" , \ "OUT_W_NEG17DOT0TONEG16DOT0" : "548fd44f-109e-4437-a844-fca66cb6eb76" , \ "OUT_U_NEG17DOT0TONEG16DOT5" : "e9d236b1-0146-429d-b74d-108a358a9c40" , \ "OUT_V_NEG17DOT0TONEG16DOT5" : "5f825279-4f9c-4a83-9a5c-9396bedd64e3" , \ "OUT_W_NEG17DOT0TONEG16DOT5" : "b7092d48-93b2-4400-b14e-1ccf62a6a806" , \ "OUT_U_NEG16DOT5TONEG16DOT0" : "32efa15a-55ec-4487-8cc6-8afcba669fe8" , \ "OUT_V_NEG16DOT5TONEG16DOT0" : "89daa459-7965-4868-9b20-2134d3b40b13" , \ "OUT_W_NEG16DOT5TONEG16DOT0" : "d6710fa9-dbcd-4f98-9aa9-55b4fe73d9fd" , \ "OUT_U_NEG16DOT0TONEG15DOT0" : "bc3409d9-1247-4020-bf3c-a2e063a704ce" , \ "OUT_V_NEG16DOT0TONEG15DOT0" : "8247a71d-1bd7-4175-93e9-c572ab976008" , \ "OUT_W_NEG16DOT0TONEG15DOT0" : "20b086a9-4051-4869-8f4e-b48b066a6748" , \ "OUT_U_NEG16DOT0TONEG15DOT5" : "d1973de0-a4c8-4940-b302-0a04b519bb3f" , \ "OUT_V_NEG16DOT0TONEG15DOT5" : "0e2a96b9-9250-486d-91cf-d6d67d6adbd3" , \ "OUT_W_NEG16DOT0TONEG15DOT5" : "2c7dda13-e3f4-4479-825b-e5bdbd9cc798" , \ "OUT_U_NEG15DOT5TONEG15DOT0" : "d8ca6ea0-24e8-4c1a-a53a-083c9d014665" , \ "OUT_V_NEG15DOT5TONEG15DOT0" : "bdace2fc-1ee0-4469-a822-8fa1fd1a5bdf" , \ "OUT_W_NEG15DOT5TONEG15DOT0" : "a47e804d-da40-45b9-9d11-8daa20d0a015" , \ "OUT_U_NEG15DOT0TONEG14DOT0" : "2c8c772c-6bee-4c03-b67c-d06d6425a57e" , \ "OUT_V_NEG15DOT0TONEG14DOT0" : "0287b0f1-ef1f-47bb-ad75-6aa89cfc8e6c" , \ "OUT_W_NEG15DOT0TONEG14DOT0" : "3c44e588-2ac9-4c7e-a4e5-3be02961f6cb" , \ "OUT_U_NEG15DOT0TONEG14DOT5" : "6c49f8ee-988a-4dbd-be1a-b6f3dadeb092" , \ "OUT_V_NEG15DOT0TONEG14DOT5" : "69352542-5fe1-40ac-8397-2374d6ee44f1" , \ "OUT_W_NEG15DOT0TONEG14DOT5" : "91771f82-dbe6-4912-9a66-ff2995635563" , \ "OUT_U_NEG14DOT5TONEG14DOT0" : "fd9e2b84-0cc1-473e-b6b3-f8e75275fffc" , \ "OUT_V_NEG14DOT5TONEG14DOT0" : "6f83d4a3-be89-4aa3-be3d-8bf74a5ec4e6" , \ "OUT_W_NEG14DOT5TONEG14DOT0" : "e6955c0e-0c10-4ad6-bbf1-53fa18244d11" , \ "OUT_U_NEG14DOT0TONEG13DOT0" : "0855c1f1-9f6e-44b8-b9dd-8d900232dd0b" , \ "OUT_V_NEG14DOT0TONEG13DOT0" : "2a279cb0-78f6-4a66-89d8-c108b2846464" , \ "OUT_W_NEG14DOT0TONEG13DOT0" : "a694e8e9-fbed-4ed3-b847-396c60dbc8cd" , \ "OUT_U_NEG14DOT0TONEG13DOT5" : "246e8937-29d2-4b58-8b81-1fd9884c19c0" , \ "OUT_V_NEG14DOT0TONEG13DOT5" : "6ebb225e-6097-472c-8903-0d5a0c8c9346" , \ "OUT_W_NEG14DOT0TONEG13DOT5" : "4ff7b31b-1f53-4cb5-a0c0-816634d62de1" , \ "OUT_U_NEG13DOT5TONEG13DOT0" : "6cfc5a71-3bea-4c31-9211-b14eeff0b982" , \ "OUT_V_NEG13DOT5TONEG13DOT0" : "9dd2434c-6bf2-45da-8073-79af3ff5976a" , \ "OUT_W_NEG13DOT5TONEG13DOT0" : "f0524153-e1ed-4d06-831e-ab3d5536ea95" , \ "OUT_U_NEG13DOT0TONEG12DOT0" : "75d9ef62-83e1-40a3-8359-2f6ed241698f" , \ "OUT_V_NEG13DOT0TONEG12DOT0" : "45b82517-e48f-494e-9684-c21a7743407d" , \ "OUT_W_NEG13DOT0TONEG12DOT0" : "9493e477-c7b9-44a7-8e4c-980f3a9ec7b3" , \ "OUT_U_NEG13DOT0TONEG12DOT5" : "369dd09f-76c6-4f68-bea7-067120e8c2b8" , \ "OUT_V_NEG13DOT0TONEG12DOT5" : "3523d601-21f4-4613-bebc-9a41e52141a5" , \ "OUT_W_NEG13DOT0TONEG12DOT5" : "8fadf7a1-9b99-479d-ab3a-cc86cc31c904" , \ "OUT_U_NEG12DOT5TONEG12DOT0" : "945224d6-5d67-459d-a107-c4bc8a23a583" , \ "OUT_V_NEG12DOT5TONEG12DOT0" : "58be66be-7b9f-4c82-af45-1bd8c99974ab" , \ "OUT_W_NEG12DOT5TONEG12DOT0" : "c142d7e7-2c8b-4819-bfd2-bd951590cb28" , \ "OUT_U_NEG12DOT0TONEG11DOT0" : "bcf21f23-7ec4-443d-90e0-969e5c706750" , \ "OUT_V_NEG12DOT0TONEG11DOT0" : "7fc00b03-b9b7-47cd-994b-a8e8c0a463d1" , \ "OUT_W_NEG12DOT0TONEG11DOT0" : "f1caa5fa-b921-468f-acc4-7349dbd71f28" , \ "OUT_U_NEG12DOT0TONEG11DOT5" : "5dc20454-0d32-42e0-9351-da7279d2e4d2" , \ "OUT_V_NEG12DOT0TONEG11DOT5" : "7e46d2b7-fa5a-4864-9ac5-263bf9b65345" , \ "OUT_W_NEG12DOT0TONEG11DOT5" : "f0581da0-18f9-4d7b-8a08-4af8a33728a0" , \ "OUT_U_NEG11DOT5TONEG11DOT0" : "4f4cfb53-a13e-44bd-bca5-d74c72d31b16" , \ "OUT_V_NEG11DOT5TONEG11DOT0" : "0daba05e-fc39-41b9-897c-af8154baa97d" , \ "OUT_W_NEG11DOT5TONEG11DOT0" : "7ae6f85f-a5b6-4e28-a506-fa0d4929981a" , \ "OUT_U_NEG11DOT0TONEG10DOT0" : "c338cb1e-964b-4c57-9003-a7b38091f9e2" , \ "OUT_V_NEG11DOT0TONEG10DOT0" : "b2cd950c-dfd6-47c7-9a7f-6112d8eecb25" , \ "OUT_W_NEG11DOT0TONEG10DOT0" : "65f244d9-6191-4e94-a6d8-fc1e8879f89c" , \ "OUT_U_NEG11DOT0TONEG10DOT5" : "e4605504-e90a-4539-b776-b1027416e898" , \ "OUT_V_NEG11DOT0TONEG10DOT5" : "36a14614-0750-4327-a7ee-419c924377f2" , \ "OUT_W_NEG11DOT0TONEG10DOT5" : "01405bb1-a94d-4a96-8163-1f61cf524e83" , \ "OUT_U_NEG10DOT5TONEG10DOT0" : "549c1908-afe2-4cda-8e52-056e187cf173" , \ "OUT_V_NEG10DOT5TONEG10DOT0" : "4541ede1-252b-409a-9b88-091e687bf0e1" , \ "OUT_W_NEG10DOT5TONEG10DOT0" : "615e2125-40d2-4ace-a796-1db2587b6bfa" , \ "OUT_U_NEG10DOT0TONEG9DOT0" : "39404a03-5383-4674-aa5f-a5014e7880d2" , \ "OUT_V_NEG10DOT0TONEG9DOT0" : "d4d513e4-aa7f-447b-abae-ba8a67c07d8e" , \ "OUT_W_NEG10DOT0TONEG9DOT0" : "6fd8031f-2d4c-4e07-8dee-d3ce56dc3159" , \ "OUT_U_NEG10DOT0TONEG9DOT5" : "af5b2a10-0177-4a0a-860d-cc645a1f84aa" , \ "OUT_V_NEG10DOT0TONEG9DOT5" : "04d047d5-3814-4f4f-ada7-1cd3c7cf23a0" , \ "OUT_W_NEG10DOT0TONEG9DOT5" : "a9294466-8320-4f9d-a7e3-6463db94285e" , \ "OUT_U_NEG9DOT5TONEG9DOT0" : "92c3db1e-b7a2-4e99-9f1d-547955ab7170" , \ "OUT_V_NEG9DOT5TONEG9DOT0" : "9d08564f-2d8e-4732-afba-f9fb3f3e4939" , \ "OUT_W_NEG9DOT5TONEG9DOT0" : "93309ecb-e604-4a4a-93a5-7f3ae4469d5f" , \ "OUT_U_NEG9DOT0TONEG8DOT0" : "92dccc3a-f450-448a-94ea-811820789233" , \ "OUT_V_NEG9DOT0TONEG8DOT0" : "31bee0d4-a0b7-4a13-86ef-bfd2c4cb091b" , \ "OUT_W_NEG9DOT0TONEG8DOT0" : "680ddb11-987d-4cbe-84e1-10468dcb0296" , \ "OUT_U_NEG9DOT0TONEG8DOT5" : "1c40af42-b072-4c82-8a46-7a302338306a" , \ "OUT_V_NEG9DOT0TONEG8DOT5" : "f0bab64b-4fd9-4ee5-8987-0dfdb4e822f4" , \ "OUT_W_NEG9DOT0TONEG8DOT5" : "5b0aaa3d-ce85-4ac2-a17d-ba08b6a3ed3a" , \ "OUT_U_NEG8DOT5TONEG8DOT0" : "eb1d189a-9910-49da-9afc-f2a1a36fd0eb" , \ "OUT_V_NEG8DOT5TONEG8DOT0" : "1080d198-7936-4f69-83bf-0090f2ce7e4c" , \ "OUT_W_NEG8DOT5TONEG8DOT0" : "af8ac348-9db5-4127-8093-dde37965acb0" , \ "OUT_U_NEG8DOT0TONEG7DOT0" : "229c1743-d4a6-4f49-8ef9-4363025528c2" , \ "OUT_V_NEG8DOT0TONEG7DOT0" : "8d90724d-14ec-4c2e-a4e5-1b678262db99" , \ "OUT_W_NEG8DOT0TONEG7DOT0" : "c6417f5d-2151-4300-82e8-f79f73688a84" , \ "OUT_U_NEG8DOT0TONEG7DOT5" : "d1b71139-08e9-4576-97b7-d20c069033a3" , \ "OUT_V_NEG8DOT0TONEG7DOT5" : "db791c3f-9514-4b62-962c-98d4466c89fb" , \ "OUT_W_NEG8DOT0TONEG7DOT5" : "c9a4b827-76a5-41e9-8bf4-527b4c06a7ab" , \ "OUT_U_NEG7DOT5TONEG7DOT0" : "5d933298-bcbb-4944-8d39-c47a33c23303" , \ "OUT_V_NEG7DOT5TONEG7DOT0" : "027bcd31-cded-4312-ad38-a77ca29b08d0" , \ "OUT_W_NEG7DOT5TONEG7DOT0" : "1f042e6c-b4b0-441c-99c6-88840c1c00f8" , \ "OUT_U_NEG7DOT0TONEG6DOT0" : "278f7ba6-ad7d-47b2-ba68-49d544919c28" , \ "OUT_V_NEG7DOT0TONEG6DOT0" : "417ee740-d816-474f-8fc2-092e27d11d24" , \ "OUT_W_NEG7DOT0TONEG6DOT0" : "a963b075-e64e-4aac-addb-87e77ab8b918" , \ "OUT_U_NEG7DOT0TONEG6DOT5" : "c17c8894-7172-4bcb-98b9-8be0bc8f91b4" , \ "OUT_V_NEG7DOT0TONEG6DOT5" : "ef93b375-bc8b-47ae-9f48-9ed598346f57" , \ "OUT_W_NEG7DOT0TONEG6DOT5" : "9195ae09-f4a6-4749-9a42-63b64e275999" , \ "OUT_U_NEG6DOT5TONEG6DOT0" : "0d2607f2-1e4f-420d-aac0-74159adf910e" , \ "OUT_V_NEG6DOT5TONEG6DOT0" : "0c69be10-edd4-4a75-87fa-89ec5a202501" , \ "OUT_W_NEG6DOT5TONEG6DOT0" : "f9f84c2b-b1c7-4947-ab83-5d125df3216d" , \ "OUT_U_NEG6DOT0TONEG5DOT0" : "ec0e93bc-2888-46a0-8873-75490060a0e8" , \ "OUT_V_NEG6DOT0TONEG5DOT0" : "dd041b28-5cd3-4ab9-9068-e6a8364edbe2" , \ "OUT_W_NEG6DOT0TONEG5DOT0" : "89a7db08-6e02-4d65-85fb-68620422b8cc" , \ "OUT_U_NEG6DOT0TONEG5DOT5" : "cd915631-53a2-4899-968e-e45a1fdd1050" , \ "OUT_V_NEG6DOT0TONEG5DOT5" : "46966672-d27e-48d6-9f4a-7338c7e4bf92" , \ "OUT_W_NEG6DOT0TONEG5DOT5" : "fadc8105-d82b-4a1c-9546-b5d04ce5d1c6" , \ "OUT_U_NEG5DOT5TONEG5DOT0" : "327db73c-9da7-4dc8-b0b3-b63cf4e3f5d7" , \ "OUT_V_NEG5DOT5TONEG5DOT0" : "8348a655-75fa-4b24-b4d3-2c2a7c9e49e4" , \ "OUT_W_NEG5DOT5TONEG5DOT0" : "8560bece-51bd-4d42-a582-92445b33bbbc" , \ "OUT_U_NEG5DOT0TONEG4DOT0" : "1879854a-ac85-474b-882b-dfda13997ce0" , \ "OUT_V_NEG5DOT0TONEG4DOT0" : "e4cc3b91-80bd-405b-a66a-07b68295753c" , \ "OUT_W_NEG5DOT0TONEG4DOT0" : "cb3971f1-5001-4e5e-96fe-617b0522fe15" , \ "OUT_U_NEG5DOT0TONEG4DOT5" : "33eb8237-5c87-44a3-ab51-ee5ffece1a26" , \ "OUT_V_NEG5DOT0TONEG4DOT5" : "d2d7cf89-7e83-4bf9-af4c-4721a1fee66d" , \ "OUT_W_NEG5DOT0TONEG4DOT5" : "fe85a5a2-5ed9-413f-b0ef-e48a06082d63" , \ "OUT_U_NEG4DOT5TONEG4DOT0" : "92680dc0-951c-4f52-b04d-3cff9efff722" , \ "OUT_V_NEG4DOT5TONEG4DOT0" : "bb8ad89e-2ddc-4859-b659-dcba06909cc3" , \ "OUT_W_NEG4DOT5TONEG4DOT0" : "90ee2eb4-6076-4f28-94b4-05f5430b4669" , \ "OUT_U_NEG4DOT0TONEG3DOT0" : "5ee43133-56cf-49e7-8a86-a95b139de14c" , \ "OUT_V_NEG4DOT0TONEG3DOT0" : "d6c196af-2cb2-4f24-b8f5-2b60277e8c7a" , \ "OUT_W_NEG4DOT0TONEG3DOT0" : "844b86e9-031c-48a0-a666-482a9b633623" , \ "OUT_U_NEG4DOT0TONEG3DOT5" : "85a4a061-f21a-465b-9c32-75d1717e0ff9" , \ "OUT_V_NEG4DOT0TONEG3DOT5" : "ddee67e4-35bb-4b6d-8818-75222ecdb82f" , \ "OUT_W_NEG4DOT0TONEG3DOT5" : "a70e0ee6-1050-49e4-a69a-0ec31b970d79" , \ "OUT_U_NEG3DOT5TONEG3DOT0" : "9976ede5-fa72-48be-a649-7fc10b9f322e" , \ "OUT_V_NEG3DOT5TONEG3DOT0" : "e4f363c5-3e41-43f8-867a-82a40ee54210" , \ "OUT_W_NEG3DOT5TONEG3DOT0" : "8bbe7276-8c6d-4517-ae43-65d80bb5dd56" , \ "OUT_U_NEG3DOT0TONEG2DOT0" : "72e628a1-2fde-4659-a04a-c53c193d0256" , \ "OUT_V_NEG3DOT0TONEG2DOT0" : "87d2bb01-4b6f-43d5-b66d-e3aa8e00ecec" , \ "OUT_W_NEG3DOT0TONEG2DOT0" : "41146ed5-761a-43a5-8001-d524ea670ad6" , \ "OUT_U_NEG3DOT0TONEG2DOT5" : "4836a4fa-0b7e-41b6-991d-76671c763f6f" , \ "OUT_V_NEG3DOT0TONEG2DOT5" : "26a0ead7-d776-4830-8251-e49ff2b8842e" , \ "OUT_W_NEG3DOT0TONEG2DOT5" : "d29746c7-8714-4cc6-97cc-f3abcae4fc15" , \ "OUT_U_NEG2DOT5TONEG2DOT0" : "059e1d64-e3d8-42a8-9b94-e1c2ed643232" , \ "OUT_V_NEG2DOT5TONEG2DOT0" : "29253612-3504-4e66-969c-515408002817" , \ "OUT_W_NEG2DOT5TONEG2DOT0" : "47da7a90-c97a-47de-81e3-99b0d7f5ba4a" , \ "OUT_U_NEG2DOT0TONEG1DOT0" : "9d08b9d0-7a23-41de-9bc8-6017eb4a57ea" , \ "OUT_V_NEG2DOT0TONEG1DOT0" : "d69290b1-e796-4c96-8037-0f0f4798fd1e" , \ "OUT_W_NEG2DOT0TONEG1DOT0" : "8ed66cb6-ba53-427f-80b3-0d580882f126" , \ "OUT_U_NEG2DOT0TONEG1DOT5" : "c55b54c7-ed16-4a0f-ba89-7546f27f8d82" , \ "OUT_V_NEG2DOT0TONEG1DOT5" : "2a8783a5-cb88-4137-b907-5f28592124e3" , \ "OUT_W_NEG2DOT0TONEG1DOT5" : "d47ad741-f565-42fb-8731-001d6f6d00dd" , \ "OUT_U_NEG1DOT5TONEG1DOT0" : "6676d2af-a125-4bca-9016-1879b3bb6341" , \ "OUT_V_NEG1DOT5TONEG1DOT0" : "b77cda92-302d-409c-97b1-6b5c36feeaee" , \ "OUT_W_NEG1DOT5TONEG1DOT0" : "969c7c15-ad16-423e-b126-6703a36e041a" , \ "OUT_U_NEG1DOT0TO0DOT0" : "cfd992e5-738b-4573-badc-5c5510552b76" , \ "OUT_V_NEG1DOT0TO0DOT0" : "dc56ef6a-e191-4cac-a353-bd9dbb38b6b1" , \ "OUT_W_NEG1DOT0TO0DOT0" : "1ae840b1-c651-4583-a7aa-6028fbd6da57" , \ "OUT_U_NEG1DOT0TONEG0DOT5" : "a2f5e475-94ff-4436-a7bf-12da379ddda5" , \ "OUT_V_NEG1DOT0TONEG0DOT5" : "de69845b-3b49-42bc-acf9-d6f408418f24" , \ "OUT_W_NEG1DOT0TONEG0DOT5" : "0123b914-8ad3-46b1-abb8-26dc7b8a973c" , \ "OUT_U_NEG0DOT5TO0DOT0" : "fa6945b7-27ca-4a04-a0ba-44c610b9f5e4" , \ "OUT_V_NEG0DOT5TO0DOT0" : "ffa8350b-cb9a-434e-aaaf-911709b9d29b" , \ "OUT_W_NEG0DOT5TO0DOT0" : "d4163df7-d235-4756-b09d-18e8c69703b1" , \ "OUT_U_0DOT0TO1DOT0" : "057947e0-44ac-47d5-b688-3987f3ec3bb4" , \ "OUT_V_0DOT0TO1DOT0" : "6c391306-fc3a-4a6f-bac6-745c7bf8f598" , \ "OUT_W_0DOT0TO1DOT0" : "037f8d78-7ca0-4423-a288-cee3f7c0275a" , \ "OUT_U_0DOT0TO0DOT5" : "c1e80afd-949d-4f7c-a5f3-7fc4215419fb" , \ "OUT_V_0DOT0TO0DOT5" : "71b1cf6f-f229-4ec1-b2e4-f96d138f5c11" , \ "OUT_W_0DOT0TO0DOT5" : "82e2f5e9-f792-411d-abb5-b7e491cfe96f" , \ "OUT_U_0DOT5TO1DOT0" : "8a76ef57-eead-4e3e-b0e1-21bc7fb1f2bc" , \ "OUT_V_0DOT5TO1DOT0" : "c11e059b-a6a4-46ad-b68b-1c0bf3404a9c" , \ "OUT_W_0DOT5TO1DOT0" : "1ece566e-0024-480a-9f84-68d497367fe3" , \ "OUT_U_1DOT0TO2DOT0" : "1174e942-7a64-4a79-8daf-16306508f54b" , \ "OUT_V_1DOT0TO2DOT0" : "d08a11e6-73e3-4f24-818d-b42fe6f0dcb8" , \ "OUT_W_1DOT0TO2DOT0" : "2fc3f2f1-8036-4be6-9f02-e84dfcade8a7" , \ "OUT_U_1DOT0TO1DOT5" : "421d9ebf-d16f-4fa2-9460-efa7a0cc793c" , \ "OUT_V_1DOT0TO1DOT5" : "d403c072-7f7f-490c-9486-4b9aed08d8df" , \ "OUT_W_1DOT0TO1DOT5" : "0f906e40-6c24-45ec-a89c-e42cd6adccaf" , \ "OUT_U_1DOT5TO2DOT0" : "58735ae7-6f87-465a-8411-5aaecd02f1bb" , \ "OUT_V_1DOT5TO2DOT0" : "09803248-757a-4e3c-b1bf-70d5206b5243" , \ "OUT_W_1DOT5TO2DOT0" : "946acd88-2d81-4cbd-9305-4cc186a8d039" , \ "OUT_U_2DOT0TO3DOT0" : "efa21c93-d9ba-43e7-a622-f30fb8a217b5" , \ "OUT_V_2DOT0TO3DOT0" : "7fba2231-7d37-4b7b-9d01-52540eb02bf6" , \ "OUT_W_2DOT0TO3DOT0" : "f2f21f24-d52a-48a3-8c4b-197f55d96c08" , \ "OUT_U_2DOT0TO2DOT5" : "751c1a29-a137-4b7d-bfcf-f00a3610753e" , \ "OUT_V_2DOT0TO2DOT5" : "752bb1bc-2bdf-4f4d-9aa3-f9c6a572ee4f" , \ "OUT_W_2DOT0TO2DOT5" : "0c0ecfc9-674e-4341-a532-0950866e1935" , \ "OUT_U_2DOT5TO3DOT0" : "42f1e26c-44d9-49bf-b1f5-bdc76a34e837" , \ "OUT_V_2DOT5TO3DOT0" : "3c02c8be-6b92-4901-b074-bbbb18ff7c67" , \ "OUT_W_2DOT5TO3DOT0" : "73839f74-6ef9-4219-8766-5eca24da179e" , \ "OUT_U_3DOT0TO4DOT0" : "bce78e90-88f8-44e4-aea6-a45907ebb64c" , \ "OUT_V_3DOT0TO4DOT0" : "0e28f77e-5f80-406d-8193-acffc8e1885d" , \ "OUT_W_3DOT0TO4DOT0" : "658c2eb1-9f51-4e11-9134-a7a92ea032c4" , \ "OUT_U_3DOT0TO3DOT5" : "8f061f8a-4851-4619-bb35-ab87b336b85a" , \ "OUT_V_3DOT0TO3DOT5" : "c145194a-4217-4e91-94a1-646997db8f12" , \ "OUT_W_3DOT0TO3DOT5" : "966fdb0e-23b7-4964-939f-08c1e0a2f183" , \ "OUT_U_3DOT5TO4DOT0" : "51ace9de-13d8-43d6-a12c-558f6101061c" , \ "OUT_V_3DOT5TO4DOT0" : "6a8fd123-496d-4004-9b4c-c7d061046d8a" , \ "OUT_W_3DOT5TO4DOT0" : "62375996-b03e-49a1-b53f-3e838489f426" , \ "OUT_U_4DOT0TO5DOT0" : "b353eb3a-baa3-484a-b8f1-63565d7574d5" , \ "OUT_V_4DOT0TO5DOT0" : "9a454f1d-5325-4c00-98b2-c429874ee677" , \ "OUT_W_4DOT0TO5DOT0" : "5405eeb2-db49-4b0f-8ca2-7b337d2f1978" , \ "OUT_U_4DOT0TO4DOT5" : "b6159529-38ca-4d50-b9a7-8f92e19d8b5b" , \ "OUT_V_4DOT0TO4DOT5" : "c37b483d-35ef-42fe-bf6d-0d50eadb7d86" , \ "OUT_W_4DOT0TO4DOT5" : "f0bad062-b3a9-4251-952e-9393ff0f1cae" , \ "OUT_U_4DOT5TO5DOT0" : "c7cf7620-9fa3-47d5-aae7-f0dcbad3769f" , \ "OUT_V_4DOT5TO5DOT0" : "4eab4e10-bc48-44db-9f03-69f07aca2e84" , \ "OUT_W_4DOT5TO5DOT0" : "7352a600-f8e2-4838-a374-3c589616ab13" , \ "OUT_U_5DOT0TO6DOT0" : "cc72c315-1a30-4267-866b-5403ef827608" , \ "OUT_V_5DOT0TO6DOT0" : "44049260-4aa8-4846-b285-82413558bc9d" , \ "OUT_W_5DOT0TO6DOT0" : "7b9e7df0-d9fb-42c8-852f-270d73b87fb4" , \ "OUT_U_5DOT0TO5DOT5" : "088d5399-f868-4074-af30-25a4e2a9c782" , \ "OUT_V_5DOT0TO5DOT5" : "4abffdd5-2666-4a42-ae87-6b1ea91f9300" , \ "OUT_W_5DOT0TO5DOT5" : "fee4311a-9fa6-45d1-84da-a722b6669774" , \ "OUT_U_5DOT5TO6DOT0" : "97dbe86e-e522-4c8d-96b7-dcaf51e014f9" , \ "OUT_V_5DOT5TO6DOT0" : "f1bde450-2b0f-4035-bbb5-48c3f20fa806" , \ "OUT_W_5DOT5TO6DOT0" : "180aeb32-4ad8-4d12-a821-1c7d7e50810f" , \ "OUT_U_6DOT0TO7DOT0" : "6886b98c-b7bd-45e3-8236-d9e4dc984d0b" , \ "OUT_V_6DOT0TO7DOT0" : "6a4973b0-fda1-4a5a-908a-d4b9c6617c5d" , \ "OUT_W_6DOT0TO7DOT0" : "00578db1-01cb-458b-bbe6-ce5cb4e87eee" , \ "OUT_U_6DOT0TO6DOT5" : "d911a510-1763-4f42-a2b2-047a54229f24" , \ "OUT_V_6DOT0TO6DOT5" : "77220a62-ea00-4bdb-873d-3cb941ab0df2" , \ "OUT_W_6DOT0TO6DOT5" : "21eadd91-69d4-4ef1-85c1-f28d17ffe98e" , \ "OUT_U_6DOT5TO7DOT0" : "484a2ad4-a7cd-4681-9d6f-65d774b05137" , \ "OUT_V_6DOT5TO7DOT0" : "1d019214-4047-4130-8258-a92c3fa1b936" , \ "OUT_W_6DOT5TO7DOT0" : "a17a4f91-ff9c-4e28-830c-383f238281dc" , \ "OUT_U_7DOT0TO8DOT0" : "a17336ed-b098-4d6e-bc3d-78b938e43aa8" , \ "OUT_V_7DOT0TO8DOT0" : "1c4f9f45-6b7e-4a14-a81d-a22e2b3bcb30" , \ "OUT_W_7DOT0TO8DOT0" : "9e148160-60e4-47e0-9f34-eae9d02520af" , \ "OUT_U_7DOT0TO7DOT5" : "5ce17a02-4872-4203-8689-452889cab106" , \ "OUT_V_7DOT0TO7DOT5" : "f9e583ca-45f9-4891-82f2-897af56df166" , \ "OUT_W_7DOT0TO7DOT5" : "ee97787c-38c0-4c20-909d-1cf6cec3ff41" , \ "OUT_U_7DOT5TO8DOT0" : "e6023ff4-1086-461b-9323-d567fbedd8ad" , \ "OUT_V_7DOT5TO8DOT0" : "84d180cc-6ebc-4b74-97fc-547fdc4bcd3e" , \ "OUT_W_7DOT5TO8DOT0" : "92688b6c-1a3c-4a1b-a582-07abe0eb4983" , \ "OUT_U_8DOT0TO9DOT0" : "8bf2d272-5816-4d74-8ef5-14dbb816c22f" , \ "OUT_V_8DOT0TO9DOT0" : "782a139c-bb97-4155-b613-b919b2980503" , \ "OUT_W_8DOT0TO9DOT0" : "ea3bf5c2-dba1-4757-8cf1-5e3b67825dcb" , \ "OUT_U_8DOT0TO8DOT5" : "d8a413f5-c833-4260-9098-e3569738703a" , \ "OUT_V_8DOT0TO8DOT5" : "79e4551b-1bef-436b-abf6-0e06440c0868" , \ "OUT_W_8DOT0TO8DOT5" : "a6644e56-ca00-47ea-9110-17620cdadb09" , \ "OUT_U_8DOT5TO9DOT0" : "cffe29df-9a8a-403e-b67a-2c3c2897e18f" , \ "OUT_V_8DOT5TO9DOT0" : "a0ffe61a-beef-4f7c-b9e8-13655c88a3dd" , \ "OUT_W_8DOT5TO9DOT0" : "980cfa10-53fb-4626-9f13-6d6bf58cad69" , \ "OUT_U_9DOT0TO10DOT0" : "66589cb0-5ab4-4d66-bd04-e81415f12611" , \ "OUT_V_9DOT0TO10DOT0" : "2a279ef7-23b2-46c2-a820-3e21905103ed" , \ "OUT_W_9DOT0TO10DOT0" : "e0420b69-1b86-46d1-8be3-0b000db7607b" , \ "OUT_U_9DOT0TO9DOT5" : "634bd87d-e743-4f08-89cf-a582f732950d" , \ "OUT_V_9DOT0TO9DOT5" : "4304731b-9b09-4da0-bdb6-88a85e91549b" , \ "OUT_W_9DOT0TO9DOT5" : "a4b40ddd-048a-4960-9b44-d66ba2d9dc92" , \ "OUT_U_9DOT5TO10DOT0" : "c8d647a9-aeb9-4714-92f2-a34880842f76" , \ "OUT_V_9DOT5TO10DOT0" : "73505e1d-2c86-4c11-89ed-3e9e101257c8" , \ "OUT_W_9DOT5TO10DOT0" : "5e076135-920d-49c2-ba82-195e04192359" , \ "OUT_U_10DOT0TO11DOT0" : "9882ecc5-5110-4ca6-bbb3-b6047e0bd398" , \ "OUT_V_10DOT0TO11DOT0" : "dae4c3e4-d398-4a84-8fc1-bcf602d0c604" , \ "OUT_W_10DOT0TO11DOT0" : "a56fd570-a1ec-4dcc-b87c-f7f6af474479" , \ "OUT_U_10DOT0TO10DOT5" : "0897d1f0-49c1-4ee9-a536-61f3d8a18f98" , \ "OUT_V_10DOT0TO10DOT5" : "ce6aba6b-c0d4-4701-855a-52cc4cd65795" , \ "OUT_W_10DOT0TO10DOT5" : "9a2845c0-5a1f-4ce1-89e6-0e3849d2a4bc" , \ "OUT_U_10DOT5TO11DOT0" : "5f79be58-676b-48dd-ba51-074b3069bf0a" , \ "OUT_V_10DOT5TO11DOT0" : "378a4868-08b2-4271-b862-4092e2bab025" , \ "OUT_W_10DOT5TO11DOT0" : "5304f8f5-c079-4781-8ae6-6ac4aa6968b0" , \ "OUT_U_11DOT0TO12DOT0" : "39627381-01af-4146-94bd-ed938f2c67fa" , \ "OUT_V_11DOT0TO12DOT0" : "24011bcf-fbe0-4e19-8d75-315f7aa91ec7" , \ "OUT_W_11DOT0TO12DOT0" : "078f6aa7-f939-4a21-b23c-0b5066108af2" , \ "OUT_U_11DOT0TO11DOT5" : "49203555-3b30-4516-ac98-df5252ed8cbd" , \ "OUT_V_11DOT0TO11DOT5" : "c3781d3f-82e7-45ed-990b-1a79463a1b3f" , \ "OUT_W_11DOT0TO11DOT5" : "4766d48e-652f-497f-9efb-e34d92effe14" , \ "OUT_U_11DOT5TO12DOT0" : "ff3d45eb-207f-4871-8d66-6a3daedc0ba9" , \ "OUT_V_11DOT5TO12DOT0" : "8141e8ba-ebee-44ca-b998-089d6952cf61" , \ "OUT_W_11DOT5TO12DOT0" : "79da6cb3-022c-4d9e-ac14-faeb8e4668e7" , \ "OUT_U_12DOT0TO13DOT0" : "a8aa1199-6b8e-43a6-accd-a5b16b4ef194" , \ "OUT_V_12DOT0TO13DOT0" : "43c28d6b-83fd-40ab-a5f4-560bed339c19" , \ "OUT_W_12DOT0TO13DOT0" : "93e81b2a-d808-401d-b54a-48045d1361ed" , \ "OUT_U_12DOT0TO12DOT5" : "00b802cf-ccd1-40d5-8c4e-80b38da0877f" , \ "OUT_V_12DOT0TO12DOT5" : "eee42375-aff6-47b6-a1a8-7cf39f6becdf" , \ "OUT_W_12DOT0TO12DOT5" : "30299f2d-5bca-4f58-a9b3-8b32786891dc" , \ "OUT_U_12DOT5TO13DOT0" : "7a299503-c221-4ed9-8b29-c93d8d6dc869" , \ "OUT_V_12DOT5TO13DOT0" : "b63b4388-f923-4236-bf2a-6faa81317808" , \ "OUT_W_12DOT5TO13DOT0" : "b6c0aa12-ac81-4930-9ee8-5d205e557526" , \ "OUT_U_13DOT0TO14DOT0" : "84a47068-27fc-48ca-bf9b-db473c460905" , \ "OUT_V_13DOT0TO14DOT0" : "55d0b3e0-a631-4ed0-8a17-cccf29a79639" , \ "OUT_W_13DOT0TO14DOT0" : "c8e2f739-8efb-40dd-b7b2-d9a9b3ebb0aa" , \ "OUT_U_13DOT0TO13DOT5" : "1c6039df-7c7c-474d-96f1-c0eb2c737636" , \ "OUT_V_13DOT0TO13DOT5" : "a3ba5e88-a451-49e1-87b1-e7b0dc0e9d33" , \ "OUT_W_13DOT0TO13DOT5" : "68cf9a53-1342-496d-91b0-423d69641262" , \ "OUT_U_13DOT5TO14DOT0" : "2ad1ba55-49ca-4604-b37c-d1bf7e7a48ca" , \ "OUT_V_13DOT5TO14DOT0" : "84991731-cf30-4f96-9504-0ad7b1182c43" , \ "OUT_W_13DOT5TO14DOT0" : "dd05a5cc-295c-4563-b802-c8e3c9e81d3c" , \ "OUT_U_14DOT0TO15DOT0" : "7e880d4d-ba1c-4c9a-834d-21287ba94b92" , \ "OUT_V_14DOT0TO15DOT0" : "a44e2770-1066-4f5d-9291-c222e403003c" , \ "OUT_W_14DOT0TO15DOT0" : "d072b579-432c-4ace-bdca-4f74ae4e2d66" , \ "OUT_U_14DOT0TO14DOT5" : "ddeb88c3-5c1e-4dd9-b526-e16275f89cb9" , \ "OUT_V_14DOT0TO14DOT5" : "8812aea6-42bf-4535-9512-bfb6a6e17145" , \ "OUT_W_14DOT0TO14DOT5" : "9ad8333b-ca38-48f0-8192-ffa9ae8549b1" , \ "OUT_U_14DOT5TO15DOT0" : "71011be7-6eff-4e80-97bd-aa6c3941f2c6" , \ "OUT_V_14DOT5TO15DOT0" : "ee1624d5-6857-43c6-89f3-9be893f46e49" , \ "OUT_W_14DOT5TO15DOT0" : "1982c969-0c11-4df0-b521-32a9c2c5ebf7" , \ "OUT_U_15DOT0TO16DOT0" : "d5ffd4de-7bf3-401d-9577-0869d47e3210" , \ "OUT_V_15DOT0TO16DOT0" : "ffb3bb43-d01a-491f-9fb1-2bc5560b3555" , \ "OUT_W_15DOT0TO16DOT0" : "02241f3a-682d-4040-ba51-9448ba469485" , \ "OUT_U_15DOT0TO15DOT5" : "f50b2f86-e88b-4bca-83ba-7f15c86161d8" , \ "OUT_V_15DOT0TO15DOT5" : "1c278483-564b-4f98-8bfb-c1d8f6fc6597" , \ "OUT_W_15DOT0TO15DOT5" : "3570518b-33b9-4b98-b4f1-2fed49964cd6" , \ "OUT_U_15DOT5TO16DOT0" : "c3cce42c-405f-4049-9e06-6fb0e5b65660" , \ "OUT_V_15DOT5TO16DOT0" : "b30d11dd-8def-442e-81a0-391cabd6abcd" , \ "OUT_W_15DOT5TO16DOT0" : "3c1f0afe-4507-48ba-a371-e0947243981b" , \ "OUT_U_16DOT0TO17DOT0" : "ce677953-9125-4e3d-9db7-ac917a18e0ed" , \ "OUT_V_16DOT0TO17DOT0" : "672fd031-148c-45bd-b0af-b6a37a74b9eb" , \ "OUT_W_16DOT0TO17DOT0" : "8adcba33-aca1-4ba3-a06e-34801149d381" , \ "OUT_U_16DOT0TO16DOT5" : "781b3dd5-f6ad-44aa-acd0-f2a6b3e00721" , \ "OUT_V_16DOT0TO16DOT5" : "ee168de9-d0ba-4c3e-8cdb-ca5395506d88" , \ "OUT_W_16DOT0TO16DOT5" : "01d0da09-24d3-4541-ab5a-a4f3ac631c1c" , \ "OUT_U_16DOT5TO17DOT0" : "836e0587-6195-4aaf-ae1f-79d905f81cf8" , \ "OUT_V_16DOT5TO17DOT0" : "9ae82186-8e35-4796-bb94-445bc67504bb" , \ "OUT_W_16DOT5TO17DOT0" : "ec207ba0-7b9a-4e43-ae2d-a1ea54246123" , \ "OUT_U_17DOT0TO18DOT0" : "c76484d5-61f8-404a-b6ab-bc96211603af" , \ "OUT_V_17DOT0TO18DOT0" : "e45184fb-c17f-4bf0-8f80-3b5613982a8b" , \ "OUT_W_17DOT0TO18DOT0" : "9b7ef67a-d1a6-4c8c-911e-1c77a8e6311c" , \ "OUT_U_17DOT0TO17DOT5" : "8db982e0-cb05-40d1-a0a2-7704c0e58538" , \ "OUT_V_17DOT0TO17DOT5" : "8acefb84-e950-4b64-9c23-b58a255beb53" , \ "OUT_W_17DOT0TO17DOT5" : "d1a4ce6b-dca2-4fde-a725-006cae0f992e" , \ "OUT_U_17DOT5TO18DOT0" : "52c66c2a-073c-47c8-af7d-851cb3ace0b6" , \ "OUT_V_17DOT5TO18DOT0" : "0d16bfd5-b274-4a02-8c81-c6d072b6299b" , \ "OUT_W_17DOT5TO18DOT0" : "1835c505-4639-4eb9-b520-2433881fa937" , \ "OUT_U_18DOT0TO19DOT0" : "d65a1f13-bc75-4763-b07d-4e99378df8c7" , \ "OUT_V_18DOT0TO19DOT0" : "7ceeb929-d25a-4012-92cb-f298a8939f29" , \ "OUT_W_18DOT0TO19DOT0" : "8c899d90-8079-439b-869b-3586221a9010" , \ "OUT_U_18DOT0TO18DOT5" : "6f051e00-72cc-472a-a653-c74871cced90" , \ "OUT_V_18DOT0TO18DOT5" : "b2f0d489-d923-458c-8f34-23bd84a33939" , \ "OUT_W_18DOT0TO18DOT5" : "0ac838cb-2067-4e24-9065-46f31364bd60" , \ "OUT_U_18DOT5TO19DOT0" : "0ae5326b-545a-43a8-aaf7-5fa5486237ad" , \ "OUT_V_18DOT5TO19DOT0" : "a5627227-79a7-4a35-b6b5-03ba4aef7e6a" , \ "OUT_W_18DOT5TO19DOT0" : "eb3eda85-10fd-4497-89a5-61be0ddead7d" , \ "OUT_U_19DOT0TO20DOT0" : "1616e30d-60bd-4c12-8a26-02d39e0d69ae" , \ "OUT_V_19DOT0TO20DOT0" : "7b15f03c-8297-47df-9299-0e0df22d95f7" , \ "OUT_W_19DOT0TO20DOT0" : "019ec06e-cf04-4c99-9b90-40d499ffaa7e" , \ "OUT_U_19DOT0TO19DOT5" : "4328dcfe-6745-4486-aa87-79979cb38fd2" , \ "OUT_V_19DOT0TO19DOT5" : "ec54356b-5408-499c-a13a-a0d5b5106e7a" , \ "OUT_W_19DOT0TO19DOT5" : "51fa31e0-7c6c-47ad-8efc-25315d9003b1" , \ "OUT_U_19DOT5TO20DOT0" : "25d19d6d-ed22-43d1-8b0b-58861d385d43" , \ "OUT_V_19DOT5TO20DOT0" : "4ed8bdf0-6017-4d04-b8e8-97eead5b1754" , \ "OUT_W_19DOT5TO20DOT0" : "ec8065ce-7beb-400c-bd2e-49aa0a6765f6" , \ "OUT_U_20DOT0TO21DOT0" : "35ac9f5c-4768-4743-99cb-b2dfad45fa77" , \ "OUT_V_20DOT0TO21DOT0" : "c7a11023-4a6a-464a-8d18-f8c6ac526eb4" , \ "OUT_W_20DOT0TO21DOT0" : "762c147c-fb47-4088-929f-dccb374177e7" , \ "OUT_U_20DOT0TO20DOT5" : "ffe132b5-377c-455c-96f3-68c7ecfadf21" , \ "OUT_V_20DOT0TO20DOT5" : "a44e5669-f1cc-45a9-b6ca-80abb5cd7a68" , \ "OUT_W_20DOT0TO20DOT5" : "9e3ceca8-1467-4b06-9b65-2435536905bf" , \ "OUT_U_20DOT5TO21DOT0" : "eb30b936-fc65-4d16-a0c5-d44de7525102" , \ "OUT_V_20DOT5TO21DOT0" : "bbe267bc-71b8-420a-8884-d62abcce038c" , \ "OUT_W_20DOT5TO21DOT0" : "62369b32-e7a9-45b6-9938-00660c61519b" , \ "OUT_U_21DOT0TO22DOT0" : "a3d293b6-7017-4d1b-889b-ff7398eeb4c7" , \ "OUT_V_21DOT0TO22DOT0" : "7399a022-3266-43d7-9ee1-4655530081d4" , \ "OUT_W_21DOT0TO22DOT0" : "45135a50-1400-4ef7-907b-b080738f4534" , \ "OUT_U_21DOT0TO21DOT5" : "f72d4436-cd80-43c5-ad99-c9ace2862e38" , \ "OUT_V_21DOT0TO21DOT5" : "40152a78-db1c-4602-8418-2bf37f74e8ce" , \ "OUT_W_21DOT0TO21DOT5" : "2d544ab6-d3e9-4583-99fd-26eab0564ad4" , \ "OUT_U_21DOT5TO22DOT0" : "284aaac6-4d91-45f7-99b1-44dd63f44d5f" , \ "OUT_V_21DOT5TO22DOT0" : "645a822e-839f-4480-a07e-06d5b92f3654" , \ "OUT_W_21DOT5TO22DOT0" : "ad080a4e-dc86-4c18-90ff-0b7627d0914e" , \ "OUT_U_22DOT0TO23DOT0" : "0dd674fa-2943-4e4d-ae77-86866612d285" , \ "OUT_V_22DOT0TO23DOT0" : "18aca929-8f9c-4e1d-ad44-b5cb8db8a2b6" , \ "OUT_W_22DOT0TO23DOT0" : "21632f84-1aa7-45d6-a5e5-30d73531430b" , \ "OUT_U_22DOT0TO22DOT5" : "4e8c14da-85d9-4054-b721-6cd861ada0a1" , \ "OUT_V_22DOT0TO22DOT5" : "11846e41-2d90-4d12-bf89-04dc6cc53235" , \ "OUT_W_22DOT0TO22DOT5" : "aa726cde-d38e-44d9-9c8d-6ffade84e81c" , \ "OUT_U_22DOT5TO23DOT0" : "9cd0fd71-263b-48e8-a25b-e6e1e4b9fdb8" , \ "OUT_V_22DOT5TO23DOT0" : "e7948ed9-4a58-48e1-8fa1-cd28ac8b4138" , \ "OUT_W_22DOT5TO23DOT0" : "3e0dec16-decf-4b30-b9d5-abe1b10e0d78" , \ "OUT_U_23DOT0TO24DOT0" : "34b8b11f-0d5f-4a04-be7d-749500ee92e2" , \ "OUT_V_23DOT0TO24DOT0" : "97346dc3-258b-4751-bcba-5261720abb42" , \ "OUT_W_23DOT0TO24DOT0" : "c1f3434e-6f5d-4fae-8f50-e520a9710cd4" , \ "OUT_U_23DOT0TO23DOT5" : "5b10c5db-8c22-42a2-b5b2-3c85bdfd878d" , \ "OUT_V_23DOT0TO23DOT5" : "edd3c391-6415-40f6-a2b9-7bdccc3ad7c9" , \ "OUT_W_23DOT0TO23DOT5" : "95cfb82d-19c8-4f34-a3fb-e6a34fcc6115" , \ "OUT_U_23DOT5TO24DOT0" : "ffb8e085-3e79-427a-890c-9db499765672" , \ "OUT_V_23DOT5TO24DOT0" : "b33ebc35-9db7-441e-93b2-585e293237a7" , \ "OUT_W_23DOT5TO24DOT0" : "c0ec9b23-c038-48c6-a812-8ddaaf115880" , \ "OUT_U_24DOT0TO25DOT0" : "eb4bce3c-93e7-4a56-90ce-9d0eb55e3cc3" , \ "OUT_V_24DOT0TO25DOT0" : "f2763f31-031e-4ada-a5d1-75d956ef3b4e" , \ "OUT_W_24DOT0TO25DOT0" : "f5db6cd0-66d0-4232-860f-bb83e32291bb" , \ "OUT_U_24DOT0TO24DOT5" : "43af33ae-e914-435b-9e72-4d4c71e9c7dc" , \ "OUT_V_24DOT0TO24DOT5" : "ba9bcbc9-2ead-49b0-9dea-f6ba6d179a58" , \ "OUT_W_24DOT0TO24DOT5" : "e4c62288-6db1-4fa6-8a1a-a1fe97ba8775" , \ "OUT_U_24DOT5TO25DOT0" : "03cda718-804c-4383-9f66-79b27c97a342" , \ "OUT_V_24DOT5TO25DOT0" : "91208cec-4ba5-41ce-8faa-e4fb98ed237c" , \ "OUT_W_24DOT5TO25DOT0" : "a09776d7-1240-43ac-8fe0-e82af0f1f81a" \ }; # Note: manualAbstractionInformation, generally speaking, is a # structured used purely in analysis scripts (as developed for # the paper describing Fanoos); placing this information # in the class defining the domain proved to be a convieniant place to store the # information during the time of development and testing. Fanoos does not access # the information in manualAbstractionInformation when determining how to make # adjustments to respond to users. Again, it is only used in analysis scripts # used to prepare results for the paper. While this sanity-checking # code does not have results discussed in the paper at the time of # writting this comment, we needed to fill the information for # this structure; while Fanoos itself does not examin content in # manualAbstractionInformation, some code (such as checking code, e.g., contracts) # expect the structure to be present and obey basic properties such as # number of entries. # # While it was convieniant for development, clearly it is not # ideal have this data stored here or this structure required # to be present. TODO: resolve the issue just described. self.manualAbstractionInformation = {\ "predicatesAndLabels" : [\ ("IN_X_NEG20DOT0TONEG19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG20DOT0TONEG19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG20DOT0TONEG19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG20DOT0TONEG19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG19DOT5TONEG19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG19DOT5TONEG19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG19DOT0TONEG18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG19DOT0TONEG18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG19DOT0TONEG18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG19DOT0TONEG18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG18DOT5TONEG18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG18DOT5TONEG18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG18DOT0TONEG17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG18DOT0TONEG17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG18DOT0TONEG17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG18DOT0TONEG17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG17DOT5TONEG17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG17DOT5TONEG17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG17DOT0TONEG16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG17DOT0TONEG16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG17DOT0TONEG16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG17DOT0TONEG16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG16DOT5TONEG16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG16DOT5TONEG16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG16DOT0TONEG15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG16DOT0TONEG15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG16DOT0TONEG15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG16DOT0TONEG15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG15DOT5TONEG15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG15DOT5TONEG15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG15DOT0TONEG14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG15DOT0TONEG14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG15DOT0TONEG14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG15DOT0TONEG14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG14DOT5TONEG14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG14DOT5TONEG14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG14DOT0TONEG13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG14DOT0TONEG13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG14DOT0TONEG13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG14DOT0TONEG13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG13DOT5TONEG13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG13DOT5TONEG13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG13DOT0TONEG12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG13DOT0TONEG12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG13DOT0TONEG12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG13DOT0TONEG12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG12DOT5TONEG12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG12DOT5TONEG12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG12DOT0TONEG11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG12DOT0TONEG11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG12DOT0TONEG11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG12DOT0TONEG11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG11DOT5TONEG11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG11DOT5TONEG11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG11DOT0TONEG10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG11DOT0TONEG10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG11DOT0TONEG10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG11DOT0TONEG10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG10DOT5TONEG10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG10DOT5TONEG10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG10DOT0TONEG9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG10DOT0TONEG9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG10DOT0TONEG9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG10DOT0TONEG9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG9DOT5TONEG9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG9DOT5TONEG9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG9DOT0TONEG8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG9DOT0TONEG8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG9DOT0TONEG8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG9DOT0TONEG8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG8DOT5TONEG8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG8DOT5TONEG8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG8DOT0TONEG7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG8DOT0TONEG7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG8DOT0TONEG7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG8DOT0TONEG7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG7DOT5TONEG7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG7DOT5TONEG7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG7DOT0TONEG6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG7DOT0TONEG6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG7DOT0TONEG6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG7DOT0TONEG6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG6DOT5TONEG6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG6DOT5TONEG6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG6DOT0TONEG5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG6DOT0TONEG5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG6DOT0TONEG5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG6DOT0TONEG5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG5DOT5TONEG5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG5DOT5TONEG5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG5DOT0TONEG4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG5DOT0TONEG4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG5DOT0TONEG4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG5DOT0TONEG4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG4DOT5TONEG4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG4DOT5TONEG4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG4DOT0TONEG3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG4DOT0TONEG3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG4DOT0TONEG3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG4DOT0TONEG3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG3DOT5TONEG3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG3DOT5TONEG3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG3DOT0TONEG2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG3DOT0TONEG2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG3DOT0TONEG2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG3DOT0TONEG2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG2DOT5TONEG2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG2DOT5TONEG2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG2DOT0TONEG1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG2DOT0TONEG1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG2DOT0TONEG1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG2DOT0TONEG1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG1DOT5TONEG1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG1DOT5TONEG1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG1DOT0TO0DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_NEG1DOT0TO0DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_NEG1DOT0TONEG0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG1DOT0TONEG0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_NEG0DOT5TO0DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_NEG0DOT5TO0DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_0DOT0TO1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_0DOT0TO1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_0DOT0TO0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_0DOT0TO0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_0DOT5TO1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_0DOT5TO1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_1DOT0TO2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_1DOT0TO2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_1DOT0TO1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_1DOT0TO1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_1DOT5TO2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_1DOT5TO2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_2DOT0TO3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_2DOT0TO3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_2DOT0TO2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_2DOT0TO2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_2DOT5TO3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_2DOT5TO3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_3DOT0TO4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_3DOT0TO4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_3DOT0TO3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_3DOT0TO3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_3DOT5TO4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_3DOT5TO4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_4DOT0TO5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_4DOT0TO5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_4DOT0TO4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_4DOT0TO4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_4DOT5TO5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_4DOT5TO5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_5DOT0TO6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_5DOT0TO6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_5DOT0TO5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_5DOT0TO5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_5DOT5TO6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_5DOT5TO6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_6DOT0TO7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_6DOT0TO7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_6DOT0TO6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_6DOT0TO6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_6DOT5TO7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_6DOT5TO7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_7DOT0TO8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_7DOT0TO8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_7DOT0TO7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_7DOT0TO7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_7DOT5TO8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_7DOT5TO8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_8DOT0TO9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_8DOT0TO9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_8DOT0TO8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_8DOT0TO8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_8DOT5TO9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_8DOT5TO9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_9DOT0TO10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_9DOT0TO10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_9DOT0TO9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_9DOT0TO9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_9DOT5TO10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_9DOT5TO10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_10DOT0TO11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_10DOT0TO11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_10DOT0TO10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_10DOT0TO10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_10DOT5TO11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_10DOT5TO11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_11DOT0TO12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_11DOT0TO12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_11DOT0TO11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_11DOT0TO11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_11DOT5TO12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_11DOT5TO12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_12DOT0TO13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_12DOT0TO13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_12DOT0TO12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_12DOT0TO12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_12DOT5TO13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_12DOT5TO13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_13DOT0TO14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_13DOT0TO14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_13DOT0TO13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_13DOT0TO13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_13DOT5TO14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_13DOT5TO14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_14DOT0TO15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_14DOT0TO15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_14DOT0TO14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_14DOT0TO14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_14DOT5TO15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_14DOT5TO15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_15DOT0TO16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_15DOT0TO16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_15DOT0TO15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_15DOT0TO15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_15DOT5TO16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_15DOT5TO16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_16DOT0TO17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_16DOT0TO17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_16DOT0TO16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_16DOT0TO16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_16DOT5TO17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_16DOT5TO17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_17DOT0TO18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_17DOT0TO18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_17DOT0TO17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_17DOT0TO17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_17DOT5TO18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_17DOT5TO18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_18DOT0TO19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_18DOT0TO19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_18DOT0TO18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_18DOT0TO18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_18DOT5TO19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_18DOT5TO19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_19DOT0TO20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_Y_19DOT0TO20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("IN_X_19DOT0TO19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_19DOT0TO19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_X_19DOT5TO20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("IN_Y_19DOT5TO20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG25DOT0TONEG24DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG25DOT0TONEG24DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG25DOT0TONEG24DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG25DOT0TONEG24DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG25DOT0TONEG24DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG25DOT0TONEG24DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG24DOT5TONEG24DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG24DOT5TONEG24DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG24DOT5TONEG24DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG24DOT0TONEG23DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG24DOT0TONEG23DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG24DOT0TONEG23DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG24DOT0TONEG23DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG24DOT0TONEG23DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG24DOT0TONEG23DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG23DOT5TONEG23DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG23DOT5TONEG23DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG23DOT5TONEG23DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG23DOT0TONEG22DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG23DOT0TONEG22DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG23DOT0TONEG22DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG23DOT0TONEG22DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG23DOT0TONEG22DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG23DOT0TONEG22DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG22DOT5TONEG22DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG22DOT5TONEG22DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG22DOT5TONEG22DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG22DOT0TONEG21DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG22DOT0TONEG21DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG22DOT0TONEG21DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG22DOT0TONEG21DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG22DOT0TONEG21DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG22DOT0TONEG21DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG21DOT5TONEG21DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG21DOT5TONEG21DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG21DOT5TONEG21DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG21DOT0TONEG20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG21DOT0TONEG20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG21DOT0TONEG20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG21DOT0TONEG20DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG21DOT0TONEG20DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG21DOT0TONEG20DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG20DOT5TONEG20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG20DOT5TONEG20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG20DOT5TONEG20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG20DOT0TONEG19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG20DOT0TONEG19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG20DOT0TONEG19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG20DOT0TONEG19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG20DOT0TONEG19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG20DOT0TONEG19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG19DOT5TONEG19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG19DOT5TONEG19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG19DOT5TONEG19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG19DOT0TONEG18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG19DOT0TONEG18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG19DOT0TONEG18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG19DOT0TONEG18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG19DOT0TONEG18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG19DOT0TONEG18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG18DOT5TONEG18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG18DOT5TONEG18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG18DOT5TONEG18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG18DOT0TONEG17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG18DOT0TONEG17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG18DOT0TONEG17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG18DOT0TONEG17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG18DOT0TONEG17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG18DOT0TONEG17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG17DOT5TONEG17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG17DOT5TONEG17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG17DOT5TONEG17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG17DOT0TONEG16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG17DOT0TONEG16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG17DOT0TONEG16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG17DOT0TONEG16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG17DOT0TONEG16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG17DOT0TONEG16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG16DOT5TONEG16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG16DOT5TONEG16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG16DOT5TONEG16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG16DOT0TONEG15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG16DOT0TONEG15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG16DOT0TONEG15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG16DOT0TONEG15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG16DOT0TONEG15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG16DOT0TONEG15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG15DOT5TONEG15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG15DOT5TONEG15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG15DOT5TONEG15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG15DOT0TONEG14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG15DOT0TONEG14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG15DOT0TONEG14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG15DOT0TONEG14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG15DOT0TONEG14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG15DOT0TONEG14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG14DOT5TONEG14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG14DOT5TONEG14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG14DOT5TONEG14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG14DOT0TONEG13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG14DOT0TONEG13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG14DOT0TONEG13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG14DOT0TONEG13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG14DOT0TONEG13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG14DOT0TONEG13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG13DOT5TONEG13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG13DOT5TONEG13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG13DOT5TONEG13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG13DOT0TONEG12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG13DOT0TONEG12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG13DOT0TONEG12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG13DOT0TONEG12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG13DOT0TONEG12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG13DOT0TONEG12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG12DOT5TONEG12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG12DOT5TONEG12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG12DOT5TONEG12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG12DOT0TONEG11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG12DOT0TONEG11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG12DOT0TONEG11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG12DOT0TONEG11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG12DOT0TONEG11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG12DOT0TONEG11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG11DOT5TONEG11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG11DOT5TONEG11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG11DOT5TONEG11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG11DOT0TONEG10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG11DOT0TONEG10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG11DOT0TONEG10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG11DOT0TONEG10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG11DOT0TONEG10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG11DOT0TONEG10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG10DOT5TONEG10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG10DOT5TONEG10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG10DOT5TONEG10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG10DOT0TONEG9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG10DOT0TONEG9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG10DOT0TONEG9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG10DOT0TONEG9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG10DOT0TONEG9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG10DOT0TONEG9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG9DOT5TONEG9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG9DOT5TONEG9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG9DOT5TONEG9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG9DOT0TONEG8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG9DOT0TONEG8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG9DOT0TONEG8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG9DOT0TONEG8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG9DOT0TONEG8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG9DOT0TONEG8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG8DOT5TONEG8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG8DOT5TONEG8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG8DOT5TONEG8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG8DOT0TONEG7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG8DOT0TONEG7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG8DOT0TONEG7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG8DOT0TONEG7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG8DOT0TONEG7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG8DOT0TONEG7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG7DOT5TONEG7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG7DOT5TONEG7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG7DOT5TONEG7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG7DOT0TONEG6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG7DOT0TONEG6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG7DOT0TONEG6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG7DOT0TONEG6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG7DOT0TONEG6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG7DOT0TONEG6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG6DOT5TONEG6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG6DOT5TONEG6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG6DOT5TONEG6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG6DOT0TONEG5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG6DOT0TONEG5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG6DOT0TONEG5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG6DOT0TONEG5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG6DOT0TONEG5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG6DOT0TONEG5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG5DOT5TONEG5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG5DOT5TONEG5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG5DOT5TONEG5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG5DOT0TONEG4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG5DOT0TONEG4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG5DOT0TONEG4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG5DOT0TONEG4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG5DOT0TONEG4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG5DOT0TONEG4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG4DOT5TONEG4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG4DOT5TONEG4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG4DOT5TONEG4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG4DOT0TONEG3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG4DOT0TONEG3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG4DOT0TONEG3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG4DOT0TONEG3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG4DOT0TONEG3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG4DOT0TONEG3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG3DOT5TONEG3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG3DOT5TONEG3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG3DOT5TONEG3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG3DOT0TONEG2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG3DOT0TONEG2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG3DOT0TONEG2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG3DOT0TONEG2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG3DOT0TONEG2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG3DOT0TONEG2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG2DOT5TONEG2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG2DOT5TONEG2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG2DOT5TONEG2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG2DOT0TONEG1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG2DOT0TONEG1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG2DOT0TONEG1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG2DOT0TONEG1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG2DOT0TONEG1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG2DOT0TONEG1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG1DOT5TONEG1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG1DOT5TONEG1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG1DOT5TONEG1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG1DOT0TO0DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_NEG1DOT0TO0DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_NEG1DOT0TO0DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_NEG1DOT0TONEG0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG1DOT0TONEG0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG1DOT0TONEG0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_NEG0DOT5TO0DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_NEG0DOT5TO0DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_NEG0DOT5TO0DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_0DOT0TO1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_0DOT0TO1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_0DOT0TO1DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_0DOT0TO0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_0DOT0TO0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_0DOT0TO0DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_0DOT5TO1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_0DOT5TO1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_0DOT5TO1DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_1DOT0TO2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_1DOT0TO2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_1DOT0TO2DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_1DOT0TO1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_1DOT0TO1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_1DOT0TO1DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_1DOT5TO2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_1DOT5TO2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_1DOT5TO2DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_2DOT0TO3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_2DOT0TO3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_2DOT0TO3DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_2DOT0TO2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_2DOT0TO2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_2DOT0TO2DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_2DOT5TO3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_2DOT5TO3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_2DOT5TO3DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_3DOT0TO4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_3DOT0TO4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_3DOT0TO4DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_3DOT0TO3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_3DOT0TO3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_3DOT0TO3DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_3DOT5TO4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_3DOT5TO4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_3DOT5TO4DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_4DOT0TO5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_4DOT0TO5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_4DOT0TO5DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_4DOT0TO4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_4DOT0TO4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_4DOT0TO4DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_4DOT5TO5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_4DOT5TO5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_4DOT5TO5DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_5DOT0TO6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_5DOT0TO6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_5DOT0TO6DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_5DOT0TO5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_5DOT0TO5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_5DOT0TO5DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_5DOT5TO6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_5DOT5TO6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_5DOT5TO6DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_6DOT0TO7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_6DOT0TO7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_6DOT0TO7DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_6DOT0TO6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_6DOT0TO6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_6DOT0TO6DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_6DOT5TO7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_6DOT5TO7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_6DOT5TO7DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_7DOT0TO8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_7DOT0TO8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_7DOT0TO8DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_7DOT0TO7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_7DOT0TO7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_7DOT0TO7DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_7DOT5TO8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_7DOT5TO8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_7DOT5TO8DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_8DOT0TO9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_8DOT0TO9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_8DOT0TO9DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_8DOT0TO8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_8DOT0TO8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_8DOT0TO8DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_8DOT5TO9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_8DOT5TO9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_8DOT5TO9DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_9DOT0TO10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_9DOT0TO10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_9DOT0TO10DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_9DOT0TO9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_9DOT0TO9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_9DOT0TO9DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_9DOT5TO10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_9DOT5TO10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_9DOT5TO10DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_10DOT0TO11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_10DOT0TO11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_10DOT0TO11DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_10DOT0TO10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_10DOT0TO10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_10DOT0TO10DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_10DOT5TO11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_10DOT5TO11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_10DOT5TO11DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_11DOT0TO12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_11DOT0TO12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_11DOT0TO12DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_11DOT0TO11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_11DOT0TO11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_11DOT0TO11DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_11DOT5TO12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_11DOT5TO12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_11DOT5TO12DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_12DOT0TO13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_12DOT0TO13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_12DOT0TO13DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_12DOT0TO12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_12DOT0TO12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_12DOT0TO12DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_12DOT5TO13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_12DOT5TO13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_12DOT5TO13DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_13DOT0TO14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_13DOT0TO14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_13DOT0TO14DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_13DOT0TO13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_13DOT0TO13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_13DOT0TO13DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_13DOT5TO14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_13DOT5TO14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_13DOT5TO14DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_14DOT0TO15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_14DOT0TO15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_14DOT0TO15DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_14DOT0TO14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_14DOT0TO14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_14DOT0TO14DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_14DOT5TO15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_14DOT5TO15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_14DOT5TO15DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_15DOT0TO16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_15DOT0TO16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_15DOT0TO16DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_15DOT0TO15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_15DOT0TO15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_15DOT0TO15DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_15DOT5TO16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_15DOT5TO16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_15DOT5TO16DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_16DOT0TO17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_16DOT0TO17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_16DOT0TO17DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_16DOT0TO16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_16DOT0TO16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_16DOT0TO16DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_16DOT5TO17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_16DOT5TO17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_16DOT5TO17DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_17DOT0TO18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_17DOT0TO18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_17DOT0TO18DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_17DOT0TO17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_17DOT0TO17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_17DOT0TO17DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_17DOT5TO18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_17DOT5TO18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_17DOT5TO18DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_18DOT0TO19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_18DOT0TO19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_18DOT0TO19DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_18DOT0TO18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_18DOT0TO18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_18DOT0TO18DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_18DOT5TO19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_18DOT5TO19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_18DOT5TO19DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_19DOT0TO20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_19DOT0TO20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_19DOT0TO20DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_19DOT0TO19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_19DOT0TO19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_19DOT0TO19DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_19DOT5TO20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_19DOT5TO20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_19DOT5TO20DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_20DOT0TO21DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_20DOT0TO21DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_20DOT0TO21DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_20DOT0TO20DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_20DOT0TO20DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_20DOT0TO20DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_20DOT5TO21DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_20DOT5TO21DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_20DOT5TO21DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_21DOT0TO22DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_21DOT0TO22DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_21DOT0TO22DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_21DOT0TO21DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_21DOT0TO21DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_21DOT0TO21DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_21DOT5TO22DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_21DOT5TO22DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_21DOT5TO22DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_22DOT0TO23DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_22DOT0TO23DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_22DOT0TO23DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_22DOT0TO22DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_22DOT0TO22DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_22DOT0TO22DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_22DOT5TO23DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_22DOT5TO23DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_22DOT5TO23DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_23DOT0TO24DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_23DOT0TO24DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_23DOT0TO24DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_23DOT0TO23DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_23DOT0TO23DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_23DOT0TO23DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_23DOT5TO24DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_23DOT5TO24DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_23DOT5TO24DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_24DOT0TO25DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_V_24DOT0TO25DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_W_24DOT0TO25DOT0" , "4609e380-322b-4a66-b5b6-b0e5cf0dd820"), \ ("OUT_U_24DOT0TO24DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_24DOT0TO24DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_24DOT0TO24DOT5" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_U_24DOT5TO25DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_V_24DOT5TO25DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0"), \ ("OUT_W_24DOT5TO25DOT0" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0") \ ], \ "labelDag_firstParent_secondChild" : [ \ ("4609e380-322b-4a66-b5b6-b0e5cf0dd820" , "859a21b5-d37d-4a64-9dd3-779b1c14f9c0") \ ] \ }; self.manualAbstractionInformation["predicatesAndLabels"] = \ [ (dictMappingPredicateStringNameToUUID[x[0]] , x[1]) for x in self.manualAbstractionInformation["predicatesAndLabels"]]; functToGetUuidProvided = (lambda predicateObjectBeingInitialized : dictMappingPredicateStringNameToUUID[str(predicateObjectBeingInitialized)] ); self.initializedConditions = \ [CharacterizationCondition_FromPythonFunction(z3SolverInstance, DomainFor_modelForTesting_twoDimInput_threeDimOutput, x, functToGetUuidProvided=functToGetUuidProvided) \ for x in getListFunctionsToBaseCondtionsOn_forInputOfDomainThisUse() + \ getListFunctionsToBaseCondtionsOn_forOutputOfDomainThisUse() + \ getListFunctionsToBaseCondtionsOn_forJointInputAndOutputDomainsInThisUse() ]; assert(all([ (x.getID() == functToGetUuidProvided(x)) for x in self.initializedConditions])); self._writeInfoToDatabase(); return; def getBaseConditions(self): return self.initializedConditions; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # class-specific utilities for defining domains #=========================================================================== def getFiniteInterval(variableType, nameOfPredicate, lowerBound, upperBound): requires(isinstance(nameOfPredicate, str)); requires(variableType in {"input", "output"}); requires(isinstance(nameOfPredicate, str)); requires(len(nameOfPredicate) > 0); requires(len(set(nameOfPredicate).intersection([" ", "\n", "\r", "\t"])) == 0); requires(isinstance(lowerBound, float)); requires(isinstance(upperBound, float)); requires(np.isfinite(lowerBound)); requires(np.isfinite(upperBound)); requires(lowerBound <= upperBound); templateString = """ def funct_{0}({1}): \"\"\"{0}\"\"\" if(isinstance({1}, z3.z3.ArithRef)): return z3.And( {1} <= {3}, {1} >= {2} ); else: return ({1} <= {3}) and ({1} >= {2} ); raise Exception("Control should not reach here"); return; """; listOfVariableNames = ["in_x", "in_y"]; if(variableType == "output"): listOfVariableNames = ["out_u", "out_v", "out_w"]; assert(listOfVariableNames in [["in_x", "in_y"], ["out_u", "out_v", "out_w"]]); return [ \ templateString.format(\ (variableNameString.upper() + "_" + nameOfPredicate), \ variableNameString, str(lowerBound), str(upperBound) ) \ for variableNameString in listOfVariableNames ]; # The below function in principle could be done with getFiniteInterval if z3 supported infinite values, but its standard theory does not seem to support # them, which, honestly, is reasonable. def getInfiniteInterval(variableType, nameOfPredicate, boundary, aboveOrBelow): requires(isinstance(nameOfPredicate, str)); requires(variableType in {"input", "output"}); requires(isinstance(nameOfPredicate, str)); requires(len(nameOfPredicate) > 0); requires(len(set(nameOfPredicate).intersection([" ", "\n", "\r", "\t"])) == 0); requires(isinstance(boundary, float)); requires(np.isfinite(boundary)); requires(aboveOrBelow in {"lowerBound", "upperBound"}); templateString = """ def funct_{0}({1}): \"\"\"{0}\"\"\" if(isinstance({1}, z3.z3.ArithRef)): return {2} <= {3}; else: return {2} <= {3}; raise Exception("Control should not reach here"); return; """; variableNameString = "in_x"; if(variableType == "output"): variableNameString = "out_y"; assert(variableNameString in {"in_x", "out_y"}); raise Exception("TODO: update this"); stringToReturn = ""; if(aboveOrBelow == "upperBound"): stringToReturn = templateString.format(nameOfPredicate, variableNameString, variableNameString, str(boundry) ) else: assert(aboveOrBelow == "lowerBound"); stringToReturn = templateString.format(nameOfPredicate, variableNameString, str(boundry), variableNameString) return stringToReturn ; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the input domain #=========================================================================== otherInputSpaceFunctionsToUse = """ """ def getListFunctionsToBaseCondtionsOn_forInputOfDomainThisUse(): listOfFunctionCodes =[]; def boundsToName(lower, upper): A = str(lower).replace(".", "DOT").replace("-", "NEG"); B = str(upper).replace(".", "DOT").replace("-", "NEG"); return A + "TO" + B; def formPredicateHere(lower, upper): return getFiniteInterval("input", boundsToName(lower, upper), lower, upper); for thisStartIndex in range(-20, 20): thisStartIndex = float(thisStartIndex); upperIndex = thisStartIndex + 1.0; middleIndex = thisStartIndex + 0.5; # nameOfPredicate, lowerBound, upperBound) listOfFunctionCodes = listOfFunctionCodes + \ formPredicateHere(thisStartIndex, upperIndex) + \ formPredicateHere(thisStartIndex, middleIndex) + \ formPredicateHere(middleIndex, upperIndex); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the output domain #=========================================================================== otherOutputSpaceFunctionsToUse = """ """; def getListFunctionsToBaseCondtionsOn_forOutputOfDomainThisUse(): listOfFunctionCodes =[]; def boundsToName(lower, upper): A = str(lower).replace(".", "DOT").replace("-", "NEG"); B = str(upper).replace(".", "DOT").replace("-", "NEG"); return A + "TO" + B; def formPredicateHere(lower, upper): return getFiniteInterval("output", boundsToName(lower, upper), lower, upper); # We have 25 below as oppossed to 20 so that it is distinct from the # [-20,20] range used for the input space, allowing for further testing of # expected behaviour, etc. for thisStartIndex in range(-25, 25): thisStartIndex = float(thisStartIndex); upperIndex = thisStartIndex + 1.0; middleIndex = thisStartIndex + 0.5; # nameOfPredicate, lowerBound, upperBound) listOfFunctionCodes = listOfFunctionCodes + \ formPredicateHere(thisStartIndex, upperIndex) + \ formPredicateHere(thisStartIndex, middleIndex) + \ formPredicateHere(middleIndex, upperIndex); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the joint domain #=========================================================================== def getBox(nameOfPredicate, lowerBoundInput, upperBoundInput, lowerBoundOutput, upperBoundOutput): requires(isinstance(lowerBoundInput, float)); requires(isinstance(upperBoundInput, float)); requires(isinstance(lowerBoundOutput, float)); requires(isinstance(upperBoundOutput, float)); requires(np.isfinite(lowerBoundInput)); requires(np.isfinite(upperBoundInput)); requires(np.isfinite(lowerBoundOutput)); requires(np.isfinite(upperBoundOutput)); requires(lowerBoundInput <= upperBoundInput); requires(lowerBoundOutput <= upperBoundOutput); templateString = """ def funct_{0}(in_x, out_y): \"\"\"{0}\"\"\" if(isinstance(in_x, z3.z3.ArithRef)): return z3.And( in_x <= {3}, in_x >= {2}, out_y <= {5}, out_y >= {4} ); else: return (in_x <= {3}) and (in_x >= {2} ) and (out_y <= {5}) and (out_y >= {4}); raise Exception("Control should not reach here"); return; """; return templateString.format(nameOfPredicate, str(lowerBoundInput), str(upperBoundInput), str(lowerBoundOutput), str(upperBoundOutput) ); # circle, # halfplane # negation (or maybe just allow the user to pass in the inequality.. but actually negation would be useful for things later on... ) def getHalfPlane(nameOfPredicate, slope, intercept, inequality): requires(isinstance(slope, float)); requires(isinstance(intercept, float)); requires(np.isfinite(intercept)); requires(np.isfinite(slope)); requires(isinstance(inequality, str)); requires(inequality in {"=<", "=>", "<", ">"}); templateString = """ def funct_{0}(in_x, out_y): \"\"\"{0}\"\"\" return in_x * {1} + {2} {3} out_y ; raise Exception("Control should not reach here"); return; """; return templateString.format(nameOfPredicate, str(slope), str(intercept), str(inequality)); def getCicle(nameOfPredicate, in_x_center, out_y_center, radius, inequality): requires(isinstance(in_x_center, float)); requires(np.isfinite(in_x_center)); requires(isinstance(out_y_center, float)); requires(np.isfinite(out_y_center)); requires(isinstance(radius, float)); requires(np.isfinite(radius)); requires(isinstance(inequality, str)); requires(inequality in {"=<", "=>", "<", ">"}); templateString = """ def funct_{0}(in_x, out_y): \"\"\"{0}\"\"\" return (in_x - {1}) ** 2 + (out_y - {2}) {3} {4} ; raise Exception("Control should not reach here"); return; """; return templateString.format(nameOfPredicate, str(in_x_center), str(out_y_center), str(inequality), str(radius ** 2) ); def getListFunctionsToBaseCondtionsOn_forJointInputAndOutputDomainsInThisUse(): listOfFunctionCodes =[]; return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^
125,482
69.338004
181
py
Fanoos
Fanoos-master/domainsAndConditions/baseClassDomainInformation.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from boxesAndBoxOperations.getBox import isProperBox, getDimensionOfBox; from utils.contracts import *; import numpy as np; import z3; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface , executeDatabaseCommandList; class BaseClassDomainInformation(): def __init__(self): return; def _writeInfoToDatabase(self): commandList = []; # Below, I use IGNORE to avoid voliating UNIQUE constraints in the database for values # that should only appear once.... commandList.append(\ "INSERT OR IGNORE INTO domainInfo (domainUUID, typeString) VALUES ('" + \ self.getUUID() +"', '" + str(self.__class__.__name__) + "');" ); for thisPredicate in self.initializedConditions: commandList.append(\ "INSERT OR IGNORE INTO domain_predicate_relation (domainUUID, predicateUUID) VALUES ('" + \ self.getUUID() +"' , '" + thisPredicate.getID() + "');" ); commandList.append(\ "INSERT OR IGNORE INTO predicateInfo (predicateUUID, stringName, typeString) VALUES ('" + \ thisPredicate.getID() +"' , '" + str(thisPredicate) + "', '" + str(thisPredicate.__class__.__name__) + "');" ); executeDatabaseCommandList(commandList); #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # recording abstractness of predicates for various uses.... #======================================================================= def firstElems(thisListOfTuples): return set([x[0] for x in thisListOfTuples]); def secondElems(thisListOfTuples): return set([x[1] for x in thisListOfTuples]); assert(\ firstElems(self.manualAbstractionInformation["predicatesAndLabels"]) == \ set([x.getID() for x in self.initializedConditions]) \ ); for thisCheckIndex in [0, 1]: assert(\ secondElems(self.manualAbstractionInformation["predicatesAndLabels"]).issuperset( \ [x[thisCheckIndex] for x in self.manualAbstractionInformation["labelDag_firstParent_secondChild"]] \ ) \ ); #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # checking that self.manualAbstractionInformation["labelDag_keyParent_valueChild"] forms a proper DAG (e.g., no cycleS) #========================================================================= reachSet = set([x[0] for x in self.manualAbstractionInformation["labelDag_firstParent_secondChild"]]); newReachSet = set(); # Why len(set(self.manualAbstractionInformation["predicatesAndLabels"].values())) ? Because, by the prior asserts, # its an upper bound on the number of distinct labels listed anywhere in labelDag_firstParent_secondChild ... # Basically, if this is a DAG on n nodes, then there should be not path of length n + 1 (counting number of nodes # visited) - that could only occur if a cycle is present.... for thisIteration in range(0, len(secondElems(self.manualAbstractionInformation["predicatesAndLabels"])) ): newReachSet.update(\ [x[1] for x in self.manualAbstractionInformation["labelDag_firstParent_secondChild"] \ if x[0] in reachSet] \ ); reachSet = newReachSet; newReachSet = set(); if(len(reachSet) > 0): raise Exception("labelDag_firstParent_secondChild contains a cycle and thus is not a DAG."); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ commandList = []; # Below, I use IGNORE to avoid voliating UNIQUE constraints in the database for values # that should only appear once.... for thisPredicateUUIDAndLabelUUID in self.manualAbstractionInformation["predicatesAndLabels"]: commandList.append(\ "INSERT OR IGNORE INTO predicate_label_relation (predicateUUID, labelUUID) VALUES ('" + \ thisPredicateUUIDAndLabelUUID[0] +"' , '" + \ thisPredicateUUIDAndLabelUUID[1] + "');" ); for thisLabelEdge in self.manualAbstractionInformation["labelDag_firstParent_secondChild"]: commandList.append(\ "INSERT OR IGNORE INTO labelDAG (parentLabelUUID , childLabelUUID) VALUES ('" + \ thisLabelEdge[0] + "' , '" + thisLabelEdge[1] + "');" ); if(len(commandList) > 0): executeDatabaseCommandList(commandList); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ return; @staticmethod def _helper_getInputSpaceUniverseBox(orderOfVariables, dictMappingVariableToBound): requires(isinstance(orderOfVariables, list)); requires(isinstance(dictMappingVariableToBound, dict)); requires(set(dictMappingVariableToBound.keys()) == set([str(x) for x in orderOfVariables])); requires(all([isinstance(x, list) for x in dictMappingVariableToBound.values()])); requires(all([len(x) ==2 for x in dictMappingVariableToBound.values()])); requires(all([all([isinstance(y, float) or isinstance(y, int) for y in x]) \ for x in dictMappingVariableToBound.values()])); requires(all([(x[0] <= x[1]) for x in dictMappingVariableToBound.values()])); requires(len(set(orderOfVariables)) == len(orderOfVariables)); # i.e., the entires in orderOfVariables are unique... thisUniverseBox = np.array([dictMappingVariableToBound[str(x)] for x in orderOfVariables]); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ assert(isProperBox(thisUniverseBox)); ensures(getDimensionOfBox(thisUniverseBox) == len(orderOfVariables)); return thisUniverseBox; @staticmethod def getInputSpaceUniverseBox(): raise NotImplementedError(); # child classes need to override # NOTE: the output space bounds are determined by the learner / policy - # that is not to say the bounds on, say, actions are not influenced by the situation, # but instead that multiple different agents/ learners might be in the same # environment / situation and wish to enforce different output bounds. # This all noted, since we primarly care about the learner input-output relations # that may be non-physical, it is not clear that a priori bounding outputs is # beneficial. We have to do it for the input space for a variety of reasons, # the most practical explanation being that our process works by partitioning # (modulo boundaries of abstract states) the input space as necessary, and thus # must know the geometry of the space to grid.... @staticmethod def inputSpaceVariables(): raise NotImplementedError(); # child classes need to override @staticmethod def outputSpaceVariables(): raise NotImplementedError(); # child classes need to override @staticmethod def getUUID(): raise NotImplementedError(); # child classes need to override @staticmethod def getName(): raise NotImplementedError(); # child classes need to override def initialize_baseConditions(self, z3SolverInstance): raise NotImplementedError(); # child classes need to override def getBaseConditions(self): raise NotImplementedError(); # child classes need to override
11,569
61.880435
2,890
py
Fanoos
Fanoos-master/domainsAndConditions/utilsForDefiningPredicates.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import re; import z3; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass, CharacterizationCondition_FromPythonFunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; def xor(boolA, boolB): return (boolA or boolB) and (not (boolA and boolB)); def z3Abs(x): return z3.If(x >= 0,x,-x); def z3Sign(x): return z3.If(x > 0,1,0) + z3.If(x < 0,-1,0); def convertCodeListToListOfFunctions(listOfFunctionsToCreate): requires(isinstance(listOfFunctionsToCreate, list)); requires(all([isinstance(x, str) for x in listOfFunctionsToCreate])); requires(all([(len(x) > 0) for x in listOfFunctionsToCreate])); listOfFunctionsToReturn = []; for thisFunctionDefinition in listOfFunctionsToCreate: initialSetOfDefinedFunctions = set(locals().keys()); exec(thisFunctionDefinition); newSetOfDefinedFunctions = set(locals().keys()).difference(\ initialSetOfDefinedFunctions).difference(\ ["initialSetOfDefinedFunctions", "newSetOfDefinedFunctions", "thisFunctionDefinition"]); listOfNewSetOfDefinedFunctions = list(newSetOfDefinedFunctions); for thisIndex in range(0, len(listOfNewSetOfDefinedFunctions)): listOfFunctionsToReturn.append(locals()[listOfNewSetOfDefinedFunctions[thisIndex]]); ensures(isinstance(listOfFunctionsToReturn ,list)); ensures(all([ (str(type(x)) == "<class 'function'>") for x in listOfFunctionsToReturn])); return listOfFunctionsToReturn; def createFunct_absThresholdCompare(name, var, comparitorAndScalarString): nameWithSpaceRemoved = name.replace(" ", ""); codeForFunct = """ def funct_{1}({2}): \"\"\"{0}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3Abs({2}) {3}; else: return abs({2}) {3}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(name, nameWithSpaceRemoved, var, comparitorAndScalarString); def createFunct_thresholdCompare(name, var, comparitorAndScalarString): nameWithSpaceRemoved = name.replace(" ", ""); codeForFunct = """ def funct_{1}({2}): \"\"\"{0}\"\"\" return {2} {3}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(name, nameWithSpaceRemoved, var, comparitorAndScalarString); def createFunct_MultiThresholdCompare(name, varsToUse, comparitorAndScalarStrings ): requires(len(varsToUse) == len(comparitorAndScalarStrings)); nameWithSpaceRemoved = name.replace(" ", ""); uniqueVariablesToUseAsString = ", ".join([str(x) for x in set(varsToUse)]); firstVariableInList = varsToUse[0]; atomicConditions = ", ".join([(x[0] + x[1]) for x in zip(varsToUse, comparitorAndScalarStrings)]); comparitorAndScalarStringsForZ3 = "z3.And(" + atomicConditions + ")"; comparitorAndScalarStringsForPython = "all(["+ atomicConditions + "])"; codeForFunct = """ def funct_{1}({2}): \"\"\"{0}\"\"\" if(isinstance({3}, z3.z3.ArithRef)): assert(all([isinstance(x, z3.z3.ArithRef) for x in [{2}]])); return {4}; else: return {5}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(name, nameWithSpaceRemoved, \ uniqueVariablesToUseAsString, firstVariableInList, \ comparitorAndScalarStringsForZ3, comparitorAndScalarStringsForPython \ ); def createFunct_SumThresholdCompare(name, varsToUse, comparitorAndScalarString ): requires(len(varsToUse) > 0); requires(isinstance(comparitorAndScalarString, str)); nameWithSpaceRemoved = name.replace(" ", ""); uniqueVariablesToUseAsString = ", ".join([str(x) for x in set(varsToUse)]); firstVariableInList = varsToUse[0]; # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg sumForZ3 = "z3.Sum(" + (", ".join(varsToUse)) + ")"; sumForPython = "np.sum(["+ (", ".join(varsToUse)) + "])"; codeForFunct = """ def funct_{1}({2}): \"\"\"{0}\"\"\" if(isinstance({3}, z3.z3.ArithRef)): assert(all([isinstance(x, z3.z3.ArithRef) for x in [{2}]])); return {4} {6}; else: return {5} {6}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(name, nameWithSpaceRemoved, \ uniqueVariablesToUseAsString, firstVariableInList, \ sumForZ3, sumForPython, comparitorAndScalarString \ ); def getFunctionCodeBasedOnThresholdsAndIndividualVariables(inputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds, indicesOfVariablesToUseBoolFor=set()): requires(isinstance(indicesOfVariablesToUseBoolFor, set)); requires(indicesOfVariablesToUseBoolFor.issubset(range(0, len(inputSpaceVariables)))); listOfFunctionCodes =[]; for thisIndex in range(0, len(inputSpaceVariables)): thisVarName = inputSpaceVariables[thisIndex]; functionToUseHere = createFunct_thresholdCompare; stringToInclude = ""; if(thisIndex in indicesOfVariablesToUseBoolFor): functionToUseHere = createFunct_absThresholdCompare; stringToInclude = " Magnitude"; listOfFunctionCodes.append(\ functionToUseHere(thisVarName + " Very High " + stringToInclude, thisVarName, ">= " + str(quantile0Dot90[thisIndex])) \ ); listOfFunctionCodes.append(\ functionToUseHere(thisVarName + " High " + stringToInclude, thisVarName, ">= " + str(quantile0Dot75[thisIndex])) \ ); listOfFunctionCodes.append(\ functionToUseHere(thisVarName + " Low " + stringToInclude, thisVarName, "<= " + str(quantile0Dot25[thisIndex])) \ ); listOfFunctionCodes.append(\ functionToUseHere(thisVarName + " Very Low " + stringToInclude, thisVarName, "<= " + str(quantile0Dot10[thisIndex])) \ ); listOfFunctionCodes.append(\ createFunct_MultiThresholdCompare( \ thisVarName + stringToInclude + " Near Normal Levels", \ [thisVarName, thisVarName], \ [" >= " + str(medians[thisIndex] - 1.5 * stds[thisIndex]), \ " <= " + str(medians[thisIndex] + 1.5 * stds[thisIndex]) ] ) ); return listOfFunctionCodes;
10,532
47.763889
2,823
py
Fanoos
Fanoos-master/domainsAndConditions/baseClassConditionsToSpecifyPredictsWith.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox, \ boxAContainsBoxB, boxContainsVector; import uuid; import re; import z3; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; class CharacterizationConditionsBaseClass(): def __str__(self): raise NotImplementedError(); # child classes must override this. def checkInitialization(self): assert(isinstance(self.z3Solver, z3.z3.Solver)); assert(isinstance(self.expectedNumberOfDimensions, int)); assert(self.expectedNumberOfDimensions > 0); assert(isinstance(self.listMappingPositionToVariableName, list)); assert(len(self.listMappingPositionToVariableName) == self.expectedNumberOfDimensions); assert(all([isinstance(x, z3.z3.ArithRef) for x in self.listMappingPositionToVariableName])); assert(re.match("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$", \ self.getID()) != None); assert(isinstance(self.relaventVariables(), frozenset)); assert(all([isinstance(x, z3.z3.ArithRef) for x in self.relaventVariables()])); assert(self.relaventVariables().issubset(self.listMappingPositionToVariableName)); return; def setID(self, uuidProvided=None): requires((uuidProvided is None) or isinstance(uuidProvided, str)); requires((uuidProvided is None) or (len(uuidProvided) > 0)); self.uuid = uuidProvided if (uuidProvided is not None) else str(uuid.uuid4()); ensures(isinstance(self.uuid , str)); ensures(len(self.uuid) > 0); return def getID(self): return self.uuid; def __init__(self, z3SolverInstance, functToGetUuidProvided=None): raise NotImplementedError(); # child classes must implement. self.expectedNumberOfDimensions = 5; self.z3FormattedCondition = ""; self.listMappingPositionToVariableName = []; self.z3Solver = z3SolverInstance; self.setID(uuidProvided=(None if (functToGetUuidProvided is None) else (functToGetUuidProvided(self))) ); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg self.checkInitialization(); raise NotImplementedError(); # child classes must implement. def defaultSlowImplementation_pythonFormatEvaluation(self, vectorToEvaluateAgainst): requires(isinstance(vectorToEvaluateAgainst, np.ndarray)); vectorToEvaluateAgainst = self.handleCaseOfJointBox(vectorToEvaluateAgainst); # hacky to put this before a requires... TODO: fix this.... requires(vectorToEvaluateAgainst.shape == (self.expectedNumberOfDimensions, 1)); precondition = z3.And([\ float(vectorToEvaluateAgainst[thisIndex]) == self.listMappingPositionToVariableName[thisIndex] \ for thisIndex in range(0, self.expectedNumberOfDimensions) ]); self.z3Solver.reset(); formulaToCheck = z3.And(precondition, self.z3FormattedCondition); # "And" as oppossed to "Implies" since # z3 is checking for satisfiability - so "Implies" would be trivial to satisfy by failing the hypothesis.... self.z3Solver.add(formulaToCheck); verdict = (self.z3Solver.check() == z3.z3.sat); return verdict; def relaventVariables(self): raise NotImplementedError(); # child classes must implement. def pythonFormatEvaluation(self, vectorToEvaluateAgainst): return self.defaultSlowImplementation_pythonFormatEvaluation(vectorToEvaluateAgainst); def convertBoxToFormulaConstraints(self, thisBox): requires(isProperBox(thisBox)); thisBox = self.handleCaseOfJointBox(thisBox); # hacky to put this before a requires... TODO: fix this.... requires(getDimensionOfBox(thisBox) == self.expectedNumberOfDimensions); F = z3.And([ \ z3.And( float(thisBox[index, 0]) <= self.listMappingPositionToVariableName[index], \ self.listMappingPositionToVariableName[index] <= float(thisBox[index, 1]) \ ) \ for index in range(0, self.expectedNumberOfDimensions) \ if self.listMappingPositionToVariableName[index] in self.relaventVariables()]); return F; def allMembersOfBoxSatisfyCondition(self, thisBox): requires(isProperBox(thisBox)); thisBox = self.handleCaseOfJointBox(thisBox); # hacky to put this before a requires... TODO: fix this.... requires(getDimensionOfBox(thisBox) == self.expectedNumberOfDimensions); self.z3Solver.reset(); # this might be the expensive.... I have to check formulaToCheck = \ z3.ForAll(self.listMappingPositionToVariableName, \ z3.Implies(\ self.convertBoxToFormulaConstraints(thisBox), \ self.z3FormattedCondition \ ) \ ); self.z3Solver.add(formulaToCheck); verdict = (self.z3Solver.check() == z3.z3.sat); return verdict; def existsMemberOfBoxSatifyingCondition(self, thisBox): requires(isProperBox(thisBox)); thisBox = self.handleCaseOfJointBox(thisBox); # hacky to put this before a requires... TODO: fix this.... requires(getDimensionOfBox(thisBox) == self.expectedNumberOfDimensions); self.z3Solver.reset(); # this might be the expensive.... I have to check formulaToCheck = \ z3.Exists(self.listMappingPositionToVariableName, \ z3.And(\ self.convertBoxToFormulaConstraints(thisBox), \ self.z3FormattedCondition \ ) \ ); self.z3Solver.add(formulaToCheck); verdict = (self.z3Solver.check() == z3.z3.sat); return verdict; def handleCaseOfJointBox(self, thisBox): if(thisBox.shape[0] == self.expectedNumberOfDimensions): # using shape[0] here kind of bleeds the box-inferface I had set up, but doing this here also allows for handling vectors appropraitely - or, rather, should return thisBox; else: assert(len(thisBox) == self.numberOfDimensionsInAJointBox); return self.functionToMapJointBoxToRelaventVariables(thisBox); import inspect; class CharacterizationCondition_FromPythonFunction(CharacterizationConditionsBaseClass): def __init__(self, z3SolverInstance, domainClass, functionToBaseConditionOn, functToGetUuidProvided=None): requires(isinstance(z3SolverInstance, z3.Solver)); requires(issubclass(domainClass, BaseClassDomainInformation)); requires(isinstance(functionToBaseConditionOn.__doc__, str)); requires(len(functionToBaseConditionOn.__doc__) > 0); self.functionToBaseConditionOn = functionToBaseConditionOn; variablesPresentInFunctionArguments = inspect.getargspec(functionToBaseConditionOn).args; self.relaventVariablesSet = frozenset([z3.Real(x) for x in variablesPresentInFunctionArguments]); self.numberOfDimensionsInAJointBox = len(domainClass.inputSpaceVariables()) + len(domainClass.outputSpaceVariables()); if(set(self.relaventVariablesSet).issubset(domainClass.inputSpaceVariables())): self.listMappingPositionToVariableName = domainClass.inputSpaceVariables(); # Note that by using the syntax [:k,] below, it works for both boxes and vectors as intended. See the # places where self.functionToMapJointBoxToRelaventVariables is used to see why it is good both are supported. self.functionToMapJointBoxToRelaventVariables = (lambda x: x[:len(domainClass.inputSpaceVariables()),]); # assumes the order of the joint box is input-space followed by output-space elif(set(self.relaventVariablesSet).issubset(domainClass.outputSpaceVariables())): self.listMappingPositionToVariableName = domainClass.outputSpaceVariables(); # Note that by using the syntax [:k,] below, it works for both boxes and vectors as intended. See the # places where self.functionToMapJointBoxToRelaventVariables is used to see why it is good both are supported. self.functionToMapJointBoxToRelaventVariables = (lambda x: x[len(domainClass.inputSpaceVariables()):,]); # assumes the order of the join box is input-space followed by output-space elif(set(self.relaventVariablesSet).issubset(domainClass.inputSpaceVariables() + domainClass.outputSpaceVariables())): # If it is a joint box, it should always be this <================================================ self.listMappingPositionToVariableName = domainClass.inputSpaceVariables() + domainClass.outputSpaceVariables(); self.functionToMapJointBoxToRelaventVariables = (lambda x: x); # assumes the order of the join box is input-space followed by output-space else: raise Exception("The provided function uses references variables not in the domain information.\n" + \ "The relavent set of variables: " + str(variablesPresentInFunctionArguments) + "\n" + \ "Known input variables: " + str(domainClass.inputSpaceVariables()) + "\n" + \ "Known output variables: " + str(domainClass.outputSpaceVariables()) + "\n"); assert(all([isinstance(x, z3.z3.ArithRef) for x in self.listMappingPositionToVariableName])); assert(set(self.relaventVariablesSet).issubset(self.listMappingPositionToVariableName)); self.dictMappingVariableNameToIndex = dict(); for thisIndex in range(0, len(self.listMappingPositionToVariableName)): self.dictMappingVariableNameToIndex[self.listMappingPositionToVariableName[thisIndex]] = thisIndex; assert(\ all([ (self.listMappingPositionToVariableName[self.dictMappingVariableNameToIndex[varName]] == varName) \ for varName in self.listMappingPositionToVariableName]\ )); assert(set(self.dictMappingVariableNameToIndex.keys()) == set(self.listMappingPositionToVariableName)); assert(set(self.dictMappingVariableNameToIndex.keys()).issuperset(self.relaventVariablesSet)); assert(set(self.dictMappingVariableNameToIndex.values()) == set(range(0, len(self.listMappingPositionToVariableName)))); self.z3FormattedCondition = \ functionToBaseConditionOn(**{str(x) : x for x in self.relaventVariablesSet}); self.expectedNumberOfDimensions = len(self.listMappingPositionToVariableName); self.z3Solver = z3SolverInstance; self.setID(uuidProvided=(None if (functToGetUuidProvided is None) else (functToGetUuidProvided(self)))); self.checkInitialization(); return; def relaventVariables(self): return self.relaventVariablesSet; def pythonFormatEvaluation(self, vectorToEvaluateAgainst): vectorToEvaluateAgainst = self.handleCaseOfJointBox(vectorToEvaluateAgainst); requires(isinstance(vectorToEvaluateAgainst, np.ndarray)); requires(len(vectorToEvaluateAgainst.shape) == 2); requires(vectorToEvaluateAgainst.shape[0] == self.expectedNumberOfDimensions); return self.functionToBaseConditionOn(\ **{str(thisVar) : vectorToEvaluateAgainst[self.dictMappingVariableNameToIndex[thisVar],0] \ for thisVar in self.relaventVariables() }); def __str__(self): return self.functionToBaseConditionOn.__doc__; class MetaCondition_Conjunction(CharacterizationConditionsBaseClass): def __init__(self, listOfConditionsToConjunct): requires(isinstance(listOfConditionsToConjunct, list)); requires(len(listOfConditionsToConjunct) > 0); self.listOfConditionsToConjunct = listOfConditionsToConjunct; self.z3Solver = listOfConditionsToConjunct[0].z3Solver; self.z3FormattedCondition = z3.And([x.z3FormattedCondition for x in listOfConditionsToConjunct]); self.setID(); return; def pythonFormatEvaluation(self, vectorToEvaluateAgainst): for thisCondition in self.listOfConditionsToConjunct: if(not thisCondition.pythonFormatEvaluation(vectorToEvaluateAgainst)): return False; return True; def allMembersOfBoxSatisfyCondition(self, thisBox): for thisCondition in self.listOfConditionsToConjunct: if(not thisCondition.allMembersOfBoxSatisfyCondition(thisBox)): return False; return True; def existsMemberOfBoxSatifyingCondition(self, thisBox): requires(isProperBox(thisBox)); self.z3Solver.reset(); conditionToCheck = \ z3.And([\ z3.And(\ x.convertBoxToFormulaConstraints(thisBox), \ x.z3FormattedCondition \ ) \ for x in self.listOfConditionsToConjunct \ ] ); setOfVariablesToQuantifyOver = set(); for thisCondition in self.listOfConditionsToConjunct: setOfVariablesToQuantifyOver.update(thisCondition.listMappingPositionToVariableName); formulaToCheck = \ z3.Exists(list(setOfVariablesToQuantifyOver), \ conditionToCheck \ ); self.z3Solver.add(formulaToCheck); verdict = (self.z3Solver.check() == z3.z3.sat); return verdict; def relaventVariables(self): setToReturn = set(); for thisCondition in self.listOfConditionsToConjunct: setToReturn.update(thisCondition.relaventVariables()); assert(setToReturn.issuperset(thisCondition.relaventVariables())); return frozenset(setToReturn); def setID(self): self.uuid = frozenset([x.getID() for x in self.listOfConditionsToConjunct]); return def __str__(self): # We sort the below in order to ensure that the order is fixed based on the content - # two instances with the same parameters should produce the same string representation. return "And("+ (", ".join(sorted([str(x) for x in self.listOfConditionsToConjunct]))) + ")"; class Condition_TheBoxItself(CharacterizationConditionsBaseClass): def __init__(self, z3Solver, thisBox, listMappingAxisIndexToVariableInQuestion): requires(isProperBox(thisBox)); requires(isinstance(listMappingAxisIndexToVariableInQuestion, list)); requires(all([isinstance(x, z3.z3.ArithRef) for x in listMappingAxisIndexToVariableInQuestion])); requires( len(listMappingAxisIndexToVariableInQuestion) == getDimensionOfBox(thisBox)); self.expectedNumberOfDimensions = getDimensionOfBox(thisBox); self.listMappingPositionToVariableName = listMappingAxisIndexToVariableInQuestion; self.personalBox = thisBox; self.z3Solver = z3Solver; # Including for uniformity of interface so that the # metaConditions can use it, not that it is used in this class itself. self.z3FormattedCondition = self.convertBoxToFormulaConstraints(thisBox); # None; self.setID(); return; def defaultSlowImplementation_pythonFormatEvaluation(self, vectorToEvaluateAgainst): raise NotImplementedError(); def pythonFormatEvaluation(self, vectorToEvaluateAgainst): return boxContainsVector(self.personalBox, vectorToEvaluateAgainst.reshape(max(vectorToEvaluateAgainst.shape),)); def allMembersOfBoxSatisfyCondition(self, thisBox): return boxAContainsBoxB(self.personalBox, thisBox); def existsMemberOfBoxSatifyingCondition(self, thisBox): raise NotImplementedError(); def relaventVariables(self): return frozenset(self.listMappingPositionToVariableName); def __str__(self): labeledValuesOfBoxes = [\ str(self.listMappingPositionToVariableName[thisIndex]) + " : " +\ str(self.personalBox[thisIndex, :]).replace("\n", ",") \ for thisIndex in range(0, len(self.listMappingPositionToVariableName)) \ ]; return "Box(" + ( ", ".join(labeledValuesOfBoxes) ) + ")"; def convertBoxToFormulaConstraints(self, thisBox): requires(isProperBox(thisBox)); F = z3.And([ \ z3.And( float(thisBox[index, 0]) <= self.listMappingPositionToVariableName[index], \ self.listMappingPositionToVariableName[index] <= float(thisBox[index, 1]) \ ) \ for index in range(0, self.expectedNumberOfDimensions) ]); return F;
20,532
57.498575
2,897
py
Fanoos
Fanoos-master/domainsAndConditions/domainAndConditionsForCPUUse.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import re; import z3; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass, CharacterizationCondition_FromPythonFunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; from domainsAndConditions.utilsForDefiningPredicates import *; class DomainForCPUUse(BaseClassDomainInformation): def __init__(self, z3SolverInstance): requires(isinstance(z3SolverInstance, z3.z3.Solver)); self.initializedConditions = None; self.initialize_baseConditions(z3SolverInstance); assert(self.initializedConditions != None); return; @staticmethod def getUUID(): return "b8d0a274-2104-4f88-8f7c-d71380bafda3"; @staticmethod def getInputSpaceUniverseBox(): orderOfVariables = __class__.inputSpaceVariables(); # Unfortunately, each of these axis probably follow a power-law distribution, meaning that most of the # bounding box contains points that were never seen, but I suppose that shows one possible benefit of our # analysis if we do things properly - not only can we consider the actual data, but the hypotheticals as # well... dictMappingVariableToBound = {\ "lread" : [0.0, 0.03685636856368564] ,\ "scall" : [0.00952842377260982, 0.42449127906976736] ,\ "sread" : [0.0028237951807228916, 0.09920933734939759] ,\ "freemem" : [0.006097560975609756, 0.6275225526227863] ,\ "freeswap" : [0.4323552671759128, 0.8317991159890958] \ }; thisUniverseBox = __class__._helper_getInputSpaceUniverseBox(\ orderOfVariables, dictMappingVariableToBound); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg ensures(getDimensionOfBox(thisUniverseBox) == len(DomainForCPUUse.inputSpaceVariables())); return thisUniverseBox; @staticmethod def inputSpaceVariables(): return [\ z3.Real(x) for x in ['lread', 'scall', 'sread', 'freemem', 'freeswap'] ]; @staticmethod def outputSpaceVariables(): return [z3.Real(x) for x in ['lwrite', 'swrite', 'usr']]; @staticmethod def getName(): return "Domain For CPU Use"; def initialize_baseConditions(self, z3SolverInstance): dictMappingPredicateStringNameToUUID = \ { "lread Very High " : "d0473bb6-fb13-40bd-b114-deb66c5a3108", "lread High " : "f772c081-4796-4d82-803e-55c85958a092", "lread Low " : "a21bcdb0-d38b-44cc-8285-cd815efc4687", "lread Very Low " : "426c8077-2372-4222-9c19-0a6a5487e2bb", "lread Near Normal Levels" : "9b4d6b53-04c1-4f06-be2c-9d6a70bd787d", "scall Very High " : "d482ea3a-8549-4300-b454-fb4bb710fb83", "scall High " : "7a6c6206-11d9-4caf-b69a-641b9aa8908f", "scall Low " : "8787422e-5690-412c-a487-0be85dc55b87", "scall Very Low " : "c0e0e467-6c8f-4536-9a22-ddc3e8f2f4c2", "scall Near Normal Levels" : "8d8be01c-f46c-4128-8e15-15189c6138ae", "sread Very High " : "831efc45-9d8a-439d-9ca3-6cdf56696d7a", "sread High " : "89362126-a1b5-46dd-8229-9a098157eddf", "sread Low " : "c5396982-66ed-487b-a7a4-107ea056f3be", "sread Very Low " : "db76fe10-87c6-45dd-95a5-af264bfb1aaa", "sread Near Normal Levels" : "2cbd9e37-b203-4a17-9e32-c24955b1d9a2", "freemem Very High " : "074f3594-828f-4565-bca2-526bb84f3066", "freemem High " : "64525dd6-f8ad-48dc-b349-95aee377e175", "freemem Low " : "b51c470e-4c11-4332-9c5b-a9e6dcb945b3", "freemem Very Low " : "8e1935a7-ae6e-4195-ab92-625cb914b692", "freemem Near Normal Levels" : "58b5002b-647a-4fd5-a3ac-4b1b2b4e1ef2", "freeswap Very High " : "b2b18474-577a-47ca-8379-dca459da4f0a", "freeswap High " : "ce118774-a9ab-4ddb-9f9a-8c6127d9a7d2", "freeswap Low " : "4def0fe4-6726-44f9-a247-8d3ba34f3689", "freeswap Very Low " : "be17eb25-cc23-4472-bbc5-656619ed6e75", "freeswap Near Normal Levels" : "fcb4caf7-a3f0-4b4b-9f0c-44fda08525ff", "Free Space Very High" : "77f110ab-e452-417e-9302-ebcb6ab8a8d3", "Free Space High" : "7cf4b681-5b3f-4f02-9415-1004fcd3ada8", "Free Space Low" : "aa464949-13cb-4af4-8f48-2f125531f516", "Free Space Very Low" : "8d164d86-ed92-4ba2-bd37-97baaae34f06", "reads Very High" : "797cde0f-63f3-4967-840e-9ee99e22c660", "reads High" : "7ffac4ab-cf20-4453-8852-aa3743049b14", "reads Low" : "c1051f5e-bbf9-46a3-8cbf-6bfab57eba92", "reads Very Low" : "c3c4274b-57a4-4c91-804f-e7cb76af19ab", "lwrite Very High " : "52586695-b844-4091-9484-efb038816fcf", "lwrite High " : "d520174d-6e84-4d4c-808f-787537f89234", "lwrite Low " : "adf5a9d0-1653-4f30-a25a-755da7a9283c", "lwrite Very Low " : "3ec03fb8-6238-49dd-8ced-8cb86495bfcf", "lwrite Near Normal Levels" : "58c0edfd-3578-4d4f-987d-9c9f9f972a72", "swrite Very High " : "4fe47b86-26d8-4477-abef-6b3c5328a321", "swrite High " : "f988e519-e2fa-474d-8e40-72a908e16cee", "swrite Low " : "76c5e3fb-238c-446f-8a66-dacb4a850d4b", "swrite Very Low " : "64d6cbdc-28ea-47a9-be37-2e03e2bb0fdd", "swrite Near Normal Levels" : "cb4b2c43-9d7c-4a2e-8cca-d6eea9f890f0", "usr Very High " : "5da35929-fc1e-4286-85c3-cdfb80a5fb87", "usr High " : "ab57c686-648c-4234-880b-ce98ed30e9e2", "usr Low " : "977fc451-daca-486b-b92c-9b05a20b1f3b", "usr Very Low " : "b5087023-3270-45ef-95c6-27d0f67630c0", "usr Near Normal Levels" : "6e3766b5-ff6e-49c2-a6bd-44c2de12d0b5", "writes Very High" : "bf0de32a-30c9-41e8-a104-485801a7d9d9", "writes High" : "24296724-4e3c-4f75-9579-09049e66413a", "writes Low" : "3f72fc32-1929-4d5f-9883-82991ce7cf8e", "writes Very Low" : "1fbc7df3-8f9a-4ec2-ab73-cdd0868f9b02", "user land related activity Very High" : "eda11b3e-8f65-40ff-821d-05a0dd978158", "user land related activity High" : "2ecd934d-cec8-4946-9509-a5013f642027", "user land related activity Low" : "28032515-fcb2-4639-a85d-25a7b3037f03", "user land related activity Very Low" : "bfa94c13-0355-4274-98aa-524456418d29", "nonuser system activity Very High" : "e32eeafe-ffbc-4c99-b826-591f94be63db", "nonuser system activity High" : "3d9c1c7c-faec-4368-8111-c6fc28d4867a", "nonuser system activity Low" : "7ec43da1-ca18-4d9f-a6ed-6431471be456", "nonuser system activity Very Low" : "c945c093-6f3a-42d2-88b0-d00360ea3457", }; # Note: manualAbstractionInformation is used purely in analysis scripts (as developed for # the paper describing Fanoos); this proved to be a convieniant place to store the # information during the time of development and testing. Fanoos does not access # the information in manualAbstractionInformation when determining how to make # adjustments to respond to users. Again, it is only used in analysis scripts # used to prepare results for the paper. # # While it was convieniant for development, clearly it is not # ideal have this data stored here or this structure required # to be present. TODO: resolve the issue just described. self.manualAbstractionInformation = {\ "predicatesAndLabels" : [\ ("d0473bb6-fb13-40bd-b114-deb66c5a3108" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("f772c081-4796-4d82-803e-55c85958a092" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("a21bcdb0-d38b-44cc-8285-cd815efc4687" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("426c8077-2372-4222-9c19-0a6a5487e2bb" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("9b4d6b53-04c1-4f06-be2c-9d6a70bd787d" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("d482ea3a-8549-4300-b454-fb4bb710fb83" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("7a6c6206-11d9-4caf-b69a-641b9aa8908f" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("8787422e-5690-412c-a487-0be85dc55b87" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("c0e0e467-6c8f-4536-9a22-ddc3e8f2f4c2" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("8d8be01c-f46c-4128-8e15-15189c6138ae" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("831efc45-9d8a-439d-9ca3-6cdf56696d7a" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("89362126-a1b5-46dd-8229-9a098157eddf" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("c5396982-66ed-487b-a7a4-107ea056f3be" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("db76fe10-87c6-45dd-95a5-af264bfb1aaa" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("2cbd9e37-b203-4a17-9e32-c24955b1d9a2" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("074f3594-828f-4565-bca2-526bb84f3066" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("64525dd6-f8ad-48dc-b349-95aee377e175" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("b51c470e-4c11-4332-9c5b-a9e6dcb945b3" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("8e1935a7-ae6e-4195-ab92-625cb914b692" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("58b5002b-647a-4fd5-a3ac-4b1b2b4e1ef2" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("b2b18474-577a-47ca-8379-dca459da4f0a" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("ce118774-a9ab-4ddb-9f9a-8c6127d9a7d2" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("4def0fe4-6726-44f9-a247-8d3ba34f3689" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("be17eb25-cc23-4472-bbc5-656619ed6e75" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("fcb4caf7-a3f0-4b4b-9f0c-44fda08525ff" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("77f110ab-e452-417e-9302-ebcb6ab8a8d3" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("7cf4b681-5b3f-4f02-9415-1004fcd3ada8" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("aa464949-13cb-4af4-8f48-2f125531f516" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("8d164d86-ed92-4ba2-bd37-97baaae34f06" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("797cde0f-63f3-4967-840e-9ee99e22c660" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("7ffac4ab-cf20-4453-8852-aa3743049b14" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("c1051f5e-bbf9-46a3-8cbf-6bfab57eba92" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("c3c4274b-57a4-4c91-804f-e7cb76af19ab" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("52586695-b844-4091-9484-efb038816fcf" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("d520174d-6e84-4d4c-808f-787537f89234" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("adf5a9d0-1653-4f30-a25a-755da7a9283c" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("3ec03fb8-6238-49dd-8ced-8cb86495bfcf" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("58c0edfd-3578-4d4f-987d-9c9f9f972a72" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("4fe47b86-26d8-4477-abef-6b3c5328a321" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("f988e519-e2fa-474d-8e40-72a908e16cee" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("76c5e3fb-238c-446f-8a66-dacb4a850d4b" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("64d6cbdc-28ea-47a9-be37-2e03e2bb0fdd" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("cb4b2c43-9d7c-4a2e-8cca-d6eea9f890f0" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("5da35929-fc1e-4286-85c3-cdfb80a5fb87" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("ab57c686-648c-4234-880b-ce98ed30e9e2" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("977fc451-daca-486b-b92c-9b05a20b1f3b" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("b5087023-3270-45ef-95c6-27d0f67630c0" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("6e3766b5-ff6e-49c2-a6bd-44c2de12d0b5" , "326e2db9-5b3f-4ac5-b777-e0d46b301458"), \ ("bf0de32a-30c9-41e8-a104-485801a7d9d9" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("24296724-4e3c-4f75-9579-09049e66413a" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("3f72fc32-1929-4d5f-9883-82991ce7cf8e" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("1fbc7df3-8f9a-4ec2-ab73-cdd0868f9b02" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("eda11b3e-8f65-40ff-821d-05a0dd978158" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("2ecd934d-cec8-4946-9509-a5013f642027" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("28032515-fcb2-4639-a85d-25a7b3037f03" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("bfa94c13-0355-4274-98aa-524456418d29" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("e32eeafe-ffbc-4c99-b826-591f94be63db" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("3d9c1c7c-faec-4368-8111-c6fc28d4867a" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("7ec43da1-ca18-4d9f-a6ed-6431471be456" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ("c945c093-6f3a-42d2-88b0-d00360ea3457" , "9b5a2373-680f-413b-9e37-cc66e59e2880"), \ ], \ "labelDag_firstParent_secondChild" : [ \ ("9b5a2373-680f-413b-9e37-cc66e59e2880" , "326e2db9-5b3f-4ac5-b777-e0d46b301458") \ ] \ }; functToGetUuidProvided = (lambda predicateObjectBeingInitialized : dictMappingPredicateStringNameToUUID[str(predicateObjectBeingInitialized)] ); self.initializedConditions = \ [CharacterizationCondition_FromPythonFunction(z3SolverInstance, DomainForCPUUse, x, functToGetUuidProvided=functToGetUuidProvided) \ for x in getListFunctionsToBaseCondtionsOn_forInputOfDomainCPUUse() + \ getListFunctionsToBaseCondtionsOn_forOutputOfDomainCPUUse() + \ getListFunctionsToBaseCondtionsOn_forJointInputAndOutputDomainsInCPUUse() ]; assert(all([ (x.getID() == functToGetUuidProvided(x)) for x in self.initializedConditions])); self._writeInfoToDatabase(); return; def getBaseConditions(self): return self.initializedConditions; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # class-specific utilities for defining domains #=========================================================================== def getFunctionsCouplingValues(functionBaseName, varaiblesToUse, \ listOfVariablesInSpace, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds): requires(isinstance(varaiblesToUse, list)); requires(len(varaiblesToUse) > 0); requires(all([isinstance(x, str) for x in ([functionBaseName] + varaiblesToUse)])); requires(all([(len(x) > 0) for x in ([functionBaseName] + varaiblesToUse)])); requires(len(set(([functionBaseName] + varaiblesToUse))) == \ len(([functionBaseName] + varaiblesToUse)) ); # checking that the variable names are unique... requires(set(varaiblesToUse).issubset(listOfVariablesInSpace)); functionCodeToReturn = []; def sumBasedOnIndices(thisListOfValues): return sum([thisListOfValues[listOfVariablesInSpace.index(thisValue)] for thisValue in varaiblesToUse]); namesAndThresHolds = [\ (" Very High" , " >= " + str(sumBasedOnIndices(quantile0Dot90)) ), \ (" High" , " >= " + str(sumBasedOnIndices(quantile0Dot75)) ), \ (" Low" , " <= " + str(sumBasedOnIndices(quantile0Dot25)) ), \ (" Very Low" , " <= " + str(sumBasedOnIndices(quantile0Dot10)) ) \ ]; for thisNameAndThresHold in namesAndThresHolds: functionName = functionBaseName + thisNameAndThresHold[0]; functionName = functionName.replace(" ", " ").replace(" ", " "); # I could do this replacement # until the length converges, but doing that seems silly since this should cover the # wide majority of cases. functionNameWithSpacesRemoved = functionName.replace(" ", ""); thisFunctionCode = createFunct_SumThresholdCompare(functionName, \ varaiblesToUse, thisNameAndThresHold[1] ); functionCodeToReturn.append(thisFunctionCode); return functionCodeToReturn; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the input domain #=========================================================================== otherInputSpaceFunctionsToUse = """ def funct_{1}({2}): \"\"\"{0}\"\"\" if(isinstance({3}, z3.z3.ArithRef)): assert(all([isinstance(x, z3.z3.ArithRef) for x in [{2}]])); return {4}; else: return {5}; raise Exception("Control should not reach here"); return; """ def getListFunctionsToBaseCondtionsOn_forInputOfDomainCPUUse(): listOfFunctionCodes =[]; inputSpaceVariables = ['lread', 'scall', 'sread', 'freemem', 'freeswap']; quantile0Dot90 = [0.02439024, 0.35714632, 0.07868976, 0.53894086, 0.82341586]; quantile0Dot75 = [0.01084011, 0.25906411, 0.05139307, 0.16265035, 0.77139313]; quantile0Dot25 = [0.00108401, 0.07291667, 0.01506024, 0.01470097, 0.46479515]; quantile0Dot10 = [0.00054201, 0.02091408, 0.00623117, 0.00760107, 0.44180199]; medians = [0.00379404, 0.15685562, 0.03012048, 0.04376879, 0.57475754]; stds = [0.02891628, 0.13190549, 0.03745633, 0.20731315, 0.18812254]; listOfFunctionCodes = \ getFunctionCodeBasedOnThresholdsAndIndividualVariables(inputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); listOfFunctionCodes = \ listOfFunctionCodes + \ getFunctionsCouplingValues("Free Space", ['freemem', 'freeswap'], \ inputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); listOfFunctionCodes = \ listOfFunctionCodes + \ getFunctionsCouplingValues("reads", ['lread', 'sread'], \ inputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the output domain #=========================================================================== otherOutputSpaceFunctionsToUse = """ """; def getListFunctionsToBaseCondtionsOn_forOutputOfDomainCPUUse(): listOfFunctionCodes =[]; outputSpaceVariables = ['lwrite', 'swrite', 'usr'] quantile0Dot90 = [0.08, 0.05303725, 0.97979798]; quantile0Dot75 = [0.0173913, 0.03266654, 0.94949495]; quantile0Dot25 = [0.0, 0.01027712, 0.81818182]; quantile0Dot10 = [0.0, 0.00440448, 0.72727273]; medians = [0.00173913, 0.02018719, 0.8989899 ]; stds = [0.05198244, 0.02944929, 0.18586648]; listOfFunctionCodes = \ getFunctionCodeBasedOnThresholdsAndIndividualVariables(outputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); listOfFunctionCodes = \ listOfFunctionCodes + \ getFunctionsCouplingValues("writes", ['lwrite', 'swrite'], \ outputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the joint domain #=========================================================================== def getListFunctionsToBaseCondtionsOn_forJointInputAndOutputDomainsInCPUUse(): listOfFunctionCodes =[]; variables = ['lread', 'scall', 'sread', 'freemem', 'freeswap']; quantile0Dot90 = [0.02439024, 0.35714632, 0.07868976, 0.53894086, 0.82341586]; quantile0Dot75 = [0.01084011, 0.25906411, 0.05139307, 0.16265035, 0.77139313]; quantile0Dot25 = [0.00108401, 0.07291667, 0.01506024, 0.01470097, 0.46479515]; quantile0Dot10 = [0.00054201, 0.02091408, 0.00623117, 0.00760107, 0.44180199]; medians = [0.00379404, 0.15685562, 0.03012048, 0.04376879, 0.57475754]; stds = [0.02891628, 0.13190549, 0.03745633, 0.20731315, 0.18812254]; variables = variables + ['lwrite', 'swrite', 'usr'] quantile0Dot90 = quantile0Dot90+ [0.08, 0.05303725, 0.97979798]; quantile0Dot75 = quantile0Dot75 + [0.0173913, 0.03266654, 0.94949495]; quantile0Dot25 = quantile0Dot25 + [0.0, 0.01027712, 0.81818182]; quantile0Dot10 = quantile0Dot10 + [0.0, 0.00440448, 0.72727273]; medians = medians + [0.00173913, 0.02018719, 0.8989899 ]; stds = stds + [0.05198244, 0.02944929, 0.18586648]; listOfFunctionCodes = \ listOfFunctionCodes + \ getFunctionsCouplingValues("user land related activity ", ['lwrite', 'lread', 'usr'], \ variables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); listOfFunctionCodes = \ listOfFunctionCodes + \ getFunctionsCouplingValues("nonuser system activity", ['scall', 'sread', 'swrite'], \ variables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^
26,329
61.840095
2,861
py
Fanoos
Fanoos-master/domainsAndConditions/domainAndConditionsFor_modelForTesting_oneDimInput_oneDimOutput.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; from utils.contracts import *; from boxesAndBoxOperations.getBox import getDimensionOfBox; import z3; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass, CharacterizationCondition_FromPythonFunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; from domainsAndConditions.utilsForDefiningPredicates import *; class DomainFor_modelForTesting_oneDimInput_oneDimOutput(BaseClassDomainInformation): def __init__(self, z3SolverInstance): requires(isinstance(z3SolverInstance, z3.z3.Solver)); self.initializedConditions = None; self.initialize_baseConditions(z3SolverInstance); assert(self.initializedConditions != None); return; @staticmethod def getUUID(): return "e2773e2f-9109-4444-8a51-37ba22dd3ceb"; @staticmethod def getInputSpaceUniverseBox(): orderOfVariables = __class__.inputSpaceVariables(); dictMappingVariableToBound = {\ "in_x" : [-5.0, 5.0] \ }; thisUniverseBox = __class__._helper_getInputSpaceUniverseBox(\ orderOfVariables, dictMappingVariableToBound); ensures(getDimensionOfBox(thisUniverseBox) == len(DomainFor_modelForTesting_oneDimInput_oneDimOutput.inputSpaceVariables())); return thisUniverseBox; @staticmethod def inputSpaceVariables(): return [\ z3.Real(x) for x in ["in_x"] ]; @staticmethod def outputSpaceVariables(): return [z3.Real(x) for x in ["out_y"]]; @staticmethod def getName(): return "Domain For modelForTesting_oneDimInput_oneDimOutput"; def initialize_baseConditions(self, z3SolverInstance): dictMappingPredicateStringNameToUUID = \ { "INNEG10DOT0TONEG9DOT0" : "db2a1ddb-8088-4068-9f69-3968646858f0" , \ "INNEG10DOT0TONEG9DOT5" : "d4649337-1c95-40f0-9af3-75e3fd7db819" , \ "INNEG9DOT5TONEG9DOT0" : "fb30abf2-b099-49de-886a-3c48a1fa9387" , \ "INNEG9DOT0TONEG8DOT0" : "75db05ee-70cb-4c08-bc17-c711f1d5f7db" , \ "INNEG9DOT0TONEG8DOT5" : "975ee971-50ca-4e93-80d1-80574b3da93d" , \ "INNEG8DOT5TONEG8DOT0" : "d9b7684c-14e4-4326-8410-018e11e91709" , \ "INNEG8DOT0TONEG7DOT0" : "44adc0c5-81ce-4286-b414-86e97fac5e0f" , \ "INNEG8DOT0TONEG7DOT5" : "c6945de5-d222-4974-956c-c80eb3cdbaf5" , \ "INNEG7DOT5TONEG7DOT0" : "2f7a0541-27ec-48fb-977e-a48d7b2e4e1c" , \ "INNEG7DOT0TONEG6DOT0" : "8c1360e7-40e7-4f8d-9def-e9c81b92dfa0" , \ "INNEG7DOT0TONEG6DOT5" : "60733d02-f787-471d-b0a1-16d5f87eb4c5" , \ "INNEG6DOT5TONEG6DOT0" : "4de52ebc-4c80-43e9-9d2d-b02615d9a87c" , \ "INNEG6DOT0TONEG5DOT0" : "b88952f9-7824-4be9-b05c-32ac0e61f727" , \ "INNEG6DOT0TONEG5DOT5" : "8dad23fd-54a7-4e15-b7a4-3c00721881a6" , \ "INNEG5DOT5TONEG5DOT0" : "4e7a95d8-6ae9-4544-b9d4-4907a0830506" , \ "INNEG5DOT0TONEG4DOT0" : "1da5d485-b94b-4234-bc5f-28ff807e5e44" , \ "INNEG5DOT0TONEG4DOT5" : "bd357a9c-3c18-4134-817f-c6ced25b5bff" , \ "INNEG4DOT5TONEG4DOT0" : "1a23a2b1-a37a-480b-b5fc-83e05660f645" , \ "INNEG4DOT0TONEG3DOT0" : "f4be9bfc-bdbf-43f1-b3b1-48842ec8c929" , \ "INNEG4DOT0TONEG3DOT5" : "1def8aea-92fc-4998-95ce-2d1f76cb8d99" , \ "INNEG3DOT5TONEG3DOT0" : "ea5ed997-ca51-4da1-a344-c38e064407b4" , \ "INNEG3DOT0TONEG2DOT0" : "dfc7f432-b738-4198-afa5-a8a2c8f41ebe" , \ "INNEG3DOT0TONEG2DOT5" : "fc85d7a3-6bb7-4a76-ad4e-fa84fbfb9c2a" , \ "INNEG2DOT5TONEG2DOT0" : "70952127-d35e-4a4c-98bd-60a4d87370ff" , \ "INNEG2DOT0TONEG1DOT0" : "77dd1811-dc71-44e8-a364-e343d6a60f69" , \ "INNEG2DOT0TONEG1DOT5" : "7ca63199-3618-49ea-950c-98b061ef1226" , \ "INNEG1DOT5TONEG1DOT0" : "6bb3434e-9d46-4fb1-b507-e15f08ad4b93" , \ "INNEG1DOT0TO0DOT0" : "207bad2a-5aea-484e-8ba5-a24d26189a79" , \ "INNEG1DOT0TONEG0DOT5" : "1387a719-7347-4090-b2ee-5667510380c8" , \ "INNEG0DOT5TO0DOT0" : "a13abe92-ebd1-4b80-b266-5ef1c473b911" , \ "IN0DOT0TO1DOT0" : "a4aff326-51d9-461d-a789-031438936e49" , \ "IN0DOT0TO0DOT5" : "9f1a2c66-009f-4440-a536-6dd19fd34235" , \ "IN0DOT5TO1DOT0" : "98017e69-1e3d-4a08-99bd-50ca1eb02a6f" , \ "IN1DOT0TO2DOT0" : "aa456821-34d0-4960-b414-62aeffabcacb" , \ "IN1DOT0TO1DOT5" : "85cc3d8d-501d-4688-91d4-c2101876c781" , \ "IN1DOT5TO2DOT0" : "b46f835a-d295-40c9-a08c-c7045efa49e6" , \ "IN2DOT0TO3DOT0" : "bc9c72cd-5fca-42b5-b94f-4a4e69666ac0" , \ "IN2DOT0TO2DOT5" : "83b55d67-52f0-4869-86e1-add216bb5dae" , \ "IN2DOT5TO3DOT0" : "67d30b36-dbda-4d54-88c6-0d549ec1061d" , \ "IN3DOT0TO4DOT0" : "35328f41-ddba-4225-98f4-a26c4211cf27" , \ "IN3DOT0TO3DOT5" : "e6ed9b47-4e4b-41b0-afef-bcbe49a29bac" , \ "IN3DOT5TO4DOT0" : "d9760c1d-18cf-4c02-b821-771cba6f6bc2" , \ "IN4DOT0TO5DOT0" : "f7a6bbf3-631f-482a-b80c-c6aae74f6ea9" , \ "IN4DOT0TO4DOT5" : "4988ab15-3a26-4764-8537-97cc4862be60" , \ "IN4DOT5TO5DOT0" : "6c28d7ab-9c76-4446-a428-3758d4561c89" , \ "IN5DOT0TO6DOT0" : "7d381ad9-0c24-47d3-947c-ccc2ac803f7f" , \ "IN5DOT0TO5DOT5" : "8336a5a2-fd71-42f4-a6a1-2e15df02a203" , \ "IN5DOT5TO6DOT0" : "9230da5a-1812-4b8a-925a-5ae8cfcfa11a" , \ "IN6DOT0TO7DOT0" : "00aa1dca-1f6f-4bb0-8823-388ed399d498" , \ "IN6DOT0TO6DOT5" : "ff5197aa-c33c-4934-abee-7fa1b1cb7874" , \ "IN6DOT5TO7DOT0" : "0490c272-b652-41ff-9d20-34804d1b2d7f" , \ "IN7DOT0TO8DOT0" : "549429a0-0f23-4be5-a577-f43a186eae5a" , \ "IN7DOT0TO7DOT5" : "5316a69e-338f-4ac6-b441-e1b2a0cc86b8" , \ "IN7DOT5TO8DOT0" : "dd73dace-91fe-44a0-b086-05ebdd1f1bf1" , \ "IN8DOT0TO9DOT0" : "02e99885-9d1e-44ab-b0d5-8c419a58a7a9" , \ "IN8DOT0TO8DOT5" : "e4d03482-1bce-495b-9a5c-7ee3fb9e21b9" , \ "IN8DOT5TO9DOT0" : "f34c630d-9113-4e44-990e-4f95fd3135a6" , \ "IN9DOT0TO10DOT0" : "c3c1a22c-4652-4d1d-869b-a5dec1c3a9a1" , \ "IN9DOT0TO9DOT5" : "698ad6b8-dddb-4571-9354-0c75d6506b77" , \ "IN9DOT5TO10DOT0" : "ea758b0c-d239-469e-a663-a1ef90c9e2d5" , \ "OUTNEG20DOT0TONEG19DOT0" : "eca0431d-d6c7-41ae-ac85-8db0cc9e7fa9" , \ "OUTNEG20DOT0TONEG19DOT5" : "caa61dbf-e329-4c3a-973d-e6ae67bb110b" , \ "OUTNEG19DOT5TONEG19DOT0" : "7fba4426-b3db-41dd-9f96-553cf1180f1d" , \ "OUTNEG19DOT0TONEG18DOT0" : "e83a19ec-f3fb-4d1e-bcfa-83a2f8bb9725" , \ "OUTNEG19DOT0TONEG18DOT5" : "3bab596b-7558-49d8-aaaa-17bbddd84a09" , \ "OUTNEG18DOT5TONEG18DOT0" : "7541b77a-fbb7-4c78-a338-1e9d96215a12" , \ "OUTNEG18DOT0TONEG17DOT0" : "9afad965-b784-48c4-98df-a05a6f2b7039" , \ "OUTNEG18DOT0TONEG17DOT5" : "eca4d162-e876-4ad1-9449-bee6525375dc" , \ "OUTNEG17DOT5TONEG17DOT0" : "45dfda64-f777-4f6c-bfec-a106d765cd66" , \ "OUTNEG17DOT0TONEG16DOT0" : "5302e37b-0898-46fb-9782-3642620cfe3d" , \ "OUTNEG17DOT0TONEG16DOT5" : "b3fde4fb-cdab-4627-b6c9-f1e36a77f098" , \ "OUTNEG16DOT5TONEG16DOT0" : "06368a59-13e0-4a2a-a54a-4f04a222f5c2" , \ "OUTNEG16DOT0TONEG15DOT0" : "1ff64a31-4736-4460-b49f-5b76abcea767" , \ "OUTNEG16DOT0TONEG15DOT5" : "ecc98606-8bcf-4f20-a20a-b917775f8b4c" , \ "OUTNEG15DOT5TONEG15DOT0" : "789ad1c3-8661-4ec1-9aea-a2b79b84855e" , \ "OUTNEG15DOT0TONEG14DOT0" : "9894ac1f-7dd7-445d-94df-f8f13af9ccd8" , \ "OUTNEG15DOT0TONEG14DOT5" : "c8a07542-a614-46f5-b5bb-8d8cdb57c148" , \ "OUTNEG14DOT5TONEG14DOT0" : "9f4094ad-53f6-4e72-8c53-9284d22bb019" , \ "OUTNEG14DOT0TONEG13DOT0" : "a497b214-4bb0-47b8-b196-087f54089079" , \ "OUTNEG14DOT0TONEG13DOT5" : "010954a7-4065-4964-82c4-33436fa70ab9" , \ "OUTNEG13DOT5TONEG13DOT0" : "3f650c7d-c7c6-482f-8901-825a312fd166" , \ "OUTNEG13DOT0TONEG12DOT0" : "65668c81-91ce-4b99-b5f8-b5153c3bdf4d" , \ "OUTNEG13DOT0TONEG12DOT5" : "ad576223-b6fa-45d8-835b-ad4172a2239f" , \ "OUTNEG12DOT5TONEG12DOT0" : "7a6743e0-9e91-4c27-be77-bfc73b1b39d3" , \ "OUTNEG12DOT0TONEG11DOT0" : "b0b8086e-5d55-4e71-a5ec-87b9c98ed676" , \ "OUTNEG12DOT0TONEG11DOT5" : "67fff4e5-1faa-4f80-8b7a-c43d82442b0e" , \ "OUTNEG11DOT5TONEG11DOT0" : "35a52222-598e-4960-b492-678bba0fce11" , \ "OUTNEG11DOT0TONEG10DOT0" : "72b3c2b5-bcab-4383-999e-d51dfddb998c" , \ "OUTNEG11DOT0TONEG10DOT5" : "82f9550e-28fa-46e1-930d-f693713aebad" , \ "OUTNEG10DOT5TONEG10DOT0" : "b1c39fb2-cbf7-47d1-b2af-b77a6cc1bdf0" , \ "OUTNEG10DOT0TONEG9DOT0" : "92822b83-935e-4156-b3f7-4808a6ced03e" , \ "OUTNEG10DOT0TONEG9DOT5" : "bb34bbe3-05fa-488e-b751-eb7acb5cec0a" , \ "OUTNEG9DOT5TONEG9DOT0" : "68326877-cb25-4648-ac83-4b76b896ff18" , \ "OUTNEG9DOT0TONEG8DOT0" : "c55c3123-bf1d-4f60-81b4-76f246cf8afd" , \ "OUTNEG9DOT0TONEG8DOT5" : "bb5c30d5-6b7a-4bf5-8964-37e2df41b59f" , \ "OUTNEG8DOT5TONEG8DOT0" : "bf7cca2e-dad7-4151-9aaf-f554085dd0c0" , \ "OUTNEG8DOT0TONEG7DOT0" : "abf8c152-efc7-4080-be00-156dea19fc66" , \ "OUTNEG8DOT0TONEG7DOT5" : "33e66e3f-b892-45a7-b561-686efb3297aa" , \ "OUTNEG7DOT5TONEG7DOT0" : "2b325745-4704-49d7-9dc7-b9de6777fafe" , \ "OUTNEG7DOT0TONEG6DOT0" : "9840c2ea-4e99-466c-90ef-d6719d105a97" , \ "OUTNEG7DOT0TONEG6DOT5" : "56172fd3-ed15-49f9-a4ff-1bf926b6434b" , \ "OUTNEG6DOT5TONEG6DOT0" : "763a52a7-f2e0-43c9-9b5b-bd5f5b927302" , \ "OUTNEG6DOT0TONEG5DOT0" : "3409e94e-aea4-42a5-840f-576e3c3220b7" , \ "OUTNEG6DOT0TONEG5DOT5" : "755c8c70-e985-408a-930a-f109eb5f8745" , \ "OUTNEG5DOT5TONEG5DOT0" : "2eea665f-7545-4ded-94bb-f7955dd54884" , \ "OUTNEG5DOT0TONEG4DOT0" : "954a4686-fa13-4446-b96b-2a6303411e46" , \ "OUTNEG5DOT0TONEG4DOT5" : "6d6294c7-58a5-4800-a438-cf3dcd03d2ad" , \ "OUTNEG4DOT5TONEG4DOT0" : "7fa6139b-7c1a-4d34-be1c-8fa1d1226a1c" , \ "OUTNEG4DOT0TONEG3DOT0" : "a4cee924-b854-4ead-9083-199942afccf4" , \ "OUTNEG4DOT0TONEG3DOT5" : "551bbac7-58bf-4945-a3ef-e96f516653e2" , \ "OUTNEG3DOT5TONEG3DOT0" : "160bff8b-3357-41d3-9d63-26b8e966179f" , \ "OUTNEG3DOT0TONEG2DOT0" : "8d8293ce-89ee-498f-9880-24bfa7420b12" , \ "OUTNEG3DOT0TONEG2DOT5" : "27c78ed8-050c-4316-8a0a-ae82e7954642" , \ "OUTNEG2DOT5TONEG2DOT0" : "51977ca9-b25f-496e-b4c0-af84d07311b8" , \ "OUTNEG2DOT0TONEG1DOT0" : "7b5e5b5f-f9eb-4c77-a891-d1d880cd261c" , \ "OUTNEG2DOT0TONEG1DOT5" : "08c6c569-07c6-4d6a-9841-8789041ef762" , \ "OUTNEG1DOT5TONEG1DOT0" : "4fc97087-d6c7-4276-acf5-549d0143cd46" , \ "OUTNEG1DOT0TO0DOT0" : "4601bca8-aa1b-41ad-b221-9facc8c038d2" , \ "OUTNEG1DOT0TONEG0DOT5" : "d43cfb34-8b7b-43ba-a972-e8f655db7611" , \ "OUTNEG0DOT5TO0DOT0" : "a70a708f-4909-43ae-be7a-63422934f4e3" , \ "OUT0DOT0TO1DOT0" : "c955d7aa-b9dd-42c6-a540-b114d46d2686" , \ "OUT0DOT0TO0DOT5" : "f5d67024-ce6e-4892-9945-9f9ad513ed08" , \ "OUT0DOT5TO1DOT0" : "a5df56a9-ad7f-4faf-995e-94b69d6f0ca5" , \ "OUT1DOT0TO2DOT0" : "e6730f20-bf15-4ad1-a06e-475988465844" , \ "OUT1DOT0TO1DOT5" : "4ecad452-6426-4249-a144-c47abaf2df34" , \ "OUT1DOT5TO2DOT0" : "4a44d303-21da-4610-b425-233c24e1e7da" , \ "OUT2DOT0TO3DOT0" : "9116e981-d0c5-4c0c-b453-8c85fc68f5ba" , \ "OUT2DOT0TO2DOT5" : "47396771-b87c-4484-a283-ed71ef5b2909" , \ "OUT2DOT5TO3DOT0" : "cadf35c6-ac3e-46eb-a963-c1ca43d4ae7a" , \ "OUT3DOT0TO4DOT0" : "5e121664-7763-4f88-8455-2ed4080fcc0e" , \ "OUT3DOT0TO3DOT5" : "5bef15be-d773-4dd9-b876-f31d9dfd254f" , \ "OUT3DOT5TO4DOT0" : "30499201-8c99-4707-89e7-fdd571bfbe96" , \ "OUT4DOT0TO5DOT0" : "37ab02df-76d3-4b5c-a01a-29ccfab28e55" , \ "OUT4DOT0TO4DOT5" : "f4ab6d8d-8bd6-41a8-9b17-f0f346af873c" , \ "OUT4DOT5TO5DOT0" : "cdfb3975-e1fc-436c-8bd6-4f9269506878" , \ "OUT5DOT0TO6DOT0" : "fd110be0-5997-4ee7-8f71-8e609598ec90" , \ "OUT5DOT0TO5DOT5" : "333fb950-24a9-4631-a3b4-df89d740ac47" , \ "OUT5DOT5TO6DOT0" : "df5b287c-eef1-4555-81ef-7a1a1e96ee4e" , \ "OUT6DOT0TO7DOT0" : "c567b7d5-e8fa-491a-8ce9-66e659c6ecef" , \ "OUT6DOT0TO6DOT5" : "87cdeb87-1156-4447-b0c1-6c17f3bbfb28" , \ "OUT6DOT5TO7DOT0" : "dd106603-e8ab-410a-adcb-5e850282e02b" , \ "OUT7DOT0TO8DOT0" : "032edbd0-0772-49f5-988e-31373608f571" , \ "OUT7DOT0TO7DOT5" : "3b9e6b42-1981-40b9-887f-3634d786ad70" , \ "OUT7DOT5TO8DOT0" : "5c278635-df0a-4a9c-8ec0-b3d155903b11" , \ "OUT8DOT0TO9DOT0" : "2b1c2cd3-49e5-4615-96e0-dd1f69554db9" , \ "OUT8DOT0TO8DOT5" : "1dc11ab2-cb24-4c96-9b3a-dbcd41fccd45" , \ "OUT8DOT5TO9DOT0" : "d11e9636-8daa-4d9b-9fb4-79618909a003" , \ "OUT9DOT0TO10DOT0" : "fe1af115-3d25-45d9-bd66-34b2113998f5" , \ "OUT9DOT0TO9DOT5" : "308a0f34-44f9-4d3a-a296-7b21c012c5f5" , \ "OUT9DOT5TO10DOT0" : "7a95baff-d474-4ce8-af32-e7581b3276a1" , \ "OUT10DOT0TO11DOT0" : "3fe00ffe-d076-4e6a-b8f4-0dfc1f965a21" , \ "OUT10DOT0TO10DOT5" : "cacf032c-3ffe-4f60-be62-66da9c97d998" , \ "OUT10DOT5TO11DOT0" : "48796d8e-e9e9-49a5-b279-93ac3a809cd2" , \ "OUT11DOT0TO12DOT0" : "35e61753-bf23-4b51-a7ba-b3b948c045c6" , \ "OUT11DOT0TO11DOT5" : "7f2389a0-b496-44c7-8ab9-a16b2667e31e" , \ "OUT11DOT5TO12DOT0" : "36a2ccd5-403d-4548-a55f-858a4cfaf6e7" , \ "OUT12DOT0TO13DOT0" : "50832367-90f7-4c5b-bf55-866635619adf" , \ "OUT12DOT0TO12DOT5" : "c1f15fd9-3bfb-41bd-9ece-aacf2edc4aef" , \ "OUT12DOT5TO13DOT0" : "1879b3ed-e7b5-403c-a36c-3e0ab972bc73" , \ "OUT13DOT0TO14DOT0" : "89ddf0f4-f62b-493c-a144-beeda5d0ef7e" , \ "OUT13DOT0TO13DOT5" : "4aa53bf2-8ee6-4ee8-8af1-216e508cee9f" , \ "OUT13DOT5TO14DOT0" : "49da7445-8626-453f-ae6f-879135a41b9d" , \ "OUT14DOT0TO15DOT0" : "313318a4-4672-41ac-b6ba-341e6cee6dc3" , \ "OUT14DOT0TO14DOT5" : "56e63bd6-a515-4cea-9ca8-3fed50420ced" , \ "OUT14DOT5TO15DOT0" : "cbaab60b-aeb1-4a7a-974c-19e4fe94f1e9" , \ "OUT15DOT0TO16DOT0" : "c320030d-5099-4e63-b45a-139b8fd74e8e" , \ "OUT15DOT0TO15DOT5" : "7f814059-80b2-41a1-ba50-d5e2bf92ff4a" , \ "OUT15DOT5TO16DOT0" : "a47bc767-56a9-4b66-b24d-d50d4b6fe811" , \ "OUT16DOT0TO17DOT0" : "5a46ce6e-d2ca-4688-8a7e-88a8cb896f6c" , \ "OUT16DOT0TO16DOT5" : "5117bf37-c52e-4822-a76d-e939f4a6e001" , \ "OUT16DOT5TO17DOT0" : "a8930162-48e7-4d67-ae42-c669c57d6faf" , \ "OUT17DOT0TO18DOT0" : "9a972a7e-a831-4d13-9206-58275a771029" , \ "OUT17DOT0TO17DOT5" : "c360b4ac-84c5-4b78-87a7-c6d5ad4a4ae3" , \ "OUT17DOT5TO18DOT0" : "933c15a3-ecae-42bb-9411-9e4b7097e4e5" , \ "OUT18DOT0TO19DOT0" : "d23386ce-18e8-4c59-a4bd-57d4b6c1e9cd" , \ "OUT18DOT0TO18DOT5" : "08c56c57-5a47-43b7-a07f-dcc99ac32ead" , \ "OUT18DOT5TO19DOT0" : "73531737-bdb7-4d10-85f1-dd53502ecdd1" , \ "OUT19DOT0TO20DOT0" : "877d7acd-2021-4573-9bdb-9e7c44c8d83b" , \ "OUT19DOT0TO19DOT5" : "8cb80ff2-8ce0-4aa7-a917-ab049f29ab70" , \ "OUT19DOT5TO20DOT0" : "b73bd413-2e4a-48db-bf8a-046488a837c8" \ }; # Note: manualAbstractionInformation, generally speaking, is a # structured used purely in analysis scripts (as developed for # the paper describing Fanoos); placing this information # in the class defining the domain proved to be a convieniant place to store the # information during the time of development and testing. Fanoos does not access # the information in manualAbstractionInformation when determining how to make # adjustments to respond to users. Again, it is only used in analysis scripts # used to prepare results for the paper. While this sanity-checking # code does not have results discussed in the paper at the time of # writting this comment, we needed to fill the information for # this structure; while Fanoos itself does not examin content in # manualAbstractionInformation, some code (such as checking code, e.g., contracts) # expect the structure to be present and obey basic properties such as # number of entries. # # While it was convieniant for development, clearly it is not # ideal have this data stored here or this structure required # to be present. TODO: resolve the issue just described. self.manualAbstractionInformation = {\ "predicatesAndLabels" : [\ ("INNEG10DOT0TONEG9DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG10DOT0TONEG9DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG9DOT5TONEG9DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG9DOT0TONEG8DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG9DOT0TONEG8DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG8DOT5TONEG8DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG8DOT0TONEG7DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG8DOT0TONEG7DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG7DOT5TONEG7DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG7DOT0TONEG6DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG7DOT0TONEG6DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG6DOT5TONEG6DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG6DOT0TONEG5DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG6DOT0TONEG5DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG5DOT5TONEG5DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG5DOT0TONEG4DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG5DOT0TONEG4DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG4DOT5TONEG4DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG4DOT0TONEG3DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG4DOT0TONEG3DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG3DOT5TONEG3DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG3DOT0TONEG2DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG3DOT0TONEG2DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG2DOT5TONEG2DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG2DOT0TONEG1DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG2DOT0TONEG1DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG1DOT5TONEG1DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG1DOT0TO0DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("INNEG1DOT0TONEG0DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("INNEG0DOT5TO0DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN0DOT0TO1DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN0DOT0TO0DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN0DOT5TO1DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN1DOT0TO2DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN1DOT0TO1DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN1DOT5TO2DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN2DOT0TO3DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN2DOT0TO2DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN2DOT5TO3DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN3DOT0TO4DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN3DOT0TO3DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN3DOT5TO4DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN4DOT0TO5DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN4DOT0TO4DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN4DOT5TO5DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN5DOT0TO6DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN5DOT0TO5DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN5DOT5TO6DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN6DOT0TO7DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN6DOT0TO6DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN6DOT5TO7DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN7DOT0TO8DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN7DOT0TO7DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN7DOT5TO8DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN8DOT0TO9DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN8DOT0TO8DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN8DOT5TO9DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN9DOT0TO10DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("IN9DOT0TO9DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("IN9DOT5TO10DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG20DOT0TONEG19DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG20DOT0TONEG19DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG19DOT5TONEG19DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG19DOT0TONEG18DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG19DOT0TONEG18DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG18DOT5TONEG18DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG18DOT0TONEG17DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG18DOT0TONEG17DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG17DOT5TONEG17DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG17DOT0TONEG16DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG17DOT0TONEG16DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG16DOT5TONEG16DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG16DOT0TONEG15DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG16DOT0TONEG15DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG15DOT5TONEG15DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG15DOT0TONEG14DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG15DOT0TONEG14DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG14DOT5TONEG14DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG14DOT0TONEG13DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG14DOT0TONEG13DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG13DOT5TONEG13DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG13DOT0TONEG12DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG13DOT0TONEG12DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG12DOT5TONEG12DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG12DOT0TONEG11DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG12DOT0TONEG11DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG11DOT5TONEG11DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG11DOT0TONEG10DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG11DOT0TONEG10DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG10DOT5TONEG10DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG10DOT0TONEG9DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG10DOT0TONEG9DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG9DOT5TONEG9DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG9DOT0TONEG8DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG9DOT0TONEG8DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG8DOT5TONEG8DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG8DOT0TONEG7DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG8DOT0TONEG7DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG7DOT5TONEG7DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG7DOT0TONEG6DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG7DOT0TONEG6DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG6DOT5TONEG6DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG6DOT0TONEG5DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG6DOT0TONEG5DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG5DOT5TONEG5DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG5DOT0TONEG4DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG5DOT0TONEG4DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG4DOT5TONEG4DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG4DOT0TONEG3DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG4DOT0TONEG3DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG3DOT5TONEG3DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG3DOT0TONEG2DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG3DOT0TONEG2DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG2DOT5TONEG2DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG2DOT0TONEG1DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG2DOT0TONEG1DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG1DOT5TONEG1DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG1DOT0TO0DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUTNEG1DOT0TONEG0DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUTNEG0DOT5TO0DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT0DOT0TO1DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT0DOT0TO0DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT0DOT5TO1DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT1DOT0TO2DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT1DOT0TO1DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT1DOT5TO2DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT2DOT0TO3DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT2DOT0TO2DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT2DOT5TO3DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT3DOT0TO4DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT3DOT0TO3DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT3DOT5TO4DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT4DOT0TO5DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT4DOT0TO4DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT4DOT5TO5DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT5DOT0TO6DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT5DOT0TO5DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT5DOT5TO6DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT6DOT0TO7DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT6DOT0TO6DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT6DOT5TO7DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT7DOT0TO8DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT7DOT0TO7DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT7DOT5TO8DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT8DOT0TO9DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT8DOT0TO8DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT8DOT5TO9DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT9DOT0TO10DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT9DOT0TO9DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT9DOT5TO10DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT10DOT0TO11DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT10DOT0TO10DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT10DOT5TO11DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT11DOT0TO12DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT11DOT0TO11DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT11DOT5TO12DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT12DOT0TO13DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT12DOT0TO12DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT12DOT5TO13DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT13DOT0TO14DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT13DOT0TO13DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT13DOT5TO14DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT14DOT0TO15DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT14DOT0TO14DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT14DOT5TO15DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT15DOT0TO16DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT15DOT0TO15DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT15DOT5TO16DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT16DOT0TO17DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT16DOT0TO16DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT16DOT5TO17DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT17DOT0TO18DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT17DOT0TO17DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT17DOT5TO18DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT18DOT0TO19DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT18DOT0TO18DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT18DOT5TO19DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT19DOT0TO20DOT0" , "84c05409-74a7-4d75-8ab6-daf96f76f675"), \ ("OUT19DOT0TO19DOT5" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4"), \ ("OUT19DOT5TO20DOT0" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4") \ ], \ "labelDag_firstParent_secondChild" : [ \ ("84c05409-74a7-4d75-8ab6-daf96f76f675" , "d7082112-67c2-42c0-a9e9-df89f1ad86d4") \ ] \ }; self.manualAbstractionInformation["predicatesAndLabels"] = \ [ (dictMappingPredicateStringNameToUUID[x[0]] , x[1]) for x in self.manualAbstractionInformation["predicatesAndLabels"]]; functToGetUuidProvided = (lambda predicateObjectBeingInitialized : dictMappingPredicateStringNameToUUID[str(predicateObjectBeingInitialized)] ); self.initializedConditions = \ [CharacterizationCondition_FromPythonFunction(z3SolverInstance, DomainFor_modelForTesting_oneDimInput_oneDimOutput, x, functToGetUuidProvided=functToGetUuidProvided) \ for x in getListFunctionsToBaseCondtionsOn_forInputOfDomainThisUse() + \ getListFunctionsToBaseCondtionsOn_forOutputOfDomainThisUse() + \ getListFunctionsToBaseCondtionsOn_forJointInputAndOutputDomainsInThisUse() ]; assert(all([ (x.getID() == functToGetUuidProvided(x)) for x in self.initializedConditions])); self._writeInfoToDatabase(); return; def getBaseConditions(self): return self.initializedConditions; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # class-specific utilities for defining domains #=========================================================================== def getFiniteInterval(variableType, nameOfPredicate, lowerBound, upperBound): requires(isinstance(nameOfPredicate, str)); requires(variableType in {"input", "output"}); requires(isinstance(nameOfPredicate, str)); requires(len(nameOfPredicate) > 0); requires(len(set(nameOfPredicate).intersection([" ", "\n", "\r", "\t"])) == 0); requires(isinstance(lowerBound, float)); requires(isinstance(upperBound, float)); requires(np.isfinite(lowerBound)); requires(np.isfinite(upperBound)); requires(lowerBound <= upperBound); templateString = """ def funct_{0}({1}): \"\"\"{0}\"\"\" if(isinstance({1}, z3.z3.ArithRef)): return z3.And( {1} <= {3}, {1} >= {2} ); else: return ({1} <= {3}) and ({1} >= {2} ); raise Exception("Control should not reach here"); return; """; variableNameString = "in_x"; if(variableType == "output"): variableNameString = "out_y"; assert(variableNameString in {"in_x", "out_y"}); return templateString.format(nameOfPredicate, variableNameString, str(lowerBound), str(upperBound) ); # The below function in principle could be done with getFiniteInterval if z3 supported infinite values, but its standard theory does not seem to support # them, which, honestly, is reasonable. def getInfiniteInterval(variableType, nameOfPredicate, boundary, aboveOrBelow): requires(isinstance(nameOfPredicate, str)); requires(variableType in {"input", "output"}); requires(isinstance(nameOfPredicate, str)); requires(len(nameOfPredicate) > 0); requires(len(set(nameOfPredicate).intersection([" ", "\n", "\r", "\t"])) == 0); requires(isinstance(boundary, float)); requires(np.isfinite(boundary)); requires(aboveOrBelow in {"lowerBound", "upperBound"}); templateString = """ def funct_{0}({1}): \"\"\"{0}\"\"\" if(isinstance({1}, z3.z3.ArithRef)): return {2} <= {3}; else: return {2} <= {3}; raise Exception("Control should not reach here"); return; """; variableNameString = "in_x"; if(variableType == "output"): variableNameString = "out_y"; assert(variableNameString in {"in_x", "out_y"}); stringToReturn = ""; if(aboveOrBelow == "upperBound"): stringToReturn = templateString.format(nameOfPredicate, variableNameString, variableNameString, str(boundry) ) else: assert(aboveOrBelow == "lowerBound"); stringToReturn = templateString.format(nameOfPredicate, variableNameString, str(boundry), variableNameString) return stringToReturn ; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the input domain #=========================================================================== otherInputSpaceFunctionsToUse = """ """ def getListFunctionsToBaseCondtionsOn_forInputOfDomainThisUse(): listOfFunctionCodes =[]; def boundsToName(lower, upper): A = str(lower).replace(".", "DOT").replace("-", "NEG"); B = str(upper).replace(".", "DOT").replace("-", "NEG"); return "IN" + A + "TO" + B; def formPredicateHere(lower, upper): return getFiniteInterval("input", boundsToName(lower, upper), lower, upper); for thisStartIndex in range(-10, 10): thisStartIndex = float(thisStartIndex); upperIndex = thisStartIndex + 1.0; middleIndex = thisStartIndex + 0.5; # nameOfPredicate, lowerBound, upperBound) listOfFunctionCodes = listOfFunctionCodes + \ [ formPredicateHere(thisStartIndex, upperIndex), formPredicateHere(thisStartIndex, middleIndex), formPredicateHere(middleIndex, upperIndex) ]; return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the output domain #=========================================================================== otherOutputSpaceFunctionsToUse = """ """; def getListFunctionsToBaseCondtionsOn_forOutputOfDomainThisUse(): listOfFunctionCodes =[]; def boundsToName(lower, upper): A = str(lower).replace(".", "DOT").replace("-", "NEG"); B = str(upper).replace(".", "DOT").replace("-", "NEG"); return "OUT" + A + "TO" + B; def formPredicateHere(lower, upper): return getFiniteInterval("output", boundsToName(lower, upper), lower, upper); for thisStartIndex in range(-20, 20): thisStartIndex = float(thisStartIndex); upperIndex = thisStartIndex + 1.0; middleIndex = thisStartIndex + 0.5; # nameOfPredicate, lowerBound, upperBound) listOfFunctionCodes = listOfFunctionCodes + \ [ formPredicateHere(thisStartIndex, upperIndex), formPredicateHere(thisStartIndex, middleIndex), formPredicateHere(middleIndex, upperIndex) ]; return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the joint domain #=========================================================================== def getBox(nameOfPredicate, lowerBoundInput, upperBoundInput, lowerBoundOutput, upperBoundOutput): requires(isinstance(lowerBoundInput, float)); requires(isinstance(upperBoundInput, float)); requires(isinstance(lowerBoundOutput, float)); requires(isinstance(upperBoundOutput, float)); requires(np.isfinite(lowerBoundInput)); requires(np.isfinite(upperBoundInput)); requires(np.isfinite(lowerBoundOutput)); requires(np.isfinite(upperBoundOutput)); requires(lowerBoundInput <= upperBoundInput); requires(lowerBoundOutput <= upperBoundOutput); templateString = """ def funct_{0}(in_x, out_y): \"\"\"{0}\"\"\" if(isinstance(in_x, z3.z3.ArithRef)): return z3.And( in_x <= {3}, in_x >= {2}, out_y <= {5}, out_y >= {4} ); else: return (in_x <= {3}) and (in_x >= {2} ) and (out_y <= {5}) and (out_y >= {4}); raise Exception("Control should not reach here"); return; """; return templateString.format(nameOfPredicate, str(lowerBoundInput), str(upperBoundInput), str(lowerBoundOutput), str(upperBoundOutput) ); # circle, # halfplane # negation (or maybe just allow the user to pass in the inequality.. but actually negation would be useful for things later on... ) def getHalfPlane(nameOfPredicate, slope, intercept, inequality): requires(isinstance(slope, float)); requires(isinstance(intercept, float)); requires(np.isfinite(intercept)); requires(np.isfinite(slope)); requires(isinstance(inequality, str)); requires(inequality in {"=<", "=>", "<", ">"}); templateString = """ def funct_{0}(in_x, out_y): \"\"\"{0}\"\"\" return in_x * {1} + {2} {3} out_y ; raise Exception("Control should not reach here"); return; """; return templateString.format(nameOfPredicate, str(slope), str(intercept), str(inequality)); def getCicle(nameOfPredicate, in_x_center, out_y_center, radius, inequality): requires(isinstance(in_x_center, float)); requires(np.isfinite(in_x_center)); requires(isinstance(out_y_center, float)); requires(np.isfinite(out_y_center)); requires(isinstance(radius, float)); requires(np.isfinite(radius)); requires(isinstance(inequality, str)); requires(inequality in {"=<", "=>", "<", ">"}); templateString = """ def funct_{0}(in_x, out_y): \"\"\"{0}\"\"\" return (in_x - {1}) ** 2 + (out_y - {2}) {3} {4} ; raise Exception("Control should not reach here"); return; """; return templateString.format(nameOfPredicate, str(in_x_center), str(out_y_center), str(inequality), str(radius ** 2) ); def getListFunctionsToBaseCondtionsOn_forJointInputAndOutputDomainsInThisUse(): listOfFunctionCodes =[]; return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^
41,105
53.954545
179
py
Fanoos
Fanoos-master/domainsAndConditions/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/domainsAndConditions/domainAndConditionsForInvertedDoublePendulum.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import re; import z3; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass, CharacterizationCondition_FromPythonFunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; from domainsAndConditions.utilsForDefiningPredicates import *; import uuid; class DomainForInvertedDoublePendulum(BaseClassDomainInformation): def __init__(self, z3SolverInstance): requires(isinstance(z3SolverInstance, z3.z3.Solver)); self.initializedConditions = None; self.initialize_baseConditions(z3SolverInstance); assert(self.initializedConditions != None); return; @staticmethod def getUUID(): return "8a324d03-1391-4f4e-9589-4e4978fb4fbb"; @staticmethod def getInputSpaceUniverseBox(): # physical model specification at: bullet3/examples/pybullet/gym/pybullet_data/mjcf/inverted_double_pendulum.xml # commit 3d87fb3b84eb9faed580ef7b69bb4cb9cb693907 # state specification at: bullet3/examples/pybullet/gym/pybullet_envs/robot_pendula.py # commit 3d87fb3b84eb9faed580ef7b69bb4cb9cb693907 """ examples/pybullet/gym/pybullet_envs/robot_pendula.py def apply_action(self, a): assert (np.isfinite(a).all()) self.slider.set_motor_torque(200 * float(np.clip(a[0], -1, +1))) def calc_state(self): theta, theta_dot = self.j1.current_position() gamma, gamma_dot = self.j2.current_position() x, vx = self.slider.current_position() self.pos_x, _, self.pos_y = self.pole2.pose().xyz() assert (np.isfinite(x)) return np.array([ x, vx, self.pos_x, np.cos(theta), np.sin(theta), theta_dot, np.cos(gamma), np.sin(gamma), gamma_dot, ]) """ # Some notes on how the bounds below were considered: """ We purposefully enlarge the range of some of the variables below, so to explore what the robot would do in situations it would not normally find itself in (since it controls its environment from the beginning of the run. <redacted some code path used to generate below; aim to add in later commit> ==================================== Statistics for the input domain. Variable order: ["x", "vx", "endOfPole2_x", "pole1Angle", "pole1Angle_rateOfChange", "pole2Angle", "pole2Angle_rateOfChange" ] See the function pushBoxThrough in <redacted some code path used to generate below; aim to add in later commit> median [-9.17087654e-01 -1.01356242e-01 -9.16462040e-01 1.11590028e-03 -1.41741023e-02 6.40714566e-05 3.43284677e-03] mean [-8.36584234e-01 -5.55825160e-02 -8.36006938e-01 -3.95094858e-04 -1.10845849e-04 4.09024498e-05 1.50040207e-04] std [0.24976108 0.21574221 0.24066486 0.0217573 0.24092117 0.00551252 0.2698615 ] min [-0.97297198 -0.75622999 -0.96496589 -0.09813181 -0.3220316 -0.03076772 -0.60042353] 5% quantile [-0.96169371 -0.57010988 -0.96055381 -0.03824488 -0.28818685 -0.00612778 -0.31254109] 10% quantile [-0.95352224 -0.21671905 -0.95255887 -0.0086053 -0.28213323 -0.0053891 -0.30915195] 25% quantile [-0.92539008 -0.14963331 -0.92511681 -0.00406909 -0.249706 -0.00275462 -0.29033308] 75% quantile [-0.91029063 0.12560582 -0.91298939 0.00584924 0.244057 0.00259336 0.28829787] 90% quantile [-0.57014145 0.1653117 -0.53118565 0.01190971 0.28264124 0.00525244 0.308271 ] 95% quantile [-0.012156 0.16972019 -0.07910288 0.03293236 0.28733768 0.00616469 0.31369203] max [0.05946957 0.23917278 0.00220712 0.05317138 0.51959322 0.02087447 0.33615193] ==================================== Statistics on abs of the input domain. Variable order: ["x", "vx", "endOfPole2_x", "pole1Angle", "pole1Angle_rateOfChange", "pole2Angle", "pole2Angle_rateOfChange" ] median [0.91708765 0.14007734 0.91646204 0.00463563 0.2465699 0.00271302 0.28928026] mean [0.83944368 0.17516411 0.83607112 0.01119903 0.22970913 0.00367768 0.2486114 ] std [0.23997518 0.137665 0.24044178 0.01865791 0.07264117 0.00410661 0.10496485] min [0. 0. 0. 0. 0. 0. 0.] 5% quantile [0.05787052 0.09214573 0.07910288 0.00050955 0.04641253 0.00023093 0.01096292] 10% quantile [0.57014145 0.10333252 0.53118565 0.00103595 0.09850179 0.00058252 0.02011684] 25% quantile [0.91029063 0.11111132 0.91298939 0.00248957 0.22880583 0.0012283 0.27657566] 75% quantile [0.92539008 0.16597374 0.92511681 0.00836597 0.27732796 0.00477952 0.30694401] 90% quantile [0.95352224 0.23168947 0.95255887 0.03493198 0.28769558 0.00616469 0.31307735] 95% quantile [0.96169371 0.57010988 0.96055381 0.05266948 0.29233937 0.00957688 0.31733186] max [0.97297198 0.75622999 0.96496589 0.09813181 0.51959322 0.03076772 0.60042353] ==================================== Statistics on the input domain. Variable order: [action,value ] median [0.01559919 2.84733465] mean [4.08913092e-05 2.88180130e+00] std [0.89601406 0.07568831] min [-1. 2.8295681] 5% quantile [-1. 2.8341802] 10% quantile [-1. 2.83570324] 25% quantile [-1. 2.84043188] 75% quantile [1. 2.87173254] 90% quantile [1. 3.04098518] 95% quantile [1. 3.0745709] max [1. 3.09394765] ==================================== Statistics on abs of the input domain. Variable order: [action,value ] median [1. 2.84733465] mean [0.81950511 2.8818013 ] std [0.36228795 0.07568831] min [2.33136512e-05 2.82956810e+00] 5% quantile [0.02753463 2.8341802 ] 10% quantile [0.05367255 2.83570324] 25% quantile [0.96223713 2.84043188] 75% quantile [1. 2.87173254] 90% quantile [1. 3.04098518] 95% quantile [1. 3.0745709] max [1. 3.09394765] """ """ V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~~V~V~V~V~V~V~V~V~V SOME NOTES ABOUT endOfPole2_x : ========================================================================== The representation we feed into the policy-frontend (see convertInvertedPendulumNetworkToAlternateFormat.py) uses endOfPole2_x as the delta between the measured-x-value of pole2 - roughly the x-value of the "end" of pole2, but again see convertInvertedPendulumNetworkToAlternateFormat.py for further description - and the x-value of the robot's main cart. The policy itself, however, expects endOfPole2_x to be in respect to the global space. The conversion is trivial addition between x and endOfPole2_x - but again, here, we use endOfPole2_x as relative to x , while the policy accepts endOfPole2_x in respect to the global coordinates. A deeper issue that this illuminates is the fact that endOfPole2_x is certainly not independant from the other input values - in fact, given the phyisical parameters of the robot have been fixed already, endOfPole2_x should entirely be a function of x, pole1Angle and pole2Angle . Arguable, the rest of the input values are actually independant of the other input values - endOfPole2_x is the only one that seems to have this clear dependancy, which limits which points in the overall input space are actually possible. We note here that it should be the case that: endOfPole2_x = 0.3 * sin(pole2Angle + pole1Angle) + 0.6 * sin(pole1Angle) + x ; The 0.6 is the length of pole1. Pole 2 also has this length, but the bullet simulator returns position in respect to the center of mass roughly. One can find where there values are specified and discussion of their use in inverted_double_pendulum.xml and gym_pendulum_envs.py , particularly the function step in the class InvertedDoublePendulumBulletEnv , located in the latter file . =============== The reason we have sin here as oppossed to cos is because pole1Angle is take in respect to the vertical (e.g., pole1Angle corresponds to being parrallel with the y-axis). <redactions of some content here> So, sin(pole1Angle) actually corresponds to the normalized x-projection of pole1, and similar can be said for sin(pole2Angle + pole1Angle), noting that pole2Angle is measured relative to pole1Angle (i.e., pole2Angle == 0 means that pole2 aligns with pole1 - if pole1 is at pi*0.7 radians in respect to the global coordinates and pole2Angle is zero, then pole2 is at pi*0.7 radians in respect to the global coordinates). With this formula noted, it would be possible to remove the variable endOfPole2_x from the input space by modifying convertInvertedPendulumNetworkToAlternateFormat.py to derive a box for the value based on x, pole1Angle, and pole2Angle (similar to what is done to get the sin and cos for the angles which the policy expects) . The code to produce a box over endOfPole2_x in this fashion should be straight-forward given that code already exists in that file to produce boxes over the sin of an angle ; the formula for the box over endOfPole2_x would easily be box(endOfPole2_x) = 0.3 * box( sin(box(pole2Angle) + box(pole1Angle)) ) + 0.6 * box( sin( box(pole1Angle) ) ) + box(x) ; while the "uncertainty"/size of box(endOfPole2_x) might be larger than for the rest of the input boxes due to these summations, there is little that can be helped about it due to the deterministic nature of the boxes. Another possibliy would be to expand the interface for the domains to include a function for rejecting boxes from the CEGAR-analysis due to them not being physically possible. That would help alivate construction of the box-propagator (the thing that shoves boxes through the learned component we are considering) and what input space the box-propagator expects from issues of the input space having internal dependancies. In general, having too many factors to consider when designing the box-propagator or the input space that the reachability eats-away at some of the point of this code, which aims to handle these sort of concerns for the user - not expect the user to already to have had to solve some of them prior to using Fanoos. But this may be a softer point - any given analysis stem has its strengths, weakness, barriers, etc. ... not to excuse the weaknesses though... This easily leads to a digression about analyzing complex, some commentary on-which can be found in appendix A of the ArXiv writeup (also, the appendix of the IJCAI-XAI writeup). For now we keep endOfPole2_x in the explicit input-space for the reachability. For one thing, doing it this way removes having to explictly embed arm-length in the code; really, this is a silly point since it is almost certainly the case that the trained policy in part relies on these arm-length factors. Another benefit, however, is that it is easier to: (1) have the system pick up properties of the controller that depend on endOfPole2_x for the most-part- given how endOfPole2_x was likely included despite its technical redundancy because its informative merits (e.g., which side of the base pole2 might be extending past), how the policy reacts in respect to this parameter seems like something reasonbly likely to be worth discussing. (2) similar to (1), but mostly from a coding and z3-runtime perspective, it is easier to define predicates that use endOfPole2_x when it is an explicit member of the space. All the points I just mentioned provide a good argument for extending the interface of the domains to filter boxes based on physical realizability, as oppossed to shoving calculations under the hood to implicitly - as oppossed to explictly - consider variables of importance to the policy. A counter-argument to this, however, is the reasonable work I did on this before, keeping the reachability-end using the angle-representation below as oppossed to the sin-and-cos representation the policy itself expects. ....hm.... anyway, for now, we keep endOfPole2_x as part of the input space, just tweaking the representation we use here to be a delta-value as oppossed to a global coordinate, as mentioned above. ^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ """ orderOfVariables = __class__.inputSpaceVariables(); dictMappingVariableToBound = {\ "x" : [-1 , 1], \ "vx" : [-0.8 , 0.8], \ "endOfPole2_x" : [-0.5, 0.5], \ "pole1Angle" : [-0.2 , 0.2], \ "pole1Angle_rateOfChange" : [-0.6 , 0.6], \ "pole2Angle" : [-0.04, 0.04], \ "pole2Angle_rateOfChange" : [-0.7, 0.7]\ }; thisUniverseBox = __class__._helper_getInputSpaceUniverseBox(\ orderOfVariables, dictMappingVariableToBound); ensures(getDimensionOfBox(thisUniverseBox) == len(DomainForInvertedDoublePendulum.inputSpaceVariables())); return thisUniverseBox; @staticmethod def inputSpaceVariables(): return [\ z3.Real(x) for x in ["x", "vx", "endOfPole2_x", \ "pole1Angle", "pole1Angle_rateOfChange", \ "pole2Angle", "pole2Angle_rateOfChange" ] ]; @staticmethod def outputSpaceVariables(): return [z3.Real("outputTorque"), z3.Real("stateValueEstimate")]; @staticmethod def getName(): return "Domain For Inverted Double Pendulum" def initialize_baseConditions(self, z3SolverInstance): dictMappingPredicateStringNameToUUID = \ { "x Very High Magnitude" : "21b3fb5b-efd9-4406-a7fe-b80937881e21", "x High Magnitude" : "f3c2655f-fa5c-4020-9dc8-c20a47401796", "x Low Magnitude" : "0155f83a-ef30-4a61-adda-c1321480240f", "x Very Low Magnitude" : "8de9828f-6ec7-4240-8293-b76d3d92f3f2", "x Magnitude Near Normal Levels" : "7389a08e-03d4-4c89-8b58-3a2c7c203933", "vx Very High Magnitude" : "f3c56183-c013-4b2e-b11d-65244231dadd", "vx High Magnitude" : "e7f2daf6-b041-4f4a-88b9-999a4590cfc0", "vx Low Magnitude" : "f361d7bc-d6a7-4692-b38f-593e4f403d15", "vx Very Low Magnitude" : "23cd37d1-e48b-41f5-8f37-5988e9410d0f", "vx Magnitude Near Normal Levels" : "69e9a231-c08e-498a-b1cb-08def786cb0c", "endOfPole2_x Very High Magnitude" : "701785bd-f91f-4818-a997-e68258f56edb", \ "endOfPole2_x High Magnitude" : "af4c0da5-b6d2-4500-8fa0-117cb812e182", \ "endOfPole2_x Low Magnitude" : "44dca983-9b7f-4516-b9a7-62a47aa259d5", \ "endOfPole2_x Very Low Magnitude" : "c88184e3-6914-43ae-920b-5e7f47b5c63c", \ "endOfPole2_x Magnitude Near Normal Levels" : "f30f1e19-a9f7-4c7f-a852-14afc5a4b9af", \ "pole1Angle Very High Magnitude" : "1ea9d650-b3e8-4d07-b80d-2ee64e2ee09b", "pole1Angle High Magnitude" : "f0b1a37b-f937-41e1-907f-1faec3c9a1dd", "pole1Angle Low Magnitude" : "7661646a-34db-45a8-95d6-e8d3da8ddf4d", "pole1Angle Very Low Magnitude" : "8640c6d3-12cc-438e-adf3-b84583faab65", "pole1Angle Magnitude Near Normal Levels" : "9dfc2e8e-4915-4172-b856-b5b20946c393", "pole1Angle_rateOfChange Very High Magnitude" : "b7d2f662-2205-41a2-a920-cf674a09e8e4", "pole1Angle_rateOfChange High Magnitude" : "0bb61b1e-4816-4d98-99c7-41e02c85e011", "pole1Angle_rateOfChange Low Magnitude" : "491fc8a3-d091-48c6-a132-56c544cd181f", "pole1Angle_rateOfChange Very Low Magnitude" : "41a5302e-8e9c-4e63-8a29-fa2b11060e21", "pole1Angle_rateOfChange Magnitude Near Normal Levels" : "a5c62d0b-386c-4e1b-bc5b-40f27e7c7cca", "pole2Angle Very High Magnitude" : "960482b5-8323-40c7-8e0a-07b64ef6e75b", "pole2Angle High Magnitude" : "44c88c08-c354-401d-a100-ae94f573391d", "pole2Angle Low Magnitude" : "2a9b0032-a39f-4faa-9664-af5209745722", "pole2Angle Very Low Magnitude" : "f95a4c4a-6cec-49fd-abf1-64f9283f33a5", "pole2Angle Magnitude Near Normal Levels" : "44268c62-371f-44d4-a966-5a4e8d588d7d", "pole2Angle_rateOfChange Very High Magnitude" : "17e5a6e2-37fc-4523-aaee-9b6ac7384c0d", "pole2Angle_rateOfChange High Magnitude" : "d3d548ee-292a-43f4-99c3-1b9cbcd3ca71", "pole2Angle_rateOfChange Low Magnitude" : "4fef3e42-87f7-49d9-8e59-d9a70f52b906", "pole2Angle_rateOfChange Very Low Magnitude" : "60e5a29e-1ae8-47d2-b062-d537c089e11f", "pole2Angle_rateOfChange Magnitude Near Normal Levels" : "35c790a9-f001-4dd4-9869-f855e0e2a5c4", "pole1 Angle At of Above X Axis " : "40d2185a-0a53-46fb-82c4-2df793ac2f99", "pole1 Moving Counter-Clockwise " : "78546928-bda7-4354-be99-7bf2dc6a2f07", "Both poles pointed to the left" : "15ab5d6a-7038-4c65-b429-82de8f575486", "Anchor Point Moving Left" : "12247f40-220f-49ed-bc3c-ef38c6c54be9", "pole1 Moving Clockwise " : "2871af7f-a7cc-46ce-8527-00dfc5e0956f", "Pole2 Moving Counter-Clockwise " : "015fb99b-fd06-4eae-9b73-95449f12dc51", "pole1 Angle At of Above X Axis " : "d5886347-c798-4da7-86e5-61feb4b87ab8", "pole2 On Left" : "1757016b-4dec-4310-80fa-4be7580c7afc", "Pole1 On Left" : "9154a8c3-e99a-42f5-81aa-544b176bed74", "Anchor Point Moving Right" : "7b62d162-2820-464b-964d-5fc853f9724e", "Anchor Point Barely Moving If At All" : "ed362df9-9e7c-4d32-9249-84c8746ffd01", "Pole2 Moving Clockwise " : "41b0753a-ab87-4157-bc27-65c63c0b242b", "Pole2 Angle At of Above X Axis " : "b077175f-d944-4d1a-a1da-323677470aa8", "Pole1 On Right" : "34e8cd54-ff78-449e-9c2b-75813da7c8b2", "pole2 Close to Vertical" : "7e63953e-d315-43fc-a5e8-1c3a1fc2a43a", "Both poles pointed to the right" : "d12ad5c6-5e8e-4732-b8ae-07ee867927d9", "pole1 Moving Barely" : "e1c004da-1dcd-4ece-82b5-448a0216e50a", "Poles are Bent Like the arc in a a D" : "2dd966e1-5043-4762-bb5a-929e12c66e56", "Poles are Roughly Straight in Respect To Each Other" : "964ab2ae-c908-4e32-b933-b84b91556688", "pole2 On Right" : "90e03923-c6ff-4450-9a9f-84ea3576fa59", "Pole2 Moving Barely" : "60f1be4c-c509-403b-86e2-1e9b6623d8cd", "Pole1 Close to Vertical" : "761c320e-c078-4c1a-92f0-a8c4aa582c5d", "Poles are Bent" : "6d7ef156-be5c-4744-9a3f-22f7c7a3c2f8", "Poles are Bent Like the Arc in a C" : "ecb3670b-62aa-4ba1-9992-7eb6af738e68", "outputTorque Very High Magnitude" : "b1aaa0bf-5924-47e8-b9a6-cf56144b1b62", "outputTorque High Magnitude" : "e32a60e7-0a21-4220-a2cd-95f9cf87f8cb", "outputTorque Low Magnitude" : "42e9fef2-1928-4254-ac42-1b33d8abbf97", "outputTorque Very Low Magnitude" : "fecc6e90-4737-4648-bcf6-c70ed6433fb1", "outputTorque Magnitude Near Normal Levels" : "695de8ac-8a76-409d-8941-63f65753cce0", "stateValueEstimate Very High " : "761dc8bb-feab-439d-9ec8-a59be174ef11", "stateValueEstimate High " : "238126d9-bf56-4945-9c4f-c9e04f7dda4a", "stateValueEstimate Low " : "91982534-0a66-4530-886a-89f1be12b93a", "stateValueEstimate Very Low " : "fa4a360c-9613-484e-a212-b9cb314bf2a1", "stateValueEstimate Near Normal Levels" : "d496f7b5-6dfb-47b8-a1e9-2ead59f667fc", "Speed Constant Assuming No Friction" : "91151f23-dea7-4298-871f-b6b3f0088d95", "Speed Decreasing Assuming No Friction" : "9f21d925-53ed-413c-b633-e4026c102a1e", "Speed Increasing Assuming No Friction" : "30298f0e-3fc0-4494-8daf-d94628226276", "Pole 2 is on the right of the robot chassy" : "4370399d-f9a6-4eb2-ab5e-1eed108c58de", \ "outputTorque is less than or equal to zero" : "085b2bcd-f3ab-4b03-977e-61109e45b6b5", \ "Speed Close to Constant Assuming No Friction" : "502ab6e8-3cc3-4ace-baa3-053c05170810", \ "Pole 2 is on the left of the robot chassy" : "cdf7d079-c309-4ae2-8fd9-305156f745f7", \ "outputTorque is greater than or equal to zero" : "c32bbe98-04ce-46fc-b347-c26dc79ae051" \ } # Note: manualAbstractionInformation is used purely in analysis scripts (as developed for # the paper describing Fanoos); this proved to be a convieniant place to store the # information during the time of development and testing. Fanoos does not access # the information in manualAbstractionInformation when determining how to make # adjustments to respond to users. Again, it is only used in analysis scripts # used to prepare results for the paper. # # While it was convieniant for development, clearly it is not # ideal have this data stored here or this structure required # to be present. TODO: resolve the issue just described. self.manualAbstractionInformation = {\ "predicatesAndLabels" : [\ ("21b3fb5b-efd9-4406-a7fe-b80937881e21", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("f3c2655f-fa5c-4020-9dc8-c20a47401796", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("0155f83a-ef30-4a61-adda-c1321480240f", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("8de9828f-6ec7-4240-8293-b76d3d92f3f2", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("7389a08e-03d4-4c89-8b58-3a2c7c203933", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("f3c56183-c013-4b2e-b11d-65244231dadd", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("e7f2daf6-b041-4f4a-88b9-999a4590cfc0", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("f361d7bc-d6a7-4692-b38f-593e4f403d15", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("23cd37d1-e48b-41f5-8f37-5988e9410d0f", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("69e9a231-c08e-498a-b1cb-08def786cb0c", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("701785bd-f91f-4818-a997-e68258f56edb", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("af4c0da5-b6d2-4500-8fa0-117cb812e182", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("44dca983-9b7f-4516-b9a7-62a47aa259d5", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("c88184e3-6914-43ae-920b-5e7f47b5c63c", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("f30f1e19-a9f7-4c7f-a852-14afc5a4b9af", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("1ea9d650-b3e8-4d07-b80d-2ee64e2ee09b", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("f0b1a37b-f937-41e1-907f-1faec3c9a1dd", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("7661646a-34db-45a8-95d6-e8d3da8ddf4d", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("8640c6d3-12cc-438e-adf3-b84583faab65", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("9dfc2e8e-4915-4172-b856-b5b20946c393", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("b7d2f662-2205-41a2-a920-cf674a09e8e4", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("0bb61b1e-4816-4d98-99c7-41e02c85e011", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("491fc8a3-d091-48c6-a132-56c544cd181f", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("41a5302e-8e9c-4e63-8a29-fa2b11060e21", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("a5c62d0b-386c-4e1b-bc5b-40f27e7c7cca", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("960482b5-8323-40c7-8e0a-07b64ef6e75b", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("44c88c08-c354-401d-a100-ae94f573391d", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("2a9b0032-a39f-4faa-9664-af5209745722", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("f95a4c4a-6cec-49fd-abf1-64f9283f33a5", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("44268c62-371f-44d4-a966-5a4e8d588d7d", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("17e5a6e2-37fc-4523-aaee-9b6ac7384c0d", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("d3d548ee-292a-43f4-99c3-1b9cbcd3ca71", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("4fef3e42-87f7-49d9-8e59-d9a70f52b906", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("60e5a29e-1ae8-47d2-b062-d537c089e11f", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("35c790a9-f001-4dd4-9869-f855e0e2a5c4", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("d5886347-c798-4da7-86e5-61feb4b87ab8", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("78546928-bda7-4354-be99-7bf2dc6a2f07", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("15ab5d6a-7038-4c65-b429-82de8f575486", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("12247f40-220f-49ed-bc3c-ef38c6c54be9", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("2871af7f-a7cc-46ce-8527-00dfc5e0956f", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("015fb99b-fd06-4eae-9b73-95449f12dc51", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("1757016b-4dec-4310-80fa-4be7580c7afc", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("b077175f-d944-4d1a-a1da-323677470aa8", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("9154a8c3-e99a-42f5-81aa-544b176bed74", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("7b62d162-2820-464b-964d-5fc853f9724e", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("ed362df9-9e7c-4d32-9249-84c8746ffd01", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("41b0753a-ab87-4157-bc27-65c63c0b242b", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("34e8cd54-ff78-449e-9c2b-75813da7c8b2", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("7e63953e-d315-43fc-a5e8-1c3a1fc2a43a", "5b27a5f8-256f-4f2a-b8c1-a9d787f62677"), \ ("d12ad5c6-5e8e-4732-b8ae-07ee867927d9", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("e1c004da-1dcd-4ece-82b5-448a0216e50a", "5b27a5f8-256f-4f2a-b8c1-a9d787f62677"), \ ("2dd966e1-5043-4762-bb5a-929e12c66e56", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("964ab2ae-c908-4e32-b933-b84b91556688", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("90e03923-c6ff-4450-9a9f-84ea3576fa59", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("60f1be4c-c509-403b-86e2-1e9b6623d8cd", "5b27a5f8-256f-4f2a-b8c1-a9d787f62677"), \ ("761c320e-c078-4c1a-92f0-a8c4aa582c5d", "5b27a5f8-256f-4f2a-b8c1-a9d787f62677"), \ ("6d7ef156-be5c-4744-9a3f-22f7c7a3c2f8", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("ecb3670b-62aa-4ba1-9992-7eb6af738e68", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("b1aaa0bf-5924-47e8-b9a6-cf56144b1b62", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("e32a60e7-0a21-4220-a2cd-95f9cf87f8cb", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("42e9fef2-1928-4254-ac42-1b33d8abbf97", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("fecc6e90-4737-4648-bcf6-c70ed6433fb1", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("695de8ac-8a76-409d-8941-63f65753cce0", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("761dc8bb-feab-439d-9ec8-a59be174ef11", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("238126d9-bf56-4945-9c4f-c9e04f7dda4a", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("91982534-0a66-4530-886a-89f1be12b93a", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("fa4a360c-9613-484e-a212-b9cb314bf2a1", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("d496f7b5-6dfb-47b8-a1e9-2ead59f667fc", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("91151f23-dea7-4298-871f-b6b3f0088d95", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("9f21d925-53ed-413c-b633-e4026c102a1e", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("30298f0e-3fc0-4494-8daf-d94628226276", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("21b3fb5b-efd9-4406-a7fe-b80937881e21", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("f3c2655f-fa5c-4020-9dc8-c20a47401796", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("0155f83a-ef30-4a61-adda-c1321480240f", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("8de9828f-6ec7-4240-8293-b76d3d92f3f2", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("7389a08e-03d4-4c89-8b58-3a2c7c203933", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("f3c56183-c013-4b2e-b11d-65244231dadd", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("e7f2daf6-b041-4f4a-88b9-999a4590cfc0", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("f361d7bc-d6a7-4692-b38f-593e4f403d15", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("23cd37d1-e48b-41f5-8f37-5988e9410d0f", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("69e9a231-c08e-498a-b1cb-08def786cb0c", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("701785bd-f91f-4818-a997-e68258f56edb", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("af4c0da5-b6d2-4500-8fa0-117cb812e182", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("44dca983-9b7f-4516-b9a7-62a47aa259d5", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("c88184e3-6914-43ae-920b-5e7f47b5c63c", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("f30f1e19-a9f7-4c7f-a852-14afc5a4b9af", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("1ea9d650-b3e8-4d07-b80d-2ee64e2ee09b", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("f0b1a37b-f937-41e1-907f-1faec3c9a1dd", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("7661646a-34db-45a8-95d6-e8d3da8ddf4d", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("8640c6d3-12cc-438e-adf3-b84583faab65", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("9dfc2e8e-4915-4172-b856-b5b20946c393", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("b7d2f662-2205-41a2-a920-cf674a09e8e4", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("0bb61b1e-4816-4d98-99c7-41e02c85e011", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("491fc8a3-d091-48c6-a132-56c544cd181f", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("41a5302e-8e9c-4e63-8a29-fa2b11060e21", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("a5c62d0b-386c-4e1b-bc5b-40f27e7c7cca", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("960482b5-8323-40c7-8e0a-07b64ef6e75b", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("44c88c08-c354-401d-a100-ae94f573391d", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("2a9b0032-a39f-4faa-9664-af5209745722", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("f95a4c4a-6cec-49fd-abf1-64f9283f33a5", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("44268c62-371f-44d4-a966-5a4e8d588d7d", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("17e5a6e2-37fc-4523-aaee-9b6ac7384c0d", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("d3d548ee-292a-43f4-99c3-1b9cbcd3ca71", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("4fef3e42-87f7-49d9-8e59-d9a70f52b906", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("60e5a29e-1ae8-47d2-b062-d537c089e11f", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("35c790a9-f001-4dd4-9869-f855e0e2a5c4", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("d5886347-c798-4da7-86e5-61feb4b87ab8", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("78546928-bda7-4354-be99-7bf2dc6a2f07", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("15ab5d6a-7038-4c65-b429-82de8f575486", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("12247f40-220f-49ed-bc3c-ef38c6c54be9", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("2871af7f-a7cc-46ce-8527-00dfc5e0956f", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("015fb99b-fd06-4eae-9b73-95449f12dc51", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("1757016b-4dec-4310-80fa-4be7580c7afc", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("b077175f-d944-4d1a-a1da-323677470aa8", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("9154a8c3-e99a-42f5-81aa-544b176bed74", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("7b62d162-2820-464b-964d-5fc853f9724e", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("ed362df9-9e7c-4d32-9249-84c8746ffd01", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("41b0753a-ab87-4157-bc27-65c63c0b242b", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("34e8cd54-ff78-449e-9c2b-75813da7c8b2", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("7e63953e-d315-43fc-a5e8-1c3a1fc2a43a", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("d12ad5c6-5e8e-4732-b8ae-07ee867927d9", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("e1c004da-1dcd-4ece-82b5-448a0216e50a", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("2dd966e1-5043-4762-bb5a-929e12c66e56", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("964ab2ae-c908-4e32-b933-b84b91556688", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("90e03923-c6ff-4450-9a9f-84ea3576fa59", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("60f1be4c-c509-403b-86e2-1e9b6623d8cd", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("761c320e-c078-4c1a-92f0-a8c4aa582c5d", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("6d7ef156-be5c-4744-9a3f-22f7c7a3c2f8", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("ecb3670b-62aa-4ba1-9992-7eb6af738e68", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("b1aaa0bf-5924-47e8-b9a6-cf56144b1b62", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("e32a60e7-0a21-4220-a2cd-95f9cf87f8cb", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("42e9fef2-1928-4254-ac42-1b33d8abbf97", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("fecc6e90-4737-4648-bcf6-c70ed6433fb1", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("695de8ac-8a76-409d-8941-63f65753cce0", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("761dc8bb-feab-439d-9ec8-a59be174ef11", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("238126d9-bf56-4945-9c4f-c9e04f7dda4a", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("91982534-0a66-4530-886a-89f1be12b93a", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("fa4a360c-9613-484e-a212-b9cb314bf2a1", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("d496f7b5-6dfb-47b8-a1e9-2ead59f667fc", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("91151f23-dea7-4298-871f-b6b3f0088d95", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("9f21d925-53ed-413c-b633-e4026c102a1e", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("30298f0e-3fc0-4494-8daf-d94628226276", "f933bc57-f657-48b9-9b30-166399918e1d"), \ ("4370399d-f9a6-4eb2-ab5e-1eed108c58de", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("4370399d-f9a6-4eb2-ab5e-1eed108c58de", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("085b2bcd-f3ab-4b03-977e-61109e45b6b5", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("085b2bcd-f3ab-4b03-977e-61109e45b6b5", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("cdf7d079-c309-4ae2-8fd9-305156f745f7", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("cdf7d079-c309-4ae2-8fd9-305156f745f7", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("c32bbe98-04ce-46fc-b347-c26dc79ae051", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("c32bbe98-04ce-46fc-b347-c26dc79ae051", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ("502ab6e8-3cc3-4ace-baa3-053c05170810", "1b541080-2d54-488f-876b-c1308d73b452"), \ ("502ab6e8-3cc3-4ace-baa3-053c05170810", "f933bc57-f657-48b9-9b30-166399918e1d") \ ], \ "labelDag_firstParent_secondChild" : [ \ ("5b27a5f8-256f-4f2a-b8c1-a9d787f62677", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("5b27a5f8-256f-4f2a-b8c1-a9d787f62677", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("1b541080-2d54-488f-876b-c1308d73b452", "347e9ab6-6712-43b8-9aff-737b9f1774d2"), \ ("1b541080-2d54-488f-876b-c1308d73b452", "84fc6cc0-0107-4c6c-8f64-d49f165118ad"), \ ("1b541080-2d54-488f-876b-c1308d73b452", "5b27a5f8-256f-4f2a-b8c1-a9d787f62677"), \ ("f933bc57-f657-48b9-9b30-166399918e1d", "aeca763c-cec3-4c0c-ba03-da90d0f8d1cb"), \ ] \ }; functToGetUuidProvided = (lambda predicateObjectBeingInitialized : dictMappingPredicateStringNameToUUID[str(predicateObjectBeingInitialized)] ); listFunctionsToBaseCondtionsOn = \ getListFunctionsToBaseCondtionsOn_forInputOfDomainInvertedDoublePendulum() + \ getListFunctionsToBaseCondtionsOn_forOutputOfDomainInvertedDoublePendulum()+ \ getListFunctionsToBaseCondtionsOn_forJointOfDomainInvertedDoublePendulum(); self.initializedConditions = \ [CharacterizationCondition_FromPythonFunction(z3SolverInstance, DomainForInvertedDoublePendulum, x, functToGetUuidProvided=functToGetUuidProvided) \ for x in listFunctionsToBaseCondtionsOn]; assert(all([ (x.getID() == functToGetUuidProvided(x)) for x in self.initializedConditions])); self._writeInfoToDatabase(); return; def getBaseConditions(self): return self.initializedConditions; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the input domain #=========================================================================== otherFunctionsForInputDomain = """ # Note that below is in respect to the input space representation we are dealing # with here, not the space the policy is expecting... def funct_endPole2OnLeft(endOfPole2_x): \"\"\"Pole 2 is on the left of the robot chassy\"\"\" return endOfPole2_x <= 0; # Note that below is in respect to the input space representation we are dealing # with here, not the space the policy is expecting... def funct_endPole2OnRight(endOfPole2_x): \"\"\"Pole 2 is on the right of the robot chassy\"\"\" return endOfPole2_x >= 0; # moving left or right===================================== def funct_vxMovingLeft(vx): \"\"\"Anchor Point Moving Left\"\"\" return vx < 0; def funct_vxMovingRight(vx): \"\"\"Anchor Point Moving Right\"\"\" return vx > 0; def funct_vxBarelyMovingIfAtAll(vx): \"\"\"Anchor Point Barely Moving If At All\"\"\" valueOf5PercentQuantileOnAbsVx = 0.09214573; if(isinstance(vx, z3.z3.ArithRef)): assert(isinstance(vx, z3.z3.ArithRef)); return z3Abs(vx) < valueOf5PercentQuantileOnAbsVx; else: return abs(vx) < valueOf5PercentQuantileOnAbsVx; raise Exception("Control should not reach here"); return; def funct_pole1CloseToVertical(pole1Angle): \"\"\"Pole1 Close to Vertical\"\"\" valueOf5PercentQuantileOnAbsPole1Angle = 0.00050955; if(isinstance(pole1Angle, z3.z3.ArithRef)): assert(isinstance(pole1Angle, z3.z3.ArithRef)); return z3Abs(pole1Angle) < valueOf5PercentQuantileOnAbsPole1Angle; else: return abs(pole1Angle) < valueOf5PercentQuantileOnAbsPole1Angle; raise Exception("Control should not reach here"); return; def funct_pole1OnLeft(pole1Angle): \"\"\"Pole1 On Left\"\"\" if(isinstance(pole1Angle, z3.z3.ArithRef)): return pole1Angle < 0; else: return pole1Angle < 0; raise Exception("Control should not reach here"); return; def funct_pole1OnRight(pole1Angle): \"\"\"Pole1 On Right\"\"\" if(isinstance(pole1Angle, z3.z3.ArithRef)): return pole1Angle > 0; else: return pole1Angle > 0; raise Exception("Control should not reach here"); return; def funct_pole2CloseToVertical(pole2Angle, pole1Angle): \"\"\"pole2 Close to Vertical\"\"\" valueOf5PercentQuantileOnAbspole2Angle = 0.00023093; if(isinstance(pole2Angle, z3.z3.ArithRef)): assert(isinstance(pole2Angle, z3.z3.ArithRef)); return z3.Or( z3Abs(pole2Angle + pole1Angle) < valueOf5PercentQuantileOnAbspole2Angle, \ 2* np.pi - z3Abs(pole2Angle + pole1Angle) < valueOf5PercentQuantileOnAbspole2Angle ); else: return (abs(pole2Angle + pole1Angle) < valueOf5PercentQuantileOnAbspole2Angle) or \ ( 2 * np.pi - abs(pole2Angle + pole1Angle) < valueOf5PercentQuantileOnAbspole2Angle); raise Exception("Control should not reach here"); return; def funct_pole2OnLeft(pole2Angle, pole1Angle): \"\"\"pole2 On Left\"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3.Xor(pole2Angle + pole1Angle < 0, z3Abs(pole2Angle + pole1Angle) > np.pi); else: return xor(pole2Angle + pole1Angle < 0, abs(pole2Angle + pole1Angle) > np.pi); raise Exception("Control should not reach here"); return; def funct_pole2OnRight(pole2Angle, pole1Angle): \"\"\"pole2 On Right\"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3.Xor(pole2Angle + pole1Angle > 0, z3Abs(pole2Angle + pole1Angle) > np.pi); else: return xor(pole2Angle + pole1Angle > 0, abs(pole2Angle + pole1Angle) > np.pi); raise Exception("Control should not reach here"); return; #================================================================ # poles moving clockwise or counter-clockwise==== def funct_pole2MovingClockwise(pole2Angle_rateOfChange): \"\"\"Pole2 Moving Clockwise \"\"\" return pole2Angle_rateOfChange > 0; def funct_pole2MovingCounterClockwise(pole2Angle_rateOfChange): \"\"\"Pole2 Moving Counter-Clockwise \"\"\" return pole2Angle_rateOfChange < 0; def funct_pole2MovingBarely(pole2Angle_rateOfChange): \"\"\"Pole2 Moving Barely\"\"\" valueOf5PercentQuantileOnAbsPole2AngleRateOfChange = 0.01096292; if(isinstance(pole2Angle_rateOfChange, z3.z3.ArithRef)): return z3Abs(pole2Angle_rateOfChange) < valueOf5PercentQuantileOnAbsPole2AngleRateOfChange; else: return abs(pole2Angle_rateOfChange) < valueOf5PercentQuantileOnAbsPole2AngleRateOfChange; raise Exception("Control should not reach here"); return; def funct_pole1MovingClockwise(pole1Angle_rateOfChange): \"\"\"pole1 Moving Clockwise \"\"\" return pole1Angle_rateOfChange > 0; def funct_pole1MovingCounterClockwise(pole1Angle_rateOfChange): \"\"\"pole1 Moving Counter-Clockwise \"\"\" return pole1Angle_rateOfChange < 0; def funct_pole1MovingBarely(pole1Angle_rateOfChange): \"\"\"pole1 Moving Barely\"\"\" valueOf5PercentQuantileOnAbspole1AngleRateOfChange = 0.04641253; if(isinstance(pole1Angle_rateOfChange, z3.z3.ArithRef)): return z3Abs(pole1Angle_rateOfChange) < valueOf5PercentQuantileOnAbspole1AngleRateOfChange; else: return abs(pole1Angle_rateOfChange) < valueOf5PercentQuantileOnAbspole1AngleRateOfChange; raise Exception("Control should not reach here"); return; def funct_bothPolesPointedToLeft(pole1Angle, pole2Angle): \"\"\"Both poles pointed to the left\"\"\" def funct_pole2OnLeft(pole2Angle, pole1Angle): \"\"\"pole2 On Left\"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3.Xor(pole2Angle + pole1Angle < 0, z3Abs(pole2Angle + pole1Angle) > np.pi); else: return xor(pole2Angle + pole1Angle < 0, abs(pole2Angle + pole1Angle) > np.pi); raise Exception("Control should not reach here"); return; def funct_pole1OnLeft(pole1Angle): \"\"\"Pole1 On Left\"\"\" if(isinstance(pole1Angle, z3.z3.ArithRef)): return pole1Angle < 0; else: return pole1Angle < 0; raise Exception("Control should not reach here"); return; if(isinstance(pole2Angle, z3.z3.ArithRef)): assert(isinstance(pole1Angle, z3.z3.ArithRef)); return z3.And( funct_pole1OnLeft(pole1Angle) , funct_pole2OnLeft(pole2Angle, pole1Angle) ); else: return (funct_pole1OnLeft(pole1Angle) and funct_pole2OnLeft(pole2Angle, pole1Angle) ); raise Exception("Control should not reach here"); return; def funct_bothPolesPointedToRight(pole1Angle, pole2Angle): \"\"\"Both poles pointed to the right\"\"\" def funct_pole1OnRight(pole1Angle): \"\"\"Pole1 On Right\"\"\" if(isinstance(pole1Angle, z3.z3.ArithRef)): return pole1Angle > 0; else: return pole1Angle > 0; raise Exception("Control should not reach here"); return; def funct_pole2OnRight(pole2Angle, pole1Angle): \"\"\"pole2 On Right\"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3.Xor(pole2Angle + pole1Angle > 0, z3Abs(pole2Angle + pole1Angle) > np.pi); else: return xor(pole2Angle + pole1Angle > 0, abs(pole2Angle + pole1Angle) > np.pi); raise Exception("Control should not reach here"); return; if(isinstance(pole2Angle, z3.z3.ArithRef)): assert(isinstance(pole1Angle, z3.z3.ArithRef)); return z3.And( funct_pole1OnRight(pole1Angle) , funct_pole2OnRight(pole2Angle, pole1Angle) ); else: return ( funct_pole1OnRight(pole1Angle) and funct_pole2OnRight(pole2Angle, pole1Angle) ); raise Exception("Control should not reach here"); return; def funct_polesRoughtlyStraight(pole1Angle, pole2Angle): \"\"\"Poles are Roughly Straight in Respect To Each Other\"\"\" IQROfPole2Angles=0.0052; if(isinstance(pole2Angle, z3.z3.ArithRef)): assert(isinstance(pole1Angle, z3.z3.ArithRef)); return z3Abs(pole2Angle) <= IQROfPole2Angles; else: return abs(pole2Angle) <= IQROfPole2Angles; raise Exception("Control should not reach here"); return; def funct_polesInCBend(pole1Angle, pole2Angle): \"\"\"Poles are Bent Like the Arc in a C\"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): assert(isinstance(pole1Angle, z3.z3.ArithRef)); return z3.Or(\ z3.And(pole1Angle < 0, pole1Angle > -np.pi/2, pole2Angle + pole1Angle > 0 , pole2Angle + pole1Angle <= np.pi) , \ z3.And(pole1Angle < -np.pi/2, pole2Angle + pole1Angle < -np.pi , pole2Angle + pole1Angle >= -1.5 * np.pi) , \ ); else: return any([\ all([pole1Angle < 0, pole1Angle > -np.pi/2, pole2Angle + pole1Angle > 0 , pole2Angle + pole1Angle <= np.pi]) , \ all([pole1Angle < -np.pi/2, pole2Angle + pole1Angle < -np.pi , pole2Angle + pole1Angle >= -1.5 * np.pi]) , \ ]); raise Exception("Control should not reach here"); return; def funct_polesInDBend(pole1Angle, pole2Angle): \"\"\"Poles are Bent Like the arc in a a D\"\"\" # below, recall that our angles exist in [-np.pi, np,pi] (at least on of these bounds might be exclusive...)... if(isinstance(pole2Angle, z3.z3.ArithRef)): assert(isinstance(pole1Angle, z3.z3.ArithRef)); return z3.Or(\ z3.And(pole1Angle > 0, pole1Angle < np.pi/2, pole2Angle + pole1Angle > 0 , pole2Angle + pole1Angle >= -np.pi) , \ z3.And(pole1Angle > np.pi/2, pole2Angle + pole1Angle > np.pi , pole2Angle + pole1Angle <= 1.5 * np.pi) , \ ); else: return any([\ all([pole1Angle > 0, pole1Angle < np.pi/2, pole2Angle + pole1Angle > 0 , pole2Angle + pole1Angle >= -np.pi]) , \ all([pole1Angle > np.pi/2, pole2Angle + pole1Angle > np.pi , pole2Angle + pole1Angle <= 1.5 * np.pi]) , \ ]); raise Exception("Control should not reach here"); return; def funct_polesAreBent(pole2Angle): \"\"\"Poles are Bent\"\"\" IQROfPole2Angles=0.0052; if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3Abs(pole2Angle) > IQROfPole2Angles; else: return abs(pole2Angle) > IQROfPole2Angles; raise Exception("Control should not reach here"); return; # poles above or below the line x=0....==== def funct_pole2AtOrAboveXAxis(pole2Angle, pole1Angle ): \"\"\"Pole2 Angle At of Above X Axis \"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3Abs(pole2Angle + pole1Angle ) >= np.pi / 2; else: return abs(pole2Angle + pole1Angle ) >= np.pi / 2; raise Exception("Control should not reach here"); return; def funct_pole2AtOrBelowXAxis(pole2Angle, pole1Angle): \"\"\"Pole2 Angle At of Above X Axis \"\"\" if(isinstance(pole2Angle, z3.z3.ArithRef)): return z3Abs(pole2Angle + pole1Angle ) <= np.pi / 2; else: return abs(pole2Angle + pole1Angle ) <= np.pi / 2; raise Exception("Control should not reach here"); return; def funct_pole1AtOrAboveXAxis(pole1Angle): \"\"\"pole1 Angle At of Above X Axis \"\"\" if(isinstance(pole1Angle, z3.z3.ArithRef)): return z3Abs(pole1Angle) >= np.pi / 2; else: return abs(pole1Angle) >= np.pi / 2; raise Exception("Control should not reach here"); return; def funct_pole1AtOrBelowXAxis(pole1Angle): \"\"\"pole1 Angle At of Above X Axis \"\"\" if(isinstance(pole1Angle, z3.z3.ArithRef)): return z3Abs(pole1Angle) <= np.pi / 2; else: return abs(pole1Angle) <= np.pi / 2; raise Exception("Control should not reach here"); return; """ def getListFunctionsToBaseCondtionsOn_forInputOfDomainInvertedDoublePendulum(): listOfFunctionCodes =[]; inputSpaceVariables = ["x", "vx", "endOfPole2_x", \ "pole1Angle", "pole1Angle_rateOfChange", \ "pole2Angle", "pole2Angle_rateOfChange" ]; medians = [0.91708765, 0.14007734, 0.00365422, 0.00463563, 0.2465699, 0.00271302, \ 0.28928026] stds = [0.23997518, 0.137665, 0.01593154, 0.01865791, 0.07264117, 0.00410661, \ 0.10496485] quantile0Dot10 = [0.57014145, 0.10333252, 0.00075904, 0.00103595, 0.09850179, 0.00058252,\ 0.02011684] quantile0Dot25 = [0.91029063, 0.11111132, 0.0019229, 0.00248957, 0.22880583, 0.0012283,\ 0.27657566] quantile0Dot75 = [0.92539008, 0.16597374, 0.00588772, 0.00836597, 0.27732796, 0.00477952, \ 0.30694401] quantile0Dot90 = [0.95352224, 0.23168947, 0.03269801, 0.03493198, 0.28769558, 0.00616469, \ 0.31307735] listOfFunctionCodes = \ getFunctionCodeBasedOnThresholdsAndIndividualVariables(inputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds, indicesOfVariablesToUseBoolFor=set([0,1,2,3,4,5,6])); listOfFunctionCodes.append(otherFunctionsForInputDomain); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the output domain #=========================================================================== otherFunctionsForOutputDomain = """ def funct_outputTorquePositive(outputTorque): \"\"\"outputTorque is greater than or equal to zero\"\"\" return outputTorque >= 0; def funct_outputTorqueNegative(outputTorque): \"\"\"outputTorque is less than or equal to zero\"\"\" return outputTorque <= 0; """; def getListFunctionsToBaseCondtionsOn_forOutputOfDomainInvertedDoublePendulum(): listOfFunctionCodes =[]; inputSpaceVariables = ["outputTorque", "stateValueEstimate"]; # We modify the bounds to account for the fact that we want the abs to be applied to # to the output torque... dictMappingVariableToBound = {\ "outputTorque" : [0.0 , 1.0], \ "stateValueEstimate" : [-3.98 , 1.42] \ }; quantile0Dot90 = [1.0, 3.04098518]; quantile0Dot75 = [1.0, 2.87173254]; quantile0Dot25 = [0.96223713, 2.84043188]; quantile0Dot10 = [0.05367255, 2.83570324]; medians = [1.0, 2.84733465]; stds = [0.36228795, 0.07568831]; listOfFunctionCodes = \ getFunctionCodeBasedOnThresholdsAndIndividualVariables(inputSpaceVariables, quantile0Dot90, \ quantile0Dot75, quantile0Dot25, quantile0Dot10, medians, stds, indicesOfVariablesToUseBoolFor=set([0])); listOfFunctionCodes.append(otherFunctionsForOutputDomain); return convertCodeListToListOfFunctions(listOfFunctionCodes); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Conditions over the joint domain #=========================================================================== otherFunctionsForJointDomain = """ def funct_speedIncreasing(vx, outputTorque): \"\"\"Speed Increasing Assuming No Friction\"\"\" return vx * outputTorque > 0; def funct_speedDecreasing(vx, outputTorque): \"\"\"Speed Decreasing Assuming No Friction\"\"\" return vx * outputTorque < 0; def funct_speedConstant(vx, outputTorque): \"\"\"Speed Constant Assuming No Friction\"\"\" return outputTorque == 0; # This predicate will probably never be used since it # specifies a lower dimensional space... it is a sanity check, then.... def funct_speedCloseToConstant(vx, outputTorque): \"\"\"Speed Close to Constant Assuming No Friction\"\"\" quartileTenPercentForAbsOutputTorque = 0.05367255; if(isinstance(outputTorque, z3.z3.ArithRef)): return z3Abs(outputTorque) < quartileTenPercentForAbsOutputTorque; else: return abs(outputTorque) < quartileTenPercentForAbsOutputTorque; raise Exception("Control should not reach here."); return; """; def getListFunctionsToBaseCondtionsOn_forJointOfDomainInvertedDoublePendulum(): listOfFunctionCodes = []; listOfFunctionCodes.append(otherFunctionsForJointDomain); # TODO: joint conditions that joint together certain subsets of the different operators return convertCodeListToListOfFunctions(listOfFunctionCodes); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg #================================================================ #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^
59,947
55.715232
2,849
py
Fanoos
Fanoos-master/domainsAndConditions/domainAndConditionsForCircleFollowing.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; import sys; from utils.contracts import *; from boxesAndBoxOperations.getBox import getDimensionOfBox ; import re; import z3; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import CharacterizationConditionsBaseClass, CharacterizationCondition_FromPythonFunction; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation ; class DomainForCircleFollowing(BaseClassDomainInformation): def __init__(self, z3SolverInstance): requires(isinstance(z3SolverInstance, z3.z3.Solver)); raise Exception("The domain for the Circle Following controllerd " + \ "is not supported in this code release. See the README.txt file. "+\ "Inclusion of this domain's source code is purely for " + \ "reference for the curious - we selected to show an " + \ "implementation that is old in respect to updates of Fanoos."); self.initializedConditions = None; self.initialize_baseConditions(z3SolverInstance); assert(self.initializedConditions != None); return; @staticmethod def getUUID(): return "8827094e-c421-463f-98f5-36dc411f29e7"; @staticmethod def getInputSpaceUniverseBox(): orderOfVariables = __class__.inputSpaceVariables(); dictMappingVariableToBound = {\ "inputDx" : [-0.25 , 0.25],\ "inputTheta" : [-np.pi /2.0 , np.pi /2.0],\ "inputDxDot" : [-1.5 , 1.5],\ "inputThetaDot" : [-3.625 , 3.625] \ }; thisUniverseBox = __class__._helper_getInputSpaceUniverseBox(\ orderOfVariables, dictMappingVariableToBound); ensures(getDimensionOfBox(thisUniverseBox) == len(DomainForCircleFollowing.inputSpaceVariables())); return thisUniverseBox; @staticmethod def inputSpaceVariables(): # see the function _state_to_relative in ./envs/circle_env.py , line 113. return [z3.Real("inputDx"), z3.Real("inputTheta"), z3.Real("inputDxDot"), z3.Real("inputThetaDot")]; @staticmethod def outputSpaceVariables(): return [z3.Real("outputVelocity"), z3.Real("outputSteeringAngle")]; @staticmethod def getName(): return "Domain for Circle Following [not supported in this release; see README.txt]"; def initialize_baseConditions(self, z3SolverInstance): dictMappingPredicateStringNameToUUID = \ { "Close To Target Position" : "9ee7ee55-c605-4e58-a7a3-aba6973de71c", "NotClose To Target Position" : "49d3939c-2bdc-42f2-b97b-e6c223afcd45", "NotFar To Target Position" : "47148b78-cea3-4f6e-8174-b578bc8ba924", "Far To Target Position" : "c93e7cd5-d2a1-4544-b48c-ac80b69f7686", "Moving Toward Target Position" : "fb18ac46-407b-4e32-a4ed-b0e8ee565778", "Moving Quickly Toward Target Position" : "24795eb1-067e-4e8d-864a-d94150da9658", "Moving Slowly Toward Target Position" : "18555dce-9f90-454f-a206-357e6cd39da6", "Moving Away From Target Position" : "2d226265-9a7e-4c35-863d-5ace4304bdca", "Moving Quickly Away From Target Position" : "440e8ab1-8db4-4921-bb7f-4b95c2feed69", "Moving Slowly Away From Target Position" : "4cb7ed4f-762f-4964-877b-5d26398833e8", "Close To Target Orientation" : "efc4aded-b408-4bfa-8b95-17f381faa155", "NotClose To Target Orientation" : "025472ee-f2f0-4fd2-853b-d651305d7443", "NotFar To Target Orientation" : "3ff7e1c6-966e-4170-91f4-4db8b02d6e9f", "Far To Target Orientation" : "b7d26e79-2f04-49ff-a73f-3ce68c52dc4a", "Moving Toward Target Orientation" : "977d6db2-26a4-4808-a44c-3930f35303d3", "Moving Quickly Toward Target Orientation" : "102a4a1f-728e-4084-9ddd-bfcc5d556e56", "Moving Slowly Toward Target Orientation" : "f4f6c19e-63c5-4e40-8163-ed0bac55605a", "Moving Away From Target Orientation" : "d6687f9c-f937-45b4-a9c6-83ac73b9b519", "Moving Quickly Away From Target Orientation" : "23c33313-6238-491a-bfbf-a13a50241c86", "Moving Slowly Away From Target Orientation" : "99722d5a-472e-49d6-b2a5-ce9fd238135c", "Moving at Low Speed" : "3ffc400e-4d80-42e8-8f04-5e63c9fed46e", "Moving at Moderate Speed" : "20a2e808-ba75-4711-8180-21ac2220b02f", "Moving at High Speed" : "0f39b35a-9742-48a8-a6fc-aef30b07ca30", "Steering Close to Center" : "98e04364-4d18-4086-9d03-faeff2983771", "Steering to Right" : "63d94834-6936-42ab-8bc4-007486059a4a", "Steering to Left" : "8029d02d-e719-4016-8231-b31301f112d6", "Steering Far to Right" : "97adc025-0842-455c-b4a5-d98fe6d06bf9", "Steering Far to Left" : "7b554b6a-1800-4a9b-ab34-58d5fac2fada" }; functToGetUuidProvided = (lambda predicateObjectBeingInitialized : dictMappingPredicateStringNameToUUID[str(predicateObjectBeingInitialized)] ); self.initializedConditions = \ [CharacterizationCondition_FromPythonFunction(z3SolverInstance, DomainForCircleFollowing, x, functToGetUuidProvided=functToGetUuidProvided) \ for x in getListFunctionsToBaseCondtionsOn_forInputOfDomainCircleFollowing() + \ getListFunctionsToBaseCondtionsOn_forOutputOfDomainCircleFollowing() ]; assert(all([ (x.getID() == functToGetUuidProvided(x)) for x in self.initializedConditions])); self._writeInfoToDatabase(); return; def getBaseConditions(self): return self.initializedConditions; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~VV~V~V~V~V~V~V~V~V~V~~V~V~V # Input domain #----------------------------------------------------------------------------- def z3Abs(x): return z3.If(x >= 0,x,-x) def z3Sign(x): return z3.If(x > 0,1,0) + z3.If(x < 0,-1,0); def createFunctToUseInInputDescription_RelativeDistanceToTarget(endPartOfName, adjectiveForDescription, var, comparitorAndScalarString): codeForFunct = """ def funct_{0}ToTarget{1}({2}): \"\"\"{0} To Target {1}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3Abs({2}) {3}; else: return np.abs({2}) {3}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(adjectiveForDescription, endPartOfName, var, comparitorAndScalarString); def createFunctToUseInInputDescription_MovingTowardTargetWithNoCommentOnSpeed(endPartOfName, var, varDot): codeForFunct = """ def funct_MovingTowardTarget{1}({2}, {3}): \"\"\"Moving Toward Target {1}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return (z3Sign({2}) + z3Sign({3}) == 0); else: return (np.sign({2}) + np.sign({3}) == 0); raise Exception("Control should not reach here"); return; """ return codeForFunct.format("", endPartOfName, var, varDot); def createFunctToUseInInputDescription_MovingTowardTargetAtSpeed(endPartOfName, adjectiveForDescription, var, varDot, comparitorAndScalarString): codeForFunct = """ def funct_Moving{0}TowardTarget{1}({2}, {3}): \"\"\"Moving {0} Toward Target {1}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3.And( z3Sign({2}) + z3Sign({3}) == 0, \\ z3Abs({3}) {4}); else: return (np.sign({2}) + np.sign({3}) == 0) and \\ (np.abs({3}) {4} ); raise Exception("Control should not reach here"); return; """ return codeForFunct.format(adjectiveForDescription, endPartOfName, var, varDot, comparitorAndScalarString); def createFunctToUseInInputDescription_MovingAwayFromTargetWithNoCommentOnSpeed(endPartOfName, var, varDot): codeForFunct = """ def funct_MovingAwayFromTarget{1}({2}, {3}): \"\"\"Moving Away From Target {1}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3.Not(z3Sign({2}) + z3Sign({3}) == 0); else: return (np.sign({2}) + np.sign({3}) != 0); raise Exception("Control should not reach here"); return; """ return codeForFunct.format("", endPartOfName, var, varDot); def createFunctToUseInInputDescription_MovingAwayFromTargetAtSpeed(endPartOfName, adjectiveForDescription, var, varDot, comparitorAndScalarString): codeForFunct = """ def funct_Moving{0}AwayFromTarget{1}({2}, {3}): \"\"\"Moving {0} Away From Target {1}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3.And( z3.Not(z3Sign({2}) + z3Sign({3}) == 0), \\ z3Abs({3}) {4}); else: return (np.sign({2}) + np.sign({3}) != 0) and \\ (np.abs({3}) {4} ); raise Exception("Control should not reach here"); return; """ return codeForFunct.format(adjectiveForDescription, endPartOfName, var, varDot, comparitorAndScalarString); def getListFunctionsToBaseCondtionsOn_forInputOfDomainCircleFollowing(): listOfFunctionsToCreate = [\ createFunctToUseInInputDescription_RelativeDistanceToTarget("Position", "Close", "inputDx", "< 0.025"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Position", "NotClose", "inputDx", ">= 0.025"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Position", "NotFar", "inputDx", "< 0.1"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Position", "Far", "inputDx", ">= 0.1"), \ createFunctToUseInInputDescription_MovingTowardTargetWithNoCommentOnSpeed("Position", "inputDx", "inputDxDot"), \ createFunctToUseInInputDescription_MovingTowardTargetAtSpeed("Position", "Quickly", "inputDx", "inputDxDot", " > 0.75"), \ createFunctToUseInInputDescription_MovingTowardTargetAtSpeed("Position", "Slowly", "inputDx", "inputDxDot", " <= 0.1"), \ createFunctToUseInInputDescription_MovingAwayFromTargetWithNoCommentOnSpeed("Position", "inputDx", "inputDxDot"), \ createFunctToUseInInputDescription_MovingAwayFromTargetAtSpeed("Position", "Quickly", "inputDx", "inputDxDot", " > 0.75"), \ createFunctToUseInInputDescription_MovingAwayFromTargetAtSpeed("Position", "Slowly", "inputDx", "inputDxDot", " <= 0.1"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Orientation", "Close", "inputTheta", "< 0.1"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Orientation", "NotClose", "inputTheta", ">= 0.1"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Orientation", "NotFar", "inputTheta", "< 0.3"), \ createFunctToUseInInputDescription_RelativeDistanceToTarget("Orientation", "Far", "inputTheta", ">= 0.3"), \ createFunctToUseInInputDescription_MovingTowardTargetWithNoCommentOnSpeed("Orientation", "inputTheta", "inputThetaDot"), \ createFunctToUseInInputDescription_MovingTowardTargetAtSpeed("Orientation", "Quickly", "inputTheta", "inputThetaDot", " > 0.75"), \ createFunctToUseInInputDescription_MovingTowardTargetAtSpeed("Orientation", "Slowly", "inputTheta", "inputThetaDot", " <= 0.1"), \ createFunctToUseInInputDescription_MovingAwayFromTargetWithNoCommentOnSpeed("Orientation", "inputTheta", "inputThetaDot"), \ createFunctToUseInInputDescription_MovingAwayFromTargetAtSpeed("Orientation", "Quickly", "inputTheta", "inputThetaDot", " > 0.75"), \ createFunctToUseInInputDescription_MovingAwayFromTargetAtSpeed("Orientation", "Slowly", "inputTheta", "inputThetaDot", " <= 0.1") \ ]; listOfFunctionsToReturn = []; for thisFunctionDefinition in listOfFunctionsToCreate: initialSetOfDefinedFunctions = set(locals().keys()); exec(thisFunctionDefinition); newSetOfDefinedFunctions = set(locals().keys()).difference(\ initialSetOfDefinedFunctions).difference(\ ["initialSetOfDefinedFunctions", "newSetOfDefinedFunctions", "thisFunctionDefinition"]); assert(len(newSetOfDefinedFunctions) == 1); listOfFunctionsToReturn.append(locals()[list(newSetOfDefinedFunctions)[0]]); assert(len(listOfFunctionsToReturn) == len(listOfFunctionsToCreate)); return listOfFunctionsToReturn; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^^_^_^_^^_^_^_^^_^_^_^_^_^^_^_^_^_^_^^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~VV~V~V~V~V~V~V~V~V~V~~V~V~V # Output domain #----------------------------------------------------------------------------- def createFunctToUseInOutputDescription_absThresholdCompare(name, var, comparitorAndScalarString): nameWithSpaceRemoved = name.replace(" ", ""); codeForFunct = """ def funct_{1}({2}): \"\"\"{0}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3Abs({2}) {3}; else: return abs({2}) {3}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(name, nameWithSpaceRemoved, var, comparitorAndScalarString); def createFunctToUseInOutputDescription_thresholdCompare(name, var, comparitorAndScalarString): nameWithSpaceRemoved = name.replace(" ", ""); codeForFunct = """ def funct_{1}({2}): \"\"\"{0}\"\"\" if(isinstance({2}, z3.z3.ArithRef)): return z3Abs({2}) {3}; else: return abs({2}) {3}; raise Exception("Control should not reach here"); return; """ return codeForFunct.format(name, nameWithSpaceRemoved, var, comparitorAndScalarString); def getListFunctionsToBaseCondtionsOn_forOutputOfDomainCircleFollowing(): listOfFunctionsToCreate = [\ createFunctToUseInOutputDescription_absThresholdCompare("Moving at Low Speed", "outputVelocity", " < 0.1 "), \ createFunctToUseInOutputDescription_absThresholdCompare("Moving at Moderate Speed", "outputVelocity", " < 0.65 "), \ createFunctToUseInOutputDescription_absThresholdCompare("Moving at High Speed", "outputVelocity", " >= 0.65 "), \ createFunctToUseInOutputDescription_absThresholdCompare("Steering Close to Center", "outputSteeringAngle", " <= 0.1 "), \ createFunctToUseInOutputDescription_thresholdCompare("Steering to Right", "outputSteeringAngle", " > 0"), \ createFunctToUseInOutputDescription_thresholdCompare("Steering to Left", "outputSteeringAngle", " < 0"), \ createFunctToUseInOutputDescription_thresholdCompare("Steering Far to Right", "outputSteeringAngle", " > np.pi / 8.0"), \ createFunctToUseInOutputDescription_thresholdCompare("Steering Far to Left", "outputSteeringAngle", " < -np.pi / 8.0") \ ]; listOfFunctionsToReturn = []; for thisFunctionDefinition in listOfFunctionsToCreate: initialSetOfDefinedFunctions = set(locals().keys()); exec(thisFunctionDefinition); newSetOfDefinedFunctions = set(locals().keys()).difference(\ initialSetOfDefinedFunctions).difference(\ ["initialSetOfDefinedFunctions", "newSetOfDefinedFunctions", "thisFunctionDefinition"]); assert(len(newSetOfDefinedFunctions) == 1); listOfFunctionsToReturn.append(locals()[list(newSetOfDefinedFunctions)[0]]); assert(len(listOfFunctionsToReturn) == len(listOfFunctionsToCreate)); return listOfFunctionsToReturn; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^^_^_^_^^_^_^_^^_^_^_^_^_^^_^_^_^_^_^^_^_^_^_^_^_^_^
16,568
44.770718
165
py
Fanoos
Fanoos-master/propagateBoxThroughLearnedSystem/inputBoxToActivationFunctionOutputContainingInterval.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import numpy as np; def tanhActivationFunction(thisInputValue, weightVector): return np.tanh(np.matmul(thisInputValue, weightVector)); def linearActivationFunction(thisInputValue, weightVector): return np.matmul(thisInputValue, weightVector); def inputBoxToActivationFunctionOutputContainingInterval(thisBox, weightVector, activationFunctionName ="tanh"): requires(isinstance(weightVector, np.ndarray)); requires(len(weightVector.shape) == 1); requires(all(isinstance(x, float) for x in weightVector)); requires(isProperBox(thisBox)); requires(getDimensionOfBox(thisBox) == weightVector.shape[0]); requires(isinstance(activationFunctionName, str)); requires(activationFunctionName in {"tanh", "linear"}); maximizingInputValues = np.where(weightVector > 0, thisBox[:,1], thisBox[:,0]); minimizingInputValues = np.where(weightVector <= 0, thisBox[:,1], thisBox[:,0]); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg activationFunction = None; if(activationFunctionName == "tanh"): activationFunction = tanhActivationFunction; elif(activationFunctionName == "linear"): activationFunction = linearActivationFunction; # for the output layer. else: raise Exception("In the current version of the code, control should never reach here."); assert(activationFunction != None); minimumValue = activationFunction(minimizingInputValues, weightVector); maximumValue = activationFunction(maximizingInputValues, weightVector); boxContainingOutputs = getBox(np.array([minimumValue]), np.array([maximumValue])); ensures(isProperBox(boxContainingOutputs)); ensures(getDimensionOfBox(boxContainingOutputs) == 1); return boxContainingOutputs;
5,794
72.35443
2,787
py
Fanoos
Fanoos-master/propagateBoxThroughLearnedSystem/inputBoxToNetwork.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; import numpy as np; from propagateBoxThroughLearnedSystem.inputBoxToActivationFunctionOutputContainingInterval import inputBoxToActivationFunctionOutputContainingInterval; def helper_getUnitOutput(inputBoxWithoutInterceptTerm, weightVector, bias, activationFunctionName="tanh"): #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Below requires basically copied from inputBoxToActivationFunctionOutputContainingInterval # after some variable renaming. #--------------------------------------------------------------- requires(isinstance(weightVector, np.ndarray)); requires(len(weightVector.shape) == 1); requires(all(isinstance(x, float) for x in weightVector)); requires(isProperBox(inputBoxWithoutInterceptTerm,)); requires(getDimensionOfBox(inputBoxWithoutInterceptTerm) == weightVector.shape[0]); requires(isinstance(activationFunctionName, str)); requires(activationFunctionName in {"tanh", "linear"}); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ requires(isinstance(bias, float)); inputBoxWithInterceptTerm = getJointBox([inputBoxWithoutInterceptTerm, \ getBox(np.array([1]), np.array([1]))]); assert(isProperBox(inputBoxWithInterceptTerm)); weightAndBias = np.concatenate([weightVector, np.array([bias])]); assert(len(weightAndBias.shape) == 1); assert(weightAndBias.shape[0] == weightVector.shape[0] + 1); # Checking that I placed the terms necessary to make the bias work as expected in the # appropraite locations assert(weightAndBias[-1] == bias); assert(np.all(inputBoxWithInterceptTerm[-1, :] == np.array([1.0, 1.0]))); return inputBoxToActivationFunctionOutputContainingInterval(\ inputBoxWithInterceptTerm, weightAndBias, activationFunctionName=activationFunctionName); def computeBoxOutputForLayer(inputBoxWithoutInterceptTerm, weight2DArrayForLayer, biasVector, activationFunctionName="tanh"): #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Below requires basically copied from inputBoxToActivationFunctionOutputContainingInterval # after some variable renaming. #--------------------------------------------------------------- requires(isinstance(weight2DArrayForLayer, np.ndarray)); requires(len(weight2DArrayForLayer.shape) == 2); requires(all(isinstance(x, np.ndarray) for x in weight2DArrayForLayer,)); requires(isProperBox(inputBoxWithoutInterceptTerm,)); requires(all( (getDimensionOfBox(inputBoxWithoutInterceptTerm) == weight2DArrayForLayer[:, index].shape[0]) for \ index in range(0, weight2DArrayForLayer.shape[1]) )); requires(isinstance(activationFunctionName, str)); requires(activationFunctionName in {"tanh", "linear"}); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ requires(isinstance(biasVector, np.ndarray)); requires(len(biasVector.shape) == 1); requires(weight2DArrayForLayer.shape[1] == biasVector.shape[0]); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg listOfBoxesToForJiontOver = []; # THE ORDER WE PROCESS THE NODES IN (AND PLACE THEM IN listOfBoxesToForJiontOver SUBSEQUENTLY) # IS IMPORTANTO TO ENSURE WE DON'T ACCIDENTALLY SCREW UP THE VARIABLE ORDERING. for thisIndex in range(0, biasVector.shape[0]): thisWeightVector = weight2DArrayForLayer[:, thisIndex]; thisBias = biasVector[thisIndex]; thisBox = helper_getUnitOutput(inputBoxWithoutInterceptTerm, thisWeightVector, thisBias, \ activationFunctionName=activationFunctionName); listOfBoxesToForJiontOver.append(thisBox); boxToSendForward = getJointBox(listOfBoxesToForJiontOver); assert(isProperBox(boxToSendForward)); return boxToSendForward;
7,887
68.192982
2,787
py
Fanoos
Fanoos-master/propagateBoxThroughLearnedSystem/classesToPropogateBoxThroughModels.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; from propagateBoxThroughLearnedSystem.inputBoxToNetwork import computeBoxOutputForLayer; from propagateBoxThroughLearnedSystem.propogateValuesThroughRegression import propogateBound as primitive_PropogatorForPolyFeatLinearModel; from propagateBoxThroughLearnedSystem.propogateValuesThroughRegression import loadModel as loadModel_linearReg; import numpy as np; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; class ModelBoxProgatorManager(): def __init__(self, pathToFile): self.fileNameOfLoadedModel = pathToFile; return; def pushBoxThrough(self, thisBox): return; class PropogatorForPolyFeatLinearModel(ModelBoxProgatorManager): def __init__(self, pathToFileToLoad): self.thisModel = loadModel_linearReg(pathToFileToLoad); ModelBoxProgatorManager.__init__(self, pathToFileToLoad); return; def pushBoxThrough(self, thisBox): return primitive_PropogatorForPolyFeatLinearModel(thisBox, self.thisModel); import pickle; class PropogatorForNueralNet(ModelBoxProgatorManager): def __init__(self, pathToFileToLoad): self.thisModel = self.getTrainedModelWeightAndBiases(pathToFileToLoad); exec(self.thisModel["clippingFunctionCode"], globals(), locals()); setattr(self, "clipResult_oneVector", locals()["properlyTransformAction"]); exec(self.thisModel["functionToPushBoxThroughModel"], globals(), locals()); setattr(PropogatorForNueralNet, "pushBoxThrough", locals()["pushBoxThrough"]); ModelBoxProgatorManager.__init__(self, pathToFileToLoad); return; def getTrainedModelWeightAndBiases(self, nameOfFileToLoad): requires(isinstance(nameOfFileToLoad, str)); fh = open(nameOfFileToLoad, "rb"); A = pickle.load(fh); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg fh.close(); return A; def clipAndScaleResultBox(self, thisBox): # We assume that clipResult_oneVector is non-decreasing and works independently on # all of its components.... lb = self.clipResult_oneVector(thisBox[:, 0]); ub = self.clipResult_oneVector(thisBox[:, 1]); return getBox(lb, ub);
6,192
62.193878
2,791
py
Fanoos
Fanoos-master/propagateBoxThroughLearnedSystem/propogateValuesThroughRegression.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; from utils.contracts import *; def convertToInformativePair(thisMonomial, dictMappingStringVariableNameToIndex): breakDown = [ y.split("^") for y in thisMonomial.split(" ")]; return [ ( int(1 if (len(w) < 2) else w[1]), \ dictMappingStringVariableNameToIndex[w[0]]) \ for w in breakDown]; def convertToUsablePairs(listOfNonFeaturizedObservationNames, stringMonomials): requires("1" not in listOfNonFeaturizedObservationNames); dictMappingStringVariableNameToIndex = \ { listOfNonFeaturizedObservationNames[index] : index for index in range(0, len(listOfNonFeaturizedObservationNames))}; return [convertToInformativePair(thisMonomial, dictMappingStringVariableNameToIndex) \ for thisMonomial in stringMonomials if thisMonomial != "1"]; def informativePairsToValueBounds(thisInformativePairList, valueBounds): listToReturn = []; for thisPair in thisInformativePairList: firstBounds = valueBounds[thisPair[1], :] ** thisPair[0]; boundsContainZero = (valueBounds[thisPair[1], 0] <= 0.0 and valueBounds[thisPair[1], 1] >= 0.0); minVal = np.min(firstBounds); maxVal = np.max(firstBounds); if(boundsContainZero): minVal = min(minVal, 0.0); maxVal = max(maxVal, 0.0); listToReturn.append([minVal, maxVal]); return listToReturn; def findBoundsForMonomial(valueBoundsPerPower): finalValueBounds = [1.0, 1.0]; def intervalFromAllCombinations(A, B): allCombinations = [A[0] * B[0], A[0] * B[1], A[1] * B[0], A[1] * B[1]]; return [np.min(allCombinations), np.max(allCombinations)]; for thisPair in valueBoundsPerPower: finalValueBounds = intervalFromAllCombinations(finalValueBounds, thisPair); return finalValueBounds; def getBoundsForEachMonomial(valueBounds, listOfNonFeaturizedObservationNames, stringMonomials): usefulPairs = convertToUsablePairs(\ listOfNonFeaturizedObservationNames, \ stringMonomials \ ); A = [ informativePairsToValueBounds(x, valueBounds ) for x in usefulPairs]; B = [ findBoundsForMonomial(y) for y in A]; return B; import pickle; def loadModel(pathToFile): requires(isinstance(pathToFile, str)); requires(len(pathToFile)> 0); fh = open(pathToFile, "rb"); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg A = pickle.load(fh); fh.close(); return A; def propogateBound(thisBox, thisModel): boundsForFeatureValues = \ getBoundsForEachMonomial( \ thisBox, \ thisModel["orderOfNonFeaturizedObservationNames"], \ thisModel["newFeatureNames"]); outputBoundingBox = np.zeros((len(thisModel["namesOfTargetValues"]) ,2)) for thisIndex in range(0, len(thisModel["namesOfTargetValues"])): A = np.array(boundsForFeatureValues) * (\ (thisModel["coefficients"][thisIndex,:]).reshape(\ len(thisModel["newFeatureNames"]),1) \ ); assert(A.shape[1] == 2); boundsOnCoordinatesAfterLinearTransform = np.array([np.min(A, axis=1), np.max(A, axis=1)]); outputBoundingBox[thisIndex, :] = np.sum(boundsOnCoordinatesAfterLinearTransform, axis=1) + thisModel["intercept"][thisIndex]; return outputBoundingBox;
7,219
62.893805
2,787
py
Fanoos
Fanoos-master/propagateBoxThroughLearnedSystem/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/statesAndOperatorsAndSelection/descriptionOperator.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; import re; import uuid; from statesAndOperatorsAndSelection.descriptionState import DescriptionState, FirstState_DescriptionState; from domainsAndConditions.baseClassDomainInformation import BaseClassDomainInformation; from domainsAndConditions.classesDefiningQuestions import QuestionBaseClass; from descriptionGeneration.generateDescription import generateDescription; from propagateBoxThroughLearnedSystem.classesToPropogateBoxThroughModels import ModelBoxProgatorManager; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface, executeDatabaseCommandList; from CEGARLikeAnalysis.CEGARLikeAnalysisMain import timingInfoForLocation_2e048534_BoxTest ; from utils.distributionStatics import distributionStatics; import time as timePackageToUseForSleep; import config; class DescriptionOperator(): def recordTimeStats(self, timeMeasurements, locationLabel): requires(isinstance(timeMeasurements, list)); requires(all([isinstance(x, float) for x in timeMeasurements])); requires(isinstance(locationLabel, str)); requires(len(locationLabel) > 0); # The below requires basically checks that the location label contains a UUID... requires(re.match("^.*[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}.*$", locationLabel) is not None); def helper_recordInDatabse(fieldName, valueToRecord): requires(isinstance(fieldName, str)); requires(len(fieldName) > 0); commandToExecute = \ "INSERT INTO QAOperatorValues (" + \ " questionInstanceUUID, QAOperatorUUID, startingAnswerIndex, fieldName, fieldValue ) VALUES (" + \ "\'" + ObjDatabaseValueTracker.get_questionInstanceUUID() + "\', " + \ "\'" + self.getID() + "\', " + \ str(self.indexIntoQA) + ", " + \ "\'" + fieldName + "\', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [valueToRecord] ); helper_recordInDatabse("allTimings:" + locationLabel, timeMeasurements); distributionStaticsForTimeValues = distributionStatics(timeMeasurements); baseLabelToUse = "timingStats:" + locationLabel + ":"; for thisKey in distributionStaticsForTimeValues: if((distributionStaticsForTimeValues["numberOfDataPoints"] == 0) and (thisKey != "numberOfDataPoints")): continue; if((distributionStaticsForTimeValues["numberOfDataPoints"] == 1) and (thisKey not in {"numberOfDataPoints", "median"})): continue; helper_recordInDatabse(baseLabelToUse + thisKey, \ distributionStaticsForTimeValues[thisKey]); objDatabaseInterface.commit(); return @staticmethod def getID(): uuid = "1b29f22c-eaf8-458a-b419-ce38c152df8e"; return uuid; def setID(self): raise Exception("""This class does not support assigning new UUIDs."""); def __init__(self): raise Exception("Child classes must overwrite..."); return; def internal_apply(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory): raise Exception("Child classes must overwrite..."); return; def apply(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory, indexIntoQA): requires(isinstance(parsedUserQuestion, QuestionBaseClass)); requires(isinstance(domainInformation, BaseClassDomainInformation)); requires(isinstance(loadedLearnedModel, ModelBoxProgatorManager)); requires(isinstance(state, DescriptionState)); requires(isinstance(objectForHistory, list)); requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); self.indexIntoQA = indexIntoQA; stateToReturn = \ self.internal_apply(parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory); ensures(isinstance(stateToReturn, DescriptionState)); return stateToReturn; import traceback; import cmd; from UI.genericUIFunctions import myLinuxStyleMoreCommand, displayForUser, promptToSelectFromList; class Operator_ManualPredicateReview(DescriptionOperator): class ManualPredicateReviewPrompt(cmd.Cmd): prompt="(Fanoos: Manual Predicate Review)"; @staticmethod def _standardWait(): timeForSleep=config.defaultValues.responceDelayTimeForUnexpectedInputes; print("Sleeping " + str(timeForSleep) + " seconds before responding....", flush=True); timePackageToUseForSleep.sleep(timeForSleep); return; def default(*x, **kwargs): __class__._standardWait(); cmd.Cmd.default(*x, **kwargs); return; def emptyline(*x, **kwargs): __class__._standardWait(); print("Ignoring Empty Line.", flush=True); return; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ def secondInit(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory, hackyListToContainStateToReturn): requires(isinstance(hackyListToContainStateToReturn, list)); requires(len(hackyListToContainStateToReturn) == 0); self.parsedUserQuestion = parsedUserQuestion; self.domainInformation = domainInformation; self.loadedLearnedModel = loadedLearnedModel; self.currentState = state; self.objectForHistory = objectForHistory; self.historyForThisQuestion = self.objectForHistory[-1][1]; self.statesVisitedInThisQuestion = [x[0].getID() for x in self.historyForThisQuestion]; self.hackyListToContainStateToReturn = hackyListToContainStateToReturn; self.newlyAllowedPredicates = set(); self.newlyDisallowedPredicate = set(); self.allNamedPredicatesInDomain = self.domainInformation.getBaseConditions().copy(); self.UUIDsOfAllNamedPredicatesInDomain = [x.getID() for x in self.allNamedPredicatesInDomain]; self.infoOnPredicatesThatOccurInCurrentState = self.forInternalUse_getNamedPredicateUUIDsInCurrentState(); return; def do_exit(self, arg): requires(isinstance(self.hackyListToContainStateToReturn, list)); requires(len(self.hackyListToContainStateToReturn) == 0); newState = DescriptionState(); print("Updating allowed/disallowed predicates. NOTE: changes will not be reflected " + \ "until the next time a new description is generated (such as by requesting a " + \ "description that is more abstract as oppossed to one that is less abstract).", flush =True); newState.setDescription(self.currentState.getDescription()); newState.setRawInputDomainBoxes(self.currentState.getRawInputDomainBoxes()); newState.setBoxes(self.currentState.getBoxes()); newState.sideInformationDict = self.currentState.sideInformationDict; newState.internalDictionary["mostRecentOperatorParameters"] = self.currentState.getCopyOfParameters(); newState.internalDictionary["mostRecentOperatorParameters"]["removedPredicates"].difference_update(\ self.newlyAllowedPredicates); newState.internalDictionary["mostRecentOperatorParameters"]["removedPredicates"].update(\ self.newlyDisallowedPredicate ); self.hackyListToContainStateToReturn.append(newState); ensures(isinstance(self.hackyListToContainStateToReturn, list)); ensures(len(self.hackyListToContainStateToReturn) == 1); return True; def forInternalUse_getNamedPredicateUUIDsInCurrentState(self): commandToExecute = "SELECT A.childUUID as predUUID, count(*) as numOccurances " + \ "FROM QAStateValues as A " + \ "WHERE " + \ " (A.fieldName = 'd:root' or A.fieldName = 'd:parent_child') " + \ " AND (A.childUUID NOT LIKE \"frozenset%\") " + \ " AND QAStateUUID = \"" + self.currentState.getID() + "\" " + \ " AND EXISTS (SELECT * FROM predicateInfo WHERE predicateUUID=childUUID) GROUP BY childUUID; "; results = objDatabaseInterface.exec(commandToExecute); valuesToReturn = [(x["numOccurances"], x["predUUID"], str([y for y in self.allNamedPredicatesInDomain if y.getID() == x["predUUID"]][0]) ) \ for x in results ]; valuesToReturn.sort(reverse=True); # I reverse them so to show the most frequently occuring predicates first. return valuesToReturn; def do_list_named_predicates_in_current_state(self, args): print("================", flush=True); print("predicate UUID,\tnumber of occurances in description,\tstring description of predicate", flush=True); valuesToPrint = self.infoOnPredicatesThatOccurInCurrentState; valuesToPrint = [ (str(x[1]) + "\t" + str(x[0]) + "\t" + str(x[2])) for x in valuesToPrint]; myLinuxStyleMoreCommand(valuesToPrint); return; def forInternalUse_completeState(self, thisString, setToRemove): requires(isinstance(thisString, str)); return [x for x in self.UUIDsOfAllNamedPredicatesInDomain if (str(x).startswith(thisString) and (x not in setToRemove))]; def do_print_current_state(self, args): displayForUser(self.currentState, \ self.currentState.getSideInformation("dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered")\ ); # there should only be one state - though it may have been visited multiple # times - that has the user-specified UUID; return; def do_list_disallowed_predicates(self, args): def getPredDescriptionGivenUUID(thisUUID): return [y for y in self.allNamedPredicatesInDomain if y.getID() == thisUUID][0]; originalDisallowed = self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; originalDisallowed = [ (x + "\t" + str(getPredDescriptionGivenUUID(x))) for x in originalDisallowed.difference(self.newlyAllowedPredicates)]; newlyDisallowed = [ ("newly disallowed: " + x + "\t" + str(getPredDescriptionGivenUUID(x))) for x in self.newlyDisallowedPredicate]; print("===================\nDisallowed predicates:\npredicate UUID,\tstring description of predicate", flush=True); myLinuxStyleMoreCommand(newlyDisallowed + originalDisallowed); return; def do_list_allowed_predicates(self, args): def getPredDescriptionGivenUUID(thisUUID): return [y for y in self.allNamedPredicatesInDomain if y.getID() == thisUUID][0]; originalAllowed = set(self.UUIDsOfAllNamedPredicatesInDomain ).difference(self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]); originalAllowed = [ (x + "\t" + str(getPredDescriptionGivenUUID(x))) for x in originalAllowed.difference(self.newlyDisallowedPredicate)]; newlyAllowed = [ ("newly disallowed: " + x + "\t" + str(getPredDescriptionGivenUUID(x))) for x in self.newlyAllowedPredicates]; print("===================\nDisallowed predicates:\npredicate UUID,\tstring description of predicate", flush=True); myLinuxStyleMoreCommand(newlyAllowed + originalAllowed); return; def do_allow_predicate(self, args): args = args.replace(" " , ""); if(args not in self.UUIDsOfAllNamedPredicatesInDomain): print("Error: allow_predicate requires one argument: the UUID of the predicate to disallow. However, " + \ " the argument(s) provided do not seem to be a UUID for a predicate in the domain.", flush=True); originallyDisallowed = self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; originallyAllowed = set(self.UUIDsOfAllNamedPredicatesInDomain).difference(originallyDisallowed); if(args in originallyAllowed): print("Predicate that was requested to be made newly allowed was already allowed. Doing nothing.", flush=True); return; self.newlyAllowedPredicates.update([args]); self.newlyDisallowedPredicate.difference_update([args]); ensures(args in self.newlyAllowedPredicates); ensures(args not in self.newlyDisallowedPredicate); ensures(self.newlyAllowedPredicates.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(self.newlyDisallowedPredicate.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(len(self.newlyAllowedPredicates.difference(self.newlyDisallowedPredicate)) == len(self.newlyAllowedPredicates)); # i.e., that the sets are disjiont... return def complete_allow_predicate(s, text, line, start_index, end_index): try: originalAllowed = set(s.UUIDsOfAllNamedPredicatesInDomain ).difference(s.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]); return s.forInternalUse_completeState(text, originalAllowed); except: errorMessageIndented = " " + traceback.format_exc().replace("\n", "\n "); print(errorMessageIndented, flush=True); return; def do_disallow_predicate(self, args): args = args.replace(" " , ""); if(args not in self.UUIDsOfAllNamedPredicatesInDomain): print("Error: disallow_predicate requires one argument: the UUID of the predicate to disallow. However, " + \ " the argument(s) provided do not seem to be a UUID for a predicate in the domain.", flush=True); originallyDisallowed = self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; originallyAllowed = set(self.UUIDsOfAllNamedPredicatesInDomain).difference(originallyDisallowed); if(args in originallyDisallowed): print("Predicate that was requested to be made newly disallowed was already disallowed. Doing nothing.", flush=True); return; self.newlyAllowedPredicates.difference_update([args]); self.newlyDisallowedPredicate.update([args]); ensures(args not in self.newlyAllowedPredicates); ensures(args in self.newlyDisallowedPredicate); ensures(self.newlyAllowedPredicates.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(self.newlyDisallowedPredicate.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(len(self.newlyAllowedPredicates.difference(self.newlyDisallowedPredicate)) == len(self.newlyAllowedPredicates)); # i.e., that the sets are disjiont... return def complete_disallow_predicate(s, text, line, start_index, end_index): try: originalDisallowed = s.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; return s.forInternalUse_completeState(text, originalDisallowed); except: errorMessageIndented = " " + traceback.format_exc().replace("\n", "\n "); print(errorMessageIndented, flush=True); return; @staticmethod def getID(): uuid = "8502c81c-1afb-4d6c-b910-05328c4350d1"; return uuid; def __init__(self): assert(self.getID() == "8502c81c-1afb-4d6c-b910-05328c4350d1"); return; def internal_apply(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory): hackyListToContainStateToReturn = []; operatorPrompt = self.ManualPredicateReviewPrompt(); operatorPrompt.secondInit(parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory, hackyListToContainStateToReturn); operatorPrompt.cmdloop(); ensures(isinstance(hackyListToContainStateToReturn, list)); ensures(len(hackyListToContainStateToReturn) == 1); return hackyListToContainStateToReturn[0]; class Operator_HistoryExamination(DescriptionOperator): class HistoryExaminationPrompt(cmd.Cmd): prompt="(Fanoos: History Travel)"; @staticmethod def _standardWait(): timeForSleep=config.defaultValues.responceDelayTimeForUnexpectedInputes; print("Sleeping " + str(timeForSleep) + " seconds before responding....", flush=True); timePackageToUseForSleep.sleep(timeForSleep); return; def default(*x, **kwargs): __class__._standardWait(); cmd.Cmd.default(*x, **kwargs); return; def emptyline(*x, **kwargs): __class__._standardWait(); print("Ignoring Empty Line.", flush=True); return; def secondInit(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory, hackyListToContainStateToReturn): requires(isinstance(hackyListToContainStateToReturn, list)); requires(len(hackyListToContainStateToReturn) == 0); self.parsedUserQuestion = parsedUserQuestion; self.domainInformation = domainInformation; self.loadedLearnedModel = loadedLearnedModel; self.currentState = state; self.objectForHistory = objectForHistory; self.historyForThisQuestion = self.objectForHistory[-1][1]; self.statesVisitedInThisQuestion = [x[0].getID() for x in self.historyForThisQuestion]; self.hackyListToContainStateToReturn = hackyListToContainStateToReturn; return; def do_show_history(self, args): if(len(args)> 0): print("warning: show_history does not take any arguments. Ignorng.", flush=True); return; def convertStatesToString(thisStateUUID): if(thisStateUUID == self.currentState.getID()): return str(thisStateUUID) + " <=== (current state)"; else: return str(thisStateUUID); myLinuxStyleMoreCommand([convertStatesToString(x) for x in self.statesVisitedInThisQuestion]); return; def forInternalUse_getStateMatchingUUID(self, args, functionName): requires(isinstance(functionName, str)); requires(len(functionName) > 0); if(isinstance(args, list) and len(args) != 1): print(functionName + " requires one argument: the uuid of the state to print. Ignorng.", flush=True); return None; thisUUID = args; if(isinstance(args, list)): thisUUID = args[0]; stateWithUUID = [x[0] for x in self.historyForThisQuestion if x[0].getID() == thisUUID]; if(len(stateWithUUID) == 0): print("No state has been visited while answering this question that has that UUID", flush=True); return None; # NOTE: len(stateWithUUID) may be greater than one, since we may, by travelling in the history, # revisit the same state multiple times. return stateWithUUID[0]; def do_print_state(self, args): thisState = self.forInternalUse_getStateMatchingUUID(args, "print_state"); if(thisState != None): displayForUser(thisState, \ thisState.getSideInformation("dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered")\ ); # there should only be one state - though it may have been visited multiple # times - that has the user-specified UUID; return; def forInternalUse_completeState(self, thisString): requires(isinstance(thisString, str)); return [x for x in self.statesVisitedInThisQuestion if (str(x).startswith(thisString))]; def complete_print_state(s, text, line, start_index, end_index): try: return s.forInternalUse_completeState(text); except: errorMessageIndented = " " + traceback.format_exc().replace("\n", "\n "); print(errorMessageIndented, flush=True); return; def do_return_to_state(self,args): thisState = self.forInternalUse_getStateMatchingUUID(args, "return_to_state"); if(thisState != None): self.currentState = thisState; return; def complete_return_to_state(s, text, line, start_index, end_index): try: return s.forInternalUse_completeState(text); except: errorMessageIndented = " " + traceback.format_exc().replace("\n", "\n "); print(errorMessageIndented, flush=True); return; def do_exit(self, arg): requires(isinstance(self.hackyListToContainStateToReturn, list)); requires(len(self.hackyListToContainStateToReturn) == 0); self.hackyListToContainStateToReturn.append(self.currentState); ensures(isinstance(self.hackyListToContainStateToReturn, list)); ensures(len(self.hackyListToContainStateToReturn) == 1); return True; @staticmethod def getID(): uuid = "38928f9a-0bef-4c48-92f9-f2a07c0831c2"; return uuid; def __init__(self): assert(self.getID() == "38928f9a-0bef-4c48-92f9-f2a07c0831c2"); return; def internal_apply(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory): hackyListToContainStateToReturn = []; operatorPrompt = self.HistoryExaminationPrompt(); operatorPrompt.secondInit(parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory, hackyListToContainStateToReturn); operatorPrompt.cmdloop(); ensures(isinstance(hackyListToContainStateToReturn, list)); ensures(len(hackyListToContainStateToReturn) == 1); return hackyListToContainStateToReturn[0]; import time; class Operator_FreshGenerateAllBoxes(DescriptionOperator): @staticmethod def getID(): uuid = "f883e116-398c-45b3-b0d4-638d3b875190"; return uuid; def __init__(self): assert(self.getID() == "f883e116-398c-45b3-b0d4-638d3b875190"); raise Exception("Child classes must override"); return; def changeParameter(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): raise Exception("Child classes must override"); ensures(isinstance(dictToReturn, dict)); return dictToReturn; def internal_apply(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): newState = DescriptionState(); dictOfNewParameters = self.changeParameter(parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory); newState.internalDictionary["mostRecentOperatorParameters"] = dictOfNewParameters; startTime = time.process_time(); #location79441a4e-30de-4e64-9ac7-27b2b4b7b503_CEGARLikeAnalysis (listOfBoxesToDescribe, listMappingAxisIndexToVariableInQuestion, rawInputDomainBoxes) = \ parsedUserQuestion.getBoxesToDescribe(loadedLearnedModel, newState ); endTime = time.process_time(); self.recordTimeStats(timingInfoForLocation_2e048534_BoxTest, "location2e048534-c79b-4177-a79d-cc0ef71384d4_boxTest"); self.recordTimeStats([endTime - startTime], "location79441a4e-30de-4e64-9ac7-27b2b4b7b503_CEGARLikeAnalysis"); newState.setBoxes(listOfBoxesToDescribe); newState.setRawInputDomainBoxes(rawInputDomainBoxes); conditionsList = domainInformation.getBaseConditions().copy(); # filtering for those conditions that only discuss the relavent variables... conditionsList = [x for x in conditionsList if x.relaventVariables().issubset(listMappingAxisIndexToVariableInQuestion)]; conditionsList = [x for x in conditionsList \ if x.getID() not in newState.internalDictionary["mostRecentOperatorParameters"]["removedPredicates"]]; startTime = time.process_time(); # location6e888203-1d65-4f09-9601-fefc60dbb13a_generateDescription temp = generateDescription(\ listOfBoxesToDescribe, \ conditionsList, \ listMappingAxisIndexToVariableInQuestion, \ newState \ ); endTime = time.process_time(); self.recordTimeStats([endTime - startTime], "location6e888203-1d65-4f09-9601-fefc60dbb13a_generateDescription"); description = temp["description"]; newState.setDescription(description); newState.setSideInformation("dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered", \ temp["dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered"] ); ensures(isinstance(newState, DescriptionState)); return newState; class Operator_StartOperator(Operator_FreshGenerateAllBoxes): """This operator just uses the parameters provided by the first state to generate the boxes, etc., for the first state....""" @staticmethod def getID(): uuid = "31497f97-4fa9-4ecd-bbce-0b267d847bf4"; return uuid def __init__(self): assert(self.getID() == "31497f97-4fa9-4ecd-bbce-0b267d847bf4"); return; def changeParameter(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): requires(isinstance(stateAppliedTo ,FirstState_DescriptionState)); dictToReturn = stateAppliedTo.getCopyOfParameters(); ensures(isinstance(dictToReturn, dict)); return dictToReturn; class Operator_IncreaseAbstractionLevel(): def __init__(self): raise NotImplementedError("Child classes must override."); return; class Operator_DecreaseAbstractionLevel(): def __init__(self): raise NotImplementedError("Child classes must override."); return; class Operator_IAL_7b16e7a5(Operator_IncreaseAbstractionLevel, Operator_FreshGenerateAllBoxes): @staticmethod def getID(): uuid = "4475bdbc-0fbf-4dbd-aa1c-c19b9c8bf1f0"; return uuid; def __init__(self): assert(self.getID() == "4475bdbc-0fbf-4dbd-aa1c-c19b9c8bf1f0"); return; def changeParameter(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): oldParameters = stateAppliedTo.getCopyOfParameters(); parameters = stateAppliedTo.getCopyOfParameters(); parameters["floatValueForBoxDivisionCutoff"] = oldParameters["floatValueForBoxDivisionCutoff"] * 2.0; parameters["limitOnNumberOfTimesToMerge"] = 0; parameters["splitOnlyOnRelaventVariables"] = False; parameters["precisionForMerging"] = max(1.0, oldParameters["precisionForMerging"] - 1.0); parameters["numberOfSamplesToTry"] = oldParameters["numberOfSamplesToTry"]; parameters["produceGreaterAbstraction"] = True; parameters["exponentialComponent"] = oldParameters["exponentialComponent"] * 1.25; ensures(isinstance(parameters, dict)); ensures(set(parameters.keys()).issuperset(oldParameters.keys())); return parameters; class Operator_DAL_7b16e7a5(Operator_DecreaseAbstractionLevel, Operator_FreshGenerateAllBoxes): @staticmethod def getID(): uuid = "e374ce99-215e-4cfd-8724-da1c2e6ace44"; return uuid; def __init__(self): assert(self.getID() == "e374ce99-215e-4cfd-8724-da1c2e6ace44"); return; def changeParameter(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): oldParameters = stateAppliedTo.getCopyOfParameters(); parameters = stateAppliedTo.getCopyOfParameters(); parameters["floatValueForBoxDivisionCutoff"] = oldParameters["floatValueForBoxDivisionCutoff"] / 2.0; parameters["limitOnNumberOfTimesToMerge"] = 0; parameters["splitOnlyOnRelaventVariables"] = False; parameters["precisionForMerging"] = oldParameters["precisionForMerging"] + 1.0; parameters["numberOfSamplesToTry"] = oldParameters["numberOfSamplesToTry"]; parameters["produceGreaterAbstraction"] = False; parameters["exponentialComponent"] = 0.0; ensures(isinstance(parameters, dict)); ensures(set(parameters.keys()).issuperset(oldParameters.keys())); return parameters; # TODO: refactor the code to allow the other function to inherit these... class ManagerForReviewingAndChangingAllowedStatusOfPredicates(): """ Largely a copy of the content in ManualPredicateReviewPrompt internal-class of Operator_ManualPredicateReview """ def __init__(self, parsedUserQuestion, domainInformation, loadedLearnedModel, state, objectForHistory): # TODO: requires and ensures.... self.parsedUserQuestion = parsedUserQuestion; self.domainInformation = domainInformation; self.loadedLearnedModel = loadedLearnedModel; self.currentState = state; self.objectForHistory = objectForHistory; self.historyForThisQuestion = self.objectForHistory[-1][1]; self.statesVisitedInThisQuestion = [x[0].getID() for x in self.historyForThisQuestion]; self.newlyAllowedPredicates = set(); self.newlyDisallowedPredicate = set(); self.allNamedPredicatesInDomain = self.domainInformation.getBaseConditions().copy(); self.UUIDsOfAllNamedPredicatesInDomain = [x.getID() for x in self.allNamedPredicatesInDomain]; self.infoOnPredicatesThatOccurInCurrentState = self.forInternalUse_getNamedPredicateUUIDsInCurrentState(); return; def applyAndExit(self): removedPredicates = self.currentState.getCopyOfParameters()["removedPredicates"].copy(); forCheckPurposes_removedPredicates_asString = str(removedPredicates); # to check that the below operations # do not modify the original state... removedPredicates.difference_update(\ self.newlyAllowedPredicates); removedPredicates.update(\ self.newlyDisallowedPredicate ); # Below are sorted to address issues with, say, sets not matching up in positions. # If the counts of each letter is the same, then very likely the originals were the # same... assert(sorted(forCheckPurposes_removedPredicates_asString) == \ sorted(str(self.currentState.getCopyOfParameters()["removedPredicates"]))); return removedPredicates; def forInternalUse_getNamedPredicateUUIDsInCurrentState(self): commandToExecute = "SELECT A.childUUID as predUUID, count(*) as numOccurances " + \ "FROM QAStateValues as A " + \ "WHERE " + \ " (A.fieldName = 'd:root' or A.fieldName = 'd:parent_child') " + \ " AND (A.childUUID NOT LIKE \"frozenset%\") " + \ " AND QAStateUUID = \"" + self.currentState.getID() + "\" " + \ " AND EXISTS (SELECT * FROM predicateInfo WHERE predicateUUID=childUUID) GROUP BY childUUID; "; results = objDatabaseInterface.exec(commandToExecute); valuesToReturn = [(x["numOccurances"], x["predUUID"], str([y for y in self.allNamedPredicatesInDomain if y.getID() == x["predUUID"]][0]) ) \ for x in results ]; valuesToReturn.sort(reverse=True); # I reverse them so to show the most frequently occuring predicates first. return valuesToReturn; def do_list_disallowed_predicates(self): # TODO: checks on the sets def getPredDescriptionGivenUUID(thisUUID): return [y for y in self.allNamedPredicatesInDomain if y.getID() == thisUUID][0]; originalDisallowed = self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; originalDisallowed = originalDisallowed.difference(self.newlyAllowedPredicates); newlyDisallowed = self.newlyDisallowedPredicate; assert(len(newlyDisallowed.intersection(originalDisallowed)) == 0); return (newlyDisallowed + originalDisallowed); def do_list_allowed_predicates(self): # TODO: checks on the sets def getPredDescriptionGivenUUID(thisUUID): return [y for y in self.allNamedPredicatesInDomain if y.getID() == thisUUID][0]; originalAllowed = set(self.UUIDsOfAllNamedPredicatesInDomain ).difference(self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]); originalAllowed = originalAllowed.difference(self.newlyDisallowedPredicate); newlyAllowed = self.newlyAllowedPredicates; assert(len(newlyAllowed.intersection(originalAllowed)) == 0); return newlyAllowed + originalAllowed; def do_allow_predicate(self, args): # TODO: check that the arguments are of the expected form with a regular expression (i.e., that the arguments are a single UUID) args = args.replace(" " , ""); if(args not in self.UUIDsOfAllNamedPredicatesInDomain): print("Error: allow_predicate requires one argument: the UUID of the predicate to disallow. However, " + \ " the argument(s) provided do not seem to be a UUID for a predicate in the domain.", flush=True); originallyDisallowed = self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; originallyAllowed = set(self.UUIDsOfAllNamedPredicatesInDomain).difference(originallyDisallowed); if(args in originallyAllowed): print("Predicate that was requested to be made newly allowed was already allowed. Doing nothing.", flush=True); return; self.newlyAllowedPredicates.update([args]); self.newlyDisallowedPredicate.difference_update([args]); ensures(args in self.newlyAllowedPredicates); ensures(args not in self.newlyDisallowedPredicate); ensures(self.newlyAllowedPredicates.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(self.newlyDisallowedPredicate.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(len(self.newlyAllowedPredicates.difference(self.newlyDisallowedPredicate)) == len(self.newlyAllowedPredicates)); # i.e., that the sets are disjiont... return; def do_disallow_predicate(self, args): # TODO: check that the arguments are of the expected form with a regular expression (i.e., that the arguments are a single UUID) args = args.replace(" " , ""); if(args not in self.UUIDsOfAllNamedPredicatesInDomain): print("Error: disallow_predicate requires one argument: the UUID of the predicate to disallow. However, " + \ " the argument(s) provided do not seem to be a UUID for a predicate in the domain.", flush=True); originallyDisallowed = self.currentState.internalDictionary[\ "mostRecentOperatorParameters"]["removedPredicates"]; originallyAllowed = set(self.UUIDsOfAllNamedPredicatesInDomain).difference(originallyDisallowed); if(args in originallyDisallowed): print("Predicate that was requested to be made newly disallowed was already disallowed. Doing nothing.", flush=True); return; self.newlyAllowedPredicates.difference_update([args]); self.newlyDisallowedPredicate.update([args]); ensures(args not in self.newlyAllowedPredicates); ensures(args in self.newlyDisallowedPredicate); ensures(self.newlyAllowedPredicates.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(self.newlyDisallowedPredicate.issubset(self.UUIDsOfAllNamedPredicatesInDomain)); ensures(len(self.newlyAllowedPredicates.difference(self.newlyDisallowedPredicate)) == len(self.newlyAllowedPredicates)); # i.e., that the sets are disjiont... return; class Operator_AutoRevisePredicates_BaseOperator(Operator_FreshGenerateAllBoxes): def computeTableOfPredicateUseInfoTempTable(self): objDatabaseInterface.executeScriptFile(\ "./statesAndOperatorsAndSelection/setupTableForPredicateSelectionAlgorithm.sql"); return; def cleanUpPredicateUseInfoTempTable(self): objDatabaseInterface.exec("DROP TABLE predicateCountsAndResponces;"); return; def determineNewPredicatesToAllowOrDisallowAutomatically(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): assert(self.getID() == "dc5a4db9-b02e-4a7a-a34d-539b7bc28b58"); raise Exception("Child classes must override."); return; def forInternalUse_extraInternalParameters(self): self.abstractionLevelDirection = "increase"; assert(self.getID() == "dc5a4db9-b02e-4a7a-a34d-539b7bc28b58"); raise Exception("Child classes must override"); return; @staticmethod def getID(): uuid = "dc5a4db9-b02e-4a7a-a34d-539b7bc28b58"; return uuid; def changeParameter(self, parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory): oldParameters = stateAppliedTo.getCopyOfParameters(); # Notice that calling the below function here (then cleaning it at the end of # this function) causes the table to be recomputed once each time this # code is run. This is in contrast to, say, running it in the init function, # which would cause it to run only when the operator is formed, which is # just once at the start of loading of Fanoos.... Obviously, improvements # that better support online updates are on the agenda, but for a vanilla # implementation to release, its not a bad start self.computeTableOfPredicateUseInfoTempTable(); self.thisPredicateManager= \ ManagerForReviewingAndChangingAllowedStatusOfPredicates(\ parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory); # The value returned by the below function currently is set to use # the value tracking, etc., in self.thisPredicateManager ... but that # might prove repetative and unnecessarly add more code with minimal # benefit, so we might just have the function determineNewPredicatesToAllowOrDisallowAutomatically # directly return the values for predicates to remore later. self.determineNewPredicatesToAllowOrDisallowAutomatically(\ parsedUserQuestion, domainInformation, loadedLearnedModel, stateAppliedTo, objectForHistory); parameters = stateAppliedTo.getCopyOfParameters(); parameters["removedPredicates"] = self.thisPredicateManager.applyAndExit(); parameters["produceGreaterAbstraction"] = (self.abstractionLevelDirection == "increase"); self.cleanUpPredicateUseInfoTempTable(); ensures(isinstance(parameters, dict)); ensures(set(parameters.keys()).issuperset(oldParameters.keys())); return parameters; def __init__(self): self.forInternalUse_extraInternalParameters(); return; class Operator_AutoRevisePredicates_7b16e7a5(\ Operator_AutoRevisePredicates_BaseOperator, Operator_IncreaseAbstractionLevel): @staticmethod def upperConfidenceBoundSelector(listOfNumberOfSuccesses, listOfNumberOfTimesTried): requires(isinstance(listOfNumberOfSuccesses, list)); requires(isinstance(listOfNumberOfTimesTried, list)); requires(len(listOfNumberOfTimesTried) == len(listOfNumberOfSuccesses)); requires(all([ listOfNumberOfSuccesses[thisIndex] <= listOfNumberOfTimesTried[thisIndex] \ for thisIndex in range(0, len(listOfNumberOfTimesTried))])); requires(all([ isinstance(x, int) for x in listOfNumberOfTimesTried ])); requires(all([ isinstance(x, int) for x in listOfNumberOfSuccesses ])); requires(all([ (x >= 0) for x in listOfNumberOfSuccesses ])); A = np.array(listOfNumberOfSuccesses); B = np.array(listOfNumberOfTimesTried); indexToReturn = np.argmax( (A / B) + np.sqrt( 2 * np.log(np.sum(B)) / B) ); assert(isinstance(indexToReturn, np.int64)); indexToReturn = int(indexToReturn); ensures(isinstance(indexToReturn, int)); ensures(indexToReturn >= 0); ensures(indexToReturn < len(listOfNumberOfTimesTried)); return indexToReturn ; @staticmethod def convertStringSetToSQLStringList(thisStringList): requires(isinstance(thisStringList, list) or isinstance(thisStringList, set)); requires(all([("'" not in x) for x in thisStringList])); return ", ".join(["'" + x + "'" for x in thisStringList]); def determineNewPredicatesToAllowOrDisallowAutomatically(self, parsedUserQuestion, domainInformation,\ loadedLearnedModel, stateAppliedTo, objectForHistory): predicatesAvailable = [x[1] for x in self.thisPredicateManager.infoOnPredicatesThatOccurInCurrentState]; if(len(predicatesAvailable) > 0): templateForQueryToIssue = \ """ SELECT count(*) AS numberOccurances, sum(userSecondResponce IN ({0})) AS numberSuccesses, predicateUUID FROM temp.predicateCountsAndResponces WHERE userFirstResponce in ({1}) AND predicateUUID in ({2}) AND numFirst > numSecond GROUP BY predicateUUID ORDER BY predicateUUID; """; queryToIssue = templateForQueryToIssue.format( self.convertStringSetToSQLStringList(self.desiredEndingUserInputs), \ self.convertStringSetToSQLStringList(self.startingUserInputs), \ self.convertStringSetToSQLStringList(predicatesAvailable) \ ); results = objDatabaseInterface.exec(queryToIssue); numberOccurancesPerPredicate = [x["numberOccurances"] for x in results]; numberSuccessesPerPredicate = [x["numberSuccesses"] for x in results]; predicateList = [x["predicateUUID"] for x in results]; predicatesNotYetTried = set(predicatesAvailable).difference(predicateList); for thisPredicatesNotYetTried in predicatesNotYetTried: numberOccurancesPerPredicate.append(1); numberSuccessesPerPredicate.append(0); predicateList.append(thisPredicatesNotYetTried); indexOfPredicateToRemove = self.upperConfidenceBoundSelector(\ numberSuccessesPerPredicate, numberOccurancesPerPredicate); self.thisPredicateManager.do_disallow_predicate(\ predicateList[indexOfPredicateToRemove]); else: print("auto-predicate review: no named predicates present.", flush=True); return; @staticmethod def getID(): uuid = "553060df-18ca-46d5-926d-2918b808a9fb"; return uuid; def forInternalUse_extraInternalParameters(self): raise Exception("Child classes must override."); self.abstractionLevelDirection = "increase"; self.startingUserInputs = set([""]); self.desiredEndingUserInputs = set([""]); assert(self.getID() == "553060df-18ca-46d5-926d-2918b808a9fb"); return; class Operator_AutoRevisePredicates_IAL_7b16e7a5(Operator_AutoRevisePredicates_7b16e7a5): def forInternalUse_extraInternalParameters(self): self.abstractionLevelDirection = "increase"; self.startingUserInputs = set(["m", "auto_u_m"]); self.desiredEndingUserInputs = set(["l", "b", "auto_u_l"]); assert(self.getID() == "98fbcb97-45b1-4841-9087-19bc63934e38"); return; @staticmethod def getID(): uuid = "98fbcb97-45b1-4841-9087-19bc63934e38"; return uuid; class Operator_AutoRevisePredicates_DAL_7b16e7a5(Operator_AutoRevisePredicates_7b16e7a5): def forInternalUse_extraInternalParameters(self): self.abstractionLevelDirection = "decrease"; self.startingUserInputs = set(["l", "auto_u_l"]); self.desiredEndingUserInputs = set(["m", "b", "auto_u_m"]); assert(self.getID() == "5fd1f949-d803-4188-a4db-e29451f868ca"); return; @staticmethod def getID(): uuid = "5fd1f949-d803-4188-a4db-e29451f868ca"; return uuid;
49,775
51.340694
2,781
py
Fanoos
Fanoos-master/statesAndOperatorsAndSelection/chooseOperatorToApply.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; from statesAndOperatorsAndSelection.descriptionState import DescriptionState; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox, boxSize; import re; import time as timePackageToUseForSleep; import inspect; from statesAndOperatorsAndSelection.automaticOperatorSelection.operationSelectionManagers import \ SelectorManagerBase; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface; import config; def executeDatabaseCommandList(commandsToExecute): for thisCommand in commandsToExecute: objDatabaseInterface.exec(thisCommand); objDatabaseInterface.commit(); return; def recordStateUserResponded(stateToRecord, indexIntoQA, userResponce): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(stateToRecord ,DescriptionState)); requires(isinstance(userResponce, str)); executeDatabaseCommandList([ "UPDATE questionInstance_QAState_relation SET dateAndTimeUserResponded = CURRENT_TIMESTAMP , userResponce = '" + userResponce + "' WHERE " + \ " answerIndex = " + str(indexIntoQA) + " and " + \ " questionInstanceUUID = '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';" ]); return; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~VV~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # Below largely based on the functions recordOperatorGeneral, recordOperatorComputationStarted, # and recordOperatorComputationFinished from UI/cycleToRespondToUserQuestion.py # As such, probably refactoring should be done at some point to reduce the # code redundancy #================================================================================ def recordOpSelectorManagerGeneral(selectorManagerToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(issubclass(type(selectorManagerToRecord) , SelectorManagerBase) or isinstance(selectorManagerToRecord , SelectorManagerBase) ); executeDatabaseCommandList([ "INSERT INTO questionInstance_OpSelectorManager_relation (questionInstanceUUID, startingAnswerIndex, OpSelectorManagerUUID)" +\ "VALUES ('" + \ str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "', " + \ str(indexIntoQA) + ", '" + \ str(selectorManagerToRecord.getID()) + "');", \ "INSERT OR IGNORE INTO OpSelectorManagerInfo (OpSelectorManagerUUID) VALUES ('" + str(selectorManagerToRecord.getID()) + "');" \ ]); return; def recordOpSelectorManagerComputationStarted(selectorManagerToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(issubclass(type(selectorManagerToRecord), SelectorManagerBase) or isinstance(selectorManagerToRecord , SelectorManagerBase) ); executeDatabaseCommandList([ "UPDATE questionInstance_OpSelectorManager_relation SET dateAndTimeComputationStarted = CURRENT_TIMESTAMP WHERE " + \ " startingAnswerIndex = " + str(indexIntoQA) + " and " + \ " questionInstanceUUID = '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';" ]); return; def recordOpSelectorManagerComputationFinished(selectorManagerToRecord, indexIntoQA): requires(isinstance(indexIntoQA, int)); requires(indexIntoQA >= 0); requires(isinstance(type(selectorManagerToRecord), SelectorManagerBase) or isinstance(selectorManagerToRecord , SelectorManagerBase) ); executeDatabaseCommandList([ "UPDATE questionInstance_OpSelectorManager_relation SET dateAndTimeComputationFinished = CURRENT_TIMESTAMP WHERE " + \ " startingAnswerIndex = " + str(indexIntoQA) + " and " + \ " questionInstanceUUID = '" + str(ObjDatabaseValueTracker.get_questionInstanceUUID()) + "';" ]); return; #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ def helper_chooseOperatorToApply(thisOpSelectorManager, typeOfBoxesToGet, \ state, objectForHistory, indexIntoQA, userResponce): recordOpSelectorManagerGeneral(thisOpSelectorManager, indexIntoQA); # Notice we record the start time PRIOR to initializing the class as an instance, # since the initialization process might be expensive - for example, if it # involves pre-computing a table... in the current set up, since a new selector # manager is created each time, doing something expensive at initialization # or in the place where the operator is chosen doesn't make much of a difference... # but in general, this is good to do to ensure the full time is captured... recordOpSelectorManagerComputationStarted(thisOpSelectorManager, indexIntoQA); operatorChosen = thisOpSelectorManager.getOperatorToUse(typeOfBoxesToGet, state, \ objectForHistory, indexIntoQA, userResponce); recordOpSelectorManagerComputationFinished(thisOpSelectorManager, indexIntoQA); return operatorChosen; def chooseOperatorToApply(typeOfBoxesToGet, domainInformation, loadedLearnedModel, state, objectForHistory, indexIntoQA, manualSelectionManager, autoSelectionManager): operatorChosen = None; print("\n\ntype letter followed by enter key:\n" + \ " b - break and ask a different question,\n" + \ " l - less abstract,\n" + \ " m - more abstract,\n" + \ " u - manual operator selection\n------------------", flush=True); while(True): # TODO: replace this loop with utils.promptToSelectFromList , after modifying that fucntion to take in a list # of strings corresponding to values, or a dict mappinging input-value to what it means, for display purposes... thisLine = sys.stdin.readline(); userResponceToRecord = thisLine[:-1];# The -1 to remove the final newline... # We put the below line (recording the user responce) in this file as oppossed to cycleToRespondToUserQuestion.py because # in this file we might conduct the (potentially expensive) hueristic search for an operator, and we want to record # the time the user responded without encorporating the time taken for the hueristic search based on that responce. recordStateUserResponded(state, indexIntoQA, userResponceToRecord); if(thisLine == "b\n"): return None; # TODO: implement the handling of this case better elif(userResponceToRecord in {"l", "m"}): operatorChosen = helper_chooseOperatorToApply(autoSelectionManager, typeOfBoxesToGet, \ state, objectForHistory, indexIntoQA, userResponceToRecord); break; elif(userResponceToRecord == "u"): operatorChosen = helper_chooseOperatorToApply(manualSelectionManager, typeOfBoxesToGet, \ state, objectForHistory, indexIntoQA, userResponceToRecord); break; else: timeForSleep=config.defaultValues.responceDelayTimeForUnexpectedInputes; print("Sleeping " + str(timeForSleep) + " seconds before responding....", flush=True); timePackageToUseForSleep.sleep(timeForSleep); print("Unrecognized Input. Try Again.\n------------------", flush=True); assert(operatorChosen != None); return operatorChosen;
11,494
59.183246
2,781
py
Fanoos
Fanoos-master/statesAndOperatorsAndSelection/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/statesAndOperatorsAndSelection/descriptionState.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; import config; import uuid; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface, executeDatabaseCommandList; from domainsAndConditions.baseClassConditionsToSpecifyPredictsWith import \ MetaCondition_Conjunction, \ Condition_TheBoxItself ; from boxesAndBoxOperations.getBox import boxSize, getDimensionOfBox; from utils.distributionStatics import distributionStatics; import copy; class DescriptionState(): @staticmethod def convertDescriptionStateIDToStringUniformly(thisID): # TODO: write requires to check it is a proper ID... valueToReturn = ""; if(isinstance(thisID, str)): assert("frozen" not in str(thisID)); valueToReturn = thisID; else: assert(isinstance(thisID, frozenset)); valueToReturn = str(sorted(list(thisID))).replace("'", "\"").replace("[", "frozenset({").replace("]", "})"); assert(eval(valueToReturn) == thisID); ensures(isinstance(valueToReturn, str)); ensures(len(valueToReturn) == len(str(thisID)) ); ensures( (set(valueToReturn).symmetric_difference(set(str(thisID)))).issubset({"'", "\""}) ); return valueToReturn; def setID(self): self.uuid = str(uuid.uuid4()); ObjDatabaseValueTracker.set_QAStateUUID_mostRecentBeingComputed(self.uuid);# In a lot of ways, this is # not ideal to set like this, since multiple states could and will exist in use at any particular time # (for example of a pair of such states: prior state and state being computed) return def getID(self): return self.uuid; def __init__(self): self.internalDictionary = dict(); self.internalDictionary["description"] = list(); self.internalDictionary["boxes"]= list(); self.internalDictionary["mostRecentOperatorParameters"] = dict(); self.internalDictionary["mostRecentOperatorParameters"]["removedPredicates"] = set(); self.internalDictionary["rawInputDomainBoxes"] = dict(); # I dislike storing this due to memory consumption.... self.setID(); self.sideInformationDict = dict(); self.sideInformationDict["dictMappingConditionIDToVolumeCoveredAndUniqueVolumeCovered"] = dict(); return; def readParameter(self, parameterName): requires(isinstance(parameterName, str)); requires(len(parameterName) > 0); if(parameterName not in self.internalDictionary["mostRecentOperatorParameters"]): raise Exception("Tried to access non-existant parameter: " + str(parameterName)); # Below we return a copy to help prevent accidental modifications. Note that, while this # is helpful, it does not ensure/prevent accidental modifications since the objects # returned might contain references to other, primary objects, among other potential # difficulties. We do not do a deep copy to (1) avoid various sorts of potential errors # with duplicating objects, as well as (2) the time / memory expense of it. Basically, # a shallow copy seems to be the appropraite level of diligence. return copy.copy(self.internalDictionary["mostRecentOperatorParameters"][parameterName]); def setDescription(self, thisDescription): requires(isinstance(thisDescription, list)); self.internalDictionary["description"] = thisDescription; return; def getDescription(self): return self.internalDictionary["description"]; def setRawInputDomainBoxes(self, theseBoxes): requires(isinstance(theseBoxes, list)); self.internalDictionary["rawInputDomainBoxes"] = theseBoxes; return; def getRawInputDomainBoxes(self): return self.internalDictionary["rawInputDomainBoxes"]; def setBoxes(self, theseBoxes): requires(isinstance(theseBoxes, list)); self.internalDictionary["boxes"] = theseBoxes; return; def getBoxes(self): return self.internalDictionary["boxes"]; def getCopyOfParameters(self): return self.internalDictionary["mostRecentOperatorParameters"].copy(); def setSideInformation(self, key, value): requires(isinstance(key, str)); requires(len(key) > 0); self.sideInformationDict[key] = value; return; def getSideInformation(self, key): requires(isinstance(key, str)); requires(len(key) > 0); if(key not in self.sideInformationDict): raise Exception("Side information of the sort requested does not exist." + \ " HINT: check that you spelled the key correctly."); return self.sideInformationDict[key]; def convertBoxAndLabelToList(self, label, d2Array, listOfAxisForBox): requires(d2Array.shape[0] == len(listOfAxisForBox)); for thisIndex in range(0, len(listOfAxisForBox)): for upperOrLowerSelector in [0, 1]: specificLabel = (\ label + ":" + str(listOfAxisForBox[thisIndex]) + ":" + \ ("u" if (upperOrLowerSelector == 1) else "l") \ ); commandToExecute = \ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, fieldValue) VALUES ('" + \ self.getID() + "', '" + specificLabel + "', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [d2Array[thisIndex, upperOrLowerSelector]] ); objDatabaseInterface.commit(); return @staticmethod def _getAnalysisResult(thisData): tempDict = distributionStatics(thisData); tempDict.pop("numberOfDataPoints"); # We want the number of data points only # to be recorded once for the boxes, namely under the general, non-domain # specific category, not for each item that we might find a distribution for. return tempDict; def _convertBoxesToSQLCommandsToRecordSummaryStatistics(self, variablesBoxesProducedMayBeOver, universeForBoxes): labelBeginning = "bstats"; # short for "box statistics"; #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # domain specific statistics - i.e., those that are most clearly applicale to this specific domain.... #============================================================= labelForDomainSpecificStats = labelBeginning + ":d"; variablesBoxesProducedMayBeOver = [str(x) for x in variablesBoxesProducedMayBeOver]; resultValue = self._getAnalysisResult(self.internalDictionary["boxes"]); for thisKey in resultValue: self.convertBoxAndLabelToList(\ labelForDomainSpecificStats + ":" + thisKey, \ resultValue[thisKey], \ variablesBoxesProducedMayBeOver ); generalSummaryFunctionsAndLabelsForThem = [\ ( (lambda A : np.prod(np.diff(A, axis=1)) ), "bvolume"), \ ( (lambda A : np.min(np.diff(A, axis=1)) ), "bminSideLength"), \ ( (lambda A : np.max(np.diff(A, axis=1)) ), "bmaxSideLength"), \ ( (lambda A : np.sum(np.diff(A, axis=1)) ), "bsumSideLengths"), \ ]; for thisFunctAndLabel in generalSummaryFunctionsAndLabelsForThem: theseValues = [ thisFunctAndLabel[0](x) for x in self.getBoxes()]; resultValue = self._getAnalysisResult(theseValues); specificLabel = labelForDomainSpecificStats + ":" + thisFunctAndLabel[1]; for thisKey in resultValue: commandToExecute = \ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, fieldValue) VALUES ('" + \ self.getID() + "', '" + (specificLabel + ":" + thisKey) + "', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [resultValue[thisKey]] ); objDatabaseInterface.commit(); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ #V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V~V # general statistics - i.e., those that may be easily useable across domains... #============================================================= labelForGeneralStats = labelBeginning + ":g"; numberOfBoxes = len(self.internalDictionary["boxes"]); commandToExecute = \ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, fieldValue) VALUES ('" + \ self.getID() + "', '" + (labelForGeneralStats + ":numberOfDataPoints") + "', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [numberOfBoxes] ); scalingFactors = np.diff(universeForBoxes, axis=1); # NOTE: below, we find the NORMALIZED side-lengths. Yes, this is normalized # even though the constant offset term (subtracting the minimal value) is not shown - those terms would # simply cancel when we compute the interval length.... getScaledSideLengths = \ (lambda A: (np.diff(A, axis=1) / scalingFactors)); labelForGeneralStats = labelForGeneralStats + ":normalizedInputBoxes"; generalSummaryFunctionsAndLabelsForThem = [\ ( (lambda A : np.prod(getScaledSideLengths(A)) ), "bvolume"), \ ( (lambda A : np.min(getScaledSideLengths(A)) ), "bminSideLength"), \ ( (lambda A : np.max(getScaledSideLengths(A)) ), "bmaxSideLength"), \ ( (lambda A : np.sum(getScaledSideLengths(A)) ), "bsumSideLengths"), \ ]; for thisFunctAndLabel in generalSummaryFunctionsAndLabelsForThem: theseValues = [ thisFunctAndLabel[0](x) for x in self.getRawInputDomainBoxes()]; resultValue = self._getAnalysisResult(theseValues); specificLabel = labelForGeneralStats + ":" + thisFunctAndLabel[1]; for thisKey in resultValue: commandToExecute = \ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, fieldValue) VALUES ('" + \ self.getID() + "', '" + (specificLabel + ":" + thisKey) + "', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [resultValue[thisKey]] ); objDatabaseInterface.commit(); #^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^_^ return; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ def _helper_convertDescriptionToSQLCommands(self, thisObj): thisObjUUIDAsString = DescriptionState.convertDescriptionStateIDToStringUniformly(thisObj.getID()); # need for the case of a MetaCondition_Conjunction, but does the proper thing in other cases as well if(isinstance(thisObj, MetaCondition_Conjunction)): for childObj in thisObj.listOfConditionsToConjunct: objDatabaseInterface.exec(\ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, parentUUID , childUUID) VALUES ('" + \ self.getID() + "', 'd:parent_child', '" + thisObjUUIDAsString + "', '" + childObj.getID() + "' );" \ ); self._helper_convertDescriptionToSQLCommands(childObj); elif(isinstance(thisObj, Condition_TheBoxItself)): commandToExecute = \ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, childUUID, fieldValue) VALUES ('" + \ self.getID() + "', 'd:box:value', '" + thisObjUUIDAsString + "', ? );"; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ commandToExecute, [str(thisObj)] ); return; def _convertDescriptionToSQLCommands(self): for thisObj in self.getDescription(): thisObjUUIDAsString = DescriptionState.convertDescriptionStateIDToStringUniformly(thisObj.getID()); # need for the case of a MetaCondition_Conjunction, but does the proper thing in other cases as well objDatabaseInterface.exec(\ "INSERT INTO QAStateValues ( QAStateUUID , fieldName, childUUID) VALUES ('" + \ self.getID() + "', 'd:root', '" + thisObjUUIDAsString + "' );" \ ); self._helper_convertDescriptionToSQLCommands(thisObj); objDatabaseInterface.commit(); return; def _convertDictionaryToSQLCommands(self,leadingTag, dictToRecord): requires(isinstance(leadingTag, str)); requires(len(leadingTag) > 0); requires(isinstance(dictToRecord, dict)); requires(all([isinstance(x, str) for x in dictToRecord.keys()])); requires(all([(len(x) > 0) for x in dictToRecord.keys()])); for keyName in dictToRecord: assert(isinstance(keyName, str)); fieldName = leadingTag + ":" + keyName; objDatabaseInterface.interfaceBleed_insertValuesForBlob(\ "INSERT INTO QAStateValues (QAStateUUID, fieldName, fieldValue) VALUES ( '" + self.getID() + "', '" + fieldName + "' , ?);", \ [dictToRecord[keyName]] \ ); objDatabaseInterface.commit(); return; def recordInDatabase(self, variablesBoxesProducedMayBeOver, universeForBoxes): self._convertDictionaryToSQLCommands("p", self.internalDictionary["mostRecentOperatorParameters"]) self._convertDictionaryToSQLCommands("s", self.sideInformationDict); self._convertDescriptionToSQLCommands(); self._convertBoxesToSQLCommandsToRecordSummaryStatistics(variablesBoxesProducedMayBeOver, universeForBoxes); return; class FirstState_DescriptionState(DescriptionState): def getFloatValueForBoxDivisionCutoff(self): print("Enter a fraction of the universe box length to limit refinement to at the beginning. " + \ "Value must be a positive real number less than or equal to one.", flush=True); thisLine = sys.stdin.readline(); thisLineMissingNewLine = thisLine.replace("\n", ""); floatValueForBoxDivisionCutoff = None; try: floatValueForBoxDivisionCutoff = float(thisLineMissingNewLine); except: raise Exception("Unable to convert input into a floating point value: " + \ str(thisLineMissingNewLine)); assert(floatValueForBoxDivisionCutoff != None); if(floatValueForBoxDivisionCutoff <= 0 or floatValueForBoxDivisionCutoff > 1): raise Exception("Provided value is outside of the range (0.0, 1.0]. Value provided: " + \ str(thisLineMissingNewLine)); return floatValueForBoxDivisionCutoff; def __init__(self): DescriptionState.__init__(self); parameterDict = self.internalDictionary["mostRecentOperatorParameters"]; self.internalDictionary["mostRecentOperatorParameters"]["removedPredicates"] = set(); parameterDict["floatValueForBoxDivisionCutoff"] = self.getFloatValueForBoxDivisionCutoff(); parameterDict["limitOnNumberOfTimesToMerge"] = config.defaultValues.limitOnNumberOfTimesToMerge; parameterDict["splitOnlyOnRelaventVariables"]=False; parameterDict["precisionForMerging"]= config.defaultValues.precisionForMerging; parameterDict["numberOfSamplesToTry"]=int(config.defaultValues.numberOfSamplesToTry); parameterDict["produceGreaterAbstraction"]=False; parameterDict["exponentialComponent"]= 0.0; return;
19,758
54.659155
2,781
py
Fanoos
Fanoos-master/statesAndOperatorsAndSelection/automaticOperatorSelection/operationSelectionManagers.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import pickle; import numpy as np; import sys; from utils.contracts import *; import config; import uuid; from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; from databaseInterface.databaseIOManager import objDatabaseInterface, executeDatabaseCommandList; from statesAndOperatorsAndSelection.descriptionOperator import *; class SelectorManagerBase(): @staticmethod def setID(): raise Exception("This class uses hard-coded UUIDs; the setID method is not necessary to envoke"); return @staticmethod def getID(): return "4f6a2f1d-23ec-4db8-8775-68d57759cfad"; @staticmethod def getAllAvailableOperators(): listOfAvailableOperators = [\ Operator_IAL_7b16e7a5, \ Operator_DAL_7b16e7a5, \ Operator_HistoryExamination, \ Operator_ManualPredicateReview, \ Operator_AutoRevisePredicates_DAL_7b16e7a5, \ Operator_AutoRevisePredicates_IAL_7b16e7a5 \ ]; return {x.getID() : x for x in listOfAvailableOperators}; def _prepare(self): return; def prepareForSelectors(self): # TODO: this should probably be private but over-rideable by child classes... return; def prepareForWeigher(self): # TODO: this should probably be private but over-rideable by child classes.. return; def __init__(self, domainInformation, loadedLearnedModel): # These two parameters should # be constant over the entire run.... self._prepare(); self.domainInformation = domainInformation; self.loadedLearnedModel = loadedLearnedModel; self.selectors = []; self.weigher = None; self.prepareForSelectors(); self.prepareForWeigher(); return; def getOperatorToUse(self, typeOfBoxesToGet, state, \ objectForHistory, indexIntoQA, userResponce): raise NotImplementedError("Child classes must overwrite this."); return; def _cleanUpForSelectors(self): requires( (self.selectors is None) or isinstance(self.selectors, list)); if(self.selectors is not None and len(self.selectors) > 0): for thisSelector in self.selectors: thisSelector.cleanUp(); return; def _cleanUpForWeigher(self): if( self.weigher is not None ): self.weigher.cleanUp(); return; def cleanUp(self): self._cleanUpForSelectors(); self._cleanUpForWeigher(); return; class Manual_SelectorManager(SelectorManagerBase): @staticmethod def getID(): return "1c612e8b-4545-4cc6-8c54-bb696caa0e48"; def prepareForSelectors(self): raise Exception("Not Used in this Class."); return; def prepareForWeigher(self): raise Exception("Not Used in this Class."); return; def __init__(self, domainInformation, loadedLearnedModel): # These two parameters should # be constant over the entire run.... self.domainInformation = domainInformation; self.loadedLearnedModel = loadedLearnedModel; self.selectors = None; # TODO: fill this in with selectors to come self.weigher = None; return; def getOperatorToUse(self, typeOfBoxesToGet, state, \ objectForHistory, indexIntoQA, userResponce): dictMappingOptionsToOperatorUUIDs = {\ "L" : Operator_DAL_7b16e7a5.getID(), \ "M" : Operator_IAL_7b16e7a5.getID(), \ "h" : Operator_HistoryExamination.getID(), \ "u" : Operator_ManualPredicateReview.getID(), \ "auto_u_L" : Operator_AutoRevisePredicates_DAL_7b16e7a5.getID(), \ "auto_u_M" : Operator_AutoRevisePredicates_IAL_7b16e7a5.getID() \ }; dictMappingOperatorUUIDsToDescriptions = {\ Operator_DAL_7b16e7a5.getID() : "less abstract", \ Operator_IAL_7b16e7a5.getID() : "more abstract", \ Operator_HistoryExamination.getID() : "history travel", \ Operator_ManualPredicateReview.getID() : "manual predicate review", \ Operator_AutoRevisePredicates_DAL_7b16e7a5.getID() : "auto predicate review: less abstract", \ Operator_AutoRevisePredicates_IAL_7b16e7a5.getID() : "auto predicate review: more abstract" \ }; print("Note: responces are case-sensative"); for thisOption in dictMappingOptionsToOperatorUUIDs: print(" " + thisOption + " - " + \ dictMappingOperatorUUIDsToDescriptions[dictMappingOptionsToOperatorUUIDs[thisOption]]); print("------------------"); thisLine = sys.stdin.readline(); userResponceToRecord = thisLine[:-1];# The -1 to remove the final newline... return self.getAllAvailableOperators()[\ dictMappingOptionsToOperatorUUIDs[userResponceToRecord]](); class OriginalMethod_SelectorManager(SelectorManagerBase): """ this function largely based on the code for class Manual_SelectorManager, uuid 1c612e8b-4545-4cc6-8c54-bb696caa0e48 """ @staticmethod def getID(): return "a6cf1bb2-bd1d-465a-9067-a2b8e2cc39de"; def prepareForSelectors(self): raise Exception("Not Used in this Class."); return; def prepareForWeigher(self): raise Exception("Not Used in this Class."); return; def __init__(self, domainInformation, loadedLearnedModel): # These two parameters should # be constant over the entire run.... self.domainInformation = domainInformation; self.loadedLearnedModel = loadedLearnedModel; self.selectors = None; # TODO: fill this in with selectors to come self.weigher = None; return; def getOperatorToUse(self, typeOfBoxesToGet, state, \ objectForHistory, indexIntoQA, userResponce): dictMappingOptionsToOperatorUUIDs = {\ "l" : Operator_DAL_7b16e7a5.getID(), \ "m" : Operator_IAL_7b16e7a5.getID() \ }; return self.getAllAvailableOperators()[\ dictMappingOptionsToOperatorUUIDs[userResponce]]();
7,336
32.35
165
py
Fanoos
Fanoos-master/statesAndOperatorsAndSelection/automaticOperatorSelection/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/databaseInterface/databaseValueTracker.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; import uuid; def isProperUUID(thisProposedUUID): if(not isinstance(thisProposedUUID, str)): return False; if(len(thisProposedUUID) <= 0): return False; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ return True; class DatabaseValueTracker(): def _produce_setGeneric(self, keyName): requires(isinstance(keyName, str)); requires(len(keyName) > 0); def setValue(self, uuidToUse=None): requires( isProperUUID( uuidToUse) or (uuidToUse is None)); self._values[keyName] = uuidToUse if (uuidToUse is not None) else str(uuid.uuid4()); return; return setValue; def _produce_getGeneric(self, keyName): requires(isinstance(keyName, str)); requires(len(keyName) > 0); def getValue(self): requires(keyName in self._values); return self._values[keyName]; return getValue; def __init__(self): self._values = dict(); self._keyNames = ["sessionUUID" , "userUUID" , "domainUUID" , "questionInstanceUUID" , "QAStateUUID_mostRecentBeingComputed" , "QAOperatorUUID" , "predicateUUID" , "labelUUID"]; for thisKeyName in self._keyNames: setattr(DatabaseValueTracker, "set_" + thisKeyName, self._produce_setGeneric(thisKeyName)); setattr(DatabaseValueTracker, "get_" + thisKeyName, self._produce_getGeneric(thisKeyName)); return; ObjDatabaseValueTracker = DatabaseValueTracker();
5,424
65.158537
2,785
py
Fanoos
Fanoos-master/databaseInterface/databaseIOManager.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; import sqlite3 import config; import pickle; import sys; class DatabaseInterface(): def open(self): raise NotImplementedError(); return; # Executes the command and returns the result. # Returns all rows given by the result - to retrieve # only a subset of the rows, the SQL command provided # must use the LIMIT clause. Result will be returned as # list of dictionaries that each map a column name to a value... def exec(sqlCommandString): requires(isinstance(sqlCommandString, str)); raise NotImplementedError(); return; def commit(self): raise NotImplementedError(); return; def rollback(self): raise NotImplementedError(); return; def close(self): raise NotImplementedError(); return; def executeScriptFile(self, scriptName): raise NotImplementedError(); return; class Sqlite3Database(DatabaseInterface): def __init__(self, databaseName=config.defaultValues.databaseName): self.databaseName = databaseName; self.executedCommandAfterLastCommit = False; self.connection= None; self.cursor = None; return def open(self): self.connection = sqlite3.connect(self.databaseName , \ timeout=config.defaultValues.databaseWriteTimeoutLimit); def dict_factory(cursor, row): d = {} for idx, col in enumerate(cursor.description): d[col[0]] = row[idx] return d; self.connection.row_factory = dict_factory; self.cursor = self.connection.cursor(); ensures(self.connection is not None); ensures(self.cursor is not None); return; def exec(self, sqlCommandString): requires(isinstance(sqlCommandString, str)); requires(self.connection is not None); # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg self.executedCommandAfterLastCommit = True; self.cursor.execute(sqlCommandString); # Note that this method can only handle one SQL # command at a time. This could be worked-around to some degree by splitting commands # on the occurance of semicolons, but that in general is not a good approach when # the content might hold strings. resultToReturn = self.cursor.fetchall(); ensures(isinstance(resultToReturn, list)); ensures(all([isinstance(x, dict) for x in resultToReturn])); return resultToReturn; def commit(self): requires(self.connection is not None); self.connection.commit(); self.executedCommandAfterLastCommit = False; return; def close(self): requires(self.connection is not None); if( self.executedCommandAfterLastCommit): sys.stderr.write("Warning: There have been commands executed since the last database commit." + \ "\n Currently closing the database now, so any changes made since the last commit " + \ "\n will be lost.\n\n"); sys.stderr.flush(); self.connection.close(); return; def rollback(self): requires(self.connection is not None); raise NotImplementedError(); return; def executeScriptFile(self, scriptName): # This function does not return anything... requires(isinstance(scriptName, str)); requires(" " not in scriptName); fh = open(scriptName, 'r'); sqlScript = fh.read(); fh.close(); self.cursor.executescript(sqlScript); return; def interfaceBleed_insertValuesForBlob(self, primarySQLCommand, blobValuesToRecord): requires(isinstance(primarySQLCommand, str)); requires(len(primarySQLCommand) > 0); requires(isinstance( blobValuesToRecord, list)); requires(len(blobValuesToRecord) == primarySQLCommand.count("?")); listOfObjectsConvertedToBinary = \ [sqlite3.Binary(pickle.dumps(x)) for x in blobValuesToRecord]; self.cursor.execute(primarySQLCommand, listOfObjectsConvertedToBinary); return; objDatabaseInterface = Sqlite3Database(); def executeDatabaseCommandList(commandsToExecute): for thisCommand in commandsToExecute: objDatabaseInterface.exec(thisCommand); objDatabaseInterface.commit(); return;
8,345
48.094118
2,791
py