text
string
size
int64
token_count
int64
from copy import copy from itertools import groupby import unittest from datetime import datetime, timedelta from typing import List from activitywatch.base import Watcher, Activity, Logger from activitywatch.settings import Settings from activitywatch.utils import floor_datetime, ceil_datetime from activitywatch.filters.split import split_by_interval, overlaps from activitywatch.filters.chunk import chunk_by_tags class MockWatcher(Watcher): def run(self): pass def wait(self): pass identifier = "mock" def __init__(self): settings = Settings() settings["watchers"][self.identifier] = {} Watcher.__init__(self) class MockLogger(Logger): def log(self, activities: List[Activity]): pass def wait(self): pass identifier = "mock" def __init__(self): settings = Settings() settings["loggers"][self.identifier] = {} Logger.__init__(self) class LoggerWatcherTest(unittest.TestCase): def test_activity_flow(self): watcher = MockWatcher() logger = MockLogger() logger.add_watcher(watcher) watcher.dispatch_activity(Activity("test", datetime.now()-timedelta(days=1), datetime.now())) activities = logger.flush_activities() self.assertTrue(len(activities) == 1) activities = logger.flush_activities() self.assertTrue(len(activities) == 0) class ActivityTest(unittest.TestCase): def test_to_zenobase(self): TAG = "something" activity = Activity(TAG, started_at=datetime.now(), ended_at=datetime.now()) event = activity.to_zenobase_event() self.assertTrue(event["tag"] == TAG) class SettingsTest(unittest.TestCase): def test_instance(self): self.assertIs(Settings(), Settings()) HOUR = timedelta(hours=1) class SplitActivityTest(unittest.TestCase): def test_by_hour(self): dt = datetime(2015, 1, 1, 8, 30) td = timedelta(hours=3, minutes=23) activity = Activity([], dt, dt+td) split = split_by_interval([copy(activity), copy(activity)], interval=HOUR) self.assertEqual(len(split), 8) activity.end += -td + timedelta(minutes=2) split = split_by_interval([copy(activity)], interval=HOUR) self.assertEqual(len(split), 1) def test_ceil_hour(self): def ceil_hour(td): return ceil_datetime(td, td=timedelta(hours=1)) self.assertEqual(ceil_hour(datetime(2015, 1, 1, 6, 2)), datetime(2015, 1, 1, 7)) self.assertEqual(ceil_hour(datetime(2015, 1, 1, 6, 2)), ceil_hour(datetime(2015, 1, 1, 6, 58))) self.assertNotEqual(ceil_hour(datetime(2015, 1, 1, 5, 2)), ceil_hour(datetime(2015, 1, 1, 6, 4))) def test_floor_hour(self): def floor_hour(td): return floor_datetime(td, td=timedelta(hours=1)) self.assertEqual(floor_hour(datetime(2015, 1, 1, 6, 2)), datetime(2015, 1, 1, 6)) self.assertEqual(floor_hour(datetime(2015, 1, 1, 6, 2)), floor_hour(datetime(2015, 1, 1, 6, 5))) def test_overlaps_hour(self): def overlaps_hours(td): return overlaps(td, interval=timedelta(hours=1)) activity = Activity([], datetime(2015, 1, 1, 5, 23), datetime(2015, 1, 1, 6, 6)) self.assertTrue(overlaps_hours(activity)) activity = Activity([], datetime(2015, 1, 1, 5, 23), datetime(2015, 1, 1, 6, 0, 0, 1)) self.assertTrue(overlaps_hours(activity)) activity = Activity([], datetime(2015, 1, 1, 6, 30), datetime(2015, 1, 1, 6, 59)) self.assertFalse(overlaps_hours(activity)) class ChunkTest(unittest.TestCase): def test_chunk_by_tags(self): interval = timedelta(minutes=5) start = floor_datetime(datetime.now(), interval) activities = [Activity(["test"], start, start+interval*0.5), Activity(["test2"], start+interval, start+interval*1.5), Activity(["test"], start+interval*2, start+interval*2.5)] self.assertEqual(3, len(activities)) activities.append(Activity(["test"], start+interval, start+interval*1.5)) self.assertEqual(4, len(activities)) self.assertEqual(2, len(chunk_by_tags(activities)))
4,253
1,474
voc_it = [ ['a', 'noun', 'c'], ['a', 'preposition', 'a'], ['abbagliante', 'pres_part', 'c'], ['abbagliante', 'adjective', 'c'], ['abbagliante', 'noun', 'c'], ['abbaiare', 'verb', 'c'], ['abbandonare', 'verb', 'a'], ['abbandonato', 'past_part', 'b'], ['abbandonato', 'adjective', 'b'], ['abbandono', 'noun', 'b'], ['abbassare', 'verb', 'a'], ['abbasso', 'adverb', 'c'], ['abbasso', 'exclamation', 'c'], ['abbastanza', 'adverb', 'a'], ['abbattere', 'verb', 'b'], ['abbeverare', 'verb', 'c'], ['abbigliamento', 'noun', 'b'], ['abbinare', 'verb', 'b'], ['abbonamento', 'noun', 'b'], ['abbonare', 'verb', 'c'], ['abbondante', 'pres_part', 'b'], ['abbondante', 'adjective', 'b'], ['abbondare', 'verb', 'c'], ['abbottonare', 'verb', 'c'], ['abbracciare', 'verb', 'a'], ['abbraccio', 'noun', 'b'], ['abbreviare', 'verb', 'c'], ['abbronzare', 'verb', 'c'], ['abete', 'noun', 'c'], ['abile', 'adjective', 'b'], ['abilità', 'noun', 'b'], ['abisso', 'noun', 'b'], ['abitante', 'pres_part', 'b'], ['abitante', 'adjective', 'b'], ['abitante', 'noun', 'b'], ['abitare', 'verb', 'a'], ['abitare', 'noun', 'a'], ['abitazione', 'noun', 'b'], ['abito', 'noun', 'a'], ['abituale', 'adjective', 'b'], ['abituare', 'verb', 'a'], ['abitudine', 'noun', 'a'], ['abolire', 'verb', 'b'], ['abortire', 'verb', 'c'], ['aborto', 'noun', 'c'], ['abruzzese', 'adjective', 'c'], ['abruzzese', 'noun', 'c'], ['abusare', 'verb', 'c'], ['abuso', 'noun', 'b'], ['acca', 'noun', 'c'], ['accademia', 'noun', 'b'], ['accademico', 'adjective', 'b'], ['accademico', 'noun', 'b'], ['accadere', 'verb', 'a'], ['accampamento', 'noun', 'c'], ['accanto', 'adverb', 'a'], ['accappatoio', 'noun', 'c'], ['accarezzare', 'verb', 'b'], ['accattone', 'noun', 'c'], ['accavallare', 'verb', 'c'], ['accecare', 'verb', 'c'], ['accedere', 'verb', 'b'], ['accelerare', 'verb', 'b'], ['acceleratore', 'adjective', 'c'], ['acceleratore', 'noun', 'c'], ['accelerazione', 'noun', 'b'], ['accendere', 'verb', 'a'], ['accendino', 'noun', 'c'], ['accennare', 'verb', 'b'], ['accenno', 'noun', 'c'], ['accentare', 'verb', 'c'], ['accertamento', 'noun', 'b'], ['accertare', 'verb', 'b'], ['acceso', 'past_part', 'b'], ['acceso', 'adjective', 'b'], ['accesso', 'noun', 'a'], ['accessorio', 'adjective', 'b'], ['accessorio', 'noun', 'b'], ['accetta', 'noun', 'c'], ['accettabile', 'adjective', 'b'], ['accettare', 'verb', 'a'], ['acchiappare', 'verb', 'c'], ['acciacco', 'noun', 'c'], ['acciaio', 'noun', 'b'], ['accidente', 'noun', 'b'], ['acciuga', 'noun', 'c'], ['accogliente', 'pres_part', 'c'], ['accogliente', 'adjective', 'c'], ['accoglienza', 'noun', 'b'], ['accogliere', 'verb', 'a'], ['accoltellare', 'verb', 'c'], ['accomodare', 'verb', 'b'], ['accompagnare', 'verb', 'a'], ['acconsentire', 'verb', 'c'], ['accontentare', 'verb', 'b'], ['accorciare', 'verb', 'c'], ['accordare', 'verb', 'b'], ['accordo', 'noun', 'a'], ['accorgersi', 'verb', 'a'], ['accorrere', 'verb', 'c'], ['accostare', 'verb', 'b'], ['accudire', 'verb', 'c'], ['accumulare', 'verb', 'b'], ['accumulatore', 'adjective', 'c'], ['accumulatore', 'noun', 'c'], ['accurato', 'past_part', 'b'], ['accurato', 'adjective', 'b'], ['accusa', 'noun', 'a'], ['accusare', 'verb', 'a'], ['accento', 'noun', 'b'], ['acerbo', 'adjective', 'c'], ['aceto', 'noun', 'c'], ['acido', 'adjective', 'b'], ['acido', 'noun', 'b'], ['acqua', 'noun', 'a'], ['acquarello', 'noun', 'c'], ['acquario', 'noun', 'c'], ['acquasanta', 'noun', 'c'], ['acquisire', 'verb', 'b'], ['acquisizione', 'noun', 'b'], ['acquistare', 'verb', 'a'], ['acquisto', 'noun', 'a'], ['acquolina', 'noun', 'c'], ['acrobata', 'noun', 'c'], ['acuto', 'adjective', 'b'], ['acuto', 'noun', 'b'], ['adattare', 'verb', 'b'], ['adattatore', 'noun', 'c'], ['adatto', 'adjective', 'a'], ['addetto', 'past_part', 'b'], ['addetto', 'adjective', 'b'], ['addetto', 'noun', 'b'], ['addio', 'exclamation', 'b'], ['addio', 'noun', 'b'], ['addirittura', 'adverb', 'a'], ['addizione', 'noun', 'c'], ['addobbare', 'verb', 'c'], ['addolcire', 'verb', 'c'], ['addomesticare', 'verb', 'c'], ['addormentarsi', 'verb', 'b'], ['addormentato', 'past_part', 'c'], ['addormentato', 'adjective', 'c'], ['addossare', 'verb', 'a'], ['addosso', 'adverb', 'c'], ['addosso', 'exclamation', 'c'], ['addrizzare', 'verb', 'c'], ['adeguare', 'verb', 'b'], ['adeguato', 'past_part', 'b'], ['adeguato', 'adjective', 'b'], ['adeguato', 'noun', 'b'], ['aderente', 'pres_part', 'c'], ['aderente', 'adjective', 'c'], ['aderente', 'noun', 'c'], ['aderire', 'verb', 'b'], ['adesione', 'noun', 'b'], ['adesso', 'adverb', 'a'], ['adolescente', 'adjective', 'a'], ['adolescente', 'noun', 'a'], ['adolescenza', 'noun', 'b'], ['adoperare', 'verb', 'b'], ['adorare', 'verb', 'a'], ['adottare', 'verb', 'a'], ['adozione', 'noun', 'b'], ['adriatico', 'adjective', 'c'], ['adulto', 'adjective', 'a'], ['adulto', 'noun', 'a'], ['aereo', 'adjective', 'a'], ['aereo', 'noun', 'a'], ['aereo', 'noun', 'b'], ['aeroplano', 'noun', 'c'], ['aeroporto', 'noun', 'b'], ['afa', 'noun', 'c'], ['affacciare', 'verb', 'b'], ['affamare', 'verb', 'c'], ['affamato', 'past_part', 'c'], ['affamato', 'adjective', 'c'], ['affamato', 'noun', 'c'], ['affannarsi', 'verb', 'c'], ['affannato', 'past_part', 'c'], ['affannato', 'adjective', 'c'], ['affanno', 'noun', 'c'], ['affare', 'noun', 'a'], ['affascinante', 'pres_part', 'b'], ['affascinante', 'adjective', 'b'], ['affascinare', 'verb', 'b'], ['affaticare', 'verb', 'c'], ['affatto', 'adverb', 'a'], ['affermare', 'verb', 'a'], ['affermazione', 'noun', 'b'], ['afferrare', 'verb', 'b'], ['affettare', 'verb', 'c'], ['affettato', 'past_part', 'c'], ['affettato', 'adjective', 'c'], ['affettato', 'noun', 'c'], ['affetto', 'noun', 'b'], ['affetto', 'adjective', 'b'], ['affettuoso', 'adjective', 'b'], ['affezionato', 'past_part', 'c'], ['affezionato', 'adjective', 'c'], ['affiancare', 'verb', 'b'], ['affidamento', 'noun', 'b'], ['affidare', 'verb', 'a'], ['affilato', 'past_part', 'c'], ['affilato', 'adjective', 'c'], ['affinché', 'conjunction', 'b'], ['affittare', 'verb', 'b'], ['affitto', 'noun', 'b'], ['affogare', 'verb', 'c'], ['affollare', 'verb', 'c'], ['affondare', 'verb', 'b'], ['affresco', 'noun', 'b'], ['affrontare', 'verb', 'a'], ['affumicare', 'verb', 'c'], ['africano', 'adjective', 'b'], ['africano', 'noun', 'b'], ['agenda', 'noun', 'b'], ['agente', 'pres_part', 'a'], ['agente', 'adjective', 'a'], ['agente', 'noun', 'a'], ['agenzia', 'noun', 'a'], ['agganciare', 'verb', 'b'], ['aggettivo', 'noun', 'b'], ['aggiornamento', 'noun', 'b'], ['aggiornare', 'verb', 'b'], ['aggirare', 'verb', 'b'], ['aggiungere', 'verb', 'a'], ['aggiustare', 'verb', 'b'], ['aggrapparsi', 'verb', 'b'], ['aggravare', 'verb', 'c'], ['aggredire', 'verb', 'b'], ['aggressione', 'noun', 'b'], ['aggressivo', 'adjective', 'b'], ['agiato', 'past_part', 'c'], ['agiato', 'adjective', 'c'], ['agile', 'adjective', 'c'], ['agio', 'noun', 'b'], ['agire', 'verb', 'a'], ['agitare', 'verb', 'b'], ['agitazione', 'noun', 'b'], ['aglio', 'noun', 'c'], ['agnello', 'noun', 'b'], ['ago', 'noun', 'b'], ['agonia', 'noun', 'c'], ['agosto', 'noun', 'a'], ['agricolo', 'adjective', 'b'], ['agricoltore', 'noun', 'c'], ['agricoltura', 'noun', 'b'], ['agrume', 'noun', 'c'], ['aguzzare', 'verb', 'c'], ['aguzzo', 'adjective', 'c'], ['aiuola', 'noun', 'c'], ['aiutare', 'verb', 'a'], ['aiuto', 'noun', 'a'], ['aiuto', 'exclamation', 'a'], ['ala', 'noun', 'a'], ['alba', 'noun', 'a'], ['albanese', 'adjective', 'b'], ['albanese', 'noun', 'b'], ['albergo', 'noun', 'a'], ['albero', 'noun', 'a'], ['albicocca', 'noun', 'c'], ['albicocca', 'adjective', 'c'], ['album', 'noun', 'a'], ['alcol', 'noun', 'b'], ['alcuno', 'adjective', 'a'], ['alcuno', 'pronoun', 'a'], ['alfabeto', 'noun', 'c'], ['alga', 'noun', 'c'], ['algerino', 'adjective', 'c'], ['algerino', 'noun', 'c'], ['alieno', 'adjective', 'b'], ['alieno', 'noun', 'b'], ['alimentare', 'adjective', 'b'], ['alimentare', 'noun', 'b'], ['alimentare', 'verb', 'b'], ['alimentari', 'noun', 'c'], ['alimentazione', 'noun', 'b'], ['alimento', 'noun', 'b'], ['alito', 'noun', 'c'], ['allacciare', 'verb', 'c'], ['allagare', 'verb', 'c'], ['allargare', 'verb', 'b'], ['allarmare', 'verb', 'c'], ['allarme', 'noun', 'b'], ['allattare', 'verb', 'c'], ['alleanza', 'noun', 'b'], ['allearsi', 'verb', 'c'], ['alleato', 'past_part', 'b'], ['alleato', 'adjective', 'b'], ['alleato', 'noun', 'b'], ['allegato', 'past_part', 'b'], ['allegato', 'adjective', 'b'], ['allegato', 'noun', 'b'], ['alleggerire', 'verb', 'c'], ['allegria', 'noun', 'b'], ['allegro', 'adjective', 'b'], ['allegro', 'adverb', 'b'], ['allegro', 'noun', 'b'], ['allenamento', 'noun', 'b'], ['allenare', 'verb', 'b'], ['allenatore', 'adjective', 'b'], ['allenatore', 'noun', 'b'], ['allentare', 'verb', 'c'], ['allergia', 'noun', 'c'], ['allevare', 'verb', 'b'], ['allievo', 'noun', 'b'], ['allineare', 'verb', 'c'], ['alloggio', 'noun', 'b'], ['allontanare', 'verb', 'a'], ['allora', 'adverb', 'a'], ['allora', 'conjunction', 'a'], ['alluce', 'noun', 'c'], ['alludere', 'verb', 'b'], ['alluminio', 'noun', 'c'], ['allungare', 'verb', 'a'], ['alluvione', 'noun', 'c'], ['almeno', 'adverb', 'a'], ['alquanto', 'adjective', 'b'], ['alquanto', 'pronoun', 'b'], ['alquanto', 'adverb', 'b'], ['altalena', 'noun', 'c'], ['altamente', 'adverb', 'b'], ['altare', 'noun', 'b'], ['alterare', 'verb', 'b'], ['alternare', 'verb', 'b'], ['alternativa', 'noun', 'b'], ['alternativo', 'adjective', 'b'], ['alterno', 'adjective', 'c'], ['altezza', 'noun', 'a'], ['alto', 'adjective', 'a'], ['alto', 'noun', 'a'], ['alto', 'adverb', 'a'], ['altoatesino', 'adjective', 'c'], ['altoatesino', 'noun', 'c'], ['altopiano', 'noun', 'c'], ['altrettanto', 'adjective', 'a'], ['altrettanto', 'pronoun', 'a'], ['altrettanto', 'adverb', 'a'], ['altrimenti', 'adverb', 'a'], ['altro', 'adjective', 'a'], ['altro', 'pronoun', 'a'], ['altro', 'adverb', 'a'], ['altrove', 'adverb', 'b'], ['altrui', 'adjective', 'b'], ['altrui', 'pronoun', 'b'], ['alunno', 'noun', 'b'], ['alveare', 'noun', 'c'], ['alzare', 'verb', 'a'], ['amante', 'pres_part', 'a'], ['amante', 'adjective', 'a'], ['amante', 'noun', 'a'], ['amare', 'verb', 'a'], ['amaro', 'adjective', 'b'], ['amaro', 'noun', 'b'], ['amato', 'past_part', 'b'], ['amato', 'adjective', 'b'], ['amato', 'noun', 'b'], ['ambasciata', 'noun', 'c'], ['ambientale', 'adjective', 'a'], ['ambientare', 'verb', 'b'], ['ambiente', 'noun', 'a'], ['ambiente', 'adjective', 'a'], ['ambito', 'noun', 'a'], ['ambizione', 'noun', 'b'], ['ambulanza', 'noun', 'b'], ['americano', 'adjective', 'a'], ['americano', 'noun', 'a'], ['amicizia', 'noun', 'a'], ['amico', 'adjective', 'a'], ['amico', 'noun', 'a'], ['ammaccare', 'verb', 'c'], ['ammalarsi', 'verb', 'b'], ['ammalato', 'past_part', 'c'], ['ammalato', 'adjective', 'c'], ['ammalato', 'noun', 'c'], ['ammanettare', 'verb', 'c'], ['ammassare', 'verb', 'c'], ['ammasso', 'noun', 'c'], ['ammazzare', 'verb', 'a'], ['ammettere', 'verb', 'a'], ['amministrativo', 'adjective', 'b'], ['amministrativo', 'noun', 'b'], ['amministratore', 'noun', 'b'], ['amministrazione', 'noun', 'a'], ['ammirare', 'verb', 'b'], ['ammissione', 'noun', 'b'], ['ammobiliare', 'verb', 'c'], ['ammoniaca', 'noun', 'c'], ['ammorbidente', 'pres_part', 'c'], ['ammorbidente', 'adjective', 'c'], ['ammorbidente', 'noun', 'c'], ['ammucchiare', 'verb', 'c'], ['ammuffire', 'verb', 'c'], ['amore', 'noun', 'a'], ['amoroso', 'adjective', 'b'], ['amoroso', 'noun', 'b'], ['ampiamente', 'adverb', 'b'], ['ampio', 'adjective', 'a'], ['ampio', 'noun', 'a'], ['amplificatore', 'adjective', 'c'], ['amplificatore', 'noun', 'c'], ['analcolico', 'adjective', 'c'], ['analcolico', 'noun', 'c'], ['analfabeta', 'adjective', 'c'], ['analfabeta', 'noun', 'c'], ['analisi', 'noun', 'a'], ['analitico', 'adjective', 'b'], ['analizzare', 'verb', 'a'], ['analogo', 'adjective', 'b'], ['ananas', 'noun', 'c'], ['anarchico', 'adjective', 'c'], ['anarchico', 'noun', 'c'], ['anatra', 'noun', 'c'], ['anche', 'conjunction', 'a'], ['anche', 'adverb', 'a'], ['anconetano', 'adjective', 'c'], ['anconetano', 'noun', 'c'], ['ancora', 'adverb', 'a'], ['ancora', 'conjunction', 'a'], ['ancorare', 'verb', 'b'], ['andamento', 'noun', 'b'], ['andare', 'verb', 'a'], ['andata', 'noun', 'c'], ['anello', 'noun', 'a'], ['angelo', 'noun', 'a'], ['angolare', 'adjective', 'b'], ['angolare', 'noun', 'b'], ['angolo', 'noun', 'a'], ['angoscia', 'noun', 'b'], ['anima', 'noun', 'a'], ['animale', 'noun', 'a'], ['animale', 'adjective', 'b'], ['animare', 'verb', 'a'], ['animato', 'past_part', 'b'], ['animato', 'adjective', 'b'], ['animato', 'adverb', 'b'], ['animo', 'noun', 'b'], ['animo', 'exclamation', 'b'], ['annacquare', 'verb', 'c'], ['annaffiare', 'verb', 'c'], ['annebbiare', 'verb', 'c'], ['anniversario', 'noun', 'b'], ['anniversario', 'adjective', 'b'], ['anno', 'noun', 'a'], ['annodare', 'verb', 'c'], ['annoiare', 'verb', 'b'], ['annotare', 'verb', 'b'], ['annuale', 'adjective', 'b'], ['annuale', 'noun', 'b'], ['annuire', 'verb', 'b'], ['annullare', 'verb', 'b'], ['annunciare', 'verb', 'a'], ['annuncio', 'noun', 'b'], ['annusare', 'verb', 'c'], ['anonimo', 'adjective', 'b'], ['anonimo', 'noun', 'b'], ['ansia', 'noun', 'a'], ['ansioso', 'adjective', 'b'], ['ansioso', 'noun', 'b'], ['antartico', 'adjective', 'c'], ['antartico', 'noun', 'c'], ['antenna', 'noun', 'b'], ['anteprima', 'noun', 'b'], ['anteriore', 'adjective', 'b'], ['anticalcare', 'adjective', 'c'], ['antichità', 'noun', 'c'], ['anticipare', 'verb', 'b'], ['anticipo', 'noun', 'b'], ['antico', 'adjective', 'a'], ['antico', 'noun', 'a'], ['antipasto', 'noun', 'c'], ['antirughe', 'adjective', 'c'], ['antirughe', 'noun', 'c'], ['antropologia', 'noun', 'b'], ['anulare', 'adjective', 'c'], ['anulare', 'noun', 'c'], ['anzi', 'adverb', 'a'], ['anzi', 'preposition', 'a'], ['anziano', 'adjective', 'a'], ['anziano', 'noun', 'a'], ['anziché', 'conjunction', 'b'], ['aostano', 'adjective', 'c'], ['aostano', 'noun', 'c'], ['ape', 'noun', 'b'], ['aperitivo', 'noun', 'c'], ['aperitivo', 'adjective', 'c'], ['aperto', 'past_part', 'a'], ['aperto', 'adjective', 'a'], ['aperto', 'noun', 'a'], ['aperto', 'adverb', 'a'], ['apertura', 'noun', 'a'], ['aspettativa', 'noun', 'b'], ['apostolo', 'noun', 'c'], ['appalto', 'noun', 'b'], ['appannare', 'verb', 'c'], ['apparato', 'noun', 'b'], ['apparecchiare', 'verb', 'c'], ['apparecchiatura', 'noun', 'c'], ['apparecchio', 'noun', 'b'], ['apparente', 'pres_part', 'b'], ['apparente', 'adjective', 'b'], ['apparentemente', 'adverb', 'b'], ['apparenza', 'noun', 'b'], ['apparire', 'verb', 'a'], ['apparizione', 'noun', 'b'], ['appartamento', 'noun', 'a'], ['appartenenza', 'noun', 'b'], ['appartenere', 'verb', 'a'], ['appassionare', 'verb', 'b'], ['appassionarsi', 'verb', 'c'], ['appassionato', 'past_part', 'b'], ['appassionato', 'adjective', 'b'], ['appassionato', 'noun', 'b'], ['appello', 'noun', 'b'], ['appena', 'adverb', 'a'], ['appena', 'conjunction', 'a'], ['appendere', 'verb', 'b'], ['appendicite', 'noun', 'c'], ['appenninico', 'adjective', 'c'], ['appeso', 'past_part', 'c'], ['appeso', 'adjective', 'c'], ['appeso', 'noun', 'c'], ['appiccicare', 'verb', 'c'], ['appiglio', 'noun', 'c'], ['applauso', 'noun', 'b'], ['applicare', 'verb', 'a'], ['applicazione', 'noun', 'b'], ['appoggiare', 'verb', 'a'], ['appoggio', 'noun', 'b'], ['apposito', 'adjective', 'b'], ['apposta', 'adverb', 'b'], ['apposta', 'adjective', 'b'], ['apprendere', 'verb', 'b'], ['apprendimento', 'noun', 'b'], ['apprendista', 'noun', 'c'], ['apprezzare', 'verb', 'a'], ['approccio', 'noun', 'b'], ['approfittare', 'verb', 'b'], ['approfondimento', 'noun', 'b'], ['approfondire', 'verb', 'b'], ['approvare', 'verb', 'b'], ['approvazione', 'noun', 'b'], ['appuntamento', 'noun', 'a'], ['appuntire', 'verb', 'c'], ['appunto', 'noun', 'b'], ['appunto', 'adverb', 'a'], ['aprile', 'noun', 'a'], ['aprire', 'verb', 'a'], ['apriscatole', 'noun', 'c'], ['aquila', 'noun', 'c'], ['aquilano', 'adjective', 'c'], ['aquilano', 'noun', 'c'], ['aquilone', 'noun', 'c'], ['arabo', 'adjective', 'a'], ['arabo', 'noun', 'a'], ['arachide', 'noun', 'c'], ['aragosta', 'noun', 'c'], ['aranciata', 'noun', 'c'], ['arancio', 'noun', 'c'], ['arare', 'verb', 'c'], ['aratro', 'noun', 'c'], ['arbitro', 'noun', 'b'], ['archeologo', 'noun', 'c'], ['architettare', 'verb', 'b'], ['architetto', 'noun', 'b'], ['architettonico', 'adjective', 'b'], ['architettura', 'noun', 'b'], ['archiviare', 'verb', 'b'], ['archivio', 'noun', 'b'], ['arco', 'noun', 'a'], ['arcobaleno', 'noun', 'c'], ['area', 'noun', 'a'], ['argentino', 'adjective', 'b'], ['argentino', 'noun', 'b'], ['argento', 'noun', 'b'], ['argomentare', 'verb', 'b'], ['argomentazione', 'noun', 'b'], ['argomento', 'noun', 'a'], ['aria', 'noun', 'a'], ['aristocratico', 'adjective', 'c'], ['aristocratico', 'noun', 'c'], ['aritmetica', 'noun', 'c'], ['aritmetico', 'adjective', 'c'], ['aritmetico', 'noun', 'c'], ['arma', 'noun', 'a'], ['armadio', 'noun', 'b'], ['armamento', 'noun', 'c'], ['armare', 'verb', 'b'], ['armato', 'past_part', 'b'], ['armato', 'adjective', 'b'], ['armato', 'noun', 'b'], ['armonia', 'noun', 'b'], ['aroma', 'noun', 'c'], ['arrabbiarsi', 'verb', 'a'], ['arrampicarsi', 'verb', 'b'], ['arredamento', 'noun', 'b'], ['arredare', 'verb', 'c'], ['arrendersi', 'verb', 'b'], ['arrendersi', 'verb', 'c'], ['arrestare', 'verb', 'a'], ['arresto', 'noun', 'b'], ['arricchire', 'verb', 'b'], ['arrivare', 'verb', 'a'], ['arrivederci', 'exclamation', 'b'], ['arrivederci', 'noun', 'b'], ['arrivo', 'noun', 'a'], ['arrosto', 'noun', 'c'], ['arrosto', 'adjective', 'c'], ['arrosto', 'adverb', 'c'], ['arrugginire', 'verb', 'c'], ['arte', 'noun', 'a'], ['arteria', 'noun', 'b'], ['artico', 'adjective', 'c'], ['artico', 'noun', 'c'], ['articolare', 'verb', 'b'], ['articolare', 'noun', 'b'], ['articolazione', 'noun', 'b'], ['articolo', 'noun', 'a'], ['artificiale', 'adjective', 'b'], ['artigianale', 'adjective', 'c'], ['artigiano', 'noun', 'b'], ['artigiano', 'adjective', 'b'], ['artiglieria', 'noun', 'c'], ['artiglio', 'noun', 'c'], ['artista', 'noun', 'a'], ['artistico', 'adjective', 'a'], ['artistico', 'noun', 'a'], ['ascella', 'noun', 'c'], ['ascensore', 'noun', 'b'], ['ascesa', 'noun', 'b'], ['ascesso', 'noun', 'c'], ['ascia', 'noun', 'c'], ['asciugamano', 'noun', 'b'], ['asciugare', 'verb', 'b'], ['asciutto', 'adjective', 'b'], ['asciutto', 'noun', 'b'], ['ascoltare', 'verb', 'a'], ['ascolto', 'noun', 'b'], ['asfaltare', 'verb', 'c'], ['asfalto', 'noun', 'c'], ['asiatico', 'adjective', 'b'], ['asiatico', 'noun', 'b'], ['asilo', 'noun', 'b'], ['asino', 'noun', 'b'], ['asma', 'noun', 'c'], ['asparago', 'noun', 'c'], ['aspettare', 'verb', 'a'], ['aspetto', 'noun', 'a'], ['aspirapolvere', 'noun', 'c'], ['aspirare', 'verb', 'b'], ['aspirazione', 'noun', 'b'], ['aspro', 'adjective', 'b'], ['aspro', 'noun', 'b'], ['assaggiare', 'verb', 'b'], ['assaggio', 'noun', 'c'], ['assai', 'adverb', 'a'], ['assai', 'adjective', 'a'], ['assai', 'noun', 'a'], ['assalire', 'verb', 'c'], ['assaltare', 'verb', 'c'], ['assalto', 'noun', 'b'], ['assaporare', 'verb', 'c'], ['assassinare', 'verb', 'b'], ['assassinio', 'noun', 'c'], ['assassino', 'noun', 'b'], ['assassino', 'adjective', 'b'], ['asse', 'noun', 'b'], ['assediare', 'verb', 'c'], ['assegnare', 'verb', 'b'], ['assegno', 'noun', 'b'], ['assemblea', 'noun', 'b'], ['assente', 'adjective', 'b'], ['assente', 'noun', 'b'], ['assenza', 'noun', 'a'], ['assicurare', 'verb', 'a'], ['assicurazione', 'noun', 'b'], ['assieme', 'adverb', 'a'], ['assieme', 'noun', 'a'], ['assistente', 'pres_part', 'b'], ['assistente', 'adjective', 'b'], ['assistente', 'noun', 'b'], ['assistenza', 'noun', 'b'], ['assistere', 'verb', 'a'], ['associare', 'verb', 'b'], ['associazione', 'noun', 'a'], ['assolutamente', 'adverb', 'a'], ['assoluto', 'adjective', 'a'], ['assoluto', 'noun', 'a'], ['assoluzione', 'noun', 'c'], ['assolvere', 'verb', 'b'], ['assomigliare', 'verb', 'b'], ['assorbente', 'pres_part', 'c'], ['assorbente', 'adjective', 'c'], ['assorbente', 'noun', 'c'], ['assorbire', 'verb', 'b'], ['assordare', 'verb', 'c'], ['assumere', 'verb', 'a'], ['assunzione', 'noun', 'b'], ['assurdo', 'adjective', 'a'], ['assurdo', 'noun', 'a'], ['asta', 'noun', 'b'], ['astemio', 'adjective', 'c'], ['astemio', 'noun', 'c'], ['astratto', 'past_part', 'b'], ['astratto', 'adjective', 'b'], ['astratto', 'noun', 'b'], ['astronave', 'noun', 'c'], ['astuccio', 'noun', 'c'], ['astuto', 'adjective', 'c'], ['astuto', 'noun', 'c'], ['astuzia', 'noun', 'c'], ['ateniese', 'adjective', 'c'], ['ateniese', 'noun', 'c'], ['ateo', 'adjective', 'b'], ['ateo', 'noun', 'b'], ['atlantico', 'adjective', 'c'], ['atleta', 'noun', 'b'], ['atmosfera', 'noun', 'a'], ['atomica', 'noun', 'c'], ['atomico', 'adjective', 'b'], ['atomo', 'noun', 'b'], ['atrio', 'noun', 'c'], ['atroce', 'adjective', 'b'], ['attaccante', 'pres_part', 'c'], ['attaccante', 'adjective', 'c'], ['attaccante', 'noun', 'c'], ['attaccapanni', 'noun', 'c'], ['attaccare', 'verb', 'a'], ['attacco', 'noun', 'a'], ['atteggiamento', 'noun', 'a'], ['atteggiare', 'verb', 'c'], ['attendere', 'verb', 'a'], ['attenere', 'verb', 'b'], ['attentamente', 'adverb', 'b'], ['attentare', 'verb', 'c'], ['attentato', 'noun', 'b'], ['attento', 'adjective', 'a'], ['attenzione', 'noun', 'a'], ['atterraggio', 'noun', 'c'], ['atterrare', 'verb', 'b'], ['attesa', 'noun', 'a'], ['attestare', 'verb', 'b'], ['attimo', 'noun', 'a'], ['attingere', 'verb', 'b'], ['attirare', 'verb', 'b'], ['attivare', 'verb', 'b'], ['attività', 'noun', 'a'], ['attivo', 'adjective', 'a'], ['attivo', 'noun', 'a'], ['atto', 'noun', 'a'], ['attore', 'noun', 'a'], ['attorno', 'adverb', 'a'], ['attrarre', 'verb', 'b'], ['attraversare', 'verb', 'a'], ['attraverso', 'preposition', 'a'], ['attraverso', 'adverb', 'a'], ['attrazione', 'noun', 'b'], ['attrezzare', 'verb', 'b'], ['attrezzatura', 'noun', 'b'], ['attrezzo', 'noun', 'b'], ['attribuire', 'verb', 'b'], ['attrice', 'noun', 'b'], ['attuale', 'adjective', 'a'], ['attualità', 'noun', 'b'], ['attualmente', 'adverb', 'b'], ['attuare', 'verb', 'b'], ['augurare', 'verb', 'b'], ['augurio', 'noun', 'b'], ['aula', 'noun', 'b'], ['aumentare', 'verb', 'a'], ['aumento', 'noun', 'a'], ['australiano', 'adjective', 'c'], ['australiano', 'noun', 'c'], ['austriaco', 'adjective', 'b'], ['austriaco', 'noun', 'b'], ['autentico', 'adjective', 'b'], ['autentico', 'noun', 'b'], ['autista', 'noun', 'b'], ['auto', 'noun', 'a'], ['autoambulanza', 'noun', 'c'], ['autobotte', 'noun', 'c'], ['autobus', 'noun', 'b'], ['autografo', 'adjective', 'c'], ['autografo', 'noun', 'c'], ['automaticamente', 'adverb', 'b'], ['automatico', 'adjective', 'b'], ['automatico', 'noun', 'b'], ['automobile', 'noun', 'b'], ['automobilista', 'noun', 'c'], ['autonomia', 'noun', 'b'], ['autonomo', 'adjective', 'b'], ['autonomo', 'noun', 'b'], ['autore', 'noun', 'a'], ['autorevole', 'adjective', 'c'], ['autorità', 'noun', 'a'], ['autorizzare', 'verb', 'a'], ['autoscontro', 'noun', 'c'], ['autoscuola', 'noun', 'c'], ['autostop', 'noun', 'c'], ['autostrada', 'noun', 'b'], ['autotreno', 'noun', 'c'], ['autunno', 'noun', 'b'], ['avambraccio', 'noun', 'c'], ['avanguardia', 'noun', 'b'], ['avanti', 'adverb', 'a'], ['avanti', 'adjective', 'a'], ['avanti', 'loc-comando', 'a'], ['avanti', 'preposition', 'a'], ['avanti', 'noun', 'a'], ['avanzare', 'verb', 'a'], ['avanzato', 'past_part', 'b'], ['avanzato', 'adjective', 'b'], ['avanzo', 'noun', 'c'], ['avarizia', 'noun', 'c'], ['avaro', 'adjective', 'c'], ['avaro', 'noun', 'c'], ['avena', 'noun', 'c'], ['avere', 'verb', 'a'], ['aviazione', 'noun', 'c'], ['avvantaggiare', 'verb', 'c'], ['avvelenare', 'verb', 'b'], ['avvelenato', 'past_part', 'c'], ['avvelenato', 'adjective', 'c'], ['avvenimento', 'noun', 'b'], ['avvenire', 'adjective', 'a'], ['avvenire', 'noun', 'a'], ['avventura', 'noun', 'a'], ['avverare', 'verb', 'c'], ['avversario', 'noun', 'b'], ['avvertire', 'verb', 'a'], ['avviamento', 'noun', 'c'], ['avviare', 'verb', 'a'], ['avvicinare', 'verb', 'a'], ['avvio', 'noun', 'b'], ['avvisare', 'verb', 'b'], ['avviso', 'noun', 'b'], ['avvitare', 'verb', 'c'], ['avvocato', 'noun', 'a'], ['avvolgere', 'verb', 'b'], ['azienda', 'noun', 'a'], ['aziendale', 'adjective', 'b'], ['azione', 'noun', 'a'], ['azione', 'noun', 'b'], ['azzardare', 'verb', 'b'], ['azzardo', 'noun', 'c'], ['azzurro', 'noun', 'a'], ['azzurro', 'adjective', 'a'], ['babbo', 'noun', 'b'], ['baby', 'noun', 'b'], ['baby', 'adjective', 'b'], ['babydoll', 'noun', 'c'], ['bacca', 'noun', 'c'], ['baccalà', 'noun', 'c'], ['bacheca', 'noun', 'b'], ['baciare', 'verb', 'a'], ['bacinella', 'noun', 'c'], ['bacino', 'noun', 'b'], ['bacio', 'noun', 'a'], ['baco', 'noun', 'c'], ['badare', 'verb', 'b'], ['baffo', 'noun', 'b'], ['bagagliaio', 'noun', 'c'], ['bagaglio', 'noun', 'b'], ['bagnare', 'verb', 'b'], ['bagnato', 'past_part', 'b'], ['bagnato', 'adjective', 'b'], ['bagnato', 'noun', 'b'], ['bagno', 'noun', 'a'], ['bagnoschiuma', 'noun', 'c'], ['balcone', 'noun', 'b'], ['balena', 'noun', 'b'], ['balia', 'noun', 'b'], ['ballare', 'verb', 'a'], ['ballerina', 'noun', 'c'], ['ballerino', 'noun', 'c'], ['ballerino', 'adjective', 'c'], ['balletto', 'noun', 'c'], ['ballo', 'noun', 'b'], ['balsamo', 'noun', 'c'], ['bambina', 'noun', 'a'], ['bambinaia', 'noun', 'c'], ['bambino', 'noun', 'a'], ['bambino', 'adjective', 'a'], ['bambola', 'noun', 'b'], ['banale', 'adjective', 'b'], ['banana', 'noun', 'c'], ['banca', 'noun', 'a'], ['bancarella', 'noun', 'c'], ['bancario', 'adjective', 'b'], ['bancario', 'noun', 'b'], ['banco', 'noun', 'b'], ['bancone', 'noun', 'b'], ['band', 'noun', 'b'], ['banda', 'noun', 'b'], ['bandiera', 'noun', 'b'], ['bando', 'noun', 'b'], ['bar', 'noun', 'a'], ['bara', 'noun', 'b'], ['baracca', 'noun', 'c'], ['barba', 'noun', 'b'], ['barbabietola', 'noun', 'c'], ['barbaro', 'adjective', 'b'], ['barbaro', 'noun', 'b'], ['barca', 'noun', 'a'], ['barella', 'noun', 'c'], ['barese', 'adjective', 'c'], ['barese', 'noun', 'c'], ['barile', 'noun', 'c'], ['barista', 'noun', 'c'], ['barriera', 'noun', 'b'], ['basare', 'verb', 'a'], ['base', 'noun', 'a'], ['basetta', 'noun', 'c'], ['basilica', 'noun', 'b'], ['basilico', 'noun', 'c'], ['basket', 'noun', 'c'], ['basso', 'adjective', 'a'], ['basso', 'noun', 'a'], ['basso', 'adverb', 'a'], ['bastardo', 'adjective', 'b'], ['bastardo', 'noun', 'b'], ['bastare', 'verb', 'a'], ['bastonare', 'verb', 'c'], ['bastone', 'noun', 'b'], ['battaglia', 'noun', 'a'], ['battello', 'noun', 'c'], ['battere', 'verb', 'a'], ['battere', 'noun', 'a'], ['batteria', 'noun', 'b'], ['batterio', 'noun', 'b'], ['batticuore', 'noun', 'c'], ['battipanni', 'noun', 'c'], ['battito', 'noun', 'c'], ['battuta', 'noun', 'a'], ['batuffolo', 'noun', 'c'], ['baule', 'noun', 'c'], ['bava', 'noun', 'c'], ['bavaglio', 'noun', 'c'], ['beato', 'past_part', 'b'], ['beato', 'adjective', 'b'], ['beato', 'noun', 'b'], ['beccare', 'verb', 'b'], ['befana', 'noun', 'c'], ['beffa', 'noun', 'c'], ['beh', 'exclamation', 'a'], ['belare', 'verb', 'c'], ['belga', 'adjective', 'c'], ['belga', 'noun', 'c'], ['bella', 'noun', 'b'], ['bellezza', 'noun', 'a'], ['bello', 'adjective', 'a'], ['bello', 'noun', 'a'], ['benché', 'conjunction', 'b'], ['benda', 'noun', 'c'], ['bene', 'adverb', 'a'], ['bene', 'exclamation', 'a'], ['bene', 'noun', 'a'], ['benedetto', 'past_part', 'b'], ['benedetto', 'adjective', 'b'], ['benedetto', 'noun', 'b'], ['beneficenza', 'noun', 'c'], ['beneficio', 'noun', 'b'], ['benessere', 'noun', 'b'], ['benestante', 'adjective', 'c'], ['benestante', 'noun', 'c'], ['bensì', 'conjunction', 'b'], ['bensì', 'adverb', 'b'], ['benvenuto', 'adjective', 'b'], ['benvenuto', 'noun', 'b'], ['benzina', 'noun', 'b'], ['benzinaio', 'noun', 'c'], ['bere', 'verb', 'a'], ['bere', 'noun', 'a'], ['berlinese', 'adjective', 'c'], ['berlinese', 'noun', 'c'], ['berretto', 'noun', 'c'], ['bersaglio', 'noun', 'b'], ['besciamella', 'noun', 'c'], ['bestemmia', 'noun', 'c'], ['bestia', 'noun', 'b'], ['bestiale', 'adjective', 'c'], ['bevanda', 'noun', 'b'], ['bevitore', 'noun', 'c'], ['bevuta', 'noun', 'c'], ['bi', 'noun', 'c'], ['bianco', 'adjective', 'a'], ['bianco', 'noun', 'a'], ['bibbia', 'noun', 'b'], ['bibita', 'noun', 'c'], ['biblico', 'adjective', 'b'], ['biblico', 'noun', 'b'], ['bibliografia', 'noun', 'b'], ['biblioteca', 'noun', 'b'], ['bicchiere', 'noun', 'a'], ['bici', 'noun', 'b'], ['bicicletta', 'noun', 'b'], ['bidè', 'noun', 'c'], ['bidello', 'noun', 'c'], ['biglia', 'noun', 'c'], ['biglietteria', 'noun', 'c'], ['biglietto', 'noun', 'a'], ['bikini', 'noun', 'c'], ['bilancia', 'noun', 'b'], ['bilancio', 'noun', 'b'], ['biliardo', 'noun', 'c'], ['bimba', 'noun', 'b'], ['bimbo', 'noun', 'b'], ['binario', 'noun', 'c'], ['biografia', 'noun', 'b'], ['biologia', 'noun', 'b'], ['biologico', 'adjective', 'b'], ['biologico', 'noun', 'b'], ['bionda', 'noun', 'b'], ['biondo', 'adjective', 'b'], ['biondo', 'noun', 'b'], ['birichino', 'noun', 'c'], ['birichino', 'adjective', 'c'], ['birillo', 'noun', 'c'], ['birra', 'noun', 'b'], ['bisbigliare', 'verb', 'c'], ['biscia', 'noun', 'c'], ['biscotto', 'adjective', 'b'], ['biscotto', 'noun', 'b'], ['bisnonno', 'noun', 'c'], ['bisognare', 'verb', 'a'], ['bisogno', 'noun', 'a'], ['bistecca', 'noun', 'c'], ['bistecchiera', 'noun', 'c'], ['bisticciare', 'verb', 'c'], ['bit', 'noun', 'b'], ['bizzarro', 'adjective', 'b'], ['bloccare', 'verb', 'a'], ['blocco', 'noun', 'b'], ['blocco', 'noun', 'b'], ['blog', 'noun', 'a'], ['blu', 'adjective', 'a'], ['blu', 'noun', 'a'], ['bocca', 'noun', 'a'], ['bocchino', 'noun', 'c'], ['boccia', 'noun', 'c'], ['bocciare', 'verb', 'b'], ['bocciatura', 'noun', 'c'], ['bocciolo', 'noun', 'c'], ['boccone', 'noun', 'c'], ['boh', 'exclamation', 'b'], ['boia', 'noun', 'c'], ['boia', 'adjective', 'c'], ['bolla', 'noun', 'b'], ['bolletta', 'noun', 'b'], ['bollito', 'past_part', 'c'], ['bollito', 'adjective', 'c'], ['bollito', 'noun', 'c'], ['bollitore', 'noun', 'c'], ['bollo', 'noun', 'c'], ['bolognese', 'adjective', 'c'], ['bolognese', 'noun', 'c'], ['bolzanino', 'adjective', 'c'], ['bolzanino', 'noun', 'c'], ['bomba', 'noun', 'b'], ['bombardare', 'verb', 'b'], ['bombola', 'noun', 'c'], ['bomboniera', 'noun', 'c'], ['bontà', 'noun', 'b'], ['bordo', 'noun', 'a'], ['borgata', 'noun', 'c'], ['borghese', 'adjective', 'b'], ['borghese', 'noun', 'b'], ['borghesia', 'noun', 'c'], ['borgo', 'noun', 'b'], ['borotalco', 'noun', 'c'], ['borsa', 'noun', 'a'], ['borsa', 'noun', 'b'], ['borsetta', 'noun', 'c'], ['bosco', 'noun', 'a'], ['bosniaco', 'adjective', 'c'], ['bosniaco', 'noun', 'c'], ['boss', 'noun', 'b'], ['bossolo', 'noun', 'c'], ['botanica', 'noun', 'c'], ['botta', 'noun', 'b'], ['botte', 'noun', 'c'], ['bottega', 'noun', 'b'], ['bottegaio', 'noun', 'c'], ['bottegaio', 'adjective', 'c'], ['bottiglia', 'noun', 'a'], ['botto', 'noun', 'c'], ['bottone', 'noun', 'b'], ['bovino', 'adjective', 'c'], ['bovino', 'noun', 'c'], ['box', 'noun', 'b'], ['boxer', 'noun', 'c'], ['braccialetto', 'noun', 'c'], ['bracciante', 'noun', 'c'], ['braccio', 'noun', 'a'], ['branco', 'noun', 'b'], ['brand', 'noun', 'b'], ['brandello', 'noun', 'c'], ['brano', 'noun', 'a'], ['brasiliano', 'adjective', 'b'], ['brasiliano', 'noun', 'b'], ['bravo', 'adjective', 'a'], ['bravo', 'noun', 'a'], ['bravo', 'exclamation', 'a'], ['bresaola', 'noun', 'c'], ['bretella', 'noun', 'c'], ['breve', 'adjective', 'a'], ['breve', 'adverb', 'a'], ['breve', 'noun', 'a'], ['briciola', 'noun', 'c'], ['brigantaggio', 'noun', 'c'], ['brigante', 'noun', 'c'], ['brillante', 'pres_part', 'b'], ['brillante', 'adjective', 'b'], ['brillante', 'noun', 'b'], ['brillantina', 'noun', 'c'], ['brillare', 'verb', 'b'], ['brina', 'noun', 'c'], ['brioche', 'noun', 'c'], ['britannico', 'adjective', 'b'], ['britannico', 'noun', 'b'], ['brivido', 'noun', 'b'], ['brocca', 'noun', 'c'], ['brogliaccio', 'noun', 'b'], ['bronchite', 'noun', 'c'], ['brontolare', 'verb', 'c'], ['bronzo', 'noun', 'b'], ['bruciare', 'verb', 'a'], ['bruciato', 'past_part', 'b'], ['bruciato', 'adjective', 'b'], ['bruciato', 'noun', 'b'], ['bruciatura', 'noun', 'c'], ['bruco', 'noun', 'c'], ['bruco', 'adjective', 'c'], ['bruschetta', 'noun', 'c'], ['brutale', 'adjective', 'c'], ['brutto', 'adjective', 'a'], ['brutto', 'noun', 'a'], ['brutto', 'adverb', 'a'], ['buca', 'noun', 'b'], ['bucare', 'verb', 'b'], ['bucato', 'noun', 'c'], ['buccia', 'noun', 'c'], ['buco', 'noun', 'a'], ['budino', 'noun', 'c'], ['bufala', 'noun', 'c'], ['bufalo', 'noun', 'c'], ['bufera', 'noun', 'c'], ['buffet', 'noun', 'c'], ['buffo', 'adjective', 'b'], ['buffo', 'noun', 'b'], ['bugia', 'noun', 'b'], ['bugiardo', 'adjective', 'b'], ['bugiardo', 'noun', 'b'], ['buio', 'adjective', 'a'], ['buio', 'noun', 'a'], ['bulgaro', 'adjective', 'c'], ['bulgaro', 'noun', 'c'], ['buonafede', 'noun', 'c'], ['buonasera', 'exclamation', 'b'], ['buongiorno', 'exclamation', 'a'], ['buongusto', 'noun', 'c'], ['buono', 'adjective', 'a'], ['buono', 'noun', 'a'], ['buono', 'adverb', 'a'], ['buonuomo', 'noun', 'c'], ['burattino', 'noun', 'c'], ['burocrazia', 'noun', 'c'], ['burrasca', 'noun', 'c'], ['burro', 'noun', 'b'], ['burrone', 'noun', 'c'], ['business', 'noun', 'b'], ['business', 'adjective', 'b'], ['bussare', 'verb', 'b'], ['bussola', 'noun', 'c'], ['busta', 'noun', 'b'], ['bustina', 'noun', 'c'], ['busto', 'noun', 'c'], ['buttare', 'verb', 'a'], ['cabina', 'noun', 'b'], ['cacao', 'noun', 'c'], ['cacca', 'noun', 'b'], ['caccia', 'noun', 'a'], ['cacciare', 'verb', 'a'], ['cacciatore', 'noun', 'b'], ['cacciavite', 'noun', 'c'], ['cadavere', 'noun', 'a'], ['cadere', 'verb', 'a'], ['cadere', 'noun', 'a'], ['caduta', 'noun', 'b'], ['caffè', 'noun', 'a'], ['caffè', 'adjective', 'a'], ['caffellatte', 'noun', 'c'], ['caffellatte', 'adjective', 'c'], ['caffettiera', 'noun', 'c'], ['cagare', 'verb', 'b'], ['cagliaritano', 'adjective', 'c'], ['cagliaritano', 'noun', 'c'], ['calabrese', 'adjective', 'c'], ['calabrese', 'noun', 'c'], ['calabrone', 'noun', 'c'], ['calamaro', 'noun', 'c'], ['calamita', 'noun', 'c'], ['calare', 'verb', 'b'], ['calcagno', 'noun', 'c'], ['calciare', 'verb', 'c'], ['calciatore', 'noun', 'b'], ['calcinaccio', 'noun', 'c'], ['calcio', 'noun', 'a'], ['calcolare', 'verb', 'b'], ['calcolatore', 'adjective', 'c'], ['calcolatore', 'noun', 'c'], ['calcolatrice', 'noun', 'c'], ['calcolo', 'noun', 'b'], ['caldo', 'adjective', 'a'], ['caldo', 'noun', 'a'], ['caldo', 'adverb', 'a'], ['calendario', 'noun', 'b'], ['calligrafia', 'noun', 'c'], ['callo', 'noun', 'c'], ['calma', 'noun', 'b'], ['calmare', 'verb', 'b'], ['calmo', 'adjective', 'b'], ['calo', 'noun', 'b'], ['calore', 'noun', 'a'], ['calpestare', 'verb', 'c'], ['calunnia', 'noun', 'c'], ['calvario', 'noun', 'c'], ['calza', 'noun', 'b'], ['calzare', 'verb', 'c'], ['calzatura', 'noun', 'c'], ['calzino', 'noun', 'c'], ['calzolaio', 'noun', 'c'], ['calzoleria', 'noun', 'c'], ['calzone', 'noun', 'c'], ['cambiamento', 'noun', 'a'], ['cambiare', 'verb', 'a'], ['cambio', 'noun', 'a'], ['camera', 'noun', 'a'], ['camerata', 'noun', 'c'], ['cameriere', 'noun', 'b'], ['camicetta', 'noun', 'c'], ['camicia', 'noun', 'b'], ['caminetto', 'noun', 'c'], ['camion', 'noun', 'a'], ['camionista', 'noun', 'c'], ['cammello', 'noun', 'c'], ['cammello', 'adjective', 'c'], ['camminare', 'verb', 'a'], ['camminata', 'noun', 'c'], ['cammino', 'noun', 'b'], ['camomilla', 'noun', 'c'], ['camorra', 'noun', 'b'], ['campagna', 'noun', 'a'], ['campana', 'noun', 'b'], ['campanella', 'noun', 'c'], ['campanello', 'noun', 'b'], ['campanile', 'noun', 'c'], ['campano', 'adjective', 'c'], ['campano', 'noun', 'c'], ['campare', 'verb', 'b'], ['campeggio', 'noun', 'c'], ['campionato', 'noun', 'b'], ['campione', 'noun', 'a'], ['campo', 'noun', 'a'], ['campobassano', 'adjective', 'c'], ['campobassano', 'noun', 'c'], ['camposanto', 'noun', 'c'], ['canadese', 'adjective', 'c'], ['canadese', 'noun', 'c'], ['canaglia', 'noun', 'c'], ['canale', 'noun', 'a'], ['canapa', 'noun', 'c'], ['canarino', 'noun', 'c'], ['canarino', 'adjective', 'c'], ['cancellare', 'verb', 'a'], ['cancellatura', 'noun', 'c'], ['cancello', 'noun', 'b'], ['cancro', 'noun', 'b'], ['candela', 'noun', 'b'], ['candeliere', 'noun', 'c'], ['candidare', 'verb', 'b'], ['candidato', 'past_part', 'a'], ['candidato', 'adjective', 'a'], ['candidato', 'noun', 'a'], ['candido', 'adjective', 'b'], ['cane', 'noun', 'a'], ['canestro', 'noun', 'c'], ['canguro', 'noun', 'c'], ['canna', 'noun', 'b'], ['cannibale', 'adjective', 'c'], ['cannibale', 'noun', 'c'], ['cannuccia', 'noun', 'c'], ['canone', 'noun', 'b'], ['canottiera', 'noun', 'c'], ['canotto', 'noun', 'c'], ['cantante', 'pres_part', 'b'], ['cantante', 'adjective', 'b'], ['cantante', 'noun', 'b'], ['cantare', 'verb', 'a'], ['cantautore', 'noun', 'c'], ['cantiere', 'noun', 'b'], ['cantilena', 'noun', 'c'], ['cantina', 'noun', 'b'], ['canto', 'noun', 'a'], ['canzone', 'noun', 'a'], ['caos', 'noun', 'b'], ['capace', 'adjective', 'a'], ['capacità', 'noun', 'a'], ['capanna', 'noun', 'b'], ['capannone', 'noun', 'b'], ['caparra', 'noun', 'c'], ['capello', 'noun', 'a'], ['capire', 'verb', 'a'], ['capitale', 'adjective', 'a'], ['capitale', 'noun', 'a'], ['capitano', 'noun', 'a'], ['capitare', 'verb', 'a'], ['capitolo', 'noun', 'a'], ['capo', 'noun', 'a'], ['capodanno', 'noun', 'c'], ['capogiro', 'noun', 'c'], ['capolavoro', 'noun', 'b'], ['capoluogo', 'noun', 'c'], ['caporale', 'noun', 'b'], ['caporale', 'adjective', 'b'], ['caposquadra', 'noun', 'c'], ['capotavola', 'noun', 'c'], ['capoufficio', 'noun', 'c'], ['cappa', 'noun', 'c'], ['cappella', 'noun', 'b'], ['cappelliera', 'noun', 'c'], ['cappello', 'noun', 'b'], ['cappero', 'noun', 'c'], ['cappotto', 'noun', 'c'], ['cappuccino', 'adjective', 'c'], ['cappuccino', 'noun', 'c'], ['cappuccino', 'adjective', 'c'], ['cappuccio', 'noun', 'c'], ['capra', 'noun', 'b'], ['capriccio', 'noun', 'b'], ['capriola', 'noun', 'c'], ['carabiniere', 'noun', 'a'], ['caramella', 'noun', 'b'], ['caramella', 'adjective', 'b'], ['carattere', 'noun', 'a'], ['caratteristica', 'noun', 'a'], ['caratteristico', 'adjective', 'b'], ['caratterizzare', 'verb', 'a'], ['carbone', 'noun', 'b'], ['carburante', 'pres_part', 'c'], ['carburante', 'adjective', 'c'], ['carburante', 'noun', 'c'], ['carcassa', 'noun', 'c'], ['carcerato', 'past_part', 'c'], ['carcerato', 'adjective', 'c'], ['carcerato', 'noun', 'c'], ['carcere', 'noun', 'a'], ['carciofino', 'noun', 'c'], ['carciofo', 'noun', 'c'], ['cardellino', 'noun', 'c'], ['cardiaco', 'adjective', 'b'], ['cardiaco', 'noun', 'b'], ['cardigan', 'noun', 'c'], ['cardinale', 'adjective', 'b'], ['cardinale', 'noun', 'b'], ['cardinale', 'adjective', 'b'], ['carenza', 'noun', 'b'], ['carica', 'noun', 'loc-comando'], ['caricare', 'verb', 'a'], ['carico', 'noun', 'a'], ['carico', 'adjective', 'b'], ['carino', 'adjective', 'a'], ['carità', 'noun', 'b'], ['carnagione', 'noun', 'c'], ['carne', 'noun', 'a'], ['carnevale', 'noun', 'c'], ['carnivoro', 'adjective', 'c'], ['carnivoro', 'noun', 'c'], ['carnoso', 'adjective', 'c'], ['carnoso', 'noun', 'c'], ['caro', 'adjective', 'a'], ['caro', 'adverb', 'a'], ['caro', 'noun', 'a'], ['carosello', 'noun', 'c'], ['carovana', 'noun', 'c'], ['carriera', 'noun', 'a'], ['carro', 'noun', 'b'], ['carrozzeria', 'noun', 'c'], ['carta', 'noun', 'a'], ['cartaceo', 'adjective', 'b'], ['cartella', 'noun', 'b'], ['cartello', 'noun', 'b'], ['cartoleria', 'noun', 'c'], ['cartolina', 'noun', 'b'], ['cartone', 'noun', 'b'], ['cartuccia', 'noun', 'c'], ['casa', 'noun', 'a'], ['casalinga', 'noun', 'c'], ['casalingo', 'adjective', 'c'], ['casalingo', 'noun', 'c'], ['cascare', 'verb', 'b'], ['cascata', 'noun', 'c'], ['casco', 'noun', 'c'], ['caserma', 'noun', 'b'], ['casetta', 'noun', 'b'], ['casino', 'noun', 'a'], ['caso', 'noun', 'a'], ['cassa', 'noun', 'a'], ['cassaforte', 'noun', 'c'], ['cassapanca', 'noun', 'c'], ['casseruola', 'noun', 'c'], ['cassetta', 'noun', 'b'], ['cassettiera', 'noun', 'c'], ['cassetto', 'noun', 'b'], ['cassiera', 'noun', 'c'], ['castagna', 'noun', 'c'], ['castagno', 'noun', 'c'], ['castano', 'adjective', 'c'], ['castello', 'noun', 'a'], ['castoro', 'noun', 'c'], ['casuale', 'adjective', 'b'], ['casuale', 'noun', 'b'], ['catalogo', 'noun', 'b'], ['catanzarese', 'adjective', 'c'], ['catanzarese', 'noun', 'c'], ['catarro', 'noun', 'c'], ['catasta', 'noun', 'c'], ['catastrofe', 'noun', 'b'], ['catechismo', 'noun', 'c'], ['categoria', 'noun', 'a'], ['catena', 'noun', 'a'], ['catenaccio', 'noun', 'c'], ['catino', 'noun', 'c'], ['catrame', 'noun', 'c'], ['cattedrale', 'adjective', 'b'], ['cattedrale', 'noun', 'b'], ['cattivo', 'adjective', 'a'], ['cattivo', 'noun', 'a'], ['cattolico', 'adjective', 'a'], ['cattolico', 'noun', 'a'], ['catturare', 'verb', 'b'], ['causa', 'noun', 'a'], ['causare', 'verb', 'a'], ['cavalcare', 'verb', 'b'], ['cavaliere', 'noun', 'a'], ['cavalletta', 'noun', 'c'], ['cavallo', 'noun', 'a'], ['cavare', 'verb', 'b'], ['cavatappi', 'noun', 'c'], ['caverna', 'noun', 'c'], ['caviglia', 'noun', 'b'], ['cavità', 'noun', 'b'], ['cavo', 'adjective', 'b'], ['cavo', 'noun', 'b'], ['cavo', 'noun', 'b'], ['cavolo', 'noun', 'b'], ['cazzata', 'noun', 'b'], ['cazzo', 'noun', 'a'], ['ce', 'pronoun', 'a'], ['ce', 'adverb', 'a'], ['cece', 'noun', 'c'], ['ceco', 'adjective', 'c'], ['ceco', 'noun', 'c'], ['cecoslovacco', 'adjective', 'c'], ['cecoslovacco', 'noun', 'c'], ['cedere', 'verb', 'a'], ['celare', 'verb', 'b'], ['celebrare', 'verb', 'b'], ['celebre', 'adjective', 'b'], ['celeste', 'adjective', 'b'], ['celeste', 'noun', 'b'], ['cella', 'noun', 'b'], ['cellula', 'noun', 'a'], ['cellulare', 'adjective', 'a'], ['cellulare', 'noun', 'a'], ['cemento', 'noun', 'b'], ['cena', 'noun', 'a'], ['cenare', 'verb', 'b'], ['cenere', 'noun', 'b'], ['cenere', 'adjective', 'b'], ['cenno', 'noun', 'b'], ['centesimo', 'adjective', 'b'], ['centesimo', 'noun', 'b'], ['centimetro', 'noun', 'b'], ['centinaio', 'noun', 'a'], ['cento', 'adjective', 'a'], ['cento', 'noun', 'a'], ['centrale', 'adjective', 'a'], ['centrale', 'noun', 'a'], ['centralino', 'noun', 'c'], ['centrare', 'verb', 'b'], ['centro', 'noun', 'a'], ['centroamericano', 'adjective', 'c'], ['centroamericano', 'noun', 'c'], ['ceramica', 'noun', 'b'], ['cercare', 'verb', 'a'], ['cerchio', 'noun', 'b'], ['cereale', 'noun', 'c'], ['cereale', 'adjective', 'c'], ['cerebrale', 'adjective', 'b'], ['cerebrale', 'noun', 'b'], ['cerimonia', 'noun', 'b'], ['cerino', 'noun', 'c'], ['cerniera', 'noun', 'c'], ['cerotto', 'noun', 'c'], ['certamente', 'adverb', 'a'], ['certezza', 'noun', 'a'], ['certificare', 'verb', 'b'], ['certificato', 'past_part', 'b'], ['certificato', 'adjective', 'b'], ['certificato', 'noun', 'b'], ['certo', 'adjective', 'a'], ['certo', 'adjective', 'a'], ['certo', 'pronoun', 'a'], ['certo', 'adverb', 'a'], ['cervello', 'noun', 'a'], ['cervo', 'noun', 'c'], ['cespuglio', 'noun', 'b'], ['cessare', 'verb', 'b'], ['cesso', 'noun', 'b'], ['cestino', 'noun', 'c'], ['cesto', 'noun', 'c'], ['cetriolo', 'noun', 'c'], ['chat', 'noun', 'b'], ['che', 'pronoun', 'a'], ['che', 'adjective', 'a'], ['che', 'noun', 'a'], ['chewingum', 'noun', 'c'], ['chi', 'pronoun', 'a'], ['chiacchiera', 'noun', 'b'], ['chiacchierare', 'verb', 'b'], ['chiamare', 'verb', 'a'], ['chiamata', 'noun', 'b'], ['chiaramente', 'adverb', 'a'], ['chiarezza', 'noun', 'b'], ['chiarire', 'verb', 'a'], ['chiaro', 'adjective', 'a'], ['chiaro', 'noun', 'a'], ['chiaro', 'adverb', 'a'], ['chiasso', 'noun', 'c'], ['chiave', 'noun', 'a'], ['chiazza', 'noun', 'c'], ['chiedere', 'verb', 'a'], ['chiesa', 'noun', 'a'], ['chilo', 'noun', 'b'], ['chilogrammo', 'noun', 'c'], ['chilometro', 'noun', 'a'], ['chimico', 'adjective', 'a'], ['chimico', 'noun', 'a'], ['china', 'noun', 'c'], ['chinare', 'verb', 'b'], ['chinotto', 'noun', 'c'], ['chiodo', 'noun', 'b'], ['chiosco', 'noun', 'b'], ['chirurgia', 'noun', 'b'], ['chirurgico', 'adjective', 'b'], ['chirurgico', 'noun', 'b'], ['chirurgo', 'noun', 'b'], ['chissà', 'adverb', 'a'], ['chitarra', 'noun', 'b'], ['chiudere', 'verb', 'a'], ['chiunque', 'pronoun', 'a'], ['chiuso', 'past_part', 'a'], ['chiuso', 'adjective', 'a'], ['chiuso', 'noun', 'a'], ['chiuso', 'adverb', 'a'], ['chiusura', 'noun', 'b'], ['ci', 'noun', 'c'], ['ci', 'pronoun', 'a'], ['ci', 'adverb', 'a'], ['ciabatta', 'noun', 'c'], ['ciambella', 'noun', 'c'], ['ciao', 'exclamation', 'a'], ['ciascuno', 'adjective', 'a'], ['ciascuno', 'pronoun', 'a'], ['cibare', 'verb', 'c'], ['cibo', 'noun', 'a'], ['cicatrice', 'noun', 'b'], ['ciclismo', 'noun', 'b'], ['ciclista', 'noun', 'c'], ['ciclo', 'noun', 'b'], ['cicogna', 'noun', 'c'], ['cicoria', 'noun', 'c'], ['cieco', 'adjective', 'b'], ['cieco', 'noun', 'b'], ['cielo', 'noun', 'a'], ['cifra', 'noun', 'a'], ['ciglio', 'noun', 'b'], ['cigno', 'noun', 'c'], ['cileno', 'adjective', 'c'], ['cileno', 'noun', 'c'], ['ciliegia', 'noun', 'c'], ['ciliegia', 'adjective', 'c'], ['ciliegio', 'noun', 'c'], ['cilindro', 'noun', 'c'], ['cima', 'noun', 'c'], ['cimice', 'noun', 'c'], ['ciminiera', 'noun', 'c'], ['cimitero', 'noun', 'b'], ['cinema', 'noun', 'a'], ['cinematografico', 'adjective', 'b'], ['cinese', 'adjective', 'a'], ['cinese', 'noun', 'a'], ['cinghia', 'noun', 'c'], ['cinghiale', 'noun', 'c'], ['cinguettare', 'verb', 'c'], ['cinguettio', 'noun', 'c'], ['cinico', 'adjective', 'c'], ['cinico', 'noun', 'c'], ['cinquanta', 'adjective', 'a'], ['cinquanta', 'noun', 'a'], ['cinque', 'adjective', 'a'], ['cinque', 'noun', 'a'], ['cinquecento', 'adjective', 'b'], ['cinquecento', 'noun', 'b'], ['cintura', 'noun', 'b'], ['cinturino', 'noun', 'c'], ['ciò', 'pronoun', 'a'], ['ciocca', 'noun', 'c'], ['cioccolatino', 'noun', 'c'], ['cioccolato', 'noun', 'b'], ['cioccolato', 'adjective', 'b'], ['cioè', 'conjunction', 'a'], ['ciotola', 'noun', 'c'], ['cipolla', 'noun', 'b'], ['cipresso', 'noun', 'c'], ['cipriota', 'adjective', 'c'], ['cipriota', 'noun', 'c'], ['circa', 'preposition', 'a'], ['circa', 'adverb', 'a'], ['circa', 'noun', 'a'], ['circo', 'noun', 'b'], ['circolare', 'adjective', 'b'], ['circolare', 'noun', 'b'], ['circolare', 'verb', 'b'], ['circolazione', 'noun', 'b'], ['circolo', 'noun', 'b'], ['circondare', 'verb', 'a'], ['circostanza', 'noun', 'a'], ['circuito', 'noun', 'b'], ['citare', 'verb', 'a'], ['citato', 'past_part', 'b'], ['citato', 'adjective', 'b'], ['citato', 'noun', 'b'], ['citazione', 'noun', 'b'], ['citofono', 'noun', 'c'], ['città', 'noun', 'a'], ['cittadina', 'noun', 'b'], ['cittadinanza', 'noun', 'b'], ['cittadino', 'adjective', 'a'], ['cittadino', 'noun', 'a'], ['ciuffo', 'noun', 'c'], ['civile', 'adjective', 'a'], ['civile', 'noun', 'a'], ['civiltà', 'noun', 'b'], ['clacson', 'noun', 'c'], ['clan', 'noun', 'b'], ['clandestino', 'adjective', 'b'], ['clandestino', 'noun', 'b'], ['classe', 'noun', 'a'], ['classico', 'adjective', 'a'], ['classico', 'noun', 'a'], ['classifica', 'noun', 'b'], ['classificare', 'verb', 'b'], ['clero', 'noun', 'c'], ['cliccare', 'verb', 'b'], ['cliente', 'noun', 'a'], ['clima', 'noun', 'b'], ['clinica', 'noun', 'b'], ['clinico', 'adjective', 'b'], ['clinico', 'noun', 'b'], ['clistere', 'noun', 'c'], ['cloro', 'noun', 'c'], ['club', 'noun', 'b'], ['cobra', 'noun', 'c'], ['cocaina', 'noun', 'b'], ['coccinella', 'noun', 'c'], ['coccio', 'noun', 'c'], ['cocciuto', 'adjective', 'c'], ['cocciuto', 'noun', 'c'], ['cocco', 'noun', 'c'], ['coccodrillo', 'noun', 'c'], ['coccola', 'noun', 'c'], ['coccolare', 'verb', 'c'], ['cocomero', 'noun', 'c'], ['coda', 'noun', 'a'], ['codice', 'noun', 'a'], ['coerente', 'adjective', 'b'], ['cofano', 'noun', 'c'], ['cogliere', 'verb', 'a'], ['coglione', 'noun', 'a'], ['cognato', 'noun', 'b'], ['cognato', 'adjective', 'b'], ['cognome', 'noun', 'b'], ['coincidenza', 'noun', 'b'], ['coincidere', 'verb', 'b'], ['coinvolgere', 'verb', 'a'], ['coinvolgimento', 'noun', 'b'], ['colare', 'verb', 'b'], ['colata', 'noun', 'c'], ['colazione', 'noun', 'b'], ['colera', 'noun', 'c'], ['colica', 'noun', 'c'], ['colino', 'noun', 'c'], ['colla', 'noun', 'c'], ['collaborare', 'verb', 'b'], ['collaboratore', 'noun', 'b'], ['collaborazione', 'noun', 'b'], ['collana', 'noun', 'b'], ['collant', 'noun', 'c'], ['collant', 'adjective', 'c'], ['collare', 'noun', 'c'], ['collasso', 'noun', 'c'], ['collaterale', 'adjective', 'b'], ['collaterale', 'noun', 'b'], ['colle', 'noun', 'c'], ['collega', 'noun', 'a'], ['collegamento', 'noun', 'b'], ['collegare', 'verb', 'a'], ['collegio', 'noun', 'b'], ['collera', 'noun', 'c'], ['colletta', 'noun', 'c'], ['collettivo', 'adjective', 'b'], ['collettivo', 'noun', 'b'], ['collezione', 'noun', 'b'], ['collina', 'noun', 'b'], ['collo', 'noun', 'a'], ['collocare', 'verb', 'b'], ['colloquio', 'noun', 'b'], ['colluttorio', 'noun', 'c'], ['colmo', 'noun', 'c'], ['colomba', 'noun', 'b'], ['colombo', 'noun', 'c'], ['colonna', 'noun', 'a'], ['colonnello', 'noun', 'b'], ['colorante', 'pres_part', 'c'], ['colorante', 'adjective', 'c'], ['colorante', 'noun', 'c'], ['colorare', 'verb', 'b'], ['colorato', 'past_part', 'b'], ['colorato', 'adjective', 'b'], ['colore', 'noun', 'a'], ['coloro', 'pronoun', 'a'], ['colosso', 'noun', 'c'], ['colpa', 'noun', 'a'], ['colpevole', 'adjective', 'b'], ['colpevole', 'noun', 'b'], ['colpire', 'verb', 'a'], ['colpo', 'noun', 'a'], ['coltellata', 'noun', 'c'], ['coltello', 'noun', 'a'], ['coltivare', 'verb', 'b'], ['coltivazione', 'noun', 'c'], ['colto', 'adjective', 'b'], ['colto', 'noun', 'b'], ['colui', 'pronoun', 'b'], ['coma', 'noun', 'b'], ['comandamento', 'noun', 'b'], ['comandante', 'pres_part', 'b'], ['comandante', 'adjective', 'b'], ['comandante', 'noun', 'b'], ['comandare', 'verb', 'b'], ['comando', 'noun', 'b'], ['combaciare', 'verb', 'c'], ['combattente', 'pres_part', 'c'], ['combattente', 'adjective', 'c'], ['combattente', 'noun', 'c'], ['combattere', 'verb', 'a'], ['combattimento', 'noun', 'b'], ['combinare', 'verb', 'b'], ['combinazione', 'noun', 'b'], ['come', 'adverb', 'a'], ['come', 'conjunction', 'a'], ['cometa', 'noun', 'c'], ['comfort', 'noun', 'c'], ['comico', 'adjective', 'b'], ['comico', 'noun', 'b'], ['cominciare', 'verb', 'a'], ['cominciare', 'noun', 'a'], ['comitato', 'noun', 'b'], ['comma', 'noun', 'b'], ['commedia', 'noun', 'b'], ['commentare', 'verb', 'a'], ['commento', 'noun', 'a'], ['commerciale', 'adjective', 'a'], ['commerciale', 'noun', 'a'], ['commerciante', 'pres_part', 'b'], ['commerciante', 'adjective', 'b'], ['commerciante', 'noun', 'b'], ['commercio', 'noun', 'b'], ['commettere', 'verb', 'a'], ['commissariato', 'noun', 'b'], ['commissario', 'noun', 'a'], ['commissione', 'noun', 'a'], ['community', 'noun', 'b'], ['commuovere', 'verb', 'b'], ['comodino', 'noun', 'c'], ['comodità', 'noun', 'c'], ['comodo', 'adjective', 'a'], ['comodo', 'noun', 'a'], ['compagnia', 'noun', 'a'], ['compagno', 'noun', 'a'], ['compagno', 'adjective', 'a'], ['comparire', 'verb', 'a'], ['comparsa', 'noun', 'b'], ['compassione', 'noun', 'c'], ['compasso', 'noun', 'c'], ['compatibile', 'adjective', 'b'], ['compatriota', 'noun', 'c'], ['compatto', 'adjective', 'b'], ['compatto', 'noun', 'b'], ['compensare', 'verb', 'b'], ['compenso', 'noun', 'b'], ['competente', 'adjective', 'b'], ['competente', 'noun', 'b'], ['competenza', 'noun', 'b'], ['competere', 'verb', 'b'], ['competizione', 'noun', 'b'], ['compiangere', 'verb', 'c'], ['compiere', 'verb', 'a'], ['compilare', 'verb', 'b'], ['compito', 'noun', 'a'], ['compleanno', 'noun', 'b'], ['complessivo', 'adjective', 'b'], ['complesso', 'noun', 'b'], ['complesso', 'adjective', 'a'], ['completamente', 'adverb', 'a'], ['completare', 'verb', 'b'], ['completo', 'adjective', 'a'], ['completo', 'noun', 'a'], ['complicare', 'verb', 'b'], ['complicato', 'past_part', 'b'], ['complicato', 'adjective', 'b'], ['complice', 'noun', 'b'], ['complice', 'adjective', 'b'], ['complimento', 'noun', 'b'], ['complotto', 'noun', 'c'], ['componente', 'pres_part', 'b'], ['componente', 'adjective', 'b'], ['componente', 'noun', 'b'], ['comporre', 'verb', 'a'], ['comportamento', 'noun', 'a'], ['comportare', 'verb', 'a'], ['composizione', 'noun', 'b'], ['composto', 'past_part', 'b'], ['composto', 'adjective', 'b'], ['composto', 'noun', 'b'], ['comprare', 'verb', 'a'], ['comprendere', 'verb', 'a'], ['comprensibile', 'adjective', 'b'], ['comprensione', 'noun', 'b'], ['comprensivo', 'adjective', 'c'], ['compreso', 'past_part', 'a'], ['compreso', 'adjective', 'a'], ['compromesso', 'noun', 'b'], ['compromettere', 'verb', 'b'], ['computer', 'noun', 'a'], ['comunale', 'adjective', 'b'], ['comunale', 'noun', 'b'], ['comune', 'adjective', 'a'], ['comune', 'noun', 'a'], ['comune', 'noun', 'a'], ['comunicare', 'verb', 'a'], ['comunicazione', 'noun', 'a'], ['comunione', 'noun', 'b'], ['comunismo', 'noun', 'b'], ['comunista', 'adjective', 'a'], ['comunista', 'noun', 'a'], ['comunità', 'noun', 'a'], ['comunque', 'adverb', 'a'], ['comunque', 'conjunction', 'a'], ['con', 'preposition', 'a'], ['conca', 'noun', 'c'], ['concedere', 'verb', 'b'], ['concentrare', 'verb', 'a'], ['concentrazione', 'noun', 'b'], ['concepire', 'noun', 'b'], ['concerto', 'noun', 'a'], ['concessione', 'noun', 'b'], ['concesso', 'past_part', 'b'], ['concesso', 'adjective', 'b'], ['concetto', 'past_part', 'a'], ['concetto', 'adjective', 'a'], ['concetto', 'noun', 'a'], ['concezione', 'noun', 'b'], ['conchiglia', 'noun', 'c'], ['concime', 'noun', 'c'], ['concludere', 'verb', 'a'], ['conclusione', 'noun', 'a'], ['concordare', 'verb', 'b'], ['concorrente', 'pres_part', 'b'], ['concorrente', 'adjective', 'b'], ['concorrente', 'noun', 'b'], ['concorrenza', 'noun', 'b'], ['concorrere', 'verb', 'b'], ['concorso', 'noun', 'b'], ['concreto', 'adjective', 'a'], ['concreto', 'noun', 'a'], ['condanna', 'noun', 'b'], ['condannare', 'verb', 'a'], ['condimento', 'noun', 'c'], ['condividere', 'verb', 'a'], ['condizionare', 'verb', 'b'], ['condizione', 'noun', 'a'], ['condoglianza', 'noun', 'c'], ['condominio', 'noun', 'b'], ['condotta', 'noun', 'b'], ['condurre', 'verb', 'a'], ['conduttore', 'adjective', 'b'], ['conduttore', 'noun', 'b'], ['conduttura', 'noun', 'c'], ['conferenza', 'noun', 'b'], ['conferire', 'verb', 'b'], ['conferma', 'noun', 'b'], ['confermare', 'verb', 'a'], ['confessare', 'verb', 'b'], ['confessione', 'noun', 'b'], ['confessore', 'noun', 'c'], ['confetto', 'noun', 'c'], ['confetto', 'adjective', 'c'], ['confettura', 'noun', 'c'], ['confezione', 'noun', 'b'], ['conficcare', 'verb', 'c'], ['confidare', 'verb', 'b'], ['confidenza', 'noun', 'b'], ['confine', 'noun', 'a'], ['conflitto', 'noun', 'b'], ['confondere', 'verb', 'a'], ['confortare', 'verb', 'c'], ['confrontare', 'verb', 'b'], ['confronto', 'noun', 'a'], ['confusione', 'noun', 'b'], ['confuso', 'past_part', 'b'], ['confuso', 'adjective', 'b'], ['congedo', 'noun', 'c'], ['congelare', 'verb', 'b'], ['congelatore', 'noun', 'c'], ['congestione', 'noun', 'c'], ['congiura', 'noun', 'c'], ['congresso', 'noun', 'b'], ['coniglio', 'noun', 'b'], ['coniugato', 'past_part', 'c'], ['coniugato', 'adjective', 'c'], ['coniugato', 'noun', 'c'], ['coniuge', 'noun', 'b'], ['connessione', 'noun', 'b'], ['connettere', 'verb', 'b'], ['cono', 'noun', 'b'], ['conoscenza', 'noun', 'a'], ['conoscere', 'verb', 'a'], ['conosciuto', 'past_part', 'b'], ['conosciuto', 'adjective', 'b'], ['conosciuto', 'noun', 'b'], ['conquista', 'noun', 'b'], ['conquistare', 'verb', 'a'], ['consapevole', 'adjective', 'b'], ['consapevolezza', 'noun', 'b'], ['consegna', 'noun', 'b'], ['consegnare', 'verb', 'a'], ['conseguente', 'pres_part', 'b'], ['conseguente', 'adjective', 'b'], ['conseguente', 'noun', 'b'], ['conseguenza', 'noun', 'a'], ['conseguire', 'verb', 'b'], ['consenso', 'noun', 'b'], ['consentire', 'verb', 'a'], ['conservare', 'verb', 'a'], ['conservazione', 'noun', 'b'], ['considerare', 'verb', 'a'], ['considerazione', 'noun', 'a'], ['consigliare', 'verb', 'a'], ['consigliere', 'noun', 'b'], ['consiglio', 'noun', 'a'], ['consistente', 'pres_part', 'b'], ['consistente', 'adjective', 'b'], ['consistenza', 'noun', 'b'], ['consistere', 'verb', 'b'], ['consolare', 'verb', 'b'], ['consonante', 'noun', 'c'], ['consorzio', 'noun', 'b'], ['constatare', 'verb', 'b'], ['consueto', 'adjective', 'b'], ['consueto', 'noun', 'b'], ['consulente', 'adjective', 'b'], ['consulente', 'noun', 'b'], ['consulenza', 'noun', 'b'], ['consultare', 'verb', 'b'], ['consumare', 'verb', 'a'], ['consumatore', 'noun', 'b'], ['consumatore', 'adjective', 'b'], ['consumazione', 'noun', 'c'], ['consumo', 'noun', 'b'], ['contachilometri', 'noun', 'c'], ['contadino', 'noun', 'b'], ['contadino', 'adjective', 'b'], ['contagiare', 'verb', 'c'], ['contagio', 'noun', 'c'], ['contagioso', 'adjective', 'c'], ['contagocce', 'noun', 'c'], ['contaminare', 'verb', 'b'], ['contante', 'pres_part', 'b'], ['contante', 'adjective', 'b'], ['contante', 'noun', 'b'], ['contare', 'verb', 'a'], ['contatore', 'noun', 'c'], ['contattare', 'verb', 'b'], ['contatto', 'noun', 'a'], ['conte', 'noun', 'b'], ['contemplare', 'verb', 'b'], ['contemporaneamente', 'adverb', 'b'], ['contemporaneo', 'adjective', 'a'], ['contemporaneo', 'noun', 'a'], ['contenere', 'verb', 'a'], ['contenitore', 'adjective', 'b'], ['contenitore', 'noun', 'b'], ['contentare', 'verb', 'b'], ['contentezza', 'noun', 'c'], ['contento', 'adjective', 'a'], ['contenuto', 'past_part', 'a'], ['contenuto', 'adjective', 'a'], ['contenuto', 'noun', 'a'], ['contestare', 'verb', 'b'], ['contestazione', 'noun', 'b'], ['contesto', 'noun', 'a'], ['continente', 'noun', 'b'], ['continuamente', 'adverb', 'b'], ['continuare', 'verb', 'a'], ['continuazione', 'noun', 'b'], ['continuità', 'noun', 'b'], ['continuo', 'adjective', 'a'], ['continuo', 'noun', 'a'], ['continuo', 'adverb', 'a'], ['conto', 'noun', 'a'], ['contorno', 'noun', 'b'], ['contrabbandiere', 'noun', 'c'], ['contrabbando', 'noun', 'c'], ['contraccambiare', 'verb', 'c'], ['contraddizione', 'noun', 'b'], ['contrario', 'adjective', 'a'], ['contrario', 'noun', 'a'], ['contrarre', 'verb', 'b'], ['contrastare', 'verb', 'b'], ['contrasto', 'noun', 'b'], ['contratto', 'noun', 'a'], ['contribuire', 'verb', 'b'], ['contributo', 'noun', 'b'], ['contro', 'preposition', 'a'], ['contro', 'adverb', 'a'], ['contro', 'noun', 'a'], ['controllare', 'verb', 'a'], ['controllo', 'noun', 'a'], ['controllore', 'noun', 'c'], ['convegno', 'noun', 'b'], ['conveniente', 'pres_part', 'b'], ['conveniente', 'adjective', 'b'], ['convenire', 'verb', 'b'], ['convenzione', 'noun', 'b'], ['conversazione', 'noun', 'a'], ['conversione', 'noun', 'b'], ['convertire', 'verb', 'b'], ['convincente', 'pres_part', 'b'], ['convincente', 'adjective', 'b'], ['convincere', 'verb', 'a'], ['convinto', 'past_part', 'b'], ['convinto', 'adjective', 'b'], ['convinzione', 'noun', 'b'], ['convivenza', 'noun', 'b'], ['convivere', 'verb', 'b'], ['convocare', 'verb', 'b'], ['convulsione', 'noun', 'c'], ['coordinamento', 'noun', 'b'], ['coordinare', 'verb', 'b'], ['coperchio', 'noun', 'c'], ['coperta', 'noun', 'b'], ['copertina', 'noun', 'b'], ['coperto', 'past_part', 'b'], ['coperto', 'adjective', 'b'], ['coperto', 'noun', 'b'], ['copertura', 'noun', 'b'], ['copia', 'noun', 'a'], ['copiare', 'verb', 'b'], ['copione', 'noun', 'b'], ['coppa', 'noun', 'b'], ['coppia', 'noun', 'a'], ['copricostume', 'noun', 'c'], ['copriletto', 'noun', 'c'], ['coprire', 'verb', 'a'], ['copyright', 'noun', 'b'], ['coraggio', 'noun', 'a'], ['coraggio', 'exclamation', 'a'], ['coraggioso', 'adjective', 'b'], ['corallo', 'noun', 'c'], ['corallo', 'adjective', 'c'], ['corazza', 'noun', 'c'], ['corazzata', 'noun', 'c'], ['corazziere', 'noun', 'c'], ['corda', 'noun', 'a'], ['coriandolo', 'noun', 'c'], ['coricare', 'verb', 'c'], ['cornacchia', 'noun', 'c'], ['cornetto', 'noun', 'c'], ['cornice', 'noun', 'b'], ['corno', 'noun', 'b'], ['cornuto', 'adjective', 'c'], ['cornuto', 'noun', 'c'], ['coro', 'noun', 'b'], ['corona', 'noun', 'b'], ['corpo', 'noun', 'a'], ['corporatura', 'noun', 'c'], ['correggere', 'verb', 'a'], ['corrente', 'pres_part', 'a'], ['corrente', 'adjective', 'a'], ['corrente', 'noun', 'a'], ['corrente', 'adverb', 'a'], ['correre', 'verb', 'a'], ['correttamente', 'adverb', 'b'], ['corretto', 'past_part', 'b'], ['corretto', 'adjective', 'b'], ['correzione', 'noun', 'c'], ['corridoio', 'noun', 'b'], ['corridore', 'adjective', 'c'], ['corridore', 'noun', 'c'], ['corriera', 'noun', 'c'], ['corriere', 'noun', 'a'], ['corrispondente', 'pres_part', 'b'], ['corrispondente', 'adjective', 'b'], ['corrispondente', 'noun', 'b'], ['corrispondenza', 'noun', 'b'], ['corrispondere', 'verb', 'a'], ['corruzione', 'noun', 'b'], ['corsa', 'noun', 'a'], ['corsia', 'noun', 'c'], ['corso', 'noun', 'a'], ['corte', 'noun', 'a'], ['corteccia', 'noun', 'c'], ['corteggiare', 'verb', 'c'], ['cortesia', 'noun', 'b'], ['cortile', 'noun', 'b'], ['corto', 'adjective', 'a'], ['corvo', 'noun', 'c'], ['cosa', 'noun', 'a'], ['coscia', 'noun', 'b'], ['cosciente', 'adjective', 'c'], ['coscienza', 'noun', 'a'], ['così', 'adverb', 'a'], ['cosiddetto', 'adjective', 'a'], ['costa', 'noun', 'a'], ['costante', 'adjective', 'b'], ['costante', 'noun', 'b'], ['costantemente', 'adverb', 'b'], ['costare', 'verb', 'a'], ['costellazione', 'noun', 'b'], ['costituire', 'verb', 'a'], ['costituzionale', 'adjective', 'b'], ['costituzione', 'noun', 'b'], ['costo', 'noun', 'a'], ['costoso', 'adjective', 'b'], ['costringere', 'verb', 'a'], ['costruire', 'verb', 'a'], ['costruttivo', 'adjective', 'b'], ['costruzione', 'noun', 'a'], ['costume', 'noun', 'a'], ['cotoletta', 'noun', 'c'], ['cotone', 'noun', 'b'], ['cottura', 'noun', 'c'], ['covare', 'verb', 'c'], ['covo', 'noun', 'c'], ['cozza', 'noun', 'c'], ['cracker', 'noun', 'c'], ['cranio', 'noun', 'b'], ['cravatta', 'noun', 'b'], ['creare', 'verb', 'a'], ['creatività', 'noun', 'b'], ['creativo', 'adjective', 'b'], ['creativo', 'noun', 'b'], ['creatura', 'noun', 'b'], ['creazione', 'noun', 'b'], ['credente', 'pres_part', 'b'], ['credente', 'adjective', 'b'], ['credente', 'noun', 'b'], ['credenza', 'noun', 'c'], ['credere', 'verb', 'a'], ['credere', 'noun', 'a'], ['credibile', 'adjective', 'b'], ['credito', 'noun', 'a'], ['creditore', 'noun', 'b'], ['credo', 'noun', 'c'], ['crema', 'noun', 'b'], ['crema', 'adjective', 'b'], ['crepaccio', 'noun', 'c'], ['crêpe', 'noun', 'c'], ['crescente', 'pres_part', 'b'], ['crescente', 'adjective', 'b'], ['crescente', 'noun', 'b'], ['crescere', 'verb', 'a'], ['crescita', 'noun', 'a'], ['cretino', 'adjective', 'b'], ['cretino', 'noun', 'b'], ['criceto', 'noun', 'c'], ['criminale', 'adjective', 'b'], ['criminale', 'noun', 'b'], ['crimine', 'noun', 'b'], ['criniera', 'noun', 'c'], ['crisantemo', 'noun', 'c'], ['crisi', 'noun', 'a'], ['cristallo', 'noun', 'b'], ['cristianesimo', 'noun', 'b'], ['cristiano', 'adjective', 'a'], ['cristiano', 'noun', 'a'], ['criterio', 'noun', 'b'], ['critica', 'noun', 'a'], ['criticare', 'verb', 'b'], ['critico', 'adjective', 'a'], ['critico', 'noun', 'a'], ['croato', 'adjective', 'c'], ['croato', 'noun', 'c'], ['croce', 'noun', 'b'], ['crocifiggere', 'verb', 'c'], ['crocifisso', 'past_part', 'c'], ['crocifisso', 'adjective', 'c'], ['crocifisso', 'noun', 'c'], ['crollare', 'verb', 'b'], ['cronaca', 'noun', 'b'], ['cronico', 'adjective', 'b'], ['cronico', 'noun', 'b'], ['cronista', 'noun', 'c'], ['crostaceo', 'noun', 'c'], ['crostino', 'noun', 'c'], ['crudele', 'adjective', 'b'], ['crudele', 'noun', 'b'], ['crudo', 'adjective', 'b'], ['crudo', 'noun', 'b'], ['cu', 'noun', 'c'], ['cubo', 'noun', 'b'], ['cubo', 'adjective', 'b'], ['cucchiaio', 'noun', 'b'], ['cuccia', 'noun', 'c'], ['cucciolo', 'noun', 'b'], ['cucina', 'noun', 'a'], ['cucinare', 'verb', 'a'], ['cucire', 'verb', 'b'], ['cucito', 'past_part', 'c'], ['cucito', 'adjective', 'c'], ['cucito', 'noun', 'c'], ['cucitura', 'noun', 'c'], ['cuffia', 'noun', 'b'], ['cugino', 'noun', 'b'], ['cui', 'pronoun', 'a'], ['cullare', 'verb', 'c'], ['culo', 'noun', 'a'], ['culto', 'noun', 'b'], ['cultura', 'noun', 'a'], ['culturale', 'adjective', 'a'], ['cumulo', 'noun', 'c'], ['cuocere', 'verb', 'b'], ['cuoco', 'noun', 'b'], ['cuore', 'noun', 'a'], ['cupo', 'adjective', 'b'], ['cupo', 'noun', 'b'], ['cura', 'noun', 'a'], ['curare', 'verb', 'a'], ['curiosare', 'verb', 'b'], ['curiosità', 'noun', 'b'], ['curioso', 'adjective', 'a'], ['curioso', 'noun', 'a'], ['curriculum', 'noun', 'b'], ['curva', 'noun', 'b'], ['curvo', 'adjective', 'b'], ['curvo', 'noun', 'b'], ['cuscino', 'noun', 'b'], ['custode', 'noun', 'b'], ['custode', 'adjective', 'b'], ['custodia', 'noun', 'b'], ['custodire', 'verb', 'b'], ['da', 'preposition', 'a'], ['dado', 'noun', 'c'], ['danese', 'adjective', 'c'], ['danese', 'noun', 'c'], ['dannato', 'past_part', 'b'], ['dannato', 'adjective', 'b'], ['dannato', 'noun', 'b'], ['danneggiare', 'verb', 'b'], ['danno', 'noun', 'a'], ['dannoso', 'adjective', 'c'], ['danza', 'noun', 'b'], ['dappertutto', 'adverb', 'b'], ['dare', 'verb', 'a'], ['dare', 'noun', 'a'], ['data', 'noun', 'a'], ['dato', 'past_part', 'a'], ['dato', 'adjective', 'a'], ['dato', 'noun', 'a'], ['dattero', 'noun', 'c'], ['davanti', 'adverb', 'a'], ['davanti', 'adjective', 'a'], ['davanti', 'noun', 'a'], ['davanzale', 'noun', 'c'], ['davvero', 'adverb', 'a'], ['dea', 'noun', 'b'], ['debito', 'noun', 'a'], ['debole', 'adjective', 'a'], ['debole', 'noun', 'a'], ['debolezza', 'noun', 'b'], ['decennio', 'noun', 'b'], ['decidere', 'verb', 'a'], ['decina', 'noun', 'a'], ['decisamente', 'adverb', 'b'], ['decisione', 'noun', 'a'], ['decisivo', 'adjective', 'b'], ['deciso', 'past_part', 'b'], ['deciso', 'adjective', 'b'], ['decorare', 'verb', 'b'], ['decorato', 'past_part', 'c'], ['decorato', 'adjective', 'c'], ['decorato', 'noun', 'c'], ['decorazione', 'noun', 'b'], ['decoroso', 'adjective', 'c'], ['decreto', 'noun', 'b'], ['dedica', 'noun', 'c'], ['dedicare', 'verb', 'a'], ['dedurre', 'verb', 'b'], ['deficiente', 'adjective', 'b'], ['deficiente', 'noun', 'b'], ['definire', 'verb', 'a'], ['definitivamente', 'adverb', 'b'], ['definitivo', 'adjective', 'a'], ['definitivo', 'noun', 'a'], ['definizione', 'noun', 'a'], ['deformare', 'verb', 'c'], ['deforme', 'adjective', 'c'], ['deforme', 'noun', 'c'], ['defunto', 'past_part', 'b'], ['defunto', 'adjective', 'b'], ['defunto', 'noun', 'b'], ['degno', 'adjective', 'b'], ['degradare', 'verb', 'b'], ['delegare', 'verb', 'b'], ['delegato', 'past_part', 'b'], ['delegato', 'adjective', 'b'], ['delegato', 'noun', 'b'], ['delegazione', 'noun', 'c'], ['delfino', 'noun', 'c'], ['delicatezza', 'noun', 'c'], ['delicato', 'adjective', 'b'], ['delicato', 'noun', 'b'], ['delinquente', 'pres_part', 'c'], ['delinquente', 'adjective', 'c'], ['delinquente', 'noun', 'c'], ['delirare', 'verb', 'c'], ['delirio', 'noun', 'b'], ['delitto', 'noun', 'b'], ['delizia', 'noun', 'c'], ['delizioso', 'adjective', 'b'], ['deludere', 'verb', 'b'], ['delusione', 'noun', 'b'], ['deluso', 'past_part', 'b'], ['deluso', 'adjective', 'b'], ['deluso', 'noun', 'b'], ['democratico', 'adjective', 'b'], ['democratico', 'noun', 'b'], ['democrazia', 'noun', 'a'], ['democristiano', 'adjective', 'c'], ['democristiano', 'noun', 'c'], ['demoralizzare', 'verb', 'c'], ['denaro', 'noun', 'a'], ['denominare', 'verb', 'b'], ['denso', 'adjective', 'b'], ['dente', 'noun', 'a'], ['dentiera', 'noun', 'c'], ['dentifricio', 'noun', 'c'], ['dentista', 'noun', 'b'], ['dentro', 'adverb', 'a'], ['dentro', 'preposition', 'a'], ['dentro', 'noun', 'a'], ['denuncia', 'noun', 'b'], ['denunciare', 'verb', 'a'], ['deodorante', 'pres_part', 'c'], ['deodorante', 'adjective', 'c'], ['deodorante', 'noun', 'c'], ['depilazione', 'noun', 'c'], ['deporre', 'verb', 'b'], ['depositare', 'verb', 'b'], ['deposito', 'noun', 'b'], ['deposizione', 'noun', 'b'], ['depressione', 'noun', 'b'], ['deprimere', 'verb', 'b'], ['depuratore', 'adjective', 'c'], ['depuratore', 'noun', 'c'], ['deputato', 'past_part', 'b'], ['deputato', 'adjective', 'b'], ['deputato', 'noun', 'b'], ['derivare', 'verb', 'a'], ['derubare', 'verb', 'c'], ['descrivere', 'verb', 'a'], ['descrizione', 'noun', 'a'], ['deserto', 'noun', 'b'], ['deserto', 'adjective', 'b'], ['desiderare', 'verb', 'a'], ['desiderio', 'noun', 'a'], ['design', 'noun', 'b'], ['dessert', 'noun', 'c'], ['destinare', 'verb', 'a'], ['destinazione', 'noun', 'b'], ['destino', 'noun', 'a'], ['destra', 'noun', 'a'], ['destro', 'adjective', 'a'], ['destro', 'noun', 'a'], ['detective', 'noun', 'b'], ['detenere', 'verb', 'b'], ['detenuto', 'past_part', 'c'], ['detenuto', 'adjective', 'c'], ['detenuto', 'noun', 'c'], ['determinare', 'verb', 'a'], ['determinato', 'past_part', 'a'], ['determinato', 'adjective', 'a'], ['determinazione', 'noun', 'b'], ['detersivo', 'adjective', 'c'], ['detersivo', 'noun', 'c'], ['dettagliato', 'past_part', 'b'], ['dettagliato', 'adjective', 'b'], ['dettaglio', 'noun', 'a'], ['dettare', 'verb', 'b'], ['dettato', 'past_part', 'c'], ['dettato', 'adjective', 'c'], ['dettato', 'noun', 'c'], ['devastare', 'verb', 'b'], ['deviare', 'verb', 'c'], ['deviazione', 'noun', 'c'], ['di', 'preposition', 'a'], ['di', 'noun', 'c'], ['diagnosi', 'noun', 'b'], ['dialetto', 'noun', 'a'], ['dialogare', 'verb', 'b'], ['dialogo', 'noun', 'a'], ['diamante', 'noun', 'a'], ['diametro', 'noun', 'b'], ['diario', 'noun', 'b'], ['diario', 'adjective', 'b'], ['diavolo', 'noun', 'a'], ['dibattito', 'noun', 'b'], ['dicembre', 'noun', 'a'], ['dichiarare', 'verb', 'a'], ['dichiarazione', 'noun', 'a'], ['diciotto', 'adjective', 'b'], ['diciotto', 'noun', 'b'], ['dieci', 'adjective', 'a'], ['dieci', 'noun', 'a'], ['diecimila', 'adjective', 'b'], ['diecimila', 'noun', 'b'], ['dieta', 'noun', 'b'], ['dietetico', 'adjective', 'c'], ['dietro', 'preposition', 'a'], ['dietro', 'adverb', 'a'], ['dietro', 'adjective', 'a'], ['dietro', 'noun', 'a'], ['difendere', 'verb', 'a'], ['difensore', 'adjective', 'b'], ['difensore', 'noun', 'b'], ['difesa', 'noun', 'a'], ['difetto', 'noun', 'b'], ['differente', 'pres_part', 'a'], ['differente', 'adjective', 'a'], ['differenza', 'noun', 'a'], ['difficile', 'adjective', 'a'], ['difficile', 'noun', 'a'], ['difficilmente', 'adverb', 'b'], ['difficoltà', 'noun', 'a'], ['diffidente', 'adjective', 'c'], ['diffidente', 'noun', 'c'], ['diffidenza', 'noun', 'c'], ['diffondere', 'verb', 'a'], ['diffusione', 'noun', 'b'], ['diffuso', 'past_part', 'b'], ['diffuso', 'adjective', 'b'], ['diga', 'noun', 'c'], ['digestione', 'noun', 'c'], ['digestivo', 'adjective', 'c'], ['digestivo', 'noun', 'c'], ['digitale', 'adjective', 'b'], ['digitale', 'noun', 'b'], ['digiunare', 'verb', 'c'], ['dignità', 'noun', 'b'], ['diluvio', 'noun', 'c'], ['dimagrante', 'pres_part', 'c'], ['dimagrante', 'adjective', 'c'], ['dimensione', 'noun', 'a'], ['dimenticare', 'verb', 'a'], ['dimettere', 'verb', 'b'], ['dimezzare', 'verb', 'c'], ['diminuire', 'verb', 'b'], ['dimostrare', 'verb', 'a'], ['dimostrazione', 'noun', 'b'], ['dinamica', 'noun', 'b'], ['dinamico', 'adjective', 'b'], ['dinosauro', 'noun', 'c'], ['dintorno', 'adverb', 'b'], ['dintorno', 'noun', 'b'], ['dio', 'noun', 'a'], ['dipartimento', 'noun', 'b'], ['dipendente', 'pres_part', 'a'], ['dipendente', 'adjective', 'a'], ['dipendente', 'noun', 'a'], ['dipendenza', 'noun', 'b'], ['dipendere', 'verb', 'a'], ['dipingere', 'verb', 'b'], ['dipinto', 'past_part', 'b'], ['dipinto', 'adjective', 'b'], ['dipinto', 'noun', 'b'], ['diploma', 'noun', 'b'], ['diplomatico', 'adjective', 'b'], ['diplomatico', 'noun', 'b'], ['dire', 'verb', 'a'], ['dire', 'noun', 'a'], ['diretta', 'noun', 'b'], ['direttamente', 'adverb', 'a'], ['diretto', 'past_part', 'a'], ['diretto', 'adjective', 'a'], ['diretto', 'noun', 'a'], ['direttore', 'noun', 'a'], ['direttore', 'adjective', 'a'], ['direttrice', 'noun', 'c'], ['direzione', 'noun', 'a'], ['dirigente', 'adjective', 'b'], ['dirigente', 'noun', 'b'], ['dirigere', 'verb', 'a'], ['diritto', 'noun', 'a'], ['disagio', 'noun', 'b'], ['disastro', 'noun', 'b'], ['disattento', 'adjective', 'c'], ['discarica', 'noun', 'b'], ['discendere', 'verb', 'b'], ['discepolo', 'noun', 'b'], ['discesa', 'noun', 'b'], ['disciplina', 'noun', 'b'], ['disco', 'noun', 'a'], ['discordia', 'noun', 'c'], ['discorso', 'noun', 'a'], ['discoteca', 'noun', 'b'], ['discreto', 'adjective', 'b'], ['discreto', 'noun', 'b'], ['discussione', 'noun', 'a'], ['discusso', 'past_part', 'b'], ['discusso', 'adjective', 'b'], ['discutere', 'verb', 'a'], ['disegnare', 'verb', 'a'], ['disegno', 'noun', 'a'], ['diseredare', 'verb', 'c'], ['disgrazia', 'noun', 'b'], ['disinfettante', 'pres_part', 'c'], ['disinfettante', 'adjective', 'c'], ['disinfettare', 'verb', 'c'], ['disinteresse', 'noun', 'c'], ['disoccupazione', 'noun', 'b'], ['disonesto', 'adjective', 'c'], ['disonesto', 'noun', 'c'], ['disordinato', 'past_part', 'c'], ['disordinato', 'adjective', 'c'], ['disordine', 'noun', 'b'], ['dispari', 'adjective', 'c'], ['dispensa', 'noun', 'c'], ['disperare', 'verb', 'b'], ['disperato', 'past_part', 'b'], ['disperato', 'adjective', 'b'], ['disperazione', 'noun', 'b'], ['disperdere', 'verb', 'b'], ['dispetto', 'noun', 'b'], ['dispettoso', 'adjective', 'c'], ['dispiacere', 'verb', 'a'], ['disponibile', 'adjective', 'a'], ['disponibile', 'noun', 'a'], ['disponibilità', 'noun', 'b'], ['disporre', 'verb', 'a'], ['dispositivo', 'adjective', 'b'], ['dispositivo', 'noun', 'b'], ['disposizione', 'noun', 'a'], ['disprezzo', 'noun', 'b'], ['dissenso', 'noun', 'c'], ['distacco', 'noun', 'b'], ['distante', 'pres_part', 'b'], ['distante', 'adjective', 'b'], ['distante', 'adverb', 'b'], ['distanza', 'noun', 'a'], ['distendere', 'verb', 'b'], ['disteso', 'past_part', 'c'], ['disteso', 'adjective', 'c'], ['disteso', 'noun', 'c'], ['distinguere', 'verb', 'a'], ['distintivo', 'adjective', 'c'], ['distintivo', 'noun', 'c'], ['distinto', 'past_part', 'b'], ['distinto', 'adjective', 'b'], ['distinto', 'noun', 'b'], ['distinzione', 'noun', 'b'], ['distrarre', 'verb', 'b'], ['distratto', 'past_part', 'c'], ['distratto', 'adjective', 'c'], ['distrazione', 'noun', 'c'], ['distretto', 'noun', 'b'], ['distribuire', 'verb', 'a'], ['distributore', 'adjective', 'b'], ['distributore', 'noun', 'b'], ['distribuzione', 'noun', 'b'], ['distruggere', 'verb', 'a'], ['distrutto', 'past_part', 'c'], ['distrutto', 'adjective', 'c'], ['distruzione', 'noun', 'b'], ['disturbare', 'verb', 'b'], ['disturbo', 'noun', 'b'], ['disubbidiente', 'pres_part', 'c'], ['disubbidiente', 'adjective', 'c'], ['disubbidienza', 'noun', 'c'], ['disubbidire', 'verb', 'c'], ['dito', 'noun', 'a'], ['ditta', 'noun', 'b'], ['dittatura', 'noun', 'b'], ['divano', 'noun', 'a'], ['divano-letto', 'noun', 'c'], ['divenire', 'verb', 'a'], ['divenire', 'noun', 'a'], ['diventare', 'verb', 'a'], ['diversamente', 'adverb', 'b'], ['diversità', 'noun', 'b'], ['diverso', 'adjective', 'a'], ['diverso', 'adjective', 'a'], ['diverso', 'pronoun', 'a'], ['divertente', 'pres_part', 'a'], ['divertente', 'adjective', 'a'], ['divertimento', 'noun', 'b'], ['divertire', 'verb', 'a'], ['divertito', 'past_part', 'b'], ['divertito', 'adjective', 'b'], ['dividere', 'verb', 'a'], ['divieto', 'noun', 'b'], ['divinità', 'noun', 'b'], ['divino', 'adjective', 'b'], ['divino', 'noun', 'b'], ['divisa', 'noun', 'b'], ['divisione', 'noun', 'b'], ['divorare', 'verb', 'b'], ['divorziare', 'verb', 'c'], ['divorzio', 'noun', 'b'], ['dizionario', 'noun', 'b'], ['do', 'noun', 'c'], ['doccia', 'noun', 'b'], ['docciaschiuma', 'noun', 'c'], ['docente', 'pres_part', 'b'], ['docente', 'adjective', 'b'], ['docente', 'noun', 'b'], ['docile', 'adjective', 'c'], ['documentare', 'verb', 'b'], ['documentario', 'adjective', 'b'], ['documentario', 'noun', 'b'], ['documentazione', 'noun', 'b'], ['documento', 'noun', 'a'], ['dodici', 'adjective', 'a'], ['dodici', 'noun', 'a'], ['dogana', 'noun', 'c'], ['dolce', 'adjective', 'a'], ['dolce', 'noun', 'a'], ['dolce', 'adverb', 'a'], ['dolcezza', 'noun', 'b'], ['dolcificante', 'pres_part', 'c'], ['dolcificante', 'adjective', 'c'], ['dolcificante', 'noun', 'c'], ['dolciume', 'noun', 'c'], ['dolere', 'verb', 'c'], ['dolersi', 'verb', 'c'], ['dollaro', 'noun', 'a'], ['dolore', 'noun', 'a'], ['doloroso', 'adjective', 'b'], ['domanda', 'noun', 'a'], ['domandare', 'verb', 'a'], ['domani', 'adverb', 'a'], ['domani', 'noun', 'a'], ['domenica', 'noun', 'a'], ['domestica', 'noun', 'c'], ['domestico', 'adjective', 'b'], ['domestico', 'noun', 'b'], ['dominante', 'pres_part', 'b'], ['dominante', 'adjective', 'b'], ['dominante', 'noun', 'b'], ['dominare', 'verb', 'b'], ['dominio', 'noun', 'b'], ['don', 'noun', 'a'], ['donare', 'verb', 'b'], ['dondolare', 'verb', 'c'], ['donna', 'noun', 'a'], ['dono', 'noun', 'b'], ['dopo', 'adverb', 'a'], ['dopo', 'preposition', 'a'], ['dopo', 'conjunction', 'a'], ['dopo', 'adjective', 'a'], ['dopo', 'noun', 'a'], ['dopobarba', 'noun', 'c'], ['doppio', 'adjective', 'a'], ['doppio', 'noun', 'a'], ['doppio', 'adverb', 'a'], ['doppione', 'noun', 'c'], ['dorato', 'past_part', 'b'], ['dorato', 'adjective', 'b'], ['dorato', 'noun', 'b'], ['dormiglione', 'adjective', 'c'], ['dormiglione', 'noun', 'c'], ['dormire', 'verb', 'a'], ['dorso', 'noun', 'b'], ['dose', 'noun', 'b'], ['dotare', 'verb', 'b'], ['dotato', 'past_part', 'b'], ['dotato', 'adjective', 'b'], ['dote', 'noun', 'b'], ['dottore', 'noun', 'a'], ['dottoressa', 'noun', 'b'], ['dottrina', 'noun', 'b'], ['dove', 'adverb', 'a'], ['dove', 'conjunction', 'a'], ['dove', 'noun', 'a'], ['dovere', 'verb', 'a'], ['dovere', 'noun', 'a'], ['dovuto', 'past_part', 'b'], ['dovuto', 'adjective', 'b'], ['dovuto', 'noun', 'b'], ['dozzina', 'noun', 'b'], ['drago', 'noun', 'b'], ['dramma', 'noun', 'b'], ['drammatico', 'adjective', 'b'], ['dritto', 'adjective', 'b'], ['dritto', 'adverb', 'b'], ['dritto', 'noun', 'b'], ['drizzare', 'verb', 'c'], ['droga', 'noun', 'a'], ['drogare', 'verb', 'b'], ['drogato', 'past_part', 'c'], ['drogato', 'adjective', 'c'], ['drogato', 'noun', 'c'], ['dubbio', 'noun', 'a'], ['dubbio', 'adjective', 'b'], ['dubitare', 'verb', 'b'], ['dublinese', 'adjective', 'c'], ['dublinese', 'noun', 'c'], ['due', 'adjective', 'a'], ['due', 'noun', 'a'], ['duecento', 'adjective', 'b'], ['duecento', 'noun', 'b'], ['duello', 'noun', 'b'], ['duemila', 'adjective', 'b'], ['duemila', 'noun', 'b'], ['dunque', 'conjunction', 'a'], ['dunque', 'noun', 'a'], ['duomo', 'noun', 'c'], ['durante', 'pres_part', 'a'], ['durante', 'preposition', 'a'], ['durante', 'noun', 'a'], ['durare', 'verb', 'a'], ['durata', 'noun', 'a'], ['duro', 'adjective', 'a'], ['duro', 'noun', 'a'], ['duro', 'adverb', 'a'], ['e', 'noun', 'c'], ['e', 'conjunction', 'a'], ['ebbene', 'conjunction', 'b'], ['ebraico', 'adjective', 'b'], ['ebraico', 'noun', 'b'], ['ebreo', 'adjective', 'a'], ['ebreo', 'noun', 'a'], ['eccellente', 'pres_part', 'b'], ['eccellente', 'adjective', 'b'], ['eccellenza', 'noun', 'b'], ['eccessivo', 'adjective', 'b'], ['eccesso', 'noun', 'b'], ['eccetera', 'adverb', 'b'], ['eccezionale', 'adjective', 'b'], ['eccezione', 'noun', 'b'], ['eccitare', 'verb', 'b'], ['ecco', 'adverb', 'a'], ['eco', 'noun', 'b'], ['ecologico', 'adjective', 'b'], ['economia', 'noun', 'a'], ['economico', 'adjective', 'a'], ['economico', 'noun', 'a'], ['economista', 'noun', 'b'], ['edicola', 'noun', 'a'], ['edificio', 'noun', 'a'], ['editore', 'noun', 'a'], ['editore', 'adjective', 'a'], ['editoriale', 'adjective', 'b'], ['editoriale', 'noun', 'b'], ['edizione', 'noun', 'a'], ['educare', 'verb', 'b'], ['educativo', 'adjective', 'b'], ['educato', 'past_part', 'c'], ['educato', 'adjective', 'c'], ['educazione', 'noun', 'a'], ['effe', 'noun', 'c'], ['effettivamente', 'adverb', 'a'], ['effettivo', 'adjective', 'b'], ['effettivo', 'noun', 'b'], ['effetto', 'noun', 'a'], ['effettuare', 'verb', 'a'], ['efficace', 'adjective', 'b'], ['efficacia', 'noun', 'b'], ['efficiente', 'adjective', 'b'], ['efficienza', 'noun', 'b'], ['egiziano', 'adjective', 'c'], ['egiziano', 'noun', 'c'], ['egli', 'pronoun', 'a'], ['elaborare', 'verb', 'b'], ['elaborazione', 'noun', 'b'], ['elastico', 'adjective', 'b'], ['elastico', 'noun', 'b'], ['elegante', 'adjective', 'a'], ['eleganza', 'noun', 'b'], ['eleggere', 'verb', 'b'], ['elementare', 'adjective', 'a'], ['elemento', 'noun', 'a'], ['elemosina', 'noun', 'c'], ['elencare', 'verb', 'b'], ['elenco', 'noun', 'a'], ['elettorale', 'adjective', 'b'], ['elettore', 'noun', 'b'], ['elettricista', 'noun', 'c'], ['elettricità', 'noun', 'c'], ['elettrico', 'adjective', 'a'], ['elettrico', 'noun', 'a'], ['elettrodomestico', 'noun', 'c'], ['elettromagnetico', 'adjective', 'b'], ['elettrone', 'noun', 'b'], ['elettronico', 'adjective', 'a'], ['elevare', 'verb', 'b'], ['elevato', 'past_part', 'b'], ['elevato', 'adjective', 'b'], ['elezione', 'noun', 'b'], ['elica', 'noun', 'c'], ['elicottero', 'noun', 'c'], ['eliminare', 'verb', 'a'], ['eliminazione', 'noun', 'b'], ['elle', 'noun', 'c'], ['elmo', 'noun', 'c'], ['e-mail', 'noun', 'a'], ['emanare', 'verb', 'b'], ['emergenza', 'noun', 'b'], ['emergere', 'verb', 'a'], ['emettere', 'verb', 'b'], ['emigrazione', 'noun', 'c'], ['emiliano', 'adjective', 'c'], ['emiliano', 'noun', 'c'], ['emissione', 'noun', 'b'], ['emme', 'noun', 'c'], ['emmenthal', 'noun', 'c'], ['emo', 'noun', 'b'], ['emotivo', 'adjective', 'b'], ['emotivo', 'noun', 'b'], ['emozionante', 'pres_part', 'c'], ['emozionante', 'adjective', 'c'], ['emozionare', 'verb', 'b'], ['emozionato', 'past_part', 'c'], ['emozionato', 'adjective', 'c'], ['emozione', 'noun', 'a'], ['enciclopedia', 'noun', 'c'], ['energetico', 'adjective', 'b'], ['energetico', 'noun', 'b'], ['energia', 'noun', 'a'], ['enne', 'noun', 'c'], ['ennesimo', 'adjective', 'b'], ['enorme', 'adjective', 'a'], ['ente', 'noun', 'a'], ['entità', 'noun', 'b'], ['entrambi', 'pronoun', 'a'], ['entrambi', 'adjective', 'a'], ['entrare', 'verb', 'a'], ['entrare', 'noun', 'a'], ['entrata', 'noun', 'a'], ['entro', 'preposition', 'a'], ['entro', 'adverb', 'a'], ['entusiasmo', 'noun', 'b'], ['entusiasta', 'adjective', 'b'], ['entusiasta', 'noun', 'b'], ['epifania', 'noun', 'c'], ['episodio', 'noun', 'a'], ['epoca', 'noun', 'a'], ['eppure', 'conjunction', 'a'], ['equazione', 'noun', 'b'], ['equilibrio', 'noun', 'a'], ['equino', 'adjective', 'c'], ['equino', 'noun', 'c'], ['equipaggio', 'noun', 'c'], ['equivalere', 'verb', 'b'], ['equivoco', 'adjective', 'b'], ['equivoco', 'noun', 'b'], ['era', 'noun', 'a'], ['erba', 'noun', 'b'], ['erede', 'noun', 'b'], ['eredità', 'noun', 'b'], ['ereditare', 'verb', 'b'], ['ergastolo', 'noun', 'c'], ['ergere', 'verb', 'b'], ['ernia', 'noun', 'c'], ['eroe', 'noun', 'a'], ['eroina', 'noun', 'c'], ['erotico', 'adjective', 'b'], ['erotico', 'noun', 'b'], ['errare', 'verb', 'b'], ['erre', 'noun', 'c'], ['errore', 'noun', 'a'], ['esagerare', 'verb', 'b'], ['esagerato', 'past_part', 'b'], ['esagerato', 'adjective', 'b'], ['esagerato', 'noun', 'b'], ['esagerazione', 'noun', 'c'], ['esagono', 'noun', 'c'], ['esagono', 'adjective', 'c'], ['esaltare', 'verb', 'b'], ['esaltazione', 'noun', 'c'], ['esame', 'noun', 'a'], ['esaminare', 'verb', 'b'], ['esattamente', 'adverb', 'a'], ['esatto', 'adjective', 'a'], ['esatto', 'adverb', 'a'], ['esaurire', 'verb', 'b'], ['esca', 'noun', 'c'], ['eschimese', 'adjective', 'c'], ['eschimese', 'noun', 'c'], ['esclamare', 'verb', 'b'], ['esclamazione', 'noun', 'c'], ['escludere', 'verb', 'a'], ['esclusione', 'noun', 'b'], ['esclusivamente', 'adverb', 'b'], ['esclusivo', 'adjective', 'b'], ['escluso', 'past_part', 'b'], ['escluso', 'adjective', 'b'], ['escluso', 'noun', 'b'], ['esecutivo', 'adjective', 'b'], ['esecutivo', 'noun', 'b'], ['esecuzione', 'noun', 'b'], ['eseguire', 'verb', 'a'], ['esempio', 'noun', 'a'], ['esemplare', 'noun', 'b'], ['esemplare', 'adjective', 'b'], ['esercitare', 'verb', 'b'], ['esercito', 'noun', 'a'], ['esercizio', 'noun', 'a'], ['esibire', 'verb', 'b'], ['esigenza', 'noun', 'a'], ['esigere', 'verb', 'b'], ['esilio', 'noun', 'c'], ['esistente', 'pres_part', 'b'], ['esistente', 'adjective', 'b'], ['esistente', 'noun', 'b'], ['esistenza', 'noun', 'a'], ['esistere', 'verb', 'a'], ['esitare', 'verb', 'b'], ['esito', 'noun', 'b'], ['esordio', 'noun', 'b'], ['espansione', 'noun', 'b'], ['espellere', 'verb', 'b'], ['esperienza', 'noun', 'a'], ['esperimento', 'noun', 'b'], ['esperto', 'past_part', 'a'], ['esperto', 'adjective', 'a'], ['esperto', 'noun', 'a'], ['esplicito', 'adjective', 'b'], ['esplodere', 'verb', 'b'], ['esplorare', 'verb', 'b'], ['esplosione', 'noun', 'b'], ['esplosivo', 'adjective', 'b'], ['esplosivo', 'noun', 'b'], ['esponente', 'pres_part', 'b'], ['esponente', 'noun', 'b'], ['esporre', 'verb', 'a'], ['esposizione', 'noun', 'b'], ['espressione', 'noun', 'a'], ['espresso', 'past_part', 'c'], ['espresso', 'adjective', 'c'], ['espresso', 'noun', 'c'], ['esprimere', 'verb', 'a'], ['essa', 'pronoun', 'a'], ['esse', 'noun', 'c'], ['esse', 'pronoun', 'b'], ['essenza', 'noun', 'b'], ['essenziale', 'adjective', 'b'], ['essenziale', 'noun', 'b'], ['essenzialmente', 'adverb', 'b'], ['essere', 'verb', 'a'], ['essere', 'noun', 'a'], ['essi', 'pronoun', 'a'], ['esso', 'pronoun', 'a'], ['est', 'noun', 'b'], ['est', 'adjective', 'b'], ['estate', 'noun', 'a'], ['estendere', 'verb', 'b'], ['estensione', 'noun', 'b'], ['esterno', 'adjective', 'a'], ['esterno', 'noun', 'a'], ['estero', 'adjective', 'a'], ['estero', 'noun', 'a'], ['estetico', 'adjective', 'b'], ['estivo', 'adjective', 'b'], ['estone', 'adjective', 'c'], ['estone', 'noun', 'c'], ['estraneo', 'adjective', 'b'], ['estraneo', 'noun', 'b'], ['estrarre', 'verb', 'b'], ['estratto', 'past_part', 'b'], ['estratto', 'adjective', 'b'], ['estratto', 'noun', 'b'], ['estrazione', 'noun', 'b'], ['estremamente', 'adverb', 'b'], ['estremità', 'noun', 'b'], ['estremo', 'adjective', 'a'], ['estremo', 'noun', 'a'], ['età', 'noun', 'a'], ['eterno', 'adjective', 'b'], ['eterno', 'noun', 'b'], ['etica', 'noun', 'b'], ['etichetta', 'noun', 'b'], ['etico', 'adjective', 'b'], ['ettaro', 'noun', 'c'], ['etto', 'noun', 'c'], ['euro', 'noun', 'a'], ['europeo', 'adjective', 'a'], ['europeo', 'noun', 'a'], ['evadere', 'verb', 'c'], ['evaporare', 'verb', 'c'], ['evasione', 'noun', 'b'], ['evento', 'noun', 'a'], ['eventuale', 'adjective', 'a'], ['eventualmente', 'adverb', 'b'], ['evidente', 'adjective', 'a'], ['evidentemente', 'adverb', 'a'], ['evidenza', 'noun', 'b'], ['evidenziare', 'verb', 'b'], ['evidenziatore', 'adjective', 'c'], ['evidenziatore', 'noun', 'c'], ['evitare', 'verb', 'a'], ['evocare', 'verb', 'b'], ['evoluzione', 'noun', 'b'], ['ex', 'adjective', 'a'], ['ex', 'noun', 'a'], ['ex', 'preposition', 'a'], ['extra', 'adjective', 'b'], ['extra', 'noun', 'b'], ['fa', 'adverb', 'a'], ['fabbrica', 'noun', 'a'], ['fabbricare', 'verb', 'b'], ['fabbro', 'noun', 'c'], ['faccenda', 'noun', 'b'], ['faccia', 'noun', 'a'], ['facciata', 'noun', 'b'], ['facile', 'adjective', 'a'], ['facile', 'adverb', 'a'], ['facilità', 'noun', 'b'], ['facilitare', 'verb', 'b'], ['facilitazione', 'noun', 'c'], ['facilmente', 'adverb', 'a'], ['facoltà', 'noun', 'b'], ['fagiano', 'noun', 'c'], ['falco', 'noun', 'c'], ['falegname', 'noun', 'c'], ['fallimento', 'noun', 'b'], ['fallire', 'verb', 'b'], ['fallito', 'past_part', 'b'], ['fallito', 'adjective', 'b'], ['fallito', 'noun', 'b'], ['falso', 'adjective', 'a'], ['falso', 'adverb', 'a'], ['falso', 'noun', 'a'], ['fama', 'noun', 'b'], ['fame', 'noun', 'a'], ['famiglia', 'noun', 'a'], ['familiare', 'adjective', 'a'], ['familiare', 'noun', 'a'], ['famoso', 'adjective', 'a'], ['fan', 'noun', 'b'], ['fanale', 'noun', 'c'], ['fanciulla', 'noun', 'b'], ['fanciullo', 'adjective', 'c'], ['fanciullo', 'noun', 'c'], ['fango', 'noun', 'b'], ['fangoso', 'adjective', 'c'], ['fantascienza', 'noun', 'b'], ['fantasia', 'noun', 'a'], ['fantasma', 'noun', 'b'], ['fantastico', 'adjective', 'a'], ['fantastico', 'noun', 'a'], ['fanteria', 'noun', 'c'], ['fantino', 'noun', 'c'], ['fantoccio', 'noun', 'c'], ['fare', 'verb', 'a'], ['fare', 'noun', 'a'], ['farfalla', 'noun', 'b'], ['farina', 'noun', 'b'], ['farmacia', 'noun', 'b'], ['farmaco', 'noun', 'b'], ['faro', 'noun', 'c'], ['fascia', 'noun', 'a'], ['fasciatoio', 'noun', 'c'], ['fascicolo', 'noun', 'b'], ['fascino', 'noun', 'b'], ['fascio', 'noun', 'b'], ['fascismo', 'noun', 'b'], ['fascista', 'adjective', 'b'], ['fascista', 'noun', 'b'], ['fase', 'noun', 'a'], ['fastidio', 'noun', 'a'], ['fastidioso', 'adjective', 'b'], ['fata', 'noun', 'b'], ['fatica', 'noun', 'a'], ['faticare', 'verb', 'b'], ['faticoso', 'adjective', 'b'], ['fatto', 'noun', 'a'], ['fattore', 'noun', 'a'], ['fattoria', 'noun', 'b'], ['fattura', 'noun', 'b'], ['fatturato', 'past_part', 'b'], ['fatturato', 'adjective', 'b'], ['fatturato', 'noun', 'b'], ['fauna', 'noun', 'c'], ['fava', 'noun', 'c'], ['favola', 'noun', 'b'], ['favoloso', 'adjective', 'b'], ['favore', 'noun', 'a'], ['favorevole', 'adjective', 'b'], ['favorire', 'verb', 'b'], ['fax', 'noun', 'b'], ['fazzoletto', 'noun', 'b'], ['febbraio', 'noun', 'a'], ['febbre', 'noun', 'b'], ['fecondare', 'verb', 'c'], ['fede', 'noun', 'a'], ['fedele', 'adjective', 'b'], ['fedele', 'noun', 'b'], ['fedeltà', 'noun', 'b'], ['federa', 'noun', 'c'], ['federale', 'adjective', 'b'], ['federale', 'noun', 'b'], ['fegato', 'noun', 'b'], ['felice', 'adjective', 'a'], ['felicità', 'noun', 'b'], ['felino', 'noun', 'c'], ['felino', 'adjective', 'c'], ['felpa', 'noun', 'c'], ['femmina', 'noun', 'a'], ['femminile', 'adjective', 'a'], ['femminile', 'noun', 'a'], ['fenomeno', 'noun', 'a'], ['feria', 'noun', 'b'], ['feriale', 'adjective', 'c'], ['ferie', 'noun', 'c'], ['ferire', 'verb', 'b'], ['ferita', 'noun', 'a'], ['ferito', 'past_part', 'b'], ['ferito', 'adjective', 'b'], ['ferito', 'noun', 'b'], ['fermaglio', 'noun', 'c'], ['fermare', 'verb', 'a'], ['fermo', 'adjective', 'a'], ['feroce', 'adjective', 'b'], ['ferragosto', 'noun', 'c'], ['ferramenta', 'noun', 'c'], ['ferro', 'noun', 'a'], ['ferrovia', 'noun', 'b'], ['ferroviario', 'adjective', 'b'], ['ferroviere', 'noun', 'c'], ['fertilizzante', 'pres_part', 'c'], ['fertilizzante', 'adjective', 'c'], ['fertilizzante', 'noun', 'c'], ['fessura', 'noun', 'c'], ['festa', 'noun', 'a'], ['festeggiare', 'verb', 'a'], ['festival', 'noun', 'b'], ['festivo', 'adjective', 'c'], ['fetta', 'noun', 'b'], ['fiaba', 'noun', 'b'], ['fiala', 'noun', 'c'], ['fiamma', 'noun', 'b'], ['fiammifero', 'noun', 'c'], ['fiammifero', 'adjective', 'c'], ['fianco', 'noun', 'a'], ['fiatare', 'verb', 'c'], ['fiato', 'noun', 'b'], ['fibbia', 'noun', 'c'], ['fibra', 'noun', 'b'], ['ficcare', 'verb', 'b'], ['fiction', 'noun', 'b'], ['fidanzamento', 'noun', 'c'], ['fidanzarsi', 'verb', 'b'], ['fidanzata', 'noun', 'b'], ['fidanzato', 'past_part', 'b'], ['fidanzato', 'adjective', 'b'], ['fidanzato', 'noun', 'b'], ['fidarsi', 'verb', 'a'], ['fiducia', 'noun', 'a'], ['fiducioso', 'adjective', 'c'], ['fieno', 'noun', 'c'], ['fiera', 'noun', 'b'], ['fiero', 'adjective', 'b'], ['figlia', 'noun', 'a'], ['figliastro', 'noun', 'c'], ['figlio', 'noun', 'a'], ['figura', 'noun', 'a'], ['figurare', 'verb', 'a'], ['figurina', 'noun', 'c'], ['fila', 'noun', 'a'], ['filante', 'pres_part', 'c'], ['filante', 'adjective', 'c'], ['filante', 'noun', 'c'], ['filare', 'verb', 'b'], ['filastrocca', 'noun', 'c'], ['file', 'noun', 'a'], ['filetto', 'noun', 'c'], ['film', 'noun', 'a'], ['filmato', 'past_part', 'b'], ['filmato', 'adjective', 'b'], ['filmato', 'noun', 'b'], ['filo', 'noun', 'a'], ['filosofia', 'noun', 'a'], ['filosofico', 'adjective', 'b'], ['filosofo', 'noun', 'b'], ['filtrare', 'verb', 'b'], ['filtro', 'noun', 'b'], ['finale', 'adjective', 'a'], ['finale', 'noun', 'a'], ['finalità', 'noun', 'b'], ['finalmente', 'adverb', 'a'], ['finanza', 'noun', 'b'], ['finanziamento', 'noun', 'b'], ['finanziare', 'verb', 'b'], ['finanziario', 'adjective', 'a'], ['finanziatore', 'adjective', 'c'], ['finanziatore', 'noun', 'c'], ['finché', 'conjunction', 'a'], ['fine', 'noun', 'a'], ['fine', 'adjective', 'b'], ['finestra', 'noun', 'a'], ['finestrino', 'noun', 'b'], ['fingere', 'verb', 'a'], ['finimondo', 'noun', 'c'], ['finire', 'verb', 'a'], ['finire', 'noun', 'a'], ['finito', 'past_part', 'b'], ['finito', 'adjective', 'b'], ['finlandese', 'adjective', 'c'], ['finlandese', 'noun', 'c'], ['fino', 'preposition', 'a'], ['fino', 'adverb', 'a'], ['finocchio', 'noun', 'c'], ['finora', 'adverb', 'b'], ['finta', 'noun', 'b'], ['finto', 'past_part', 'a'], ['finto', 'adjective', 'a'], ['fiocco', 'noun', 'c'], ['fionda', 'noun', 'c'], ['fioraio', 'noun', 'c'], ['fiore', 'noun', 'a'], ['fiorentino', 'adjective', 'b'], ['fiorentino', 'noun', 'b'], ['fiorito', 'past_part', 'c'], ['fiorito', 'adjective', 'c'], ['firma', 'noun', 'a'], ['firmare', 'verb', 'a'], ['fiscale', 'adjective', 'b'], ['fiscale', 'noun', 'b'], ['fisicamente', 'adverb', 'b'], ['fisico', 'adjective', 'a'], ['fisico', 'noun', 'a'], ['fissare', 'verb', 'a'], ['fisso', 'adjective', 'a'], ['fisso', 'adverb', 'a'], ['fisso', 'noun', 'a'], ['fitto', 'past_part', 'b'], ['fitto', 'adjective', 'b'], ['fitto', 'adverb', 'b'], ['fitto', 'noun', 'b'], ['fiume', 'noun', 'a'], ['fiuto', 'noun', 'c'], ['flash', 'noun', 'b'], ['flauto', 'noun', 'c'], ['flessibile', 'adjective', 'b'], ['flessibile', 'noun', 'b'], ['flora', 'noun', 'c'], ['fluido', 'adjective', 'b'], ['fluido', 'noun', 'b'], ['fluoro', 'noun', 'c'], ['flusso', 'noun', 'b'], ['foca', 'noun', 'c'], ['focaccia', 'noun', 'c'], ['fodera', 'noun', 'c'], ['foderare', 'verb', 'c'], ['foglia', 'noun', 'b'], ['foglio', 'noun', 'a'], ['fogna', 'noun', 'c'], ['folla', 'noun', 'b'], ['folle', 'adjective', 'b'], ['folle', 'noun', 'b'], ['follia', 'noun', 'b'], ['fondamentale', 'adjective', 'a'], ['fondamentale', 'noun', 'a'], ['fondamentalmente', 'adverb', 'b'], ['fondamento', 'noun', 'b'], ['fondare', 'verb', 'a'], ['fondatore', 'noun', 'b'], ['fondazione', 'noun', 'b'], ['fondere', 'verb', 'b'], ['fondo', 'adjective', 'loc-comando'], ['fondo', 'noun', 'loc-comando'], ['fondo', 'adverb', 'loc-comando'], ['fontana', 'noun', 'b'], ['fontanella', 'noun', 'c'], ['fonte', 'noun', 'a'], ['forare', 'verb', 'b'], ['forbice', 'noun', 'c'], ['forchetta', 'noun', 'c'], ['forcina', 'noun', 'c'], ['foresta', 'noun', 'b'], ['forestale', 'adjective', 'c'], ['forestale', 'noun', 'c'], ['forfora', 'noun', 'c'], ['forma', 'noun', 'a'], ['formaggino', 'noun', 'c'], ['formaggio', 'noun', 'b'], ['formale', 'adjective', 'b'], ['formare', 'verb', 'a'], ['formato', 'past_part', 'b'], ['formato', 'adjective', 'b'], ['formato', 'noun', 'b'], ['formazione', 'noun', 'a'], ['formula', 'noun', 'a'], ['formulare', 'verb', 'b'], ['fornace', 'noun', 'c'], ['fornaio', 'noun', 'c'], ['fornello', 'noun', 'b'], ['fornire', 'verb', 'a'], ['fornitore', 'adjective', 'b'], ['fornitore', 'noun', 'b'], ['forno', 'noun', 'b'], ['foro', 'noun', 'b'], ['forse', 'adverb', 'a'], ['forse', 'noun', 'a'], ['forte', 'adjective', 'a'], ['forte', 'adverb', 'a'], ['forte', 'noun', 'a'], ['fortemente', 'adverb', 'b'], ['fortuna', 'noun', 'a'], ['fortunatamente', 'adverb', 'b'], ['fortunato', 'adjective', 'b'], ['forum', 'noun', 'b'], ['forza', 'noun', 'a'], ['forzare', 'verb', 'b'], ['fosforescente', 'adjective', 'c'], ['fossa', 'noun', 'b'], ['fossetta', 'noun', 'c'], ['fosso', 'noun', 'c'], ['foto', 'noun', 'a'], ['fotografare', 'verb', 'b'], ['fotografia', 'noun', 'a'], ['fotografico', 'adjective', 'b'], ['fotografo', 'noun', 'b'], ['fottere', 'verb', 'b'], ['foulard', 'noun', 'c'], ['fra', 'preposition', 'a'], ['fracasso', 'noun', 'c'], ['fragile', 'adjective', 'b'], ['frammento', 'noun', 'b'], ['francamente', 'adverb', 'b'], ['francese', 'adjective', 'a'], ['francese', 'noun', 'a'], ['francobollo', 'noun', 'c'], ['frangia', 'noun', 'c'], ['frase', 'noun', 'a'], ['fratello', 'noun', 'a'], ['frazione', 'noun', 'b'], ['freccia', 'noun', 'b'], ['freddezza', 'noun', 'c'], ['freddo', 'adjective', 'a'], ['freddo', 'noun', 'a'], ['fregare', 'verb', 'a'], ['frenare', 'verb', 'b'], ['frenetico', 'adjective', 'b'], ['freno', 'noun', 'b'], ['frequentare', 'verb', 'a'], ['frequente', 'adjective', 'b'], ['frequenza', 'noun', 'b'], ['fresco', 'adjective', 'a'], ['fresco', 'noun', 'a'], ['fretta', 'noun', 'a'], ['frigo', 'noun', 'b'], ['frigorifero', 'adjective', 'b'], ['frigorifero', 'noun', 'b'], ['fringuello', 'noun', 'c'], ['frittata', 'noun', 'c'], ['fritto', 'past_part', 'c'], ['fritto', 'adjective', 'c'], ['fritto', 'noun', 'c'], ['friulano', 'adjective', 'c'], ['friulano', 'noun', 'c'], ['fronte', 'noun', 'a'], ['frontiera', 'noun', 'b'], ['frugare', 'verb', 'b'], ['frumento', 'noun', 'c'], ['fruscio', 'noun', 'c'], ['frusta', 'noun', 'c'], ['frutta', 'noun', 'b'], ['fruttivendolo', 'noun', 'c'], ['frutto', 'noun', 'a'], ['fucile', 'noun', 'b'], ['fuga', 'noun', 'a'], ['fuggire', 'verb', 'a'], ['fulmine', 'noun', 'b'], ['fumare', 'verb', 'a'], ['fumetto', 'noun', 'b'], ['fumo', 'noun', 'a'], ['fumo', 'adjective', 'a'], ['fune', 'noun', 'c'], ['funerale', 'noun', 'b'], ['funerale', 'adjective', 'b'], ['fungo', 'noun', 'b'], ['funzionale', 'adjective', 'b'], ['funzionale', 'noun', 'b'], ['funzionamento', 'noun', 'b'], ['funzionare', 'verb', 'a'], ['funzionario', 'noun', 'b'], ['funzione', 'noun', 'a'], ['fuoco', 'noun', 'loc-comando'], ['fuori', 'adverb', 'a'], ['fuori', 'preposition', 'a'], ['fuori', 'noun', 'a'], ['fuori', 'adjective', 'a'], ['furbo', 'adjective', 'b'], ['furbo', 'noun', 'b'], ['furfante', 'noun', 'c'], ['furgone', 'noun', 'b'], ['furia', 'noun', 'b'], ['furioso', 'adjective', 'b'], ['furto', 'noun', 'b'], ['fusione', 'noun', 'b'], ['fuso', 'past_part', 'b'], ['fuso', 'adjective', 'b'], ['fuso', 'noun', 'b'], ['futuro', 'adjective', 'a'], ['futuro', 'noun', 'a'], ['gabbia', 'noun', 'b'], ['galassia', 'noun', 'b'], ['galeotto', 'noun', 'c'], ['galera', 'noun', 'b'], ['galleggiare', 'verb', 'c'], ['galleria', 'noun', 'b'], ['gallese', 'adjective', 'c'], ['gallese', 'noun', 'c'], ['galletta', 'noun', 'c'], ['gallina', 'noun', 'b'], ['gallo', 'noun', 'c'], ['gamba', 'noun', 'a'], ['gambero', 'noun', 'c'], ['gambo', 'noun', 'c'], ['ganascia', 'noun', 'c'], ['gancio', 'noun', 'c'], ['gara', 'noun', 'a'], ['garage', 'noun', 'b'], ['garantire', 'verb', 'a'], ['garanzia', 'noun', 'b'], ['garbo', 'noun', 'c'], ['gargarismo', 'noun', 'c'], ['garofano', 'noun', 'c'], ['garza', 'noun', 'c'], ['gas', 'noun', 'a'], ['gasolio', 'noun', 'c'], ['gassosa', 'noun', 'c'], ['gastronomia', 'noun', 'c'], ['gatto', 'noun', 'a'], ['gavetta', 'noun', 'c'], ['gay', 'adjective', 'b'], ['gay', 'noun', 'b'], ['gazza', 'noun', 'c'], ['gelateria', 'noun', 'c'], ['gelatina', 'noun', 'c'], ['gelato', 'past_part', 'b'], ['gelato', 'adjective', 'b'], ['gelato', 'noun', 'b'], ['gelido', 'adjective', 'b'], ['gelo', 'noun', 'c'], ['gelosia', 'noun', 'b'], ['geloso', 'adjective', 'b'], ['gelsomino', 'noun', 'c'], ['gemello', 'adjective', 'b'], ['gemello', 'noun', 'b'], ['gemma', 'noun', 'c'], ['gene', 'noun', 'b'], ['generale', 'adjective', 'a'], ['generale', 'noun', 'a'], ['generalmente', 'adverb', 'b'], ['generare', 'verb', 'a'], ['generazione', 'noun', 'a'], ['genere', 'noun', 'a'], ['generico', 'adjective', 'b'], ['generico', 'noun', 'b'], ['generosità', 'noun', 'c'], ['generoso', 'adjective', 'b'], ['genetico', 'adjective', 'b'], ['gengiva', 'noun', 'c'], ['geniale', 'adjective', 'b'], ['genio', 'noun', 'b'], ['genitore', 'noun', 'a'], ['gennaio', 'noun', 'a'], ['genovese', 'adjective', 'c'], ['genovese', 'noun', 'c'], ['gente', 'noun', 'a'], ['gentile', 'adjective', 'a'], ['gentile', 'noun', 'a'], ['genuino', 'adjective', 'c'], ['geografico', 'adjective', 'b'], ['geografo', 'noun', 'c'], ['geometra', 'noun', 'c'], ['geometria', 'noun', 'c'], ['geometrico', 'adjective', 'c'], ['gesso', 'noun', 'b'], ['gestione', 'noun', 'a'], ['gestire', 'verb', 'a'], ['gesto', 'noun', 'a'], ['gestore', 'noun', 'b'], ['gettare', 'verb', 'a'], ['gettone', 'noun', 'c'], ['ghiaccio', 'noun', 'b'], ['ghiacciolo', 'noun', 'c'], ['ghianda', 'noun', 'c'], ['ghiro', 'noun', 'c'], ['gi', 'noun', 'c'], ['già', 'adverb', 'a'], ['giacca', 'noun', 'a'], ['giacere', 'verb', 'b'], ['giaguaro', 'noun', 'c'], ['giallo', 'adjective', 'a'], ['giallo', 'noun', 'a'], ['giapponese', 'adjective', 'a'], ['giapponese', 'noun', 'a'], ['giardinaggio', 'noun', 'c'], ['giardiniera', 'noun', 'c'], ['giardino', 'noun', 'a'], ['gigante', 'noun', 'b'], ['gigante', 'adjective', 'b'], ['gigantesco', 'adjective', 'b'], ['giglio', 'noun', 'b'], ['ginnastica', 'noun', 'b'], ['ginocchio', 'noun', 'a'], ['giocare', 'verb', 'a'], ['giocatore', 'noun', 'a'], ['giocattolo', 'noun', 'b'], ['gioco', 'noun', 'a'], ['gioia', 'noun', 'a'], ['gioiello', 'noun', 'b'], ['gioioso', 'adjective', 'c'], ['giordano', 'adjective', 'c'], ['giordano', 'noun', 'c'], ['giornale', 'noun', 'a'], ['giornale', 'adjective', 'a'], ['giornalino', 'noun', 'c'], ['giornalista', 'noun', 'a'], ['giornata', 'noun', 'a'], ['giorno', 'noun', 'a'], ['giostra', 'noun', 'c'], ['giovane', 'adjective', 'a'], ['giovane', 'noun', 'a'], ['giovanile', 'adjective', 'b'], ['giovedì', 'noun', 'b'], ['gioventù', 'noun', 'b'], ['giovinezza', 'noun', 'b'], ['giraffa', 'noun', 'c'], ['girare', 'verb', 'a'], ['giravite', 'noun', 'c'], ['giretto', 'noun', 'c'], ['giro', 'noun', 'a'], ['gironzolare', 'verb', 'c'], ['girotondo', 'noun', 'c'], ['gita', 'noun', 'b'], ['giù', 'adverb', 'a'], ['giù', 'adjective', 'a'], ['giubba', 'noun', 'c'], ['giubbotto', 'noun', 'c'], ['giudicare', 'verb', 'a'], ['giudice', 'noun', 'a'], ['giudiziario', 'adjective', 'b'], ['giudizio', 'noun', 'a'], ['giugno', 'noun', 'a'], ['giungere', 'verb', 'a'], ['giungla', 'noun', 'c'], ['giuramento', 'noun', 'b'], ['giurare', 'verb', 'a'], ['giuria', 'noun', 'c'], ['giuridico', 'adjective', 'b'], ['giustamente', 'adverb', 'b'], ['giustificare', 'verb', 'b'], ['giustizia', 'noun', 'a'], ['giusto', 'adjective', 'a'], ['giusto', 'noun', 'a'], ['giusto', 'adverb', 'a'], ['gli', 'pronoun', 'a'], ['glicine', 'noun', 'c'], ['global', 'adjective', 'b'], ['global', 'noun', 'b'], ['globale', 'adjective', 'b'], ['gloria', 'noun', 'b'], ['gnocco', 'noun', 'c'], ['gnomo', 'noun', 'c'], ['goal', 'noun', 'b'], ['gobbo', 'adjective', 'c'], ['gobbo', 'noun', 'c'], ['goccia', 'noun', 'b'], ['godere', 'verb', 'a'], ['gola', 'noun', 'b'], ['goloso', 'adjective', 'c'], ['gomito', 'noun', 'b'], ['gomitolo', 'noun', 'c'], ['gomma', 'noun', 'b'], ['gonfiare', 'verb', 'b'], ['gonfio', 'adjective', 'b'], ['gonfio', 'noun', 'b'], ['gonna', 'noun', 'b'], ['gorgonzola', 'noun', 'c'], ['gorilla', 'noun', 'c'], ['gossip', 'noun', 'b'], ['governare', 'verb', 'b'], ['governatore', 'noun', 'b'], ['governo', 'noun', 'a'], ['gradino', 'noun', 'b'], ['gradire', 'verb', 'b'], ['grado', 'noun', 'a'], ['graffiare', 'verb', 'c'], ['graffio', 'noun', 'c'], ['grafico', 'adjective', 'b'], ['grafico', 'noun', 'b'], ['grammatica', 'noun', 'b'], ['grammo', 'noun', 'b'], ['grana', 'noun', 'c'], ['granaio', 'noun', 'c'], ['granchio', 'noun', 'c'], ['grande', 'adjective', 'a'], ['grande', 'noun', 'a'], ['grandezza', 'noun', 'b'], ['grandine', 'noun', 'c'], ['grandioso', 'adjective', 'b'], ['grano', 'noun', 'b'], ['granturco', 'noun', 'c'], ['grappa', 'noun', 'c'], ['grasso', 'adjective', 'a'], ['grasso', 'noun', 'a'], ['gratis', 'adverb', 'b'], ['gratis', 'adjective', 'b'], ['grattare', 'verb', 'b'], ['grattugiato', 'past_part', 'c'], ['grattugiato', 'adjective', 'c'], ['gratuito', 'adjective', 'b'], ['grave', 'adjective', 'a'], ['grave', 'noun', 'a'], ['grave', 'adverb', 'a'], ['gravidanza', 'noun', 'b'], ['gravità', 'noun', 'b'], ['grazie', 'exclamation', 'a'], ['grazie', 'noun', 'a'], ['grazioso', 'adjective', 'c'], ['greco', 'adjective', 'a'], ['greco', 'noun', 'a'], ['grembiule', 'noun', 'c'], ['gridare', 'verb', 'a'], ['grido', 'noun', 'b'], ['grigio', 'adjective', 'a'], ['grigio', 'noun', 'a'], ['griglia', 'noun', 'c'], ['grinza', 'noun', 'c'], ['grissino', 'noun', 'c'], ['grossista', 'noun', 'c'], ['grosso', 'adjective', 'a'], ['grosso', 'noun', 'a'], ['grotta', 'noun', 'b'], ['gru', 'noun', 'c'], ['gruppo', 'noun', 'a'], ['guadagnare', 'verb', 'a'], ['guadagno', 'noun', 'b'], ['guaio', 'noun', 'b'], ['guaire', 'verb', 'c'], ['guancia', 'noun', 'b'], ['guanciale', 'noun', 'c'], ['guanciale', 'adjective', 'c'], ['guanto', 'noun', 'b'], ['guardare', 'verb', 'a'], ['guardaroba', 'noun', 'c'], ['guardia', 'noun', 'a'], ['guarire', 'verb', 'b'], ['guarnizione', 'noun', 'c'], ['guasto', 'noun', 'c'], ['guerra', 'noun', 'a'], ['guerriero', 'noun', 'b'], ['guerriero', 'adjective', 'b'], ['gufo', 'noun', 'c'], ['guida', 'noun', 'a'], ['guidare', 'verb', 'a'], ['guidatore', 'noun', 'c'], ['guinzaglio', 'noun', 'c'], ['gustare', 'verb', 'b'], ['gusto', 'noun', 'a'], ['gustoso', 'adjective', 'c'], ['hamburger', 'noun', 'c'], ['hobby', 'noun', 'b'], ['home', 'noun', 'b'], ['hotel', 'noun', 'b'], ['hyperlink', 'noun', 'b'], ['i', 'noun', 'c'], ['i', 'determiner', 'b'], ['icona', 'noun', 'b'], ['ics', 'noun', 'c'], ['idea', 'noun', 'a'], ['ideale', 'adjective', 'a'], ['ideale', 'noun', 'a'], ['ideare', 'verb', 'b'], ['identico', 'adjective', 'b'], ['identico', 'noun', 'b'], ['identificare', 'verb', 'a'], ['identificazione', 'noun', 'b'], ['identità', 'noun', 'a'], ['ideologia', 'noun', 'b'], ['ideologico', 'adjective', 'b'], ['idiota', 'adjective', 'a'], ['idiota', 'noun', 'a'], ['idraulico', 'adjective', 'b'], ['idraulico', 'noun', 'b'], ['idrico', 'adjective', 'b'], ['idrogeno', 'noun', 'b'], ['ieri', 'adverb', 'a'], ['ieri', 'noun', 'a'], ['igiene', 'noun', 'c'], ['ignorante', 'pres_part', 'b'], ['ignorante', 'adjective', 'b'], ['ignorante', 'noun', 'b'], ['ignoranza', 'noun', 'b'], ['ignorare', 'verb', 'a'], ['ignoto', 'adjective', 'b'], ['ignoto', 'noun', 'b'], ['il', 'determiner', 'a'], ['il', 'pronoun', 'a'], ['illecito', 'adjective', 'b'], ['illecito', 'noun', 'b'], ['illegale', 'adjective', 'b'], ['illegale', 'noun', 'b'], ['illegittimo', 'adjective', 'c'], ['illegittimo', 'noun', 'c'], ['illudere', 'verb', 'b'], ['illuminare', 'verb', 'b'], ['illuminato', 'past_part', 'b'], ['illuminato', 'adjective', 'b'], ['illuminato', 'noun', 'b'], ['illusione', 'noun', 'b'], ['illustrare', 'verb', 'b'], ['illustre', 'adjective', 'b'], ['imballare', 'verb', 'c'], ['imbarazzante', 'pres_part', 'b'], ['imbarazzante', 'adjective', 'b'], ['imbarazzato', 'past_part', 'b'], ['imbarazzato', 'adjective', 'b'], ['imbarazzo', 'noun', 'b'], ['imbattersi', 'verb', 'b'], ['imbecille', 'adjective', 'b'], ['imbecille', 'noun', 'b'], ['imbiancare', 'verb', 'c'], ['imbianchino', 'noun', 'c'], ['imbottigliare', 'verb', 'c'], ['imbrogliare', 'verb', 'c'], ['imbroglio', 'noun', 'c'], ['imbuto', 'noun', 'c'], ['imitare', 'verb', 'b'], ['immaginare', 'verb', 'a'], ['immaginare', 'noun', 'a'], ['immaginario', 'adjective', 'b'], ['immaginario', 'noun', 'b'], ['immaginazione', 'noun', 'b'], ['immagine', 'noun', 'a'], ['immaturo', 'adjective', 'c'], ['immediatamente', 'adverb', 'a'], ['immediato', 'adjective', 'b'], ['immediato', 'noun', 'b'], ['immenso', 'adjective', 'b'], ['immenso', 'noun', 'b'], ['immergere', 'verb', 'b'], ['immigrato', 'past_part', 'b'], ['immigrato', 'adjective', 'b'], ['immigrato', 'noun', 'b'], ['immobile', 'adjective', 'a'], ['immobile', 'noun', 'a'], ['immobiliare', 'adjective', 'b'], ['immobiliare', 'noun', 'b'], ['immondizia', 'noun', 'c'], ['impallidire', 'verb', 'c'], ['imparare', 'verb', 'a'], ['impastare', 'verb', 'c'], ['impatto', 'noun', 'b'], ['impaziente', 'adjective', 'c'], ['impaziente', 'noun', 'c'], ['impazzire', 'verb', 'b'], ['impedire', 'verb', 'a'], ['impegnare', 'verb', 'a'], ['impegnativo', 'adjective', 'b'], ['impegnato', 'past_part', 'c'], ['impegnato', 'adjective', 'c'], ['impegno', 'noun', 'a'], ['imperare', 'verb', 'b'], ['imperatore', 'noun', 'b'], ['imperiale', 'adjective', 'b'], ['imperiale', 'noun', 'b'], ['impermeabile', 'adjective', 'c'], ['impermeabile', 'noun', 'c'], ['impero', 'noun', 'b'], ['impero', 'adjective', 'b'], ['impianto', 'noun', 'a'], ['impiegare', 'verb', 'a'], ['impiegato', 'past_part', 'b'], ['impiegato', 'adjective', 'b'], ['impiegato', 'noun', 'b'], ['impiego', 'noun', 'b'], ['implicare', 'verb', 'b'], ['imporre', 'verb', 'a'], ['importante', 'pres_part', 'a'], ['importante', 'adjective', 'a'], ['importante', 'noun', 'a'], ['importanza', 'noun', 'a'], ['importare', 'verb', 'a'], ['importo', 'noun', 'b'], ['impossibile', 'adjective', 'a'], ['impossibile', 'noun', 'a'], ['impostare', 'verb', 'b'], ['impostazione', 'noun', 'b'], ['impreciso', 'adjective', 'c'], ['imprenditore', 'noun', 'b'], ['impresa', 'noun', 'a'], ['impressionante', 'pres_part', 'b'], ['impressionante', 'adjective', 'b'], ['impressionare', 'verb', 'b'], ['impressione', 'noun', 'a'], ['imprevisto', 'adjective', 'b'], ['imprevisto', 'noun', 'b'], ['imprigionare', 'verb', 'c'], ['improbabile', 'adjective', 'b'], ['impronta', 'noun', 'b'], ['improvvisamente', 'adverb', 'b'], ['improvvisare', 'verb', 'b'], ['improvviso', 'adjective', 'a'], ['improvviso', 'noun', 'a'], ['imprudente', 'adjective', 'c'], ['imprudente', 'noun', 'c'], ['impulsivo', 'adjective', 'c'], ['impulsivo', 'noun', 'c'], ['impulso', 'noun', 'b'], ['imputata', 'noun', 'b'], ['imputato', 'past_part', 'a'], ['imputato', 'adjective', 'a'], ['imputato', 'noun', 'a'], ['in', 'preposition', 'a'], ['inaspettato', 'adjective', 'b'], ['inaugurare', 'verb', 'b'], ['incamminare', 'verb', 'c'], ['incantare', 'verb', 'c'], ['incapace', 'adjective', 'b'], ['incapace', 'noun', 'b'], ['incapacità', 'noun', 'b'], ['incaricare', 'verb', 'b'], ['incarico', 'noun', 'b'], ['incartare', 'verb', 'c'], ['incassare', 'verb', 'b'], ['incasso', 'noun', 'c'], ['incastrare', 'verb', 'b'], ['incatenare', 'verb', 'c'], ['incazzarsi', 'verb', 'b'], ['incendio', 'noun', 'b'], ['incertezza', 'noun', 'b'], ['incerto', 'adjective', 'b'], ['incerto', 'noun', 'b'], ['inchiesta', 'noun', 'b'], ['inchiodare', 'verb', 'c'], ['incidente', 'noun', 'a'], ['incidere', 'verb', 'b'], ['incinta', 'adjective', 'b'], ['incitare', 'verb', 'c'], ['incivile', 'adjective', 'c'], ['incivile', 'noun', 'c'], ['includere', 'verb', 'b'], ['incluso', 'past_part', 'b'], ['incluso', 'adjective', 'b'], ['incluso', 'noun', 'b'], ['incollare', 'verb', 'b'], ['incominciare', 'verb', 'b'], ['incompleto', 'adjective', 'c'], ['incomprensibile', 'adjective', 'b'], ['inconsolabile', 'adjective', 'c'], ['incontentabile', 'adjective', 'c'], ['incontrare', 'verb', 'a'], ['incontro', 'noun', 'a'], ['incontro', 'adverb', 'b'], ['incoraggiare', 'verb', 'b'], ['incoronare', 'verb', 'c'], ['incorreggibile', 'adjective', 'c'], ['incredibile', 'adjective', 'a'], ['incremento', 'noun', 'b'], ['incrinare', 'verb', 'c'], ['incrociare', 'verb', 'b'], ['incrocio', 'noun', 'c'], ['incubo', 'noun', 'b'], ['incurabile', 'adjective', 'c'], ['incurabile', 'noun', 'c'], ['incuriosire', 'verb', 'b'], ['indagare', 'verb', 'b'], ['indagine', 'noun', 'a'], ['indescrivibile', 'adjective', 'c'], ['indiano', 'adjective', 'b'], ['indiano', 'noun', 'b'], ['indicare', 'verb', 'a'], ['indicazione', 'noun', 'a'], ['indice', 'noun', 'a'], ['indice', 'adjective', 'a'], ['indietreggiare', 'verb', 'c'], ['indietro', 'adverb', 'a'], ['indietro', 'adjective', 'a'], ['indietro', 'loc-comando', 'a'], ['indifeso', 'adjective', 'c'], ['indifferente', 'adjective', 'b'], ['indifferente', 'noun', 'b'], ['indifferenza', 'noun', 'b'], ['indigestione', 'noun', 'c'], ['indimenticabile', 'adjective', 'c'], ['indipendente', 'adjective', 'b'], ['indipendente', 'noun', 'b'], ['indipendentemente', 'adverb', 'b'], ['indipendenza', 'noun', 'b'], ['indiretto', 'adjective', 'b'], ['indirizzare', 'verb', 'b'], ['indirizzo', 'noun', 'a'], ['indisciplinato', 'adjective', 'c'], ['indispensabile', 'adjective', 'b'], ['indispensabile', 'noun', 'b'], ['individuale', 'adjective', 'b'], ['individuare', 'verb', 'a'], ['individuo', 'noun', 'a'], ['individuo', 'adjective', 'a'], ['indizio', 'noun', 'b'], ['indossare', 'verb', 'a'], ['indovinare', 'verb', 'b'], ['indovinello', 'noun', 'c'], ['indubbiamente', 'adverb', 'b'], ['indumento', 'noun', 'c'], ['indurre', 'verb', 'b'], ['industria', 'noun', 'a'], ['industriale', 'adjective', 'a'], ['industriale', 'noun', 'a'], ['inedito', 'adjective', 'b'], ['inefficace', 'adjective', 'c'], ['inerte', 'adjective', 'c'], ['inesistente', 'adjective', 'b'], ['inesperienza', 'noun', 'c'], ['inesperto', 'adjective', 'c'], ['inevitabile', 'adjective', 'b'], ['inevitabile', 'noun', 'b'], ['inevitabilmente', 'adverb', 'b'], ['infame', 'adjective', 'c'], ['infame', 'noun', 'c'], ['infantile', 'adjective', 'b'], ['infanzia', 'noun', 'b'], ['infarto', 'noun', 'b'], ['infatti', 'conjunction', 'a'], ['infatti', 'adverb', 'a'], ['infedele', 'adjective', 'c'], ['infedele', 'noun', 'c'], ['infelice', 'adjective', 'b'], ['infelice', 'noun', 'b'], ['inferiore', 'adjective', 'a'], ['infermiera', 'noun', 'b'], ['infermiere', 'noun', 'c'], ['inferno', 'noun', 'b'], ['inferno', 'adjective', 'b'], ['infezione', 'noun', 'b'], ['infilare', 'verb', 'a'], ['infine', 'adverb', 'a'], ['infinito', 'adjective', 'a'], ['infinito', 'noun', 'a'], ['influenza', 'noun', 'b'], ['influenzare', 'verb', 'b'], ['informare', 'verb', 'a'], ['informatica', 'noun', 'b'], ['informatico', 'adjective', 'b'], ['informatico', 'noun', 'b'], ['informativo', 'adjective', 'b'], ['informazione', 'noun', 'a'], ['infradito', 'adjective', 'c'], ['infradito', 'noun', 'c'], ['infrastruttura', 'noun', 'b'], ['infuriare', 'verb', 'b'], ['infuso', 'past_part', 'c'], ['infuso', 'adjective', 'c'], ['infuso', 'noun', 'c'], ['ingannare', 'verb', 'b'], ['inganno', 'noun', 'b'], ['ingegnere', 'noun', 'b'], ['ingegneria', 'noun', 'b'], ['ingelosire', 'verb', 'c'], ['ingenuo', 'adjective', 'b'], ['ingenuo', 'noun', 'b'], ['ingessare', 'verb', 'c'], ['ingiusto', 'adjective', 'b'], ['ingiusto', 'noun', 'b'], ['inglese', 'adjective', 'a'], ['inglese', 'noun', 'a'], ['ingoiare', 'verb', 'b'], ['ingorgo', 'noun', 'c'], ['ingrandire', 'verb', 'c'], ['ingrassare', 'verb', 'b'], ['ingrediente', 'noun', 'b'], ['ingresso', 'noun', 'a'], ['iniezione', 'noun', 'c'], ['iniziale', 'adjective', 'a'], ['iniziale', 'noun', 'a'], ['inizialmente', 'adverb', 'b'], ['iniziare', 'verb', 'a'], ['iniziativa', 'noun', 'a'], ['inizio', 'noun', 'a'], ['innamorarsi', 'verb', 'a'], ['innamorato', 'past_part', 'b'], ['innamorato', 'adjective', 'b'], ['innamorato', 'noun', 'b'], ['innanzitutto', 'adverb', 'b'], ['innervosire', 'verb', 'c'], ['innocente', 'adjective', 'b'], ['innocente', 'noun', 'b'], ['innocuo', 'adjective', 'b'], ['innovativo', 'adjective', 'b'], ['innovazione', 'noun', 'b'], ['inoltre', 'adverb', 'a'], ['inquadrare', 'verb', 'b'], ['inquietante', 'pres_part', 'b'], ['inquietante', 'adjective', 'b'], ['inquinamento', 'noun', 'b'], ['inquinare', 'verb', 'c'], ['inquinato', 'past_part', 'c'], ['inquinato', 'adjective', 'c'], ['insalata', 'noun', 'b'], ['insegna', 'noun', 'b'], ['insegnamento', 'noun', 'b'], ['insegnante', 'pres_part', 'a'], ['insegnante', 'adjective', 'a'], ['insegnante', 'noun', 'a'], ['insegnare', 'verb', 'a'], ['inseguire', 'verb', 'b'], ['inseparabile', 'adjective', 'c'], ['inseparabile', 'noun', 'c'], ['inserimento', 'noun', 'b'], ['inserire', 'verb', 'a'], ['insetticida', 'adjective', 'c'], ['insetto', 'noun', 'b'], ['insieme', 'adverb', 'a'], ['insieme', 'noun', 'a'], ['insinuare', 'verb', 'b'], ['insistere', 'verb', 'a'], ['insoddisfatto', 'adjective', 'c'], ['insolito', 'adjective', 'b'], ['insolito', 'noun', 'b'], ['insomma', 'adverb', 'a'], ['insopportabile', 'adjective', 'b'], ['insospettire', 'verb', 'c'], ['installare', 'verb', 'b'], ['insuccesso', 'noun', 'c'], ['insultare', 'verb', 'b'], ['insulto', 'noun', 'b'], ['intanto', 'adverb', 'a'], ['intasare', 'verb', 'c'], ['intatto', 'adjective', 'b'], ['integrale', 'adjective', 'b'], ['integrale', 'noun', 'b'], ['integrare', 'verb', 'b'], ['integrazione', 'noun', 'b'], ['intellettuale', 'adjective', 'b'], ['intellettuale', 'noun', 'b'], ['intelligente', 'adjective', 'a'], ['intelligenza', 'noun', 'b'], ['intendere', 'verb', 'a'], ['intensità', 'noun', 'b'], ['intenso', 'adjective', 'a'], ['intento', 'noun', 'b'], ['intenzione', 'noun', 'a'], ['interagire', 'verb', 'b'], ['interamente', 'adverb', 'b'], ['interazione', 'noun', 'b'], ['intercettare', 'verb', 'b'], ['intercettazione', 'noun', 'b'], ['interessante', 'pres_part', 'a'], ['interessante', 'adjective', 'a'], ['interessare', 'verb', 'a'], ['interessato', 'past_part', 'b'], ['interessato', 'adjective', 'b'], ['interessato', 'noun', 'b'], ['interesse', 'noun', 'a'], ['interiore', 'adjective', 'b'], ['interiore', 'noun', 'b'], ['interlocutore', 'noun', 'b'], ['internazionale', 'adjective', 'a'], ['internazionale', 'noun', 'a'], ['internet', 'noun', 'a'], ['interno', 'adjective', 'a'], ['interno', 'noun', 'a'], ['intero', 'adjective', 'a'], ['intero', 'noun', 'a'], ['interpretare', 'verb', 'a'], ['interpretazione', 'noun', 'b'], ['interprete', 'noun', 'b'], ['interrogare', 'verb', 'b'], ['interrogativo', 'adjective', 'b'], ['interrogativo', 'noun', 'b'], ['interrogatorio', 'adjective', 'b'], ['interrogatorio', 'noun', 'b'], ['interrogazione', 'noun', 'c'], ['interrompere', 'verb', 'a'], ['interruttore', 'noun', 'c'], ['interruzione', 'noun', 'b'], ['intervallo', 'noun', 'b'], ['intervenire', 'verb', 'a'], ['intervento', 'noun', 'a'], ['intervista', 'noun', 'a'], ['intesa', 'noun', 'b'], ['intestare', 'verb', 'b'], ['intestino', 'noun', 'c'], ['intimidire', 'verb', 'c'], ['intimità', 'noun', 'b'], ['intimo', 'adjective', 'b'], ['intimo', 'noun', 'b'], ['intitolare', 'verb', 'b'], ['intonaco', 'noun', 'c'], ['intorno', 'adverb', 'a'], ['intorno', 'preposition', 'a'], ['intorno', 'adjective', 'a'], ['intorno', 'noun', 'a'], ['intraprendere', 'verb', 'b'], ['intravedere', 'verb', 'b'], ['intrecciare', 'verb', 'b'], ['introdurre', 'verb', 'a'], ['introduzione', 'noun', 'b'], ['intuire', 'verb', 'b'], ['intuizione', 'noun', 'b'], ['inutile', 'adjective', 'a'], ['invadente', 'pres_part', 'c'], ['invadente', 'adjective', 'c'], ['invadente', 'noun', 'c'], ['invadere', 'verb', 'b'], ['invasione', 'noun', 'b'], ['invecchiare', 'verb', 'b'], ['invece', 'adverb', 'a'], ['inventare', 'verb', 'a'], ['invenzione', 'noun', 'b'], ['invernale', 'adjective', 'b'], ['invernale', 'noun', 'b'], ['inverno', 'noun', 'a'], ['investimento', 'noun', 'b'], ['investire', 'verb', 'a'], ['inviare', 'verb', 'a'], ['inviato', 'past_part', 'b'], ['inviato', 'adjective', 'b'], ['inviato', 'noun', 'b'], ['invidiare', 'verb', 'b'], ['invidioso', 'adjective', 'c'], ['invidioso', 'noun', 'c'], ['invincibile', 'adjective', 'c'], ['invisibile', 'adjective', 'b'], ['invisibile', 'noun', 'b'], ['invitare', 'verb', 'a'], ['invitato', 'past_part', 'b'], ['invitato', 'adjective', 'b'], ['invitato', 'noun', 'b'], ['invito', 'noun', 'b'], ['invocare', 'verb', 'b'], ['inzuppare', 'verb', 'c'], ['io', 'pronoun', 'a'], ['ionico', 'adjective', 'c'], ['ipotesi', 'noun', 'a'], ['ipotizzare', 'verb', 'b'], ['ippopotamo', 'noun', 'c'], ['ipsilon', 'noun', 'c'], ['ira', 'noun', 'b'], ['irlandese', 'adjective', 'b'], ['irlandese', 'noun', 'b'], ['ironia', 'noun', 'b'], ['ironico', 'adjective', 'b'], ['irriconoscibile', 'adjective', 'c'], ['irritare', 'verb', 'b'], ['iscritto', 'past_part', 'b'], ['iscritto', 'adjective', 'b'], ['iscritto', 'noun', 'b'], ['iscrivere', 'verb', 'a'], ['iscrizione', 'noun', 'b'], ['islamico', 'adjective', 'b'], ['islamico', 'noun', 'b'], ['islandese', 'adjective', 'c'], ['islandese', 'noun', 'c'], ['isola', 'noun', 'a'], ['isolare', 'verb', 'b'], ['isolato', 'past_part', 'b'], ['isolato', 'adjective', 'b'], ['isolato', 'noun', 'b'], ['ispettore', 'noun', 'b'], ['ispirare', 'verb', 'a'], ['ispirazione', 'noun', 'b'], ['israeliano', 'adjective', 'c'], ['israeliano', 'noun', 'c'], ['istante', 'noun', 'a'], ['istanza', 'noun', 'b'], ['istintivo', 'adjective', 'c'], ['istinto', 'noun', 'b'], ['istituto', 'noun', 'a'], ['istituzionale', 'adjective', 'b'], ['istituzione', 'noun', 'a'], ['istruttivo', 'adjective', 'c'], ['istruttore', 'noun', 'c'], ['istruzione', 'noun', 'a'], ['italiano', 'adjective', 'a'], ['italiano', 'noun', 'a'], ['iugoslavo', 'adjective', 'c'], ['iugoslavo', 'noun', 'c'], ['jeans', 'noun', 'b'], ['karatè', 'noun', 'c'], ['ketchup', 'noun', 'c'], ['killer', 'noun', 'b'], ['killer', 'adjective', 'b'], ['kit', 'noun', 'c'], ['kiwi', 'noun', 'c'], ['là', 'adverb', 'a'], ['la', 'determiner', 'a'], ['la', 'pronoun', 'a'], ['labbro', 'noun', 'a'], ['labirinto', 'noun', 'c'], ['laboratorio', 'noun', 'a'], ['laborioso', 'adjective', 'c'], ['lacca', 'noun', 'c'], ['lacca', 'adjective', 'c'], ['laccio', 'noun', 'c'], ['lacrima', 'noun', 'a'], ['laddove', 'adverb', 'b'], ['laddove', 'conjunction', 'b'], ['ladro', 'noun', 'b'], ['laggiù', 'adverb', 'b'], ['lago', 'noun', 'a'], ['laico', 'adjective', 'b'], ['laico', 'noun', 'b'], ['lama', 'noun', 'b'], ['lamentare', 'verb', 'a'], ['lamentela', 'noun', 'c'], ['lametta', 'noun', 'c'], ['lamiera', 'noun', 'c'], ['lampada', 'noun', 'b'], ['lampadario', 'noun', 'c'], ['lampo', 'noun', 'b'], ['lampo', 'adjective', 'b'], ['lampo', 'noun', 'b'], ['lana', 'noun', 'b'], ['lancetta', 'noun', 'c'], ['lanciare', 'verb', 'a'], ['lancio', 'noun', 'b'], ['lanterna', 'noun', 'c'], ['lapis', 'noun', 'c'], ['lardo', 'noun', 'c'], ['larghezza', 'noun', 'c'], ['largo', 'adjective', 'a'], ['largo', 'noun', 'a'], ['largo', 'adverb', 'a'], ['lasagna', 'noun', 'c'], ['lasciare', 'verb', 'a'], ['lassù', 'adverb', 'b'], ['lastra', 'noun', 'b'], ['laterale', 'adjective', 'b'], ['laterale', 'noun', 'b'], ['latino', 'adjective', 'b'], ['latino', 'noun', 'b'], ['lato', 'noun', 'a'], ['latta', 'noun', 'c'], ['lattante', 'pres_part', 'c'], ['lattante', 'adjective', 'c'], ['lattante', 'noun', 'c'], ['latte', 'noun', 'a'], ['latte', 'adjective', 'a'], ['latteria', 'noun', 'c'], ['lattina', 'noun', 'c'], ['lattuga', 'noun', 'c'], ['laurea', 'noun', 'b'], ['laureare', 'verb', 'b'], ['laureato', 'past_part', 'b'], ['laureato', 'adjective', 'b'], ['laureato', 'noun', 'b'], ['lava', 'noun', 'c'], ['lavabo', 'noun', 'c'], ['lavagna', 'noun', 'c'], ['lavagna', 'adjective', 'c'], ['lavanda', 'noun', 'c'], ['lavanderia', 'noun', 'c'], ['lavandino', 'noun', 'c'], ['lavapiatti', 'noun', 'c'], ['lavare', 'verb', 'a'], ['lavastoviglie', 'noun', 'c'], ['lavatrice', 'noun', 'b'], ['lavello', 'noun', 'c'], ['lavorare', 'verb', 'a'], ['lavorativo', 'adjective', 'b'], ['lavoratore', 'adjective', 'a'], ['lavoratore', 'noun', 'a'], ['lavorazione', 'noun', 'b'], ['lavoro', 'noun', 'a'], ['laziale', 'adjective', 'c'], ['laziale', 'noun', 'c'], ['le', 'determiner', 'a'], ['le', 'pronoun', 'a'], ['le', 'pronoun', 'a'], ['leader', 'noun', 'b'], ['lealtà', 'noun', 'c'], ['lebbra', 'noun', 'c'], ['leccare', 'verb', 'b'], ['leccio', 'noun', 'c'], ['lecito', 'adjective', 'b'], ['lecito', 'noun', 'b'], ['lega', 'noun', 'b'], ['legale', 'adjective', 'a'], ['legale', 'noun', 'a'], ['legame', 'noun', 'b'], ['legare', 'verb', 'a'], ['legato', 'past_part', 'a'], ['legato', 'adjective', 'a'], ['legato', 'noun', 'a'], ['legge', 'noun', 'a'], ['leggenda', 'noun', 'b'], ['leggere', 'verb', 'a'], ['leggermente', 'adverb', 'b'], ['leggero', 'adjective', 'a'], ['leggero', 'adverb', 'a'], ['leggero', 'noun', 'a'], ['legislativo', 'adjective', 'b'], ['legittimo', 'adjective', 'b'], ['legna', 'noun', 'c'], ['legno', 'noun', 'a'], ['legume', 'noun', 'c'], ['lei', 'pronoun', 'a'], ['lentamente', 'adverb', 'a'], ['lente', 'noun', 'c'], ['lenticchia', 'noun', 'c'], ['lentiggine', 'noun', 'c'], ['lento', 'adjective', 'a'], ['lento', 'noun', 'a'], ['lento', 'adverb', 'a'], ['lenza', 'noun', 'c'], ['lenzuolo', 'noun', 'b'], ['leone', 'noun', 'b'], ['leonessa', 'noun', 'c'], ['leopardo', 'noun', 'c'], ['lepre', 'noun', 'c'], ['lesione', 'noun', 'b'], ['lessare', 'verb', 'c'], ['lessema', 'noun', 'b'], ['lettera', 'noun', 'a'], ['letterale', 'adjective', 'c'], ['letteralmente', 'adverb', 'b'], ['letterario', 'adjective', 'b'], ['letteratura', 'noun', 'a'], ['letto', 'noun', 'a'], ['lettone', 'noun', 'c'], ['lettore', 'noun', 'a'], ['lettura', 'noun', 'a'], ['leva', 'noun', 'b'], ['levare', 'verb', 'a'], ['levare', 'noun', 'a'], ['lezione', 'noun', 'a'], ['lì', 'adverb', 'a'], ['li', 'pronoun', 'a'], ['libanese', 'adjective', 'b'], ['libanese', 'noun', 'b'], ['liberale', 'adjective', 'b'], ['liberale', 'noun', 'b'], ['liberamente', 'adverb', 'b'], ['liberare', 'verb', 'a'], ['liberazione', 'noun', 'b'], ['libero', 'adjective', 'a'], ['libero', 'noun', 'a'], ['libertà', 'noun', 'a'], ['libico', 'adjective', 'c'], ['libico', 'noun', 'c'], ['libraio', 'noun', 'c'], ['libreria', 'noun', 'b'], ['libretto', 'noun', 'b'], ['libro', 'noun', 'a'], ['licenza', 'noun', 'b'], ['licenziamento', 'noun', 'c'], ['licenziare', 'verb', 'b'], ['liceo', 'noun', 'b'], ['lido', 'noun', 'c'], ['lieto', 'adjective', 'b'], ['lieve', 'adjective', 'b'], ['lievito', 'noun', 'c'], ['ligure', 'adjective', 'c'], ['ligure', 'noun', 'c'], ['lima', 'noun', 'c'], ['limare', 'verb', 'c'], ['limitare', 'verb', 'a'], ['limitato', 'past_part', 'b'], ['limitato', 'adjective', 'b'], ['limite', 'noun', 'a'], ['limite', 'adjective', 'a'], ['limonata', 'noun', 'c'], ['limone', 'noun', 'b'], ['limone', 'adjective', 'b'], ['linea', 'noun', 'a'], ['lineare', 'adjective', 'b'], ['lineare', 'noun', 'b'], ['linfa', 'noun', 'b'], ['lingerie', 'noun', 'c'], ['lingua', 'noun', 'a'], ['linguaggio', 'noun', 'a'], ['linguistica', 'noun', 'b'], ['linguistico', 'adjective', 'b'], ['linguistico', 'noun', 'b'], ['link', 'noun', 'b'], ['liquido', 'adjective', 'a'], ['liquido', 'noun', 'a'], ['liquore', 'noun', 'c'], ['lira', 'noun', 'a'], ['lirico', 'adjective', 'b'], ['lisbonese', 'adjective', 'c'], ['lisbonese', 'noun', 'c'], ['liscio', 'adjective', 'b'], ['liscio', 'noun', 'b'], ['lista', 'noun', 'a'], ['lite', 'noun', 'b'], ['litigare', 'verb', 'a'], ['litigio', 'noun', 'b'], ['litro', 'noun', 'b'], ['lituano', 'adjective', 'c'], ['lituano', 'noun', 'c'], ['live', 'adjective', 'b'], ['livello', 'noun', 'a'], ['lo', 'determiner', 'a'], ['lo', 'pronoun', 'a'], ['locale', 'adjective', 'a'], ['locale', 'noun', 'a'], ['locale', 'noun', 'a'], ['località', 'noun', 'b'], ['locanda', 'noun', 'c'], ['locazione', 'noun', 'b'], ['locomotiva', 'noun', 'c'], ['logica', 'noun', 'b'], ['logico', 'adjective', 'b'], ['logico', 'noun', 'b'], ['logoro', 'past_part', 'c'], ['logoro', 'adjective', 'c'], ['lombardo', 'adjective', 'b'], ['lombardo', 'noun', 'b'], ['londinese', 'adjective', 'c'], ['londinese', 'noun', 'c'], ['lontananza', 'noun', 'b'], ['lontano', 'adjective', 'a'], ['lontano', 'adverb', 'a'], ['lontano', 'noun', 'a'], ['lonza', 'noun', 'c'], ['look', 'noun', 'b'], ['loro', 'pronoun', 'a'], ['loro', 'adjective', 'a'], ['lotta', 'noun', 'a'], ['lottare', 'verb', 'b'], ['lozione', 'noun', 'c'], ['lucano', 'adjective', 'c'], ['lucano', 'noun', 'c'], ['luccicare', 'verb', 'c'], ['lucciola', 'noun', 'c'], ['luce', 'noun', 'a'], ['lucente', 'pres_part', 'c'], ['lucente', 'adjective', 'c'], ['lucente', 'noun', 'c'], ['lucertola', 'noun', 'c'], ['lucidare', 'verb', 'c'], ['lucido', 'adjective', 'b'], ['lucido', 'noun', 'b'], ['luglio', 'noun', 'a'], ['lui', 'pronoun', 'a'], ['lumaca', 'noun', 'c'], ['luminoso', 'adjective', 'b'], ['luna', 'noun', 'a'], ['lunedì', 'noun', 'a'], ['lunghezza', 'noun', 'b'], ['lungo', 'adjective', 'a'], ['lungo', 'preposition', 'a'], ['lungo', 'noun', 'a'], ['luogo', 'noun', 'a'], ['lupo', 'noun', 'a'], ['lussemburghese', 'adjective', 'c'], ['lussemburghese', 'noun', 'c'], ['lusso', 'noun', 'b'], ['lutto', 'noun', 'b'], ['ma', 'conjunction', 'a'], ['ma', 'noun', 'a'], ['maccherone', 'noun', 'c'], ['macchia', 'noun', 'a'], ['macchina', 'noun', 'a'], ['macchinista', 'noun', 'c'], ['macedone', 'adjective', 'c'], ['macedone', 'noun', 'c'], ['macedonia', 'noun', 'c'], ['maceria', 'noun', 'b'], ['macinare', 'verb', 'c'], ['madonna', 'noun', 'b'], ['madonna', 'exclamation', 'b'], ['madre', 'noun', 'a'], ['madrileno', 'adjective', 'c'], ['madrileno', 'noun', 'c'], ['madrileno', 'adjective', 'c'], ['madrileno', 'noun', 'c'], ['madrina', 'noun', 'c'], ['maestra', 'noun', 'b'], ['maestranza', 'noun', 'c'], ['maestro', 'noun', 'a'], ['maestro', 'adjective', 'a'], ['mafia', 'noun', 'b'], ['mafioso', 'adjective', 'b'], ['mafioso', 'noun', 'b'], ['magari', 'exclamation', 'a'], ['magari', 'conjunction', 'a'], ['magari', 'adverb', 'a'], ['magazzino', 'noun', 'b'], ['maggio', 'noun', 'a'], ['maggioranza', 'noun', 'a'], ['maggiorenne', 'adjective', 'c'], ['maggiorenne', 'noun', 'c'], ['maggiormente', 'adverb', 'b'], ['magia', 'noun', 'b'], ['magico', 'adjective', 'a'], ['magistrato', 'noun', 'b'], ['magistratura', 'noun', 'b'], ['maglia', 'noun', 'a'], ['maglietta', 'noun', 'b'], ['magnetico', 'adjective', 'b'], ['magnifico', 'adjective', 'b'], ['mago', 'noun', 'b'], ['mago', 'adjective', 'b'], ['magro', 'adjective', 'b'], ['magro', 'noun', 'b'], ['mah', 'exclamation', 'b'], ['mai', 'adverb', 'a'], ['maiale', 'noun', 'b'], ['maionese', 'noun', 'c'], ['mais', 'noun', 'c'], ['maiuscola', 'noun', 'c'], ['malato', 'adjective', 'a'], ['malato', 'noun', 'a'], ['malattia', 'noun', 'a'], ['malaugurio', 'noun', 'c'], ['malavita', 'noun', 'c'], ['male', 'adverb', 'a'], ['male', 'exclamation', 'a'], ['male', 'noun', 'a'], ['maledetto', 'past_part', 'b'], ['maledetto', 'adjective', 'b'], ['maledetto', 'noun', 'b'], ['maledizione', 'noun', 'b'], ['maledizione', 'exclamation', 'b'], ['maleducato', 'adjective', 'c'], ['maleducato', 'noun', 'c'], ['maleducazione', 'noun', 'c'], ['malgrado', 'noun', 'b'], ['malgrado', 'adverb', 'b'], ['malgrado', 'conjunction', 'b'], ['malgrado', 'preposition', 'b'], ['malinconia', 'noun', 'b'], ['malinteso', 'adjective', 'c'], ['malinteso', 'noun', 'c'], ['malizia', 'noun', 'c'], ['maltempo', 'noun', 'c'], ['maltese', 'adjective', 'c'], ['maltese', 'noun', 'c'], ['maltrattamento', 'noun', 'c'], ['maltrattare', 'verb', 'c'], ['malva', 'noun', 'c'], ['malvagio', 'adjective', 'b'], ['malvagio', 'noun', 'b'], ['mamma', 'noun', 'a'], ['mammella', 'noun', 'c'], ['mammifero', 'noun', 'c'], ['manager', 'noun', 'b'], ['mancanza', 'noun', 'a'], ['mancare', 'verb', 'a'], ['mancato', 'past_part', 'b'], ['mancato', 'adjective', 'b'], ['mancino', 'adjective', 'c'], ['mancino', 'noun', 'c'], ['manco', 'adjective', 'b'], ['manco', 'adverb', 'b'], ['mandare', 'verb', 'a'], ['mandarino', 'noun', 'c'], ['mandarino', 'adjective', 'c'], ['mandato', 'past_part', 'b'], ['mandato', 'adjective', 'b'], ['mandato', 'noun', 'b'], ['mandorla', 'noun', 'c'], ['mandorlo', 'noun', 'c'], ['manganello', 'noun', 'c'], ['mangiare', 'verb', 'a'], ['mangime', 'noun', 'c'], ['mania', 'noun', 'b'], ['maniaco', 'adjective', 'c'], ['maniaco', 'noun', 'c'], ['manica', 'noun', 'b'], ['manico', 'noun', 'b'], ['maniera', 'noun', 'a'], ['manifestare', 'verb', 'a'], ['manifestazione', 'noun', 'a'], ['manifesto', 'noun', 'b'], ['mano', 'noun', 'a'], ['manodopera', 'noun', 'c'], ['manoscritto', 'adjective', 'b'], ['manoscritto', 'noun', 'b'], ['manovale', 'noun', 'c'], ['manovra', 'noun', 'b'], ['mantello', 'noun', 'b'], ['mantenere', 'verb', 'a'], ['manuale', 'adjective', 'b'], ['manuale', 'noun', 'b'], ['manuale', 'noun', 'b'], ['manutenzione', 'noun', 'b'], ['manzo', 'noun', 'c'], ['mappa', 'noun', 'b'], ['marca', 'noun', 'b'], ['marcare', 'verb', 'b'], ['marchigiano', 'adjective', 'c'], ['marchigiano', 'noun', 'c'], ['marchio', 'noun', 'b'], ['marcia', 'noun', 'b'], ['marciapiede', 'noun', 'b'], ['marcio', 'adjective', 'b'], ['marcio', 'noun', 'b'], ['marcire', 'verb', 'c'], ['marco', 'noun', 'a'], ['mare', 'noun', 'a'], ['marea', 'noun', 'b'], ['maresciallo', 'noun', 'b'], ['margherita', 'noun', 'c'], ['marginale', 'adjective', 'b'], ['marginale', 'noun', 'b'], ['margine', 'noun', 'b'], ['marinaio', 'noun', 'b'], ['marino', 'adjective', 'b'], ['marino', 'noun', 'b'], ['marionetta', 'noun', 'c'], ['marito', 'noun', 'a'], ['marketing', 'noun', 'b'], ['marmellata', 'noun', 'c'], ['marmo', 'noun', 'b'], ['marocchino', 'adjective', 'c'], ['marocchino', 'noun', 'c'], ['marrone', 'noun', 'b'], ['marrone', 'adjective', 'b'], ['martedì', 'noun', 'b'], ['marzo', 'noun', 'a'], ['mascarpone', 'noun', 'c'], ['maschera', 'noun', 'b'], ['mascherare', 'verb', 'b'], ['mascherato', 'past_part', 'c'], ['mascherato', 'adjective', 'c'], ['maschile', 'adjective', 'a'], ['maschile', 'noun', 'a'], ['maschio', 'noun', 'a'], ['maschio', 'adjective', 'a'], ['massa', 'noun', 'a'], ['massa', 'adverb', 'a'], ['massacrare', 'verb', 'b'], ['massacro', 'noun', 'c'], ['massaggio', 'noun', 'c'], ['massaia', 'noun', 'c'], ['massiccio', 'adjective', 'b'], ['massiccio', 'noun', 'b'], ['massimo', 'adjective', 'a'], ['massimo', 'noun', 'a'], ['massimo', 'adverb', 'a'], ['master', 'noun', 'b'], ['masticare', 'verb', 'b'], ['masturbare', 'verb', 'b'], ['matematica', 'noun', 'b'], ['matematico', 'adjective', 'b'], ['matematico', 'noun', 'b'], ['materasso', 'noun', 'b'], ['materia', 'noun', 'a'], ['materiale', 'adjective', 'a'], ['materiale', 'noun', 'a'], ['maternità', 'noun', 'b'], ['materno', 'adjective', 'b'], ['matita', 'noun', 'b'], ['matricola', 'noun', 'b'], ['matrimoniale', 'adjective', 'b'], ['matrimoniale', 'noun', 'b'], ['matrimonio', 'noun', 'a'], ['mattina', 'noun', 'a'], ['mattinata', 'noun', 'b'], ['mattino', 'noun', 'a'], ['matto', 'adjective', 'a'], ['matto', 'noun', 'a'], ['mattone', 'noun', 'b'], ['mattone', 'adjective', 'b'], ['mattone', 'noun', 'b'], ['maturare', 'verb', 'b'], ['maturità', 'noun', 'b'], ['maturo', 'adjective', 'b'], ['mazzo', 'noun', 'b'], ['me', 'pronoun', 'a'], ['meccanico', 'adjective', 'a'], ['meccanico', 'noun', 'a'], ['meccanismo', 'noun', 'a'], ['medaglia', 'noun', 'b'], ['medesimo', 'adjective', 'b'], ['medesimo', 'pronoun', 'b'], ['media', 'noun', 'a'], ['media', 'noun', 'b'], ['mediante', 'preposition', 'b'], ['medicare', 'verb', 'c'], ['medicina', 'noun', 'a'], ['medico', 'noun', 'a'], ['medico', 'adjective', 'b'], ['medievale', 'adjective', 'b'], ['medio', 'adjective', 'a'], ['medio', 'noun', 'a'], ['medioevo', 'noun', 'b'], ['meditare', 'verb', 'b'], ['mediterraneo', 'adjective', 'b'], ['mediterraneo', 'noun', 'b'], ['meglio', 'adverb', 'a'], ['meglio', 'adjective', 'a'], ['meglio', 'noun', 'a'], ['mela', 'noun', 'b'], ['melagrana', 'noun', 'c'], ['melanzana', 'noun', 'c'], ['melo', 'noun', 'c'], ['melograno', 'noun', 'c'], ['melone', 'noun', 'c'], ['membrana', 'noun', 'b'], ['membro', 'noun', 'a'], ['memoria', 'noun', 'a'], ['menare', 'verb', 'b'], ['mendicante', 'pres_part', 'c'], ['mendicante', 'adjective', 'c'], ['mendicante', 'noun', 'c'], ['meno', 'adverb', 'a'], ['meno', 'adjective', 'a'], ['meno', 'preposition', 'a'], ['meno', 'noun', 'a'], ['mensa', 'noun', 'b'], ['mensile', 'adjective', 'b'], ['mensile', 'noun', 'b'], ['mensola', 'noun', 'c'], ['menta', 'noun', 'c'], ['mentale', 'adjective', 'a'], ['mentalità', 'noun', 'b'], ['mente', 'noun', 'a'], ['mentire', 'verb', 'a'], ['mento', 'noun', 'b'], ['mentre', 'conjunction', 'a'], ['menu', 'noun', 'b'], ['menzogna', 'noun', 'b'], ['meraviglia', 'noun', 'b'], ['meravigliare', 'verb', 'b'], ['meraviglioso', 'adjective', 'a'], ['meraviglioso', 'noun', 'a'], ['mercante', 'noun', 'b'], ['mercato', 'noun', 'a'], ['merce', 'noun', 'b'], ['merceria', 'noun', 'c'], ['mercoledì', 'noun', 'b'], ['merda', 'noun', 'a'], ['merenda', 'noun', 'c'], ['merendina', 'noun', 'c'], ['meridiano', 'adjective', 'c'], ['meridiano', 'noun', 'c'], ['meridionale', 'adjective', 'a'], ['meridionale', 'noun', 'a'], ['meridione', 'noun', 'c'], ['meritare', 'verb', 'a'], ['merito', 'noun', 'a'], ['merlo', 'noun', 'c'], ['merluzzo', 'noun', 'c'], ['mero', 'adjective', 'b'], ['mescolare', 'verb', 'b'], ['mese', 'noun', 'a'], ['messa', 'noun', 'b'], ['messa', 'noun', 'b'], ['messaggio', 'noun', 'a'], ['messe', 'noun', 'c'], ['messicano', 'adjective', 'c'], ['messicano', 'noun', 'c'], ['mestiere', 'noun', 'a'], ['mestolo', 'noun', 'c'], ['mestruazione', 'noun', 'c'], ['metà', 'noun', 'a'], ['meta', 'noun', 'b'], ['metafora', 'noun', 'b'], ['metallico', 'adjective', 'b'], ['metallo', 'noun', 'b'], ['metalmeccanico', 'adjective', 'c'], ['metalmeccanico', 'noun', 'c'], ['meteo', 'adjective', 'b'], ['meteo', 'noun', 'b'], ['metodo', 'noun', 'a'], ['metro', 'noun', 'a'], ['metropolitano', 'adjective', 'b'], ['metropolitano', 'noun', 'b'], ['mettere', 'verb', 'a'], ['mezzanotte', 'noun', 'b'], ['mezzo', 'adjective', 'a'], ['mezzo', 'noun', 'a'], ['mezzo', 'adverb', 'a'], ['mezzogiorno', 'noun', 'b'], ['mi', 'pronoun', 'a'], ['miagolare', 'verb', 'c'], ['mica', 'noun', 'a'], ['mica', 'adverb', 'a'], ['micio', 'noun', 'c'], ['microfono', 'noun', 'b'], ['miele', 'noun', 'b'], ['miele', 'adjective', 'b'], ['mietere', 'verb', 'c'], ['migliaio', 'noun', 'c'], ['migliaio', 'noun', 'a'], ['miglioramento', 'noun', 'b'], ['migliorare', 'verb', 'a'], ['migliore', 'adjective', 'a'], ['migliore', 'noun', 'a'], ['migliore', 'adverb', 'a'], ['mignolo', 'noun', 'c'], ['mila', 'adjective', 'a'], ['milanese', 'adjective', 'b'], ['milanese', 'noun', 'b'], ['miliardo', 'noun', 'a'], ['milione', 'noun', 'a'], ['militare', 'adjective', 'a'], ['militare', 'noun', 'a'], ['mille', 'adjective', 'a'], ['mille', 'noun', 'a'], ['millennio', 'noun', 'b'], ['millimetro', 'noun', 'b'], ['mimosa', 'noun', 'c'], ['minaccia', 'noun', 'b'], ['minacciare', 'verb', 'a'], ['minchia', 'noun', 'b'], ['minestra', 'noun', 'c'], ['minestrone', 'noun', 'c'], ['mini', 'adjective', 'c'], ['miniera', 'noun', 'b'], ['minigonna', 'noun', 'c'], ['minimo', 'adjective', 'a'], ['minimo', 'noun', 'a'], ['ministero', 'noun', 'a'], ['ministro', 'noun', 'a'], ['minoranza', 'noun', 'b'], ['minore', 'adjective', 'a'], ['minore', 'noun', 'a'], ['minuscolo', 'adjective', 'b'], ['minuto', 'noun', 'a'], ['mio', 'adjective', 'a'], ['mio', 'pronoun', 'a'], ['miracolo', 'noun', 'a'], ['mirare', 'verb', 'b'], ['mischiare', 'verb', 'b'], ['miscuglio', 'noun', 'c'], ['miseria', 'noun', 'b'], ['misero', 'adjective', 'b'], ['missile', 'adjective', 'c'], ['missile', 'noun', 'c'], ['missione', 'noun', 'a'], ['mister', 'noun', 'c'], ['misterioso', 'adjective', 'b'], ['mistero', 'noun', 'a'], ['misto', 'adjective', 'b'], ['misto', 'noun', 'b'], ['misura', 'noun', 'a'], ['misurare', 'verb', 'b'], ['misurazione', 'noun', 'c'], ['mitico', 'adjective', 'b'], ['mito', 'noun', 'b'], ['mitragliatrice', 'noun', 'c'], ['mobile', 'adjective', 'a'], ['mobile', 'noun', 'a'], ['mobilio', 'noun', 'c'], ['mocassino', 'noun', 'c'], ['moda', 'noun', 'a'], ['modalità', 'noun', 'b'], ['modella', 'noun', 'b'], ['modellare', 'verb', 'c'], ['modello', 'noun', 'a'], ['moderato', 'past_part', 'b'], ['moderato', 'adjective', 'b'], ['moderato', 'adverb', 'b'], ['moderato', 'noun', 'b'], ['moderatore', 'adjective', 'b'], ['moderatore', 'noun', 'b'], ['modernità', 'noun', 'b'], ['moderno', 'adjective', 'a'], ['moderno', 'noun', 'a'], ['modestia', 'noun', 'c'], ['modesto', 'adjective', 'b'], ['modifica', 'noun', 'b'], ['modificare', 'verb', 'a'], ['modificazione', 'noun', 'b'], ['modo', 'noun', 'a'], ['modulo', 'noun', 'b'], ['moglie', 'noun', 'a'], ['molecola', 'noun', 'b'], ['molisano', 'adjective', 'c'], ['molisano', 'noun', 'c'], ['molla', 'noun', 'c'], ['mollare', 'verb', 'b'], ['mollusco', 'noun', 'c'], ['molo', 'noun', 'c'], ['moltiplicare', 'verb', 'b'], ['molto', 'adjective', 'a'], ['molto', 'pronoun', 'a'], ['molto', 'adverb', 'a'], ['molto', 'noun', 'a'], ['momento', 'noun', 'a'], ['monaca', 'noun', 'c'], ['monaco', 'noun', 'c'], ['monarchica', 'noun', 'c'], ['mondiale', 'adjective', 'a'], ['mondiale', 'noun', 'a'], ['mondo', 'noun', 'a'], ['monello', 'noun', 'c'], ['moneta', 'noun', 'a'], ['monetario', 'adjective', 'b'], ['monitor', 'noun', 'b'], ['monologo', 'noun', 'b'], ['montaggio', 'noun', 'b'], ['montagna', 'noun', 'a'], ['montare', 'verb', 'b'], ['monte', 'noun', 'a'], ['montenegrino', 'adjective', 'c'], ['montenegrino', 'noun', 'c'], ['monumento', 'noun', 'b'], ['mora', 'noun', 'b'], ['morale', 'adjective', 'a'], ['morale', 'noun', 'a'], ['morbido', 'adjective', 'b'], ['morbido', 'noun', 'b'], ['mordere', 'verb', 'b'], ['morire', 'verb', 'a'], ['moro', 'adjective', 'b'], ['moro', 'noun', 'b'], ['morsicare', 'verb', 'c'], ['morso', 'noun', 'c'], ['mortadella', 'noun', 'c'], ['mortale', 'adjective', 'b'], ['mortale', 'noun', 'b'], ['morte', 'noun', 'a'], ['morto', 'past_part', 'a'], ['morto', 'adjective', 'a'], ['morto', 'noun', 'a'], ['mosca', 'noun', 'b'], ['moscovita', 'adjective', 'c'], ['moscovita', 'noun', 'c'], ['mossa', 'noun', 'b'], ['mostarda', 'noun', 'c'], ['mostra', 'noun', 'a'], ['mostrare', 'verb', 'a'], ['mostro', 'noun', 'b'], ['motel', 'noun', 'c'], ['motivare', 'verb', 'b'], ['motivazione', 'noun', 'b'], ['motivo', 'noun', 'a'], ['moto', 'noun', 'a'], ['moto', 'noun', 'b'], ['motociclismo', 'noun', 'c'], ['motociclista', 'adjective', 'c'], ['motociclista', 'noun', 'c'], ['motore', 'adjective', 'a'], ['motore', 'noun', 'a'], ['motorino', 'noun', 'b'], ['motoscafo', 'noun', 'c'], ['mousse', 'noun', 'c'], ['movimento', 'noun', 'a'], ['mozzarella', 'noun', 'c'], ['mucca', 'noun', 'b'], ['mucchio', 'noun', 'b'], ['muggire', 'verb', 'c'], ['muggito', 'past_part', 'c'], ['muggito', 'noun', 'c'], ['mugnaio', 'noun', 'c'], ['mugolare', 'verb', 'c'], ['mulino', 'noun', 'c'], ['multa', 'noun', 'b'], ['multare', 'verb', 'c'], ['multinazionale', 'adjective', 'b'], ['multinazionale', 'noun', 'b'], ['multiplo', 'adjective', 'b'], ['multiplo', 'noun', 'b'], ['multipresa', 'noun', 'c'], ['mummia', 'noun', 'c'], ['mungere', 'verb', 'c'], ['municipio', 'noun', 'c'], ['muovere', 'verb', 'a'], ['murare', 'verb', 'c'], ['muratore', 'noun', 'c'], ['muro', 'noun', 'a'], ['muschio', 'noun', 'c'], ['muschio', 'adjective', 'c'], ['muscolare', 'adjective', 'b'], ['muscolare', 'noun', 'b'], ['muscolo', 'noun', 'a'], ['museo', 'noun', 'a'], ['musica', 'noun', 'a'], ['musicale', 'adjective', 'a'], ['musicista', 'noun', 'b'], ['muso', 'noun', 'b'], ['musulmano', 'adjective', 'b'], ['musulmano', 'noun', 'b'], ['muta', 'noun', 'c'], ['mutamento', 'noun', 'b'], ['mutanda', 'noun', 'b'], ['mutandina', 'noun', 'c'], ['mutare', 'verb', 'b'], ['mutazione', 'noun', 'b'], ['mutilato', 'past_part', 'c'], ['mutilato', 'adjective', 'c'], ['mutilato', 'noun', 'c'], ['muto', 'adjective', 'b'], ['muto', 'noun', 'b'], ['mutuo', 'noun', 'b'], ['nanna', 'noun', 'c'], ['nano', 'adjective', 'b'], ['nano', 'noun', 'b'], ['napoletano', 'adjective', 'b'], ['napoletano', 'noun', 'b'], ['narrare', 'verb', 'b'], ['narrativo', 'adjective', 'b'], ['narratore', 'noun', 'b'], ['narrazione', 'noun', 'b'], ['nasale', 'adjective', 'b'], ['nasale', 'noun', 'b'], ['nascere', 'verb', 'a'], ['nascere', 'noun', 'a'], ['nascita', 'noun', 'a'], ['nascondere', 'verb', 'a'], ['nascondiglio', 'noun', 'c'], ['nascondino', 'noun', 'c'], ['nascosto', 'past_part', 'a'], ['nascosto', 'adjective', 'a'], ['nascosto', 'noun', 'a'], ['naso', 'noun', 'a'], ['nastro', 'noun', 'a'], ['natale', 'adjective', 'a'], ['natale', 'noun', 'a'], ['natalizio', 'adjective', 'b'], ['natalizio', 'noun', 'b'], ['nato', 'past_part', 'b'], ['nato', 'adjective', 'b'], ['nato', 'noun', 'b'], ['natura', 'noun', 'a'], ['naturale', 'adjective', 'a'], ['naturale', 'noun', 'a'], ['naturalmente', 'adverb', 'a'], ['naufragio', 'noun', 'c'], ['navale', 'adjective', 'c'], ['nave', 'noun', 'a'], ['navicella', 'noun', 'c'], ['navigare', 'verb', 'b'], ['navigazione', 'noun', 'b'], ['nazionale', 'adjective', 'a'], ['nazionale', 'noun', 'a'], ['nazionalità', 'noun', 'c'], ['nazione', 'noun', 'a'], ['nazista', 'adjective', 'b'], ['nazista', 'noun', 'b'], ['ndrangheta', 'noun', 'c'], ['né', 'conjunction', 'a'], ['ne', 'pronoun', 'a'], ['ne', 'adverb', 'a'], ['neanche', 'adverb', 'a'], ['nebbia', 'noun', 'b'], ['necessariamente', 'adverb', 'b'], ['necessario', 'adjective', 'a'], ['necessario', 'noun', 'a'], ['necessità', 'noun', 'a'], ['necessitare', 'verb', 'b'], ['negare', 'verb', 'a'], ['negativo', 'adjective', 'a'], ['negativo', 'noun', 'a'], ['negativo', 'adverb', 'a'], ['negazione', 'noun', 'c'], ['negoziante', 'pres_part', 'c'], ['negoziante', 'noun', 'c'], ['negozio', 'noun', 'a'], ['negro', 'adjective', 'b'], ['negro', 'noun', 'b'], ['nemico', 'adjective', 'a'], ['nemico', 'noun', 'a'], ['nemmeno', 'adverb', 'a'], ['neo', 'noun', 'c'], ['neonato', 'noun', 'b'], ['neonato', 'adjective', 'b'], ['neppure', 'adverb', 'a'], ['nero', 'adjective', 'a'], ['nero', 'noun', 'a'], ['nervo', 'noun', 'b'], ['nervosismo', 'noun', 'c'], ['nervoso', 'adjective', 'a'], ['nervoso', 'noun', 'a'], ['nessuno', 'adjective', 'a'], ['nessuno', 'pronoun', 'a'], ['nettare', 'noun', 'c'], ['netto', 'adjective', 'b'], ['netto', 'noun', 'b'], ['netto', 'adverb', 'b'], ['network', 'noun', 'b'], ['neutro', 'adjective', 'b'], ['neutro', 'noun', 'b'], ['neve', 'noun', 'a'], ['nevicare', 'verb', 'c'], ['news', 'noun', 'b'], ['newyorkese', 'adjective', 'c'], ['newyorkese', 'noun', 'c'], ['nido', 'noun', 'b'], ['niente', 'pronoun', 'a'], ['niente', 'adjective', 'a'], ['niente', 'adverb', 'a'], ['nipote', 'noun', 'a'], ['no', 'adverb', 'a'], ['no', 'noun', 'a'], ['no', 'adjective', 'a'], ['nobile', 'adjective', 'b'], ['nobile', 'noun', 'b'], ['nocciola', 'noun', 'c'], ['nocciola', 'adjective', 'c'], ['nocciolina', 'noun', 'c'], ['nocivo', 'adjective', 'c'], ['nodo', 'noun', 'b'], ['noi', 'pronoun', 'a'], ['noia', 'noun', 'b'], ['noioso', 'adjective', 'b'], ['noleggiare', 'verb', 'c'], ['nome', 'noun', 'a'], ['nomina', 'noun', 'b'], ['nominare', 'verb', 'a'], ['non', 'adverb', 'a'], ['nonché', 'conjunction', 'b'], ['nonna', 'noun', 'a'], ['nonno', 'noun', 'a'], ['nono', 'adjective', 'b'], ['nono', 'noun', 'b'], ['nonostante', 'preposition', 'a'], ['nonostante', 'conjunction', 'a'], ['nord', 'noun', 'a'], ['nord', 'adjective', 'a'], ['nordamericano', 'adjective', 'c'], ['nordamericano', 'noun', 'c'], ['norma', 'noun', 'a'], ['normale', 'adjective', 'a'], ['normale', 'noun', 'a'], ['normalità', 'noun', 'b'], ['normalmente', 'adverb', 'b'], ['normativa', 'noun', 'b'], ['norvegese', 'adjective', 'c'], ['norvegese', 'noun', 'c'], ['nostalgia', 'noun', 'b'], ['nostro', 'adjective', 'a'], ['nostro', 'pronoun', 'a'], ['nota', 'noun', 'a'], ['notaio', 'noun', 'b'], ['notare', 'verb', 'a'], ['notevole', 'adjective', 'b'], ['notizia', 'noun', 'a'], ['noto', 'adjective', 'a'], ['noto', 'noun', 'a'], ['notte', 'noun', 'a'], ['notturno', 'adjective', 'b'], ['notturno', 'noun', 'b'], ['novanta', 'adjective', 'b'], ['novanta', 'noun', 'b'], ['nove', 'adjective', 'a'], ['nove', 'noun', 'a'], ['novella', 'noun', 'c'], ['novembre', 'noun', 'a'], ['novità', 'noun', 'a'], ['nozione', 'noun', 'b'], ['nozze', 'noun', 'b'], ['nube', 'noun', 'b'], ['nucleare', 'adjective', 'a'], ['nucleare', 'noun', 'a'], ['nucleo', 'noun', 'b'], ['nudo', 'adjective', 'a'], ['nudo', 'noun', 'a'], ['nulla', 'pronoun', 'a'], ['nulla', 'adverb', 'a'], ['numerare', 'verb', 'b'], ['numerazione', 'noun', 'c'], ['numero', 'noun', 'a'], ['numeroso', 'adjective', 'a'], ['nuora', 'noun', 'c'], ['nuotare', 'verb', 'b'], ['nuoto', 'noun', 'b'], ['nuovamente', 'adverb', 'b'], ['nuovo', 'adjective', 'a'], ['nuovo', 'noun', 'a'], ['nutrire', 'verb', 'b'], ['nuvola', 'noun', 'b'], ['nuvoloso', 'adjective', 'c'], ['nylon', 'noun', 'c'], ['o', 'noun', 'c'], ['o', 'conjunction', 'a'], ['obbedire', 'verb', 'b'], ['obbiettivo', 'adjective', 'c'], ['obbiettivo', 'noun', 'c'], ['obbligare', 'verb', 'a'], ['obbligatorio', 'adjective', 'b'], ['obbligazione', 'noun', 'b'], ['obbligo', 'noun', 'b'], ['obiettivo', 'adjective', 'a'], ['obiettivo', 'noun', 'a'], ['obiezione', 'noun', 'b'], ['oblò', 'noun', 'c'], ['occasione', 'noun', 'a'], ['occhiaia', 'noun', 'c'], ['occhiale', 'noun', 'a'], ['occhiale', 'adjective', 'a'], ['occhiata', 'noun', 'b'], ['occhiello', 'noun', 'c'], ['occhio', 'noun', 'a'], ['occidentale', 'adjective', 'a'], ['occidentale', 'noun', 'a'], ['occidente', 'noun', 'b'], ['occidente', 'adjective', 'b'], ['occorrere', 'verb', 'a'], ['occupare', 'verb', 'a'], ['occupato', 'past_part', 'c'], ['occupato', 'adjective', 'c'], ['occupato', 'noun', 'c'], ['occupazione', 'noun', 'b'], ['oceano', 'noun', 'b'], ['oculista', 'noun', 'c'], ['oddio', 'exclamation', 'b'], ['odiare', 'verb', 'a'], ['odio', 'noun', 'b'], ['odorare', 'verb', 'c'], ['odore', 'noun', 'a'], ['offendere', 'verb', 'b'], ['offerta', 'noun', 'a'], ['offesa', 'noun', 'b'], ['offeso', 'past_part', 'c'], ['offeso', 'adjective', 'c'], ['offeso', 'noun', 'c'], ['officina', 'noun', 'b'], ['offline', 'adjective', 'b'], ['offline', 'noun', 'b'], ['offrire', 'verb', 'a'], ['oggettivo', 'adjective', 'b'], ['oggetto', 'noun', 'a'], ['oggi', 'adverb', 'a'], ['oggi', 'noun', 'a'], ['ogni', 'adjective', 'a'], ['ognuno', 'pronoun', 'a'], ['ognuno', 'adjective', 'a'], ['ok', 'adverb', 'a'], ['ok', 'noun', 'a'], ['ok', 'adjective', 'a'], ['okay', 'adverb', 'a'], ['okay', 'noun', 'a'], ['okay', 'adjective', 'a'], ['olandese', 'adjective', 'b'], ['olandese', 'noun', 'b'], ['oliare', 'verb', 'c'], ['oliera', 'noun', 'c'], ['olimpico', 'adjective', 'b'], ['olio', 'noun', 'a'], ['oliva', 'noun', 'b'], ['oliva', 'adjective', 'b'], ['oltre', 'adverb', 'a'], ['oltre', 'preposition', 'a'], ['oltrepassare', 'verb', 'c'], ['oltretutto', 'adverb', 'b'], ['omaggio', 'noun', 'b'], ['ombelico', 'noun', 'c'], ['ombra', 'noun', 'a'], ['ombrellone', 'noun', 'c'], ['omicidio', 'noun', 'a'], ['omogeneizzato', 'past_part', 'c'], ['omogeneizzato', 'adjective', 'c'], ['omogeneizzato', 'noun', 'c'], ['omonimo', 'adjective', 'b'], ['omonimo', 'noun', 'b'], ['onda', 'noun', 'a'], ['ondata', 'noun', 'b'], ['ondeggiare', 'verb', 'c'], ['onere', 'noun', 'b'], ['onestamente', 'adverb', 'b'], ['onesto', 'adjective', 'b'], ['onesto', 'noun', 'b'], ['onesto', 'adverb', 'b'], ['online', 'adjective', 'b'], ['online', 'noun', 'b'], ['onorare', 'verb', 'b'], ['onore', 'noun', 'a'], ['opera', 'noun', 'a'], ['operaio', 'noun', 'a'], ['operaio', 'adjective', 'a'], ['operare', 'verb', 'a'], ['operativo', 'adjective', 'b'], ['operativo', 'noun', 'b'], ['operatore', 'adjective', 'b'], ['operatore', 'noun', 'b'], ['operazione', 'noun', 'a'], ['opinione', 'noun', 'a'], ['opporre', 'verb', 'a'], ['opportunità', 'noun', 'b'], ['opportuno', 'adjective', 'b'], ['opposizione', 'noun', 'b'], ['opposto', 'past_part', 'a'], ['opposto', 'adjective', 'a'], ['opposto', 'noun', 'a'], ['oppressivo', 'adjective', 'c'], ['oppresso', 'past_part', 'c'], ['oppresso', 'adjective', 'c'], ['oppresso', 'noun', 'c'], ['oppressore', 'adjective', 'c'], ['oppressore', 'noun', 'c'], ['oppure', 'conjunction', 'a'], ['opzione', 'noun', 'b'], ['ora', 'noun', 'a'], ['ora', 'adverb', 'a'], ['orale', 'adjective', 'b'], ['oramai', 'adverb', 'b'], ['orario', 'adjective', 'a'], ['orario', 'noun', 'a'], ['orbita', 'noun', 'b'], ['orchestra', 'noun', 'b'], ['orco', 'noun', 'b'], ['ordinamento', 'noun', 'b'], ['ordinanza', 'noun', 'b'], ['ordinare', 'verb', 'a'], ['ordinario', 'adjective', 'b'], ['ordinario', 'noun', 'b'], ['ordine', 'noun', 'a'], ['orecchino', 'noun', 'c'], ['orecchio', 'noun', 'a'], ['orefice', 'noun', 'c'], ['organico', 'adjective', 'b'], ['organico', 'noun', 'b'], ['organismo', 'noun', 'a'], ['organizzare', 'verb', 'a'], ['organizzato', 'past_part', 'b'], ['organizzato', 'adjective', 'b'], ['organizzato', 'noun', 'b'], ['organizzazione', 'noun', 'a'], ['organo', 'noun', 'a'], ['orgasmo', 'noun', 'b'], ['orgoglio', 'noun', 'b'], ['orgoglioso', 'adjective', 'b'], ['orientale', 'adjective', 'b'], ['orientale', 'noun', 'b'], ['orientamento', 'noun', 'b'], ['orientare', 'verb', 'b'], ['oriente', 'adjective', 'b'], ['oriente', 'noun', 'b'], ['origano', 'noun', 'c'], ['originale', 'adjective', 'a'], ['originale', 'noun', 'a'], ['originario', 'adjective', 'b'], ['origine', 'noun', 'a'], ['orizzontale', 'adjective', 'b'], ['orizzontale', 'noun', 'b'], ['orizzonte', 'noun', 'b'], ['orlo', 'noun', 'b'], ['orma', 'noun', 'c'], ['ormai', 'adverb', 'a'], ['ormone', 'noun', 'b'], ['oro', 'noun', 'a'], ['orologiaio', 'noun', 'c'], ['orologio', 'noun', 'a'], ['oroscopo', 'noun', 'b'], ['orribile', 'adjective', 'b'], ['orrore', 'noun', 'b'], ['orso', 'noun', 'b'], ['ortaggio', 'noun', 'c'], ['ortensia', 'noun', 'c'], ['ortica', 'noun', 'c'], ['orto', 'noun', 'b'], ['ortolano', 'noun', 'c'], ['ortolano', 'adjective', 'c'], ['orzo', 'noun', 'c'], ['osare', 'verb', 'b'], ['osceno', 'adjective', 'c'], ['oscillare', 'verb', 'b'], ['oscurare', 'verb', 'b'], ['oscuro', 'adjective', 'b'], ['oscuro', 'noun', 'b'], ['oscuro', 'adverb', 'b'], ['ospedale', 'noun', 'a'], ['ospitalità', 'noun', 'c'], ['ospitare', 'verb', 'a'], ['ospite', 'adjective', 'a'], ['ospite', 'noun', 'a'], ['ospizio', 'noun', 'c'], ['osservare', 'verb', 'a'], ['osservazione', 'noun', 'b'], ['ossessione', 'noun', 'b'], ['ossia', 'conjunction', 'b'], ['ossigeno', 'noun', 'b'], ['osso', 'noun', 'a'], ['ostacolare', 'verb', 'b'], ['ostacolo', 'noun', 'b'], ['ostaggio', 'noun', 'c'], ['oste', 'noun', 'c'], ['ostile', 'adjective', 'b'], ['ostinato', 'past_part', 'c'], ['ostinato', 'adjective', 'c'], ['ostrica', 'noun', 'c'], ['ottanta', 'adjective', 'b'], ['ottanta', 'noun', 'b'], ['ottavo', 'adjective', 'b'], ['ottavo', 'noun', 'b'], ['ottenere', 'verb', 'a'], ['ottica', 'noun', 'b'], ['ottimo', 'adjective', 'a'], ['ottimo', 'noun', 'a'], ['otto', 'adjective', 'a'], ['otto', 'noun', 'a'], ['ottobre', 'noun', 'a'], ['ottone', 'noun', 'c'], ['ovale', 'adjective', 'c'], ['ovale', 'noun', 'c'], ['ovatta', 'noun', 'c'], ['ove', 'adverb', 'b'], ['ove', 'conjunction', 'b'], ['ovest', 'noun', 'b'], ['ovest', 'adjective', 'b'], ['ovile', 'noun', 'c'], ['ovino', 'adjective', 'c'], ['ovino', 'noun', 'c'], ['ovunque', 'adverb', 'a'], ['ovunque', 'conjunction', 'a'], ['ovvero', 'conjunction', 'a'], ['ovviamente', 'adverb', 'a'], ['ovviare', 'verb', 'b'], ['ovvio', 'adjective', 'b'], ['ozono', 'noun', 'c'], ['pacchetto', 'noun', 'b'], ['pacco', 'noun', 'b'], ['pace', 'noun', 'a'], ['padella', 'noun', 'c'], ['padre', 'noun', 'a'], ['padrona', 'noun', 'b'], ['padronato', 'noun', 'c'], ['padrone', 'noun', 'a'], ['padroneggiare', 'verb', 'c'], ['paesaggio', 'noun', 'b'], ['paese', 'noun', 'a'], ['paga', 'noun', 'b'], ['pagamento', 'noun', 'a'], ['pagare', 'verb', 'a'], ['pagella', 'noun', 'c'], ['pagina', 'noun', 'a'], ['paglia', 'noun', 'b'], ['paglia', 'adjective', 'b'], ['pagliaio', 'noun', 'c'], ['pago', 'past_part', 'b'], ['pago', 'adjective', 'b'], ['paio', 'noun', 'a'], ['pala', 'noun', 'b'], ['palato', 'noun', 'c'], ['palazzina', 'noun', 'c'], ['palazzo', 'noun', 'a'], ['palco', 'noun', 'b'], ['palcoscenico', 'noun', 'b'], ['palermitano', 'adjective', 'c'], ['palermitano', 'noun', 'c'], ['palestinese', 'adjective', 'c'], ['palestinese', 'noun', 'c'], ['palestra', 'noun', 'b'], ['paletta', 'noun', 'c'], ['palla', 'noun', 'a'], ['pallacanestro', 'noun', 'c'], ['pallanuoto', 'noun', 'c'], ['pallavolo', 'noun', 'c'], ['pallido', 'adjective', 'b'], ['pallina', 'noun', 'b'], ['pallino', 'noun', 'c'], ['palloncino', 'noun', 'c'], ['pallone', 'noun', 'b'], ['pallottola', 'noun', 'c'], ['pallottoliere', 'noun', 'c'], ['palma', 'noun', 'c'], ['palo', 'noun', 'b'], ['palombaro', 'noun', 'c'], ['palpebra', 'noun', 'c'], ['palude', 'noun', 'c'], ['panca', 'noun', 'c'], ['pancarrè', 'noun', 'c'], ['pancetta', 'noun', 'c'], ['panchina', 'noun', 'b'], ['pancia', 'noun', 'b'], ['panciotto', 'noun', 'c'], ['panda', 'noun', 'c'], ['pandoro', 'noun', 'c'], ['pane', 'noun', 'a'], ['panetteria', 'noun', 'c'], ['panettiere', 'noun', 'c'], ['panettone', 'noun', 'c'], ['panico', 'adjective', 'b'], ['panico', 'noun', 'b'], ['paniere', 'noun', 'c'], ['panino', 'noun', 'b'], ['panna', 'noun', 'b'], ['pannello', 'noun', 'b'], ['panno', 'noun', 'b'], ['pannocchia', 'noun', 'c'], ['pannolino', 'noun', 'c'], ['pannolone', 'noun', 'c'], ['panorama', 'noun', 'b'], ['pantalone', 'noun', 'a'], ['pantera', 'noun', 'c'], ['pantofola', 'noun', 'c'], ['panzerotto', 'noun', 'c'], ['papa', 'noun', 'a'], ['papà', 'noun', 'a'], ['papavero', 'noun', 'c'], ['papera', 'noun', 'c'], ['papero', 'noun', 'c'], ['pappa', 'noun', 'c'], ['pappagallo', 'noun', 'c'], ['parabola', 'noun', 'c'], ['parabrezza', 'noun', 'c'], ['paracadute', 'noun', 'c'], ['paracadutista', 'noun', 'c'], ['paradiso', 'noun', 'b'], ['paradosso', 'noun', 'b'], ['paradosso', 'adjective', 'b'], ['parafulmine', 'noun', 'c'], ['paragonare', 'verb', 'b'], ['paragone', 'noun', 'b'], ['paralisi', 'noun', 'c'], ['paralizzato', 'past_part', 'c'], ['paralizzato', 'adjective', 'c'], ['parallelepipedo', 'noun', 'c'], ['parallelo', 'adjective', 'b'], ['parallelo', 'noun', 'b'], ['paralume', 'noun', 'c'], ['parametro', 'noun', 'b'], ['paraocchi', 'noun', 'c'], ['parare', 'verb', 'b'], ['paraurti', 'noun', 'c'], ['paravento', 'noun', 'c'], ['parcheggiare', 'verb', 'b'], ['parcheggio', 'noun', 'b'], ['parco', 'noun', 'a'], ['parecchio', 'adjective', 'a'], ['parecchio', 'pronoun', 'a'], ['parecchio', 'adverb', 'a'], ['parecchio', 'adjective', 'a'], ['pareggiare', 'verb', 'c'], ['pareggio', 'noun', 'c'], ['parente', 'noun', 'a'], ['parentesi', 'noun', 'b'], ['parere', 'verb', 'a'], ['parere', 'noun', 'a'], ['parete', 'noun', 'a'], ['pari', 'adjective', 'a'], ['pari', 'adverb', 'a'], ['pari', 'noun', 'a'], ['parigino', 'adjective', 'c'], ['parigino', 'noun', 'c'], ['parità', 'noun', 'c'], ['parlamentare', 'adjective', 'b'], ['parlamentare', 'noun', 'b'], ['parlamento', 'noun', 'b'], ['parlare', 'verb', 'a'], ['parmigiano', 'adjective', 'c'], ['parmigiano', 'noun', 'c'], ['parola', 'noun', 'a'], ['parquet', 'noun', 'c'], ['parroco', 'noun', 'c'], ['parrucca', 'noun', 'c'], ['parrucchiere', 'noun', 'c'], ['parte', 'noun', 'a'], ['parte', 'adverb', 'a'], ['partecipante', 'pres_part', 'b'], ['partecipante', 'adjective', 'b'], ['partecipante', 'noun', 'b'], ['partecipare', 'verb', 'a'], ['partecipazione', 'noun', 'b'], ['parteggiare', 'verb', 'c'], ['partenza', 'noun', 'a'], ['particella', 'noun', 'b'], ['particolare', 'adjective', 'a'], ['particolare', 'noun', 'a'], ['particolarmente', 'adverb', 'a'], ['partigiano', 'noun', 'b'], ['partigiano', 'adjective', 'b'], ['partire', 'verb', 'a'], ['partita', 'noun', 'a'], ['partito', 'noun', 'a'], ['partner', 'noun', 'b'], ['parto', 'noun', 'b'], ['partorire', 'verb', 'b'], ['party', 'noun', 'b'], ['parziale', 'adjective', 'b'], ['parziale', 'noun', 'b'], ['parzialmente', 'adverb', 'b'], ['pascolare', 'verb', 'c'], ['pasqua', 'noun', 'c'], ['pasquale', 'adjective', 'b'], ['passaggio', 'noun', 'a'], ['passare', 'verb', 'a'], ['passata', 'noun', 'c'], ['passatempo', 'noun', 'c'], ['passato', 'past_part', 'a'], ['passato', 'adjective', 'a'], ['passato', 'noun', 'a'], ['passeggero', 'adjective', 'b'], ['passeggero', 'noun', 'b'], ['passeggiare', 'verb', 'b'], ['passeggiata', 'noun', 'b'], ['passeggio', 'noun', 'c'], ['passero', 'noun', 'c'], ['passione', 'noun', 'a'], ['passivo', 'adjective', 'b'], ['passivo', 'noun', 'b'], ['passo', 'noun', 'a'], ['pasta', 'noun', 'a'], ['pasticca', 'noun', 'c'], ['pasticcere', 'noun', 'c'], ['pasticceria', 'noun', 'c'], ['pasticcino', 'noun', 'c'], ['pasticcio', 'noun', 'c'], ['pastiglia', 'noun', 'c'], ['pastina', 'noun', 'c'], ['pasto', 'noun', 'b'], ['pastore', 'noun', 'b'], ['patata', 'noun', 'b'], ['patatina', 'noun', 'c'], ['patè', 'noun', 'c'], ['patente', 'noun', 'b'], ['patetico', 'adjective', 'b'], ['patetico', 'noun', 'b'], ['patologia', 'noun', 'b'], ['patria', 'noun', 'b'], ['patrimonio', 'noun', 'b'], ['pattinaggio', 'noun', 'c'], ['pattinare', 'verb', 'c'], ['pattino', 'noun', 'c'], ['patto', 'noun', 'b'], ['pattumiera', 'noun', 'c'], ['paura', 'noun', 'a'], ['pauroso', 'adjective', 'c'], ['pausa', 'noun', 'a'], ['pavimento', 'noun', 'b'], ['pavone', 'noun', 'c'], ['pavone', 'adjective', 'c'], ['paziente', 'adjective', 'a'], ['paziente', 'noun', 'a'], ['pazienza', 'noun', 'a'], ['pazza', 'noun', 'c'], ['pazzesco', 'adjective', 'b'], ['pazzo', 'adjective', 'a'], ['pazzo', 'noun', 'a'], ['peccato', 'noun', 'b'], ['peccato', 'exclamation', 'b'], ['peccatore', 'noun', 'c'], ['peccatore', 'adjective', 'c'], ['pechinese', 'adjective', 'c'], ['pechinese', 'noun', 'c'], ['pecora', 'noun', 'b'], ['pecorino', 'adjective', 'c'], ['pecorino', 'noun', 'c'], ['pedalare', 'verb', 'c'], ['pedale', 'noun', 'c'], ['pedale', 'adjective', 'c'], ['pedone', 'noun', 'c'], ['pedone', 'adjective', 'c'], ['peggio', 'adverb', 'a'], ['peggio', 'adjective', 'a'], ['peggio', 'noun', 'a'], ['peggioramento', 'noun', 'c'], ['peggiorare', 'verb', 'b'], ['peggiore', 'adjective', 'b'], ['peggiore', 'noun', 'b'], ['peggiore', 'adverb', 'b'], ['pelato', 'past_part', 'c'], ['pelato', 'adjective', 'c'], ['pelato', 'noun', 'c'], ['pelle', 'noun', 'a'], ['pellegrino', 'noun', 'c'], ['pellegrino', 'adjective', 'c'], ['pellerossa', 'adjective', 'c'], ['pellerossa', 'noun', 'c'], ['pelletteria', 'noun', 'c'], ['pellicola', 'noun', 'b'], ['pelo', 'noun', 'b'], ['peloso', 'adjective', 'c'], ['peloso', 'noun', 'c'], ['peluche', 'noun', 'c'], ['pena', 'noun', 'a'], ['penale', 'adjective', 'b'], ['penale', 'noun', 'b'], ['pendere', 'verb', 'b'], ['pendolo', 'noun', 'c'], ['pene', 'noun', 'b'], ['penetrare', 'verb', 'b'], ['penisola', 'noun', 'c'], ['penna', 'noun', 'b'], ['pennarello', 'noun', 'c'], ['pensare', 'verb', 'a'], ['pensiero', 'noun', 'a'], ['pensionato', 'past_part', 'c'], ['pensionato', 'adjective', 'c'], ['pensionato', 'noun', 'c'], ['pensione', 'noun', 'a'], ['pentagono', 'noun', 'c'], ['pentirsi', 'verb', 'b'], ['pentola', 'noun', 'b'], ['penultimo', 'adjective', 'c'], ['pepe', 'noun', 'c'], ['peperoncino', 'noun', 'c'], ['peperone', 'noun', 'c'], ['per', 'preposition', 'a'], ['pera', 'noun', 'c'], ['peraltro', 'adverb', 'b'], ['percentuale', 'adjective', 'b'], ['percentuale', 'noun', 'b'], ['percepire', 'verb', 'a'], ['percezione', 'noun', 'b'], ['perché', 'adverb', 'a'], ['perché', 'conjunction', 'a'], ['perché', 'noun', 'a'], ['perciò', 'conjunction', 'a'], ['percorrere', 'verb', 'b'], ['percorso', 'past_part', 'a'], ['percorso', 'adjective', 'a'], ['percorso', 'noun', 'a'], ['perdere', 'verb', 'a'], ['perdita', 'noun', 'a'], ['perdonare', 'verb', 'a'], ['perdono', 'noun', 'b'], ['perduto', 'past_part', 'b'], ['perduto', 'adjective', 'b'], ['perfettamente', 'adverb', 'a'], ['perfetto', 'past_part', 'a'], ['perfetto', 'adjective', 'a'], ['perfetto', 'noun', 'a'], ['perfezione', 'noun', 'b'], ['perfino', 'adverb', 'a'], ['perfino', 'preposition', 'a'], ['pergola', 'noun', 'c'], ['pergolato', 'noun', 'c'], ['pergolato', 'adjective', 'c'], ['pericolo', 'noun', 'a'], ['pericoloso', 'adjective', 'a'], ['periferia', 'noun', 'b'], ['periodico', 'adjective', 'b'], ['periodico', 'noun', 'b'], ['periodo', 'noun', 'a'], ['perito', 'noun', 'b'], ['perito', 'adjective', 'b'], ['perla', 'noun', 'b'], ['perla', 'adjective', 'b'], ['permaloso', 'adjective', 'c'], ['permaloso', 'noun', 'c'], ['permanente', 'pres_part', 'b'], ['permanente', 'adjective', 'b'], ['permanente', 'noun', 'b'], ['permesso', 'past_part', 'b'], ['permesso', 'adjective', 'b'], ['permesso', 'noun', 'b'], ['permettere', 'verb', 'a'], ['pero', 'noun', 'c'], ['però', 'conjunction', 'a'], ['perpendicolare', 'adjective', 'c'], ['perpendicolare', 'noun', 'c'], ['perplesso', 'adjective', 'b'], ['perquisizione', 'noun', 'b'], ['perseguire', 'verb', 'b'], ['persiana', 'noun', 'c'], ['persiano', 'adjective', 'b'], ['persiano', 'noun', 'b'], ['persino', 'adverb', 'a'], ['perso', 'past_part', 'b'], ['perso', 'adjective', 'b'], ['persona', 'noun', 'a'], ['personaggio', 'noun', 'a'], ['personale', 'adjective', 'a'], ['personale', 'noun', 'a'], ['personale', 'noun', 'a'], ['personalità', 'noun', 'b'], ['personalmente', 'adverb', 'a'], ['pertanto', 'conjunction', 'b'], ['perugino', 'adjective', 'c'], ['perugino', 'noun', 'c'], ['peruviano', 'adjective', 'c'], ['peruviano', 'noun', 'c'], ['pervenire', 'verb', 'b'], ['pesante', 'pres_part', 'a'], ['pesante', 'adjective', 'a'], ['pesante', 'adverb', 'a'], ['pesare', 'verb', 'b'], ['pesca', 'noun', 'c'], ['pesca', 'adjective', 'c'], ['pesca', 'noun', 'b'], ['pescare', 'verb', 'b'], ['pescatore', 'noun', 'b'], ['pescatore', 'adjective', 'b'], ['pesce', 'noun', 'a'], ['peschereccio', 'noun', 'c'], ['peschereccio', 'adjective', 'c'], ['pescheria', 'noun', 'c'], ['pesco', 'noun', 'c'], ['peso', 'noun', 'a'], ['pessimo', 'adjective', 'b'], ['pestare', 'verb', 'c'], ['peste', 'noun', 'c'], ['pesto', 'past_part', 'c'], ['pesto', 'adjective', 'c'], ['pesto', 'noun', 'c'], ['petalo', 'noun', 'c'], ['petardo', 'noun', 'c'], ['petroliera', 'noun', 'c'], ['petrolio', 'noun', 'b'], ['pettegolezzo', 'noun', 'c'], ['pettegolo', 'adjective', 'c'], ['pettegolo', 'noun', 'c'], ['pettinare', 'verb', 'c'], ['pettinatura', 'noun', 'c'], ['pettine', 'noun', 'c'], ['pettirosso', 'noun', 'c'], ['petto', 'noun', 'a'], ['pezza', 'noun', 'c'], ['pezzetto', 'noun', 'b'], ['pezzo', 'noun', 'a'], ['pezzuola', 'noun', 'c'], ['pi', 'noun', 'c'], ['piacere', 'verb', 'a'], ['piacere', 'noun', 'a'], ['piacevole', 'adjective', 'b'], ['piadina', 'noun', 'c'], ['piaga', 'noun', 'c'], ['pialla', 'noun', 'c'], ['piallare', 'verb', 'c'], ['pianeggiante', 'pres_part', 'c'], ['pianeggiante', 'adjective', 'c'], ['pianerottolo', 'noun', 'b'], ['pianeta', 'noun', 'a'], ['piangere', 'verb', 'a'], ['piangere', 'noun', 'a'], ['piano', 'noun', 'a'], ['piano', 'noun', 'a'], ['piano', 'adjective', 'a'], ['piano', 'adverb', 'a'], ['pianoforte', 'noun', 'b'], ['pianoterra', 'noun', 'c'], ['pianta', 'noun', 'a'], ['piantare', 'verb', 'b'], ['pianto', 'noun', 'b'], ['pianura', 'noun', 'b'], ['piastra', 'noun', 'c'], ['piattaforma', 'noun', 'b'], ['piatto', 'adjective', 'a'], ['piatto', 'noun', 'a'], ['piazza', 'noun', 'a'], ['piazzale', 'noun', 'b'], ['piazzare', 'verb', 'b'], ['piccante', 'adjective', 'c'], ['picchiare', 'verb', 'b'], ['piccino', 'adjective', 'c'], ['piccino', 'noun', 'c'], ['piccione', 'noun', 'c'], ['picco', 'noun', 'b'], ['piccolo', 'adjective', 'a'], ['piccolo', 'noun', 'a'], ['piccone', 'noun', 'c'], ['picnic', 'noun', 'c'], ['pidocchio', 'noun', 'c'], ['piede', 'noun', 'a'], ['piega', 'noun', 'b'], ['piegare', 'verb', 'b'], ['pieghevole', 'adjective', 'c'], ['pieghevole', 'noun', 'c'], ['piemontese', 'adjective', 'b'], ['piemontese', 'noun', 'b'], ['piena', 'noun', 'c'], ['pienamente', 'adverb', 'b'], ['pieno', 'adjective', 'a'], ['pieno', 'noun', 'a'], ['pietà', 'noun', 'b'], ['pietra', 'noun', 'a'], ['pigiama', 'noun', 'c'], ['pigione', 'noun', 'c'], ['pigliare', 'verb', 'b'], ['pigna', 'noun', 'c'], ['pigrizia', 'noun', 'c'], ['pigro', 'adjective', 'c'], ['pigro', 'noun', 'c'], ['pila', 'noun', 'b'], ['pillola', 'noun', 'b'], ['pilota', 'noun', 'b'], ['pineta', 'noun', 'c'], ['ping-pong', 'noun', 'c'], ['pinguino', 'noun', 'c'], ['pinna', 'noun', 'c'], ['pinolo', 'noun', 'c'], ['pinza', 'noun', 'c'], ['pinzetta', 'noun', 'c'], ['pioggia', 'noun', 'a'], ['piombo', 'noun', 'b'], ['piombo', 'adjective', 'b'], ['piombo', 'noun', 'b'], ['pioppo', 'noun', 'c'], ['piovere', 'verb', 'b'], ['piovoso', 'adjective', 'c'], ['piovoso', 'noun', 'c'], ['pipì', 'noun', 'c'], ['pipistrello', 'noun', 'c'], ['pirata', 'noun', 'b'], ['piscina', 'noun', 'b'], ['pisello', 'noun', 'c'], ['pisello', 'adjective', 'c'], ['pisolino', 'noun', 'c'], ['pista', 'noun', 'b'], ['pistacchio', 'noun', 'c'], ['pistacchio', 'adjective', 'c'], ['pistola', 'noun', 'a'], ['pittare', 'verb', 'c'], ['pittore', 'noun', 'b'], ['pittore', 'adjective', 'b'], ['pittura', 'noun', 'b'], ['pitturare', 'verb', 'c'], ['più', 'adverb', 'a'], ['più', 'adjective', 'a'], ['più', 'preposition', 'a'], ['più', 'noun', 'a'], ['piuma', 'noun', 'c'], ['piumino', 'noun', 'c'], ['piuttosto', 'adverb', 'a'], ['pizza', 'noun', 'b'], ['pizzeria', 'noun', 'c'], ['pizzetta', 'noun', 'c'], ['pizzicare', 'verb', 'c'], ['pizzo', 'noun', 'c'], ['plaid', 'noun', 'c'], ['plastica', 'noun', 'b'], ['plastico', 'adjective', 'b'], ['plastico', 'noun', 'b'], ['platano', 'noun', 'c'], ['platino', 'noun', 'c'], ['platino', 'adjective', 'c'], ['plurale', 'noun', 'c'], ['plurale', 'adjective', 'c'], ['pneumatico', 'noun', 'c'], ['pochino', 'noun', 'b'], ['poco', 'adjective', 'a'], ['poco', 'pronoun', 'a'], ['poco', 'adverb', 'a'], ['podere', 'noun', 'c'], ['poema', 'noun', 'b'], ['poesia', 'noun', 'a'], ['poeta', 'noun', 'a'], ['poetico', 'adjective', 'b'], ['poetico', 'noun', 'b'], ['poggiapiedi', 'noun', 'c'], ['poggiare', 'verb', 'c'], ['poi', 'adverb', 'a'], ['poiché', 'conjunction', 'a'], ['poker', 'noun', 'b'], ['polacco', 'adjective', 'b'], ['polacco', 'noun', 'b'], ['polemica', 'noun', 'b'], ['polenta', 'noun', 'c'], ['polipo', 'noun', 'c'], ['politica', 'noun', 'a'], ['politico', 'adjective', 'a'], ['politico', 'noun', 'a'], ['polizia', 'noun', 'a'], ['poliziotto', 'noun', 'a'], ['pollaio', 'noun', 'c'], ['pollame', 'noun', 'c'], ['pollice', 'noun', 'b'], ['pollo', 'noun', 'c'], ['polmone', 'noun', 'b'], ['polo', 'noun', 'b'], ['polpa', 'noun', 'c'], ['polpastrello', 'noun', 'c'], ['polpetta', 'noun', 'c'], ['polpo', 'noun', 'c'], ['polsino', 'noun', 'c'], ['polso', 'noun', 'b'], ['poltrona', 'noun', 'b'], ['polvere', 'noun', 'a'], ['polverina', 'noun', 'c'], ['polveroso', 'adjective', 'c'], ['pomata', 'noun', 'c'], ['pomello', 'noun', 'c'], ['pomeriggio', 'noun', 'a'], ['pomodoro', 'noun', 'b'], ['pompa', 'noun', 'b'], ['pompelmo', 'noun', 'c'], ['pompiere', 'noun', 'c'], ['ponte', 'noun', 'a'], ['pony', 'noun', 'c'], ['pop', 'adjective', 'b'], ['pop', 'noun', 'b'], ['popolare', 'adjective', 'a'], ['popolare', 'noun', 'a'], ['popolare', 'verb', 'b'], ['popolarità', 'noun', 'c'], ['popolazione', 'noun', 'a'], ['popolo', 'noun', 'a'], ['porcellana', 'noun', 'c'], ['porcheria', 'noun', 'c'], ['porco', 'noun', 'b'], ['porco', 'adjective', 'b'], ['porgere', 'verb', 'b'], ['porno', 'adjective', 'b'], ['porno', 'noun', 'b'], ['porre', 'verb', 'a'], ['porta', 'noun', 'a'], ['portabagagli', 'noun', 'c'], ['portabagagli', 'adjective', 'c'], ['portacenere', 'noun', 'c'], ['portachiavi', 'noun', 'c'], ['portacipria', 'noun', 'c'], ['portaerei', 'noun', 'c'], ['portafinestra', 'noun', 'c'], ['portafoglio', 'noun', 'b'], ['portafortuna', 'noun', 'c'], ['portale', 'noun', 'b'], ['portamonete', 'noun', 'c'], ['portaombrelli', 'noun', 'c'], ['portare', 'verb', 'a'], ['portata', 'noun', 'b'], ['portatore', 'adjective', 'b'], ['portatore', 'noun', 'b'], ['portiere', 'noun', 'b'], ['portineria', 'noun', 'c'], ['porto', 'noun', 'a'], ['portoghese', 'adjective', 'b'], ['portoghese', 'noun', 'b'], ['portone', 'noun', 'b'], ['porzione', 'noun', 'b'], ['posa', 'noun', 'b'], ['posacenere', 'noun', 'c'], ['posare', 'verb', 'b'], ['posata', 'noun', 'c'], ['positivo', 'adjective', 'a'], ['positivo', 'noun', 'a'], ['positivo', 'adverb', 'a'], ['posizionare', 'verb', 'b'], ['posizione', 'noun', 'a'], ['possedere', 'verb', 'a'], ['possesso', 'noun', 'b'], ['possibile', 'adjective', 'a'], ['possibile', 'noun', 'a'], ['possibilità', 'noun', 'a'], ['post', 'noun', 'b'], ['posta', 'noun', 'a'], ['postale', 'adjective', 'b'], ['postare', 'verb', 'b'], ['posteggiatore', 'noun', 'c'], ['posteriore', 'adjective', 'b'], ['posteriore', 'noun', 'b'], ['postino', 'noun', 'c'], ['postino', 'adjective', 'c'], ['posto', 'noun', 'a'], ['potare', 'verb', 'c'], ['potente', 'pres_part', 'a'], ['potente', 'adjective', 'a'], ['potente', 'noun', 'a'], ['potentino', 'adjective', 'c'], ['potentino', 'noun', 'c'], ['potenza', 'noun', 'b'], ['potenziale', 'adjective', 'b'], ['potenziale', 'noun', 'b'], ['potere', 'verb', 'a'], ['potere', 'noun', 'a'], ['povero', 'adjective', 'a'], ['povertà', 'noun', 'b'], ['pozzanghera', 'noun', 'c'], ['pozzo', 'noun', 'b'], ['praghese', 'adjective', 'c'], ['praghese', 'noun', 'c'], ['pranzo', 'noun', 'a'], ['prassi', 'noun', 'b'], ['pratica', 'noun', 'a'], ['praticamente', 'adverb', 'a'], ['praticare', 'verb', 'b'], ['pratico', 'adjective', 'a'], ['prato', 'noun', 'b'], ['precario', 'adjective', 'b'], ['precedente', 'pres_part', 'a'], ['precedente', 'adjective', 'a'], ['precedente', 'noun', 'a'], ['precedentemente', 'adverb', 'b'], ['precedenza', 'noun', 'b'], ['precedere', 'verb', 'b'], ['precipitare', 'verb', 'b'], ['precisamente', 'adverb', 'b'], ['precisare', 'verb', 'a'], ['precisione', 'noun', 'b'], ['preciso', 'adjective', 'a'], ['preciso', 'adverb', 'a'], ['preda', 'noun', 'b'], ['predisporre', 'verb', 'b'], ['preferenza', 'noun', 'b'], ['preferire', 'verb', 'a'], ['preferito', 'past_part', 'b'], ['preferito', 'adjective', 'b'], ['preferito', 'noun', 'b'], ['pregare', 'verb', 'a'], ['preghiera', 'noun', 'b'], ['pregiato', 'past_part', 'c'], ['pregiato', 'adjective', 'c'], ['pregio', 'noun', 'b'], ['pregiudizio', 'noun', 'b'], ['prego', 'exclamation', 'a'], ['prelevare', 'verb', 'b'], ['preliminare', 'adjective', 'b'], ['preliminare', 'noun', 'b'], ['prémaman', 'adjective', 'c'], ['premere', 'verb', 'b'], ['premessa', 'noun', 'b'], ['premiare', 'verb', 'b'], ['premier', 'noun', 'b'], ['premio', 'noun', 'a'], ['premio', 'adjective', 'a'], ['prendere', 'verb', 'a'], ['prenotare', 'verb', 'b'], ['prenotazione', 'noun', 'c'], ['preoccupare', 'verb', 'a'], ['preoccupato', 'past_part', 'b'], ['preoccupato', 'adjective', 'b'], ['preoccupazione', 'noun', 'b'], ['preparare', 'verb', 'a'], ['preparazione', 'noun', 'b'], ['prepotente', 'adjective', 'c'], ['prepotente', 'noun', 'c'], ['presa', 'noun', 'a'], ['prescindere', 'verb', 'b'], ['prescrivere', 'verb', 'b'], ['prescrizione', 'noun', 'b'], ['presentare', 'verb', 'a'], ['presentazione', 'noun', 'b'], ['presente', 'adjective', 'a'], ['presente', 'noun', 'a'], ['presente', 'adverb', 'a'], ['presenza', 'noun', 'a'], ['presepe', 'noun', 'b'], ['preside', 'noun', 'c'], ['presidente', 'noun', 'a'], ['presidente', 'adjective', 'a'], ['presidenza', 'noun', 'b'], ['pressione', 'noun', 'a'], ['presso', 'adverb', 'a'], ['presso', 'preposition', 'a'], ['presso', 'noun', 'a'], ['presso', 'adjective', 'a'], ['prestare', 'verb', 'a'], ['prestazione', 'noun', 'b'], ['prestigio', 'noun', 'b'], ['prestigioso', 'adjective', 'b'], ['prestito', 'noun', 'b'], ['presto', 'adverb', 'a'], ['presto', 'exclamation', 'a'], ['presto', 'adjective', 'a'], ['presumere', 'verb', 'b'], ['presunto', 'past_part', 'b'], ['presunto', 'adjective', 'b'], ['presupposto', 'past_part', 'b'], ['presupposto', 'adjective', 'b'], ['presupposto', 'noun', 'b'], ['prete', 'noun', 'a'], ['pretendere', 'verb', 'a'], ['pretesa', 'noun', 'b'], ['pretesto', 'noun', 'b'], ['prevalentemente', 'adverb', 'b'], ['prevalere', 'verb', 'b'], ['prevedere', 'verb', 'a'], ['prevedibile', 'adjective', 'b'], ['prevenire', 'verb', 'b'], ['preventivo', 'adjective', 'b'], ['preventivo', 'noun', 'b'], ['prevenzione', 'noun', 'b'], ['previdenza', 'noun', 'c'], ['previsione', 'noun', 'b'], ['previsto', 'past_part', 'a'], ['previsto', 'adjective', 'a'], ['previsto', 'noun', 'a'], ['prezioso', 'adjective', 'a'], ['prezioso', 'noun', 'a'], ['prezzemolo', 'noun', 'c'], ['prezzo', 'noun', 'a'], ['prigione', 'noun', 'b'], ['prigioniero', 'adjective', 'b'], ['prigioniero', 'noun', 'b'], ['prima', 'adverb', 'a'], ['prima', 'adjective', 'a'], ['prima', 'noun', 'a'], ['prima', 'noun', 'a'], ['primario', 'adjective', 'b'], ['primario', 'noun', 'b'], ['primavera', 'noun', 'a'], ['primizia', 'noun', 'c'], ['primo', 'adjective', 'a'], ['primo', 'noun', 'a'], ['primo', 'adverb', 'a'], ['primula', 'noun', 'c'], ['principale', 'adjective', 'a'], ['principale', 'noun', 'a'], ['principalmente', 'adverb', 'b'], ['principe', 'noun', 'a'], ['principe', 'adjective', 'a'], ['principessa', 'noun', 'b'], ['principio', 'noun', 'a'], ['priorità', 'noun', 'b'], ['privacy', 'noun', 'b'], ['privare', 'verb', 'b'], ['privato', 'adjective', 'a'], ['privato', 'noun', 'a'], ['privilegio', 'noun', 'b'], ['privo', 'adjective', 'b'], ['privo', 'preposition', 'b'], ['privo', 'noun', 'b'], ['probabile', 'adjective', 'b'], ['probabilità', 'noun', 'b'], ['probabilmente', 'adverb', 'a'], ['problema', 'noun', 'a'], ['problematico', 'adjective', 'b'], ['procedere', 'verb', 'a'], ['procedimento', 'noun', 'b'], ['procedura', 'noun', 'a'], ['processo', 'noun', 'a'], ['proclamare', 'verb', 'b'], ['procura', 'noun', 'b'], ['procurare', 'verb', 'b'], ['procuratore', 'noun', 'b'], ['prodotto', 'past_part', 'a'], ['prodotto', 'adjective', 'a'], ['prodotto', 'noun', 'a'], ['produrre', 'verb', 'a'], ['produttivo', 'adjective', 'b'], ['produttore', 'adjective', 'b'], ['produttore', 'noun', 'b'], ['produzione', 'noun', 'a'], ['prof', 'noun', 'b'], ['professionale', 'adjective', 'a'], ['professione', 'noun', 'b'], ['professionista', 'noun', 'b'], ['professore', 'noun', 'a'], ['professoressa', 'noun', 'b'], ['profeta', 'noun', 'b'], ['profilattico', 'adjective', 'c'], ['profilattico', 'noun', 'c'], ['profilo', 'noun', 'a'], ['profitto', 'noun', 'b'], ['profondamente', 'adverb', 'b'], ['profondità', 'noun', 'b'], ['profondo', 'adjective', 'a'], ['profondo', 'noun', 'a'], ['profondo', 'adverb', 'a'], ['profumare', 'verb', 'b'], ['profumato', 'past_part', 'c'], ['profumato', 'adjective', 'c'], ['profumo', 'noun', 'b'], ['progettare', 'verb', 'b'], ['progettazione', 'noun', 'b'], ['progetto', 'noun', 'a'], ['programma', 'noun', 'a'], ['programmare', 'verb', 'b'], ['programmazione', 'noun', 'b'], ['progressista', 'adjective', 'c'], ['progressista', 'noun', 'c'], ['progressivo', 'adjective', 'b'], ['progresso', 'noun', 'b'], ['proibire', 'verb', 'b'], ['proiettare', 'verb', 'b'], ['proiettile', 'noun', 'b'], ['proiezione', 'noun', 'b'], ['prolunga', 'noun', 'c'], ['promessa', 'noun', 'b'], ['promettere', 'verb', 'a'], ['promozione', 'noun', 'b'], ['promuovere', 'verb', 'b'], ['pronto', 'adjective', 'a'], ['pronuncia', 'noun', 'c'], ['pronunciare', 'verb', 'a'], ['propaganda', 'noun', 'b'], ['propagandare', 'verb', 'c'], ['proporre', 'verb', 'a'], ['proporzione', 'noun', 'b'], ['proposito', 'noun', 'a'], ['proposizione', 'noun', 'c'], ['proposta', 'noun', 'a'], ['proprietà', 'noun', 'a'], ['proprietario', 'adjective', 'a'], ['proprietario', 'noun', 'a'], ['proprio', 'adjective', 'a'], ['proprio', 'adverb', 'a'], ['proprio', 'noun', 'a'], ['prosa', 'noun', 'b'], ['prosciugare', 'verb', 'c'], ['prosciutto', 'noun', 'b'], ['prosecco', 'noun', 'c'], ['proseguire', 'verb', 'a'], ['prospettiva', 'noun', 'b'], ['prossimo', 'adjective', 'a'], ['prossimo', 'noun', 'a'], ['prostituta', 'noun', 'b'], ['protagonista', 'adjective', 'a'], ['protagonista', 'noun', 'a'], ['proteggere', 'verb', 'a'], ['proteina', 'noun', 'b'], ['protesta', 'noun', 'b'], ['protestare', 'verb', 'b'], ['protetto', 'past_part', 'b'], ['protetto', 'adjective', 'b'], ['protetto', 'noun', 'b'], ['protezione', 'noun', 'b'], ['protocollo', 'noun', 'b'], ['prova', 'noun', 'a'], ['provare', 'verb', 'a'], ['provenienza', 'noun', 'b'], ['provenire', 'verb', 'a'], ['provincia', 'noun', 'a'], ['provinciale', 'adjective', 'b'], ['provinciale', 'noun', 'b'], ['provocare', 'verb', 'a'], ['provola', 'noun', 'c'], ['provolone', 'noun', 'c'], ['provvedere', 'verb', 'b'], ['provvedimento', 'noun', 'b'], ['provvisorio', 'adjective', 'b'], ['prudere', 'verb', 'c'], ['prugna', 'noun', 'c'], ['prugna', 'adjective', 'c'], ['prurito', 'noun', 'c'], ['pseudonimo', 'noun', 'b'], ['pseudonimo', 'adjective', 'b'], ['psichiatra', 'noun', 'b'], ['psichiatria', 'noun', 'c'], ['psichico', 'adjective', 'b'], ['psicologia', 'noun', 'b'], ['psicologico', 'adjective', 'b'], ['psicologo', 'noun', 'b'], ['pub', 'noun', 'b'], ['pubblicare', 'verb', 'a'], ['pubblicazione', 'noun', 'b'], ['pubblicità', 'noun', 'a'], ['pubblicitario', 'adjective', 'b'], ['pubblicitario', 'noun', 'b'], ['pubblico', 'adjective', 'a'], ['pubblico', 'noun', 'a'], ['pugilato', 'noun', 'c'], ['pugliese', 'adjective', 'c'], ['pugliese', 'noun', 'c'], ['pugno', 'noun', 'a'], ['pulce', 'noun', 'c'], ['pulce', 'adjective', 'c'], ['pulcino', 'noun', 'c'], ['puledro', 'noun', 'c'], ['pulire', 'verb', 'a'], ['pulito', 'past_part', 'b'], ['pulito', 'adjective', 'b'], ['pulito', 'noun', 'b'], ['pulizia', 'noun', 'b'], ['pullman', 'noun', 'b'], ['pullover', 'noun', 'c'], ['pulmino', 'noun', 'c'], ['pulsante', 'pres_part', 'b'], ['pulsante', 'adjective', 'b'], ['pulsante', 'noun', 'b'], ['puma', 'noun', 'c'], ['pungere', 'verb', 'c'], ['punire', 'verb', 'b'], ['punizione', 'noun', 'b'], ['punk', 'adjective', 'c'], ['punk', 'noun', 'c'], ['punta', 'noun', 'a'], ['puntare', 'verb', 'a'], ['puntata', 'noun', 'b'], ['puntato', 'past_part', 'b'], ['puntato', 'adjective', 'b'], ['punteggio', 'noun', 'c'], ['puntiglio', 'noun', 'c'], ['puntino', 'noun', 'b'], ['punto', 'noun', 'a'], ['puntuale', 'adjective', 'b'], ['puntura', 'noun', 'c'], ['pupa', 'noun', 'b'], ['pupazzo', 'noun', 'c'], ['pupo', 'noun', 'c'], ['purché', 'conjunction', 'b'], ['pure', 'adverb', 'a'], ['pure', 'conjunction', 'a'], ['purè', 'noun', 'c'], ['purga', 'noun', 'c'], ['puro', 'adjective', 'a'], ['puro', 'noun', 'a'], ['purtroppo', 'adverb', 'a'], ['puttana', 'noun', 'b'], ['puzza', 'noun', 'b'], ['puzzare', 'verb', 'b'], ['puzzle', 'noun', 'c'], ['qua', 'adverb', 'a'], ['quaderno', 'noun', 'b'], ['quadrato', 'past_part', 'b'], ['quadrato', 'adjective', 'b'], ['quadrato', 'noun', 'b'], ['quadrifoglio', 'noun', 'c'], ['quadro', 'adjective', 'a'], ['quadro', 'noun', 'a'], ['quaglia', 'noun', 'c'], ['qualche', 'adjective', 'a'], ['qualche', 'adverb', 'a'], ['qualcosa', 'pronoun', 'a'], ['qualcuno', 'pronoun', 'a'], ['qualcuno', 'adjective', 'a'], ['qualcuno', 'noun', 'a'], ['quale', 'adjective', 'a'], ['quale', 'pronoun', 'a'], ['quale', 'adverb', 'a'], ['quale', 'noun', 'a'], ['qualificare', 'verb', 'b'], ['qualità', 'noun', 'a'], ['qualora', 'conjunction', 'b'], ['qualsiasi', 'adjective', 'a'], ['qualunque', 'adjective', 'a'], ['qualunque', 'pronoun', 'a'], ['quando', 'conjunction', 'a'], ['quando', 'adverb', 'a'], ['quando', 'noun', 'a'], ['quantità', 'noun', 'a'], ['quantitativo', 'adjective', 'b'], ['quantitativo', 'noun', 'b'], ['quanto', 'adjective', 'a'], ['quanto', 'pronoun', 'a'], ['quanto', 'adverb', 'a'], ['quanto', 'noun', 'a'], ['quaranta', 'adjective', 'a'], ['quaranta', 'noun', 'a'], ['quarta', 'noun', 'b'], ['quartiere', 'noun', 'a'], ['quarto', 'adjective', 'a'], ['quarto', 'noun', 'a'], ['quasi', 'adverb', 'a'], ['quasi', 'conjunction', 'a'], ['quattordici', 'adjective', 'b'], ['quattordici', 'noun', 'b'], ['quattro', 'adjective', 'a'], ['quattro', 'noun', 'a'], ['quello', 'adjective', 'a'], ['quello', 'pronoun', 'a'], ['quercia', 'noun', 'c'], ['questione', 'noun', 'a'], ['questo', 'adjective', 'a'], ['questo', 'pronoun', 'a'], ['questura', 'noun', 'b'], ['qui', 'adverb', 'a'], ['quindi', 'adverb', 'a'], ['quindi', 'conjunction', 'a'], ['quindici', 'adjective', 'a'], ['quindici', 'noun', 'a'], ['quinta', 'noun', 'b'], ['quinto', 'adjective', 'b'], ['quinto', 'noun', 'b'], ['quiz', 'noun', 'a'], ['quota', 'noun', 'a'], ['quotidiano', 'adjective', 'a'], ['quotidiano', 'noun', 'a'], ['rabbia', 'noun', 'a'], ['racchetta', 'noun', 'c'], ['racchiudere', 'verb', 'b'], ['raccogliere', 'verb', 'a'], ['raccolta', 'noun', 'a'], ['raccomandare', 'verb', 'b'], ['raccomandazione', 'noun', 'c'], ['raccontare', 'verb', 'a'], ['racconto', 'noun', 'a'], ['raddoppiare', 'verb', 'b'], ['raddrizzare', 'verb', 'c'], ['radere', 'verb', 'c'], ['radiazione', 'noun', 'b'], ['radicale', 'adjective', 'b'], ['radicale', 'noun', 'b'], ['radicchio', 'noun', 'c'], ['radice', 'noun', 'a'], ['radio', 'noun', 'a'], ['radio', 'adjective', 'a'], ['rado', 'adjective', 'b'], ['rado', 'adverb', 'b'], ['raffigurare', 'verb', 'b'], ['raffinato', 'past_part', 'b'], ['raffinato', 'adjective', 'b'], ['raffinato', 'noun', 'b'], ['rafforzamento', 'noun', 'c'], ['rafforzare', 'verb', 'b'], ['raffreddore', 'noun', 'c'], ['ragazza', 'noun', 'a'], ['ragazzino', 'noun', 'a'], ['ragazzo', 'noun', 'a'], ['raggio', 'noun', 'a'], ['raggiungere', 'verb', 'a'], ['ragionamento', 'noun', 'b'], ['ragionare', 'verb', 'b'], ['ragione', 'noun', 'a'], ['ragionevole', 'adjective', 'b'], ['ragioniere', 'noun', 'b'], ['ragnatela', 'noun', 'c'], ['ragno', 'noun', 'c'], ['ragù', 'noun', 'c'], ['rallegrare', 'verb', 'c'], ['rallentare', 'verb', 'b'], ['rame', 'noun', 'b'], ['rammendo', 'noun', 'c'], ['ramo', 'noun', 'b'], ['rampicante', 'pres_part', 'c'], ['rampicante', 'adjective', 'c'], ['rampicante', 'noun', 'c'], ['rana', 'noun', 'c'], ['rancio', 'noun', 'c'], ['rapa', 'noun', 'c'], ['rapidamente', 'adverb', 'b'], ['rapido', 'adjective', 'a'], ['rapido', 'noun', 'a'], ['rapimento', 'noun', 'c'], ['rapina', 'noun', 'b'], ['rapinatore', 'adjective', 'c'], ['rapinatore', 'noun', 'c'], ['rapire', 'verb', 'b'], ['rapporto', 'noun', 'a'], ['rappresentante', 'pres_part', 'b'], ['rappresentante', 'adjective', 'b'], ['rappresentante', 'noun', 'b'], ['rappresentanza', 'noun', 'b'], ['rappresentare', 'verb', 'a'], ['rappresentazione', 'noun', 'b'], ['raramente', 'adverb', 'b'], ['raro', 'adjective', 'a'], ['raro', 'noun', 'a'], ['raro', 'adverb', 'a'], ['rasare', 'verb', 'c'], ['rasoio', 'noun', 'c'], ['rassegna', 'noun', 'b'], ['rassegnare', 'verb', 'b'], ['rassegnazione', 'noun', 'c'], ['rasserenare', 'verb', 'c'], ['rassicurare', 'verb', 'b'], ['rastrello', 'noun', 'c'], ['rata', 'noun', 'c'], ['rateale', 'adjective', 'c'], ['rattristare', 'verb', 'c'], ['rauco', 'adjective', 'c'], ['ravanello', 'noun', 'c'], ['razionale', 'adjective', 'b'], ['razionale', 'noun', 'b'], ['razza', 'noun', 'b'], ['razzo', 'noun', 'c'], ['re', 'noun', 'a'], ['reagire', 'verb', 'a'], ['reale', 'adjective', 'a'], ['reale', 'noun', 'a'], ['realistico', 'adjective', 'b'], ['realizzare', 'verb', 'a'], ['realizzazione', 'noun', 'b'], ['realmente', 'adverb', 'b'], ['realtà', 'noun', 'a'], ['reato', 'noun', 'a'], ['reazione', 'noun', 'a'], ['recare', 'verb', 'a'], ['recensione', 'noun', 'b'], ['recente', 'adjective', 'a'], ['recentemente', 'adverb', 'b'], ['recintare', 'verb', 'c'], ['recinto', 'past_part', 'c'], ['recinto', 'adjective', 'c'], ['recinto', 'noun', 'c'], ['recipiente', 'adjective', 'c'], ['recipiente', 'noun', 'c'], ['reciproco', 'adjective', 'b'], ['reciproco', 'noun', 'b'], ['recita', 'noun', 'c'], ['recitare', 'verb', 'a'], ['reclame', 'noun', 'c'], ['reclame', 'adjective', 'c'], ['reclamo', 'noun', 'c'], ['recluta', 'noun', 'c'], ['record', 'noun', 'b'], ['recuperare', 'verb', 'a'], ['recupero', 'noun', 'b'], ['redazione', 'noun', 'b'], ['reddito', 'noun', 'b'], ['redigere', 'verb', 'b'], ['referendum', 'noun', 'b'], ['regalare', 'verb', 'a'], ['regale', 'adjective', 'b'], ['regalo', 'noun', 'a'], ['reggere', 'verb', 'a'], ['reggimento', 'noun', 'c'], ['reggiseno', 'noun', 'b'], ['regia', 'noun', 'b'], ['regime', 'noun', 'a'], ['regina', 'noun', 'a'], ['regionale', 'adjective', 'b'], ['regionale', 'noun', 'b'], ['regione', 'noun', 'a'], ['regista', 'noun', 'a'], ['registrare', 'verb', 'a'], ['registratore', 'adjective', 'c'], ['registratore', 'noun', 'c'], ['registrazione', 'noun', 'a'], ['registro', 'noun', 'b'], ['regnare', 'verb', 'b'], ['regno', 'noun', 'a'], ['regola', 'noun', 'a'], ['regolamento', 'noun', 'b'], ['regolare', 'adjective', 'b'], ['regolare', 'noun', 'b'], ['regolare', 'verb', 'b'], ['regolarmente', 'adverb', 'b'], ['relativamente', 'adverb', 'b'], ['relativo', 'adjective', 'a'], ['relazione', 'noun', 'a'], ['religione', 'noun', 'a'], ['religioso', 'adjective', 'a'], ['religioso', 'noun', 'a'], ['remare', 'verb', 'c'], ['remo', 'noun', 'c'], ['remoto', 'adjective', 'b'], ['rendere', 'verb', 'a'], ['rene', 'noun', 'b'], ['reparto', 'noun', 'b'], ['repertorio', 'noun', 'b'], ['replica', 'noun', 'b'], ['replicare', 'verb', 'b'], ['repressione', 'noun', 'c'], ['reprimere', 'verb', 'c'], ['repubblica', 'noun', 'a'], ['repubblicano', 'adjective', 'b'], ['repubblicano', 'noun', 'b'], ['requisito', 'noun', 'b'], ['resa', 'noun', 'b'], ['residente', 'adjective', 'b'], ['residente', 'noun', 'b'], ['residenza', 'noun', 'b'], ['residuo', 'adjective', 'b'], ['residuo', 'noun', 'b'], ['resistente', 'pres_part', 'b'], ['resistente', 'adjective', 'b'], ['resistente', 'noun', 'b'], ['resistenza', 'noun', 'b'], ['resistere', 'verb', 'a'], ['resoconto', 'noun', 'c'], ['respingere', 'verb', 'b'], ['respirare', 'verb', 'a'], ['respirazione', 'noun', 'c'], ['respiro', 'noun', 'b'], ['responsabile', 'adjective', 'a'], ['responsabile', 'noun', 'a'], ['responsabilità', 'noun', 'a'], ['restare', 'verb', 'a'], ['restituire', 'verb', 'b'], ['resto', 'noun', 'a'], ['restringere', 'verb', 'b'], ['rete', 'noun', 'a'], ['retorica', 'noun', 'b'], ['retro', 'adverb', 'b'], ['retro', 'noun', 'b'], ['retta', 'noun', 'b'], ['rettangolare', 'adjective', 'c'], ['rettile', 'noun', 'c'], ['rettile', 'adjective', 'c'], ['retto', 'adjective', 'b'], ['retto', 'noun', 'b'], ['revisione', 'noun', 'b'], ['rialzare', 'verb', 'b'], ['riaprire', 'verb', 'b'], ['riassumere', 'verb', 'b'], ['ribadire', 'verb', 'b'], ['ribattere', 'verb', 'b'], ['ribellare', 'verb', 'b'], ['ribelle', 'adjective', 'b'], ['ribelle', 'noun', 'b'], ['ricadere', 'verb', 'b'], ['ricaduta', 'noun', 'c'], ['ricalcare', 'verb', 'c'], ['ricamare', 'verb', 'c'], ['ricambiare', 'verb', 'b'], ['ricambio', 'noun', 'c'], ['ricamo', 'noun', 'c'], ['ricarica', 'noun', 'c'], ['ricavare', 'verb', 'b'], ['ricchezza', 'noun', 'b'], ['riccio', 'adjective', 'c'], ['riccio', 'noun', 'c'], ['ricciolo', 'adjective', 'c'], ['ricciolo', 'noun', 'c'], ['ricco', 'adjective', 'a'], ['ricerca', 'noun', 'a'], ['ricercare', 'verb', 'b'], ['ricercatore', 'adjective', 'b'], ['ricercatore', 'noun', 'b'], ['ricetta', 'noun', 'a'], ['ricevere', 'verb', 'a'], ['ricevimento', 'noun', 'c'], ['ricevuta', 'noun', 'b'], ['richiamare', 'verb', 'a'], ['richiamo', 'noun', 'b'], ['richiedere', 'verb', 'a'], ['richiesta', 'noun', 'a'], ['richiudere', 'verb', 'b'], ['ricominciare', 'verb', 'a'], ['ricompensa', 'noun', 'c'], ['ricompensare', 'verb', 'c'], ['riconciliarsi', 'verb', 'c'], ['riconoscere', 'verb', 'a'], ['riconoscimento', 'noun', 'b'], ['ricopiare', 'verb', 'c'], ['ricoprire', 'verb', 'b'], ['ricordare', 'verb', 'a'], ['ricordo', 'noun', 'a'], ['ricorrere', 'verb', 'b'], ['ricorso', 'noun', 'b'], ['ricostruire', 'verb', 'b'], ['ricostruzione', 'noun', 'b'], ['ricotta', 'noun', 'c'], ['ricoverare', 'verb', 'b'], ['ricovero', 'noun', 'c'], ['ricreazione', 'noun', 'c'], ['ridare', 'verb', 'b'], ['ridere', 'verb', 'a'], ['ridere', 'noun', 'a'], ['ridicolo', 'adjective', 'b'], ['ridicolo', 'noun', 'b'], ['ridotto', 'past_part', 'b'], ['ridotto', 'adjective', 'b'], ['ridotto', 'noun', 'b'], ['ridurre', 'verb', 'a'], ['riduzione', 'noun', 'b'], ['riempire', 'verb', 'a'], ['rientrare', 'verb', 'a'], ['rientro', 'noun', 'b'], ['rifare', 'verb', 'a'], ['riferimento', 'noun', 'a'], ['riferire', 'verb', 'a'], ['rifinire', 'verb', 'c'], ['rifiutare', 'verb', 'a'], ['rifiuto', 'noun', 'a'], ['riflessione', 'noun', 'a'], ['riflesso', 'noun', 'b'], ['riflettere', 'verb', 'a'], ['riflettore', 'noun', 'c'], ['riflettore', 'adjective', 'c'], ['riforma', 'noun', 'b'], ['rifornimento', 'noun', 'c'], ['rifugiare', 'verb', 'b'], ['rifugio', 'noun', 'b'], ['riga', 'noun', 'a'], ['rigattiere', 'noun', 'c'], ['rigido', 'adjective', 'b'], ['rigore', 'noun', 'b'], ['rigoroso', 'adjective', 'b'], ['rigovernare', 'verb', 'c'], ['riguardare', 'verb', 'a'], ['riguardo', 'noun', 'a'], ['rilasciare', 'verb', 'b'], ['rilassare', 'verb', 'a'], ['rilegare', 'verb', 'c'], ['rileggere', 'verb', 'b'], ['rilevante', 'pres_part', 'b'], ['rilevante', 'adjective', 'b'], ['rilevare', 'verb', 'b'], ['rilievo', 'noun', 'b'], ['rima', 'noun', 'b'], ['rimandare', 'verb', 'b'], ['rimanenza', 'noun', 'c'], ['rimanere', 'verb', 'a'], ['rimbombare', 'verb', 'c'], ['rimborsare', 'verb', 'c'], ['rimediare', 'verb', 'b'], ['rimedio', 'noun', 'b'], ['rimettere', 'verb', 'a'], ['rimodernare', 'verb', 'c'], ['rimorchio', 'noun', 'c'], ['rimpiangere', 'verb', 'b'], ['rimproverare', 'verb', 'b'], ['rimprovero', 'noun', 'c'], ['rimuovere', 'verb', 'b'], ['rinascere', 'verb', 'b'], ['rinascimento', 'noun', 'b'], ['rinascimento', 'adjective', 'b'], ['rincarare', 'verb', 'c'], ['rinchiudere', 'verb', 'b'], ['rincorsa', 'noun', 'c'], ['rinforzo', 'noun', 'c'], ['rinfresco', 'noun', 'c'], ['ringhiare', 'verb', 'c'], ['ringhiera', 'noun', 'c'], ['ringhio', 'noun', 'c'], ['ringiovanire', 'verb', 'c'], ['ringraziare', 'verb', 'a'], ['rinnegare', 'verb', 'c'], ['rinnovare', 'verb', 'b'], ['rinoceronte', 'noun', 'c'], ['rintracciare', 'verb', 'b'], ['rinuncia', 'noun', 'c'], ['rinunciare', 'verb', 'a'], ['rinvenire', 'verb', 'b'], ['rinviare', 'verb', 'b'], ['rinvio', 'noun', 'c'], ['rione', 'noun', 'c'], ['riordinare', 'verb', 'c'], ['riparare', 'verb', 'b'], ['riparo', 'noun', 'b'], ['ripartire', 'verb', 'b'], ['ripartire', 'verb', 'b'], ['ripensamento', 'noun', 'c'], ['ripensare', 'verb', 'b'], ['ripetente', 'pres_part', 'c'], ['ripetente', 'adjective', 'c'], ['ripetente', 'noun', 'c'], ['ripetere', 'verb', 'a'], ['ripetizione', 'noun', 'b'], ['ripido', 'adjective', 'c'], ['ripiego', 'noun', 'c'], ['ripieno', 'adjective', 'c'], ['ripieno', 'noun', 'c'], ['riportare', 'verb', 'a'], ['riposare', 'verb', 'b'], ['riposo', 'noun', 'b'], ['riposo', 'loc-comando', 'b'], ['riposo', 'noun', 'b'], ['riprendere', 'verb', 'a'], ['ripresa', 'noun', 'b'], ['riprodurre', 'verb', 'b'], ['riproduzione', 'noun', 'a'], ['riproporre', 'verb', 'b'], ['riprovare', 'verb', 'b'], ['ripulire', 'verb', 'b'], ['risaia', 'noun', 'c'], ['risalire', 'verb', 'a'], ['risarcimento', 'noun', 'b'], ['risata', 'noun', 'b'], ['riscaldamento', 'noun', 'b'], ['riscaldare', 'verb', 'b'], ['riscattare', 'verb', 'c'], ['riscatto', 'noun', 'c'], ['rischiare', 'verb', 'a'], ['rischio', 'noun', 'a'], ['rischioso', 'adjective', 'b'], ['risciacquare', 'verb', 'c'], ['riscontrare', 'verb', 'b'], ['riscontro', 'noun', 'b'], ['riscuotere', 'verb', 'b'], ['risentimento', 'noun', 'c'], ['risentire', 'verb', 'b'], ['riserva', 'noun', 'b'], ['riservare', 'verb', 'a'], ['riservato', 'past_part', 'a'], ['riservato', 'adjective', 'a'], ['risiedere', 'verb', 'b'], ['riso', 'noun', 'b'], ['risoluzione', 'noun', 'b'], ['risolvere', 'verb', 'a'], ['risonanza', 'noun', 'b'], ['risorsa', 'noun', 'a'], ['risparmiare', 'verb', 'b'], ['risparmio', 'noun', 'b'], ['rispettare', 'verb', 'a'], ['rispettivamente', 'adverb', 'b'], ['rispettivo', 'adjective', 'b'], ['rispetto', 'noun', 'a'], ['risplendere', 'verb', 'c'], ['rispondere', 'verb', 'a'], ['risposta', 'noun', 'a'], ['rissa', 'noun', 'b'], ['ristampare', 'verb', 'c'], ['ristorante', 'noun', 'a'], ['ristretto', 'past_part', 'b'], ['ristretto', 'adjective', 'b'], ['ristretto', 'noun', 'b'], ['risultare', 'verb', 'a'], ['risultato', 'past_part', 'a'], ['risultato', 'adjective', 'a'], ['risultato', 'noun', 'a'], ['risvegliare', 'verb', 'b'], ['risveglio', 'noun', 'b'], ['ritagliare', 'verb', 'b'], ['ritardare', 'verb', 'b'], ['ritardo', 'noun', 'a'], ['ritenere', 'verb', 'a'], ['ritirare', 'verb', 'a'], ['ritirata', 'noun', 'c'], ['ritiro', 'noun', 'b'], ['ritmo', 'noun', 'a'], ['rito', 'noun', 'b'], ['ritoccare', 'verb', 'c'], ['ritornare', 'verb', 'a'], ['ritornello', 'noun', 'c'], ['ritorno', 'noun', 'a'], ['ritrarre', 'verb', 'b'], ['ritratto', 'past_part', 'b'], ['ritratto', 'adjective', 'b'], ['ritratto', 'noun', 'b'], ['ritrovare', 'verb', 'a'], ['ritrovo', 'noun', 'c'], ['ritto', 'adjective', 'c'], ['ritto', 'noun', 'c'], ['ritto', 'adverb', 'c'], ['ritto', 'preposition', 'c'], ['rituale', 'adjective', 'b'], ['rituale', 'noun', 'b'], ['riunione', 'noun', 'a'], ['riunire', 'verb', 'a'], ['riunito', 'past_part', 'c'], ['riunito', 'adjective', 'c'], ['riunito', 'noun', 'c'], ['riuscire', 'verb', 'a'], ['riuscita', 'noun', 'c'], ['riva', 'noun', 'b'], ['rivale', 'adjective', 'b'], ['rivale', 'noun', 'b'], ['rivedere', 'verb', 'a'], ['rivelare', 'verb', 'a'], ['rivelazione', 'noun', 'b'], ['rivendicare', 'verb', 'b'], ['rivendita', 'noun', 'c'], ['rivestimento', 'noun', 'c'], ['rivestire', 'verb', 'b'], ['rivincita', 'noun', 'c'], ['rivista', 'noun', 'a'], ['rivisto', 'past_part', 'b'], ['rivisto', 'adjective', 'b'], ['rivolgere', 'verb', 'a'], ['rivolta', 'noun', 'b'], ['rivoltare', 'verb', 'c'], ['rivoluzionario', 'adjective', 'b'], ['rivoluzionario', 'noun', 'b'], ['rivoluzione', 'noun', 'a'], ['roba', 'noun', 'a'], ['robot', 'noun', 'b'], ['robusto', 'adjective', 'b'], ['rocca', 'noun', 'c'], ['rocchetto', 'noun', 'c'], ['roccia', 'noun', 'b'], ['roccioso', 'adjective', 'c'], ['rock', 'noun', 'b'], ['rock', 'adjective', 'b'], ['rodaggio', 'noun', 'c'], ['rodere', 'verb', 'c'], ['romagnolo', 'adjective', 'c'], ['romagnolo', 'noun', 'c'], ['romano', 'adjective', 'a'], ['romano', 'noun', 'a'], ['romantico', 'adjective', 'b'], ['romantico', 'noun', 'b'], ['romanzo', 'noun', 'a'], ['rombo', 'noun', 'c'], ['romeno', 'adjective', 'c'], ['romeno', 'noun', 'c'], ['rompere', 'verb', 'a'], ['rondine', 'noun', 'c'], ['ronzare', 'verb', 'c'], ['ronzio', 'noun', 'c'], ['rosa', 'noun', 'a'], ['rosa', 'adjective', 'a'], ['rosario', 'noun', 'c'], ['rosato', 'adjective', 'c'], ['rosato', 'noun', 'c'], ['roseo', 'adjective', 'c'], ['roseo', 'noun', 'c'], ['rosetta', 'noun', 'c'], ['rosmarino', 'noun', 'c'], ['rosolia', 'noun', 'c'], ['rosso', 'adjective', 'a'], ['rosso', 'noun', 'a'], ['rossore', 'noun', 'c'], ['rosticceria', 'noun', 'c'], ['rotaia', 'noun', 'c'], ['rotella', 'noun', 'c'], ['rotolare', 'verb', 'c'], ['rotondo', 'adjective', 'b'], ['rotondo', 'noun', 'b'], ['rotta', 'noun', 'b'], ['rotto', 'past_part', 'b'], ['rotto', 'adjective', 'b'], ['rotto', 'noun', 'b'], ['rottura', 'noun', 'b'], ['roulotte', 'noun', 'c'], ['rovesciare', 'verb', 'b'], ['rovescio', 'adjective', 'b'], ['rovescio', 'noun', 'b'], ['rovina', 'noun', 'b'], ['rovinare', 'verb', 'a'], ['rovo', 'noun', 'c'], ['rozzo', 'adjective', 'c'], ['rubare', 'verb', 'a'], ['rubinetto', 'noun', 'c'], ['rubrica', 'noun', 'b'], ['rude', 'adjective', 'c'], ['ruga', 'noun', 'c'], ['ruggine', 'noun', 'c'], ['ruggine', 'adjective', 'c'], ['ruggire', 'verb', 'c'], ['ruggito', 'past_part', 'c'], ['ruggito', 'noun', 'c'], ['rullo', 'noun', 'c'], ['rumeno', 'adjective', 'c'], ['rumeno', 'noun', 'c'], ['ruminante', 'pres_part', 'c'], ['ruminante', 'adjective', 'c'], ['ruminante', 'noun', 'c'], ['rumore', 'noun', 'a'], ['ruolo', 'noun', 'a'], ['ruota', 'noun', 'b'], ['ruotare', 'verb', 'b'], ['ruscello', 'noun', 'c'], ['ruspa', 'noun', 'c'], ['russare', 'verb', 'c'], ['russo', 'adjective', 'a'], ['russo', 'noun', 'a'], ['rustico', 'adjective', 'c'], ['rustico', 'noun', 'c'], ['ruttare', 'verb', 'c'], ['rutto', 'noun', 'c'], ['sabato', 'noun', 'a'], ['sabbia', 'noun', 'b'], ['sabbia', 'adjective', 'b'], ['sabotare', 'verb', 'c'], ['saccheggiare', 'verb', 'c'], ['sacchetto', 'noun', 'b'], ['sacco', 'noun', 'a'], ['sacerdote', 'noun', 'b'], ['sacrificare', 'verb', 'b'], ['sacrificio', 'noun', 'b'], ['sacro', 'adjective', 'b'], ['sacro', 'noun', 'b'], ['safari', 'noun', 'c'], ['saga', 'noun', 'b'], ['saggezza', 'noun', 'b'], ['saggio', 'adjective', 'b'], ['saggio', 'noun', 'b'], ['saggio', 'noun', 'b'], ['sagra', 'noun', 'c'], ['sagrestano', 'noun', 'c'], ['sagrestano', 'adjective', 'c'], ['sala', 'noun', 'a'], ['salame', 'noun', 'c'], ['salare', 'verb', 'c'], ['salario', 'adjective', 'b'], ['salario', 'noun', 'b'], ['salatino', 'noun', 'c'], ['salato', 'past_part', 'b'], ['salato', 'adjective', 'b'], ['salato', 'noun', 'b'], ['saldatura', 'noun', 'c'], ['sale', 'noun', 'b'], ['salice', 'noun', 'c'], ['saliera', 'noun', 'c'], ['salire', 'verb', 'a'], ['salita', 'noun', 'b'], ['saliva', 'noun', 'c'], ['salmone', 'noun', 'c'], ['salmone', 'adjective', 'c'], ['salone', 'noun', 'b'], ['salotto', 'noun', 'b'], ['salsa', 'noun', 'b'], ['salsiccia', 'noun', 'c'], ['saltare', 'verb', 'a'], ['saltellare', 'verb', 'c'], ['salto', 'noun', 'b'], ['salume', 'noun', 'c'], ['salutare', 'verb', 'a'], ['salutare', 'noun', 'a'], ['salute', 'noun', 'a'], ['salute', 'exclamation', 'a'], ['saluto', 'noun', 'a'], ['salvadanaio', 'noun', 'c'], ['salvagente', 'noun', 'c'], ['salvare', 'verb', 'a'], ['salvaslip', 'noun', 'c'], ['salvatore', 'adjective', 'b'], ['salvatore', 'noun', 'b'], ['salve', 'exclamation', 'b'], ['salvezza', 'noun', 'b'], ['salvia', 'noun', 'c'], ['salvietta', 'noun', 'c'], ['salvo', 'adjective', 'a'], ['salvo', 'preposition', 'a'], ['sandalo', 'noun', 'c'], ['sangue', 'noun', 'a'], ['sangue', 'adjective', 'a'], ['sanguinare', 'verb', 'c'], ['sanguisuga', 'noun', 'c'], ['sanità', 'noun', 'b'], ['sanitaria', 'noun', 'c'], ['sanitario', 'adjective', 'b'], ['sanitario', 'noun', 'b'], ['sano', 'adjective', 'a'], ['santo', 'adjective', 'a'], ['santo', 'noun', 'a'], ['sanzione', 'noun', 'b'], ['sapere', 'verb', 'a'], ['sapere', 'noun', 'b'], ['sapiente', 'adjective', 'c'], ['sapiente', 'noun', 'c'], ['sapone', 'noun', 'b'], ['saponetta', 'noun', 'c'], ['sapore', 'noun', 'b'], ['saporito', 'past_part', 'c'], ['saporito', 'adjective', 'c'], ['sardina', 'noun', 'c'], ['sardo', 'adjective', 'b'], ['sardo', 'noun', 'b'], ['sarto', 'noun', 'c'], ['sasso', 'noun', 'b'], ['satellite', 'noun', 'b'], ['sazio', 'past_part', 'c'], ['sazio', 'adjective', 'c'], ['sbadato', 'adjective', 'c'], ['sbadato', 'noun', 'c'], ['sbadigliare', 'verb', 'c'], ['sbadiglio', 'noun', 'c'], ['sbagliare', 'verb', 'a'], ['sbagliato', 'past_part', 'a'], ['sbagliato', 'adjective', 'a'], ['sbaglio', 'noun', 'b'], ['sbarbare', 'verb', 'c'], ['sbarcare', 'verb', 'b'], ['sbarra', 'noun', 'c'], ['sbarramento', 'noun', 'c'], ['sbattere', 'verb', 'a'], ['sberla', 'noun', 'c'], ['sbiadire', 'verb', 'c'], ['sbiancare', 'verb', 'c'], ['sbigottire', 'verb', 'c'], ['sbloccare', 'verb', 'c'], ['sboccare', 'verb', 'c'], ['sbocciare', 'verb', 'c'], ['sbocco', 'noun', 'c'], ['sbornia', 'noun', 'c'], ['sbottonare', 'verb', 'c'], ['sbriciolare', 'verb', 'c'], ['sbrigare', 'verb', 'b'], ['sbronza', 'noun', 'c'], ['sbronzo', 'adjective', 'c'], ['sbucciare', 'verb', 'c'], ['sbuffare', 'verb', 'c'], ['scacchiera', 'noun', 'c'], ['scadenza', 'noun', 'b'], ['scadere', 'verb', 'b'], ['scaffale', 'noun', 'b'], ['scafo', 'noun', 'c'], ['scala', 'noun', 'a'], ['scalare', 'verb', 'b'], ['scalata', 'noun', 'c'], ['scaldabagno', 'noun', 'c'], ['scaldare', 'verb', 'b'], ['scalinata', 'noun', 'c'], ['scalino', 'noun', 'c'], ['scalpello', 'noun', 'c'], ['scalzo', 'adjective', 'c'], ['scambiare', 'verb', 'a'], ['scambio', 'noun', 'a'], ['scamorza', 'noun', 'c'], ['scampagnata', 'noun', 'c'], ['scampo', 'noun', 'c'], ['scandalizzare', 'verb', 'c'], ['scandalo', 'noun', 'b'], ['scandire', 'verb', 'b'], ['scansare', 'verb', 'c'], ['scapito', 'noun', 'c'], ['scappamento', 'noun', 'c'], ['scappare', 'verb', 'a'], ['scappatoia', 'noun', 'c'], ['scarabocchiare', 'verb', 'c'], ['scarabocchio', 'noun', 'c'], ['scarafaggio', 'noun', 'c'], ['scarcerare', 'verb', 'c'], ['scaricare', 'verb', 'a'], ['scaricatore', 'noun', 'c'], ['scarico', 'noun', 'b'], ['scarlattina', 'noun', 'c'], ['scarpa', 'noun', 'a'], ['scarpiera', 'noun', 'c'], ['scarpone', 'noun', 'c'], ['scarso', 'adjective', 'b'], ['scartare', 'verb', 'b'], ['scatenare', 'verb', 'b'], ['scatola', 'noun', 'a'], ['scattare', 'verb', 'a'], ['scatto', 'noun', 'b'], ['scavalcare', 'verb', 'c'], ['scavare', 'verb', 'b'], ['scavo', 'noun', 'c'], ['scegliere', 'verb', 'a'], ['scelta', 'noun', 'a'], ['scemo', 'past_part', 'b'], ['scemo', 'adjective', 'b'], ['scemo', 'noun', 'b'], ['scena', 'noun', 'a'], ['scenario', 'noun', 'b'], ['scendere', 'verb', 'a'], ['sceneggiatura', 'noun', 'b'], ['sceriffo', 'noun', 'c'], ['scheda', 'noun', 'b'], ['schedario', 'noun', 'c'], ['scheggia', 'noun', 'c'], ['scheletro', 'noun', 'c'], ['schema', 'noun', 'b'], ['schermo', 'noun', 'a'], ['scherzare', 'verb', 'a'], ['scherzo', 'noun', 'b'], ['scherzoso', 'adjective', 'c'], ['schiacciare', 'verb', 'b'], ['schiacciato', 'past_part', 'c'], ['schiacciato', 'adjective', 'c'], ['schiaffo', 'noun', 'b'], ['schiavo', 'adjective', 'b'], ['schiavo', 'noun', 'b'], ['schiena', 'noun', 'a'], ['schierare', 'verb', 'b'], ['schietto', 'adjective', 'c'], ['schifo', 'noun', 'a'], ['schifo', 'adjective', 'a'], ['schiuma', 'noun', 'c'], ['schizzare', 'verb', 'b'], ['schizzo', 'noun', 'b'], ['sci', 'noun', 'b'], ['scia', 'noun', 'b'], ['sciacquare', 'verb', 'c'], ['scialle', 'noun', 'c'], ['sciame', 'noun', 'c'], ['sciare', 'verb', 'c'], ['sciarpa', 'noun', 'c'], ['sciatore', 'noun', 'c'], ['scientifico', 'adjective', 'a'], ['scientifico', 'noun', 'a'], ['scienza', 'noun', 'a'], ['scienziato', 'noun', 'b'], ['scienziato', 'adjective', 'b'], ['scimmia', 'noun', 'b'], ['scintilla', 'noun', 'b'], ['sciocchezza', 'noun', 'b'], ['sciocco', 'adjective', 'b'], ['sciocco', 'noun', 'b'], ['sciogliere', 'verb', 'b'], ['scioperare', 'verb', 'c'], ['sciopero', 'noun', 'b'], ['scirocco', 'noun', 'c'], ['sciroppo', 'noun', 'c'], ['scivolare', 'verb', 'b'], ['scivolata', 'noun', 'c'], ['scivolo', 'noun', 'c'], ['scocciare', 'verb', 'c'], ['scodella', 'noun', 'c'], ['scodinzolare', 'verb', 'c'], ['scoglio', 'noun', 'c'], ['scoiattolo', 'noun', 'c'], ['scolapiatti', 'noun', 'c'], ['scolaro', 'noun', 'c'], ['scolastico', 'adjective', 'b'], ['scolastico', 'noun', 'b'], ['scolpire', 'verb', 'c'], ['scommessa', 'noun', 'b'], ['scommettere', 'verb', 'b'], ['scomodo', 'adjective', 'c'], ['scomparire', 'verb', 'a'], ['scomparsa', 'noun', 'b'], ['scompartimento', 'noun', 'c'], ['sconfiggere', 'verb', 'b'], ['sconfitta', 'noun', 'b'], ['scongelare', 'verb', 'c'], ['sconosciuto', 'past_part', 'a'], ['sconosciuto', 'adjective', 'a'], ['sconsigliare', 'verb', 'c'], ['scontato', 'past_part', 'b'], ['scontato', 'adjective', 'b'], ['scontento', 'adjective', 'c'], ['sconto', 'noun', 'b'], ['scontrare', 'verb', 'b'], ['scontro', 'noun', 'b'], ['sconvolgere', 'verb', 'b'], ['scopa', 'noun', 'c'], ['scopare', 'verb', 'b'], ['scoperta', 'noun', 'a'], ['scopo', 'noun', 'a'], ['scoppiare', 'verb', 'a'], ['scoprire', 'verb', 'a'], ['scordare', 'verb', 'b'], ['scorgere', 'verb', 'b'], ['scorpione', 'noun', 'c'], ['scorrere', 'verb', 'a'], ['scorretto', 'adjective', 'c'], ['scorso', 'past_part', 'a'], ['scorso', 'adjective', 'a'], ['scorso', 'noun', 'a'], ['scorta', 'noun', 'b'], ['scortese', 'adjective', 'c'], ['scossa', 'noun', 'c'], ['scout', 'noun', 'c'], ['scout', 'adjective', 'c'], ['scozzese', 'adjective', 'c'], ['scozzese', 'noun', 'c'], ['screpolare', 'verb', 'c'], ['scricchiolare', 'verb', 'c'], ['scritta', 'noun', 'b'], ['scritto', 'past_part', 'b'], ['scritto', 'adjective', 'b'], ['scritto', 'noun', 'b'], ['scrittore', 'noun', 'a'], ['scrittura', 'noun', 'a'], ['scrivania', 'noun', 'b'], ['scrivere', 'verb', 'a'], ['scrofa', 'noun', 'c'], ['scrupolo', 'noun', 'c'], ['scudetto', 'noun', 'c'], ['scudo', 'noun', 'b'], ['scultore', 'noun', 'c'], ['scultura', 'noun', 'b'], ['scuola', 'noun', 'a'], ['scuotere', 'verb', 'b'], ['scure', 'noun', 'c'], ['scurire', 'verb', 'c'], ['scuro', 'adjective', 'b'], ['scuro', 'noun', 'b'], ['scuro', 'adverb', 'b'], ['scusa', 'noun', 'a'], ['scusare', 'verb', 'a'], ['sdebitarsi', 'verb', 'c'], ['sdegnare', 'verb', 'c'], ['sdraiare', 'verb', 'b'], ['sdraiato', 'past_part', 'c'], ['sdraiato', 'adjective', 'c'], ['se', 'pronoun', 'a'], ['se', 'conjunction', 'a'], ['se', 'noun', 'a'], ['sebbene', 'conjunction', 'b'], ['seccare', 'verb', 'b'], ['seccatura', 'noun', 'c'], ['secchio', 'noun', 'b'], ['secchione', 'noun', 'b'], ['secco', 'adjective', 'a'], ['secco', 'noun', 'a'], ['secolo', 'noun', 'a'], ['seconda', 'noun', 'b'], ['secondario', 'adjective', 'b'], ['secondario', 'noun', 'b'], ['secondo', 'adjective', 'a'], ['secondo', 'noun', 'a'], ['secondo', 'adverb', 'a'], ['secondo', 'preposition', 'a'], ['secondo', 'conjunction', 'a'], ['sedano', 'noun', 'c'], ['sede', 'noun', 'a'], ['sedere', 'verb', 'a'], ['sedia', 'noun', 'a'], ['sedici', 'adjective', 'b'], ['sedici', 'noun', 'b'], ['sedile', 'noun', 'b'], ['sedurre', 'verb', 'b'], ['seduta', 'noun', 'b'], ['seduttore', 'adjective', 'c'], ['seduttore', 'noun', 'c'], ['seggiolino', 'noun', 'c'], ['seggiovia', 'noun', 'c'], ['segheria', 'noun', 'c'], ['segmento', 'noun', 'b'], ['segnalare', 'verb', 'a'], ['segnalazione', 'noun', 'b'], ['segnale', 'noun', 'a'], ['segnare', 'verb', 'a'], ['segno', 'noun', 'a'], ['segretaria', 'noun', 'b'], ['segretario', 'noun', 'b'], ['segreteria', 'noun', 'b'], ['segreto', 'noun', 'a'], ['segreto', 'adjective', 'a'], ['segreto', 'noun', 'a'], ['segreto', 'adverb', 'a'], ['seguente', 'pres_part', 'a'], ['seguente', 'adjective', 'a'], ['seguente', 'noun', 'a'], ['seguire', 'verb', 'a'], ['seguito', 'noun', 'a'], ['sei', 'adjective', 'a'], ['sei', 'noun', 'a'], ['selezionare', 'verb', 'b'], ['selezione', 'noun', 'b'], ['selva', 'noun', 'c'], ['selvaggina', 'noun', 'c'], ['selvaggio', 'adjective', 'b'], ['selvaggio', 'noun', 'b'], ['semaforo', 'noun', 'c'], ['semantico', 'adjective', 'b'], ['sembrare', 'verb', 'a'], ['seme', 'noun', 'b'], ['semestre', 'noun', 'c'], ['semifreddo', 'adjective', 'c'], ['semifreddo', 'noun', 'c'], ['seminare', 'verb', 'b'], ['semmai', 'conjunction', 'b'], ['semmai', 'adverb', 'b'], ['semolino', 'noun', 'c'], ['semplice', 'adjective', 'a'], ['semplice', 'noun', 'a'], ['semplicemente', 'adverb', 'a'], ['semplicità', 'noun', 'b'], ['semplificare', 'verb', 'b'], ['sempre', 'adverb', 'a'], ['senape', 'noun', 'c'], ['senape', 'adjective', 'c'], ['senato', 'noun', 'b'], ['senatore', 'noun', 'b'], ['sennò', 'adverb', 'b'], ['seno', 'noun', 'a'], ['sensazione', 'noun', 'a'], ['sensibile', 'adjective', 'b'], ['sensibile', 'noun', 'b'], ['sensibilità', 'noun', 'b'], ['senso', 'noun', 'a'], ['sensuale', 'adjective', 'b'], ['sentenza', 'noun', 'a'], ['sentiero', 'noun', 'b'], ['sentimentale', 'adjective', 'b'], ['sentimentale', 'noun', 'b'], ['sentimento', 'noun', 'a'], ['sentire', 'verb', 'a'], ['sentito', 'past_part', 'b'], ['sentito', 'adjective', 'b'], ['senza', 'preposition', 'a'], ['senza', 'conjunction', 'a'], ['separare', 'verb', 'a'], ['separato', 'past_part', 'b'], ['separato', 'adjective', 'b'], ['separato', 'noun', 'b'], ['separazione', 'noun', 'b'], ['sepolto', 'past_part', 'b'], ['sepolto', 'adjective', 'b'], ['sepolto', 'noun', 'b'], ['seppellire', 'verb', 'b'], ['seppia', 'noun', 'c'], ['seppia', 'adjective', 'c'], ['seppia', 'noun', 'c'], ['sequenza', 'noun', 'b'], ['sequestrare', 'verb', 'b'], ['sequestro', 'noun', 'b'], ['sera', 'noun', 'a'], ['serata', 'noun', 'a'], ['serbo', 'adjective', 'c'], ['serbo', 'noun', 'c'], ['serenata', 'noun', 'c'], ['serenità', 'noun', 'b'], ['sereno', 'adjective', 'a'], ['sereno', 'noun', 'a'], ['sergente', 'noun', 'b'], ['seriamente', 'adverb', 'b'], ['serie', 'noun', 'a'], ['serietà', 'noun', 'c'], ['serio', 'adjective', 'a'], ['serio', 'noun', 'a'], ['serpente', 'noun', 'b'], ['serra', 'noun', 'b'], ['servire', 'verb', 'a'], ['servizio', 'noun', 'a'], ['servo', 'noun', 'b'], ['servo', 'adjective', 'b'], ['sessanta', 'adjective', 'b'], ['sessanta', 'noun', 'b'], ['sesso', 'noun', 'a'], ['sessuale', 'adjective', 'a'], ['sesto', 'adjective', 'b'], ['sesto', 'noun', 'b'], ['set', 'noun', 'b'], ['seta', 'noun', 'b'], ['sete', 'noun', 'b'], ['setta', 'noun', 'b'], ['settanta', 'adjective', 'b'], ['settanta', 'noun', 'b'], ['sette', 'adjective', 'a'], ['sette', 'noun', 'a'], ['settembre', 'noun', 'a'], ['settentrione', 'noun', 'c'], ['settimana', 'noun', 'a'], ['settimanale', 'adjective', 'b'], ['settimanale', 'noun', 'b'], ['settimo', 'adjective', 'b'], ['settimo', 'noun', 'b'], ['settore', 'noun', 'a'], ['severo', 'adjective', 'b'], ['sexy', 'adjective', 'b'], ['sezione', 'noun', 'a'], ['sfera', 'noun', 'b'], ['sfida', 'noun', 'a'], ['sfidare', 'verb', 'b'], ['sfiducia', 'noun', 'c'], ['sfigato', 'adjective', 'b'], ['sfigato', 'noun', 'b'], ['sfilare', 'verb', 'b'], ['sfilata', 'noun', 'b'], ['sfinire', 'verb', 'c'], ['sfiorare', 'verb', 'b'], ['sfociare', 'verb', 'c'], ['sfogare', 'verb', 'b'], ['sfoglia', 'noun', 'c'], ['sfogliare', 'verb', 'b'], ['sfogo', 'noun', 'b'], ['sfollamento', 'noun', 'c'], ['sfollare', 'verb', 'c'], ['sfondare', 'verb', 'b'], ['sfondo', 'noun', 'b'], ['sfortunato', 'adjective', 'c'], ['sforzare', 'verb', 'b'], ['sforzo', 'noun', 'a'], ['sfrenato', 'past_part', 'c'], ['sfrenato', 'adjective', 'c'], ['sfruttare', 'verb', 'a'], ['sfuggire', 'verb', 'a'], ['sgabello', 'noun', 'c'], ['sganciare', 'verb', 'c'], ['sgarbato', 'adjective', 'c'], ['sgarbato', 'noun', 'c'], ['sgarbo', 'noun', 'c'], ['sgombro', 'noun', 'c'], ['sgomento', 'noun', 'c'], ['sgonfiare', 'verb', 'c'], ['sgozzare', 'verb', 'c'], ['sgrassare', 'verb', 'c'], ['sgrassatore', 'noun', 'c'], ['sgridare', 'verb', 'c'], ['sguardo', 'noun', 'a'], ['shampoo', 'noun', 'c'], ['share', 'noun', 'b'], ['shopping', 'noun', 'b'], ['shorts', 'noun', 'c'], ['show', 'noun', 'b'], ['sì', 'adverb', 'a'], ['sì', 'noun', 'a'], ['sì', 'adjective', 'a'], ['si', 'pronoun', 'a'], ['sia', 'conjunction', 'a'], ['siamese', 'adjective', 'c'], ['siamese', 'noun', 'c'], ['sicché', 'conjunction', 'b'], ['siccità', 'noun', 'c'], ['siccome', 'conjunction', 'a'], ['siccome', 'adverb', 'a'], ['siciliano', 'adjective', 'b'], ['siciliano', 'noun', 'b'], ['sicuramente', 'adverb', 'a'], ['sicurezza', 'noun', 'a'], ['sicuro', 'adjective', 'a'], ['sicuro', 'noun', 'a'], ['sicuro', 'adverb', 'a'], ['siepe', 'noun', 'c'], ['sigaretta', 'noun', 'a'], ['sigaro', 'noun', 'c'], ['sigla', 'noun', 'b'], ['significare', 'verb', 'a'], ['significativo', 'adjective', 'b'], ['significato', 'past_part', 'a'], ['significato', 'noun', 'a'], ['signora', 'noun', 'a'], ['signore', 'noun', 'a'], ['signorina', 'noun', 'a'], ['silenzio', 'noun', 'a'], ['silenzioso', 'adjective', 'b'], ['sillaba', 'noun', 'c'], ['simbolico', 'adjective', 'b'], ['simbolo', 'noun', 'a'], ['simile', 'adjective', 'a'], ['simile', 'adjective', 'a'], ['simile', 'noun', 'a'], ['simile', 'adverb', 'a'], ['simpatia', 'noun', 'b'], ['simpatico', 'adjective', 'a'], ['simulare', 'verb', 'b'], ['sinceramente', 'adverb', 'b'], ['sincero', 'adjective', 'b'], ['sindacale', 'adjective', 'b'], ['sindacato', 'noun', 'b'], ['sindaco', 'noun', 'b'], ['sindrome', 'noun', 'b'], ['single', 'noun', 'b'], ['singolare', 'adjective', 'b'], ['singolare', 'noun', 'b'], ['singolo', 'adjective', 'a'], ['singolo', 'noun', 'a'], ['sinistra', 'noun', 'a'], ['sinistro', 'adjective', 'a'], ['sinistro', 'noun', 'a'], ['sino', 'preposition', 'a'], ['sino', 'adverb', 'a'], ['sinonimo', 'noun', 'b'], ['sintesi', 'noun', 'b'], ['sintetico', 'adjective', 'b'], ['sintetizzare', 'verb', 'b'], ['sintomo', 'noun', 'b'], ['sir', 'noun', 'b'], ['siriano', 'adjective', 'c'], ['siriano', 'noun', 'c'], ['siringa', 'noun', 'c'], ['sistema', 'noun', 'a'], ['sistemare', 'verb', 'a'], ['sito', 'noun', 'a'], ['sito', 'adjective', 'a'], ['situare', 'verb', 'b'], ['situazione', 'noun', 'a'], ['slacciare', 'verb', 'c'], ['slanciato', 'past_part', 'c'], ['slanciato', 'adjective', 'c'], ['slavo', 'adjective', 'c'], ['slavo', 'noun', 'c'], ['slegare', 'verb', 'c'], ['slip', 'noun', 'c'], ['slitta', 'noun', 'c'], ['slogan', 'noun', 'b'], ['slogare', 'verb', 'c'], ['slogatura', 'noun', 'c'], ['slovacco', 'adjective', 'c'], ['slovacco', 'noun', 'c'], ['sloveno', 'adjective', 'c'], ['sloveno', 'noun', 'c'], ['smacchiare', 'verb', 'c'], ['smacchiatore', 'adjective', 'c'], ['smacchiatore', 'noun', 'c'], ['smaltimento', 'noun', 'b'], ['smalto', 'noun', 'c'], ['smascherare', 'verb', 'c'], ['smentire', 'verb', 'b'], ['smettere', 'verb', 'a'], ['smisurato', 'past_part', 'c'], ['smisurato', 'adjective', 'c'], ['smog', 'noun', 'c'], ['smontare', 'verb', 'b'], ['smorfia', 'noun', 'c'], ['smuovere', 'verb', 'c'], ['snack', 'noun', 'c'], ['sneaker', 'noun', 'c'], ['snello', 'adjective', 'c'], ['soccorrere', 'verb', 'c'], ['soccorso', 'noun', 'b'], ['socialdemocratico', 'adjective', 'c'], ['socialdemocratico', 'noun', 'c'], ['sociale', 'adjective', 'a'], ['sociale', 'noun', 'a'], ['socialista', 'adjective', 'b'], ['socialista', 'noun', 'b'], ['società', 'noun', 'a'], ['socievole', 'adjective', 'c'], ['socio', 'noun', 'b'], ['soddisfare', 'verb', 'a'], ['soddisfatto', 'past_part', 'b'], ['soddisfatto', 'adjective', 'b'], ['soddisfazione', 'noun', 'a'], ['sodo', 'adjective', 'b'], ['sodo', 'noun', 'b'], ['sodo', 'adverb', 'b'], ['sofà', 'noun', 'c'], ['sofferenza', 'noun', 'a'], ['soffermare', 'verb', 'b'], ['soffiare', 'verb', 'b'], ['soffice', 'adjective', 'c'], ['soffitta', 'noun', 'c'], ['soffitto', 'noun', 'b'], ['soffocare', 'verb', 'b'], ['soffriggere', 'verb', 'c'], ['soffrire', 'verb', 'a'], ['sofisticato', 'past_part', 'b'], ['sofisticato', 'adjective', 'b'], ['software', 'noun', 'b'], ['soggettivo', 'adjective', 'b'], ['soggetto', 'noun', 'a'], ['soggetto', 'adjective', 'b'], ['soggezione', 'noun', 'c'], ['soggiorno', 'noun', 'a'], ['soglia', 'noun', 'b'], ['sogliola', 'noun', 'c'], ['sognare', 'verb', 'a'], ['sogno', 'noun', 'a'], ['sol', 'noun', 'c'], ['solaio', 'noun', 'c'], ['solamente', 'adverb', 'a'], ['solamente', 'conjunction', 'a'], ['solare', 'adjective', 'b'], ['solare', 'noun', 'b'], ['solco', 'noun', 'b'], ['soldato', 'noun', 'a'], ['soldo', 'noun', 'a'], ['sole', 'noun', 'a'], ['solenne', 'adjective', 'b'], ['solidarietà', 'noun', 'b'], ['solido', 'adjective', 'b'], ['solido', 'noun', 'b'], ['solitamente', 'adverb', 'b'], ['solitario', 'adjective', 'b'], ['solitario', 'noun', 'b'], ['solito', 'adjective', 'a'], ['solito', 'noun', 'a'], ['solitudine', 'noun', 'b'], ['solletico', 'noun', 'c'], ['sollevare', 'verb', 'a'], ['sollievo', 'noun', 'b'], ['solo', 'adjective', 'a'], ['solo', 'noun', 'a'], ['solo', 'adverb', 'a'], ['solo', 'conjunction', 'a'], ['soltanto', 'adverb', 'a'], ['soltanto', 'conjunction', 'a'], ['soluzione', 'noun', 'a'], ['somigliare', 'verb', 'b'], ['somma', 'noun', 'a'], ['sommare', 'verb', 'b'], ['sondaggio', 'noun', 'a'], ['sonno', 'noun', 'a'], ['sonoro', 'adjective', 'b'], ['sonoro', 'noun', 'b'], ['soppalco', 'noun', 'c'], ['sopportare', 'verb', 'a'], ['sopra', 'preposition', 'a'], ['sopra', 'adverb', 'a'], ['sopra', 'adjective', 'a'], ['sopra', 'noun', 'a'], ['soprabito', 'noun', 'c'], ['sopracciglio', 'noun', 'c'], ['soprammobile', 'noun', 'c'], ['soprannome', 'noun', 'c'], ['soprattutto', 'adverb', 'a'], ['sopravvalutare', 'verb', 'c'], ['sopravvivenza', 'noun', 'b'], ['sopravvivere', 'verb', 'a'], ['sorcio', 'noun', 'c'], ['sordo', 'adjective', 'b'], ['sordo', 'noun', 'b'], ['sorella', 'noun', 'a'], ['sorgente', 'pres_part', 'b'], ['sorgente', 'adjective', 'b'], ['sorgente', 'noun', 'b'], ['sorgere', 'verb', 'b'], ['sorpassare', 'verb', 'c'], ['sorpasso', 'noun', 'c'], ['sorprendente', 'pres_part', 'b'], ['sorprendente', 'adjective', 'b'], ['sorprendere', 'verb', 'b'], ['sorpresa', 'noun', 'a'], ['sorridente', 'pres_part', 'c'], ['sorridente', 'adjective', 'c'], ['sorridere', 'verb', 'a'], ['sorriso', 'noun', 'a'], ['sorso', 'noun', 'c'], ['sorta', 'noun', 'a'], ['sorte', 'noun', 'b'], ['sorteggiare', 'verb', 'c'], ['sorteggio', 'noun', 'c'], ['sorvegliare', 'verb', 'b'], ['sospendere', 'verb', 'b'], ['sospensione', 'noun', 'b'], ['sospeso', 'past_part', 'b'], ['sospeso', 'adjective', 'b'], ['sospeso', 'noun', 'b'], ['sospettare', 'verb', 'b'], ['sospetto', 'noun', 'a'], ['sospetto', 'adjective', 'a'], ['sospetto', 'noun', 'a'], ['sospirare', 'verb', 'b'], ['sospiro', 'noun', 'b'], ['sosta', 'noun', 'b'], ['sostanza', 'noun', 'a'], ['sostanzialmente', 'adverb', 'b'], ['sostare', 'verb', 'c'], ['sostegno', 'noun', 'b'], ['sostenere', 'verb', 'a'], ['sostenitore', 'adjective', 'b'], ['sostenitore', 'noun', 'b'], ['sostituire', 'verb', 'a'], ['sostituzione', 'noun', 'b'], ['sottaceto', 'adjective', 'c'], ['sottaceto', 'adverb', 'c'], ['sottaceto', 'noun', 'c'], ['sotterraneo', 'adjective', 'b'], ['sotterraneo', 'noun', 'b'], ['sottile', 'adjective', 'a'], ['sottile', 'noun', 'a'], ['sottile', 'adverb', 'a'], ['sottinteso', 'past_part', 'c'], ['sottinteso', 'adjective', 'c'], ['sottinteso', 'noun', 'c'], ['sotto', 'preposition', 'a'], ['sotto', 'adverb', 'a'], ['sotto', 'adjective', 'a'], ['sotto', 'noun', 'a'], ['sottofondo', 'noun', 'b'], ['sottolineare', 'verb', 'a'], ['sottolio', 'adverb', 'c'], ['sottolio', 'adjective', 'c'], ['sottomarino', 'adjective', 'c'], ['sottomarino', 'noun', 'c'], ['sottopassaggio', 'noun', 'c'], ['sottoporre', 'verb', 'a'], ['sottoscrivere', 'verb', 'b'], ['sottovalutare', 'verb', 'b'], ['sottrarre', 'verb', 'b'], ['sovietico', 'adjective', 'b'], ['sovietico', 'noun', 'b'], ['sovrano', 'adjective', 'b'], ['sovrano', 'noun', 'b'], ['sovrapporre', 'verb', 'b'], ['spaccare', 'verb', 'b'], ['spaccatura', 'noun', 'c'], ['spacciare', 'verb', 'b'], ['spacciatore', 'noun', 'c'], ['spaccio', 'noun', 'c'], ['spada', 'noun', 'b'], ['spaghetto', 'noun', 'b'], ['spagnolo', 'adjective', 'a'], ['spagnolo', 'noun', 'a'], ['spago', 'noun', 'c'], ['spalancare', 'verb', 'b'], ['spalla', 'noun', 'a'], ['spalmabile', 'adjective', 'c'], ['spalmare', 'verb', 'c'], ['spam', 'noun', 'b'], ['sparare', 'verb', 'a'], ['sparecchiare', 'verb', 'c'], ['spargere', 'verb', 'b'], ['sparire', 'verb', 'a'], ['sparo', 'noun', 'b'], ['sparso', 'past_part', 'b'], ['sparso', 'adjective', 'b'], ['spassare', 'verb', 'b'], ['spasso', 'noun', 'c'], ['spavaldo', 'adjective', 'c'], ['spaventare', 'verb', 'a'], ['spaventato', 'past_part', 'b'], ['spaventato', 'adjective', 'b'], ['spaventoso', 'adjective', 'b'], ['spaziale', 'adjective', 'b'], ['spazio', 'noun', 'a'], ['spazioso', 'adjective', 'c'], ['spazzare', 'verb', 'b'], ['spazzatura', 'noun', 'b'], ['spazzino', 'noun', 'c'], ['spazzola', 'noun', 'c'], ['spazzolare', 'verb', 'c'], ['spazzolino', 'noun', 'c'], ['spazzolone', 'noun', 'c'], ['specchiarsi', 'verb', 'c'], ['specchio', 'noun', 'a'], ['speciale', 'adjective', 'a'], ['speciale', 'noun', 'a'], ['specialista', 'noun', 'b'], ['specializzato', 'past_part', 'b'], ['specializzato', 'adjective', 'b'], ['specializzato', 'noun', 'b'], ['specialmente', 'adverb', 'b'], ['specie', 'noun', 'a'], ['specie', 'adverb', 'a'], ['specificare', 'verb', 'b'], ['specifico', 'adjective', 'a'], ['specifico', 'noun', 'a'], ['speck', 'noun', 'c'], ['spedire', 'verb', 'b'], ['spedizione', 'noun', 'b'], ['spegnere', 'verb', 'a'], ['spellare', 'verb', 'c'], ['spendere', 'verb', 'a'], ['spennare', 'verb', 'c'], ['spensierato', 'adjective', 'c'], ['spento', 'past_part', 'b'], ['spento', 'adjective', 'b'], ['speranza', 'noun', 'a'], ['sperare', 'verb', 'a'], ['sperimentale', 'adjective', 'b'], ['sperimentare', 'verb', 'b'], ['sperimentazione', 'noun', 'b'], ['sperone', 'noun', 'c'], ['spesa', 'noun', 'a'], ['spesso', 'adjective', 'b'], ['spesso', 'adverb', 'a'], ['spessore', 'noun', 'b'], ['spettacolare', 'adjective', 'b'], ['spettacolo', 'noun', 'a'], ['spettare', 'verb', 'b'], ['spettatore', 'noun', 'b'], ['spettinare', 'verb', 'c'], ['spettro', 'noun', 'b'], ['spezia', 'noun', 'c'], ['spezzare', 'verb', 'b'], ['spia', 'noun', 'b'], ['spiacere', 'verb', 'b'], ['spiaggia', 'noun', 'a'], ['spianare', 'verb', 'c'], ['spiare', 'verb', 'b'], ['spiazzo', 'noun', 'c'], ['spiccare', 'verb', 'b'], ['spicciolo', 'adjective', 'c'], ['spicciolo', 'noun', 'c'], ['spiedino', 'noun', 'c'], ['spiedo', 'noun', 'c'], ['spiegare', 'verb', 'a'], ['spiegazione', 'noun', 'a'], ['spietato', 'adjective', 'b'], ['spiga', 'noun', 'c'], ['spigolo', 'noun', 'c'], ['spillo', 'noun', 'c'], ['spina', 'noun', 'b'], ['spinacio', 'noun', 'c'], ['spingere', 'verb', 'a'], ['spinta', 'noun', 'b'], ['spionaggio', 'noun', 'c'], ['spirito', 'noun', 'a'], ['spiritoso', 'adjective', 'c'], ['spirituale', 'adjective', 'b'], ['spirituale', 'noun', 'b'], ['splendente', 'pres_part', 'c'], ['splendente', 'adjective', 'c'], ['splendere', 'verb', 'b'], ['splendido', 'adjective', 'b'], ['splendore', 'noun', 'b'], ['spogliare', 'verb', 'b'], ['spogliatoio', 'noun', 'c'], ['spoglio', 'noun', 'c'], ['spolverare', 'verb', 'c'], ['sponda', 'noun', 'b'], ['spontaneo', 'adjective', 'b'], ['sporcare', 'verb', 'b'], ['sporcizia', 'noun', 'c'], ['sporco', 'adjective', 'a'], ['sporco', 'noun', 'a'], ['sporgente', 'pres_part', 'c'], ['sporgente', 'adjective', 'c'], ['sporgente', 'noun', 'c'], ['sporgere', 'verb', 'b'], ['sport', 'noun', 'a'], ['sport', 'adjective', 'a'], ['sportello', 'noun', 'b'], ['sportivo', 'adjective', 'a'], ['sportivo', 'noun', 'a'], ['sposare', 'verb', 'a'], ['sposato', 'past_part', 'b'], ['sposato', 'adjective', 'b'], ['sposato', 'noun', 'b'], ['sposo', 'noun', 'b'], ['spostamento', 'noun', 'b'], ['spostare', 'verb', 'a'], ['spot', 'noun', 'b'], ['spranga', 'noun', 'c'], ['spray', 'adjective', 'c'], ['spray', 'noun', 'c'], ['sprecare', 'verb', 'b'], ['spreco', 'noun', 'c'], ['spremere', 'verb', 'c'], ['spremuta', 'noun', 'c'], ['sprofondare', 'verb', 'b'], ['sproposito', 'noun', 'c'], ['spruzzare', 'verb', 'c'], ['spuma', 'noun', 'c'], ['spumante', 'pres_part', 'c'], ['spumante', 'adjective', 'c'], ['spumante', 'noun', 'c'], ['spuntare', 'verb', 'b'], ['spuntino', 'noun', 'c'], ['spunto', 'noun', 'b'], ['sputare', 'verb', 'b'], ['sputo', 'noun', 'c'], ['squadra', 'noun', 'a'], ['squallido', 'adjective', 'c'], ['squalo', 'noun', 'c'], ['squarcio', 'noun', 'c'], ['squillare', 'verb', 'b'], ['squisito', 'adjective', 'c'], ['stabile', 'adjective', 'b'], ['stabile', 'noun', 'b'], ['stabilire', 'verb', 'a'], ['stabilità', 'noun', 'b'], ['staccare', 'verb', 'a'], ['stacco', 'noun', 'c'], ['stadio', 'noun', 'b'], ['staffa', 'noun', 'c'], ['stagione', 'noun', 'a'], ['stagno', 'noun', 'c'], ['stalla', 'noun', 'b'], ['stallone', 'noun', 'c'], ['stamattina', 'adverb', 'b'], ['stampa', 'noun', 'a'], ['stampare', 'verb', 'b'], ['stampatello', 'noun', 'c'], ['stampato', 'past_part', 'b'], ['stampato', 'adjective', 'b'], ['stampato', 'noun', 'b'], ['stampella', 'noun', 'c'], ['stampo', 'noun', 'c'], ['stancare', 'verb', 'b'], ['stanchezza', 'noun', 'b'], ['stanco', 'adjective', 'a'], ['standard', 'noun', 'b'], ['standard', 'adjective', 'b'], ['stanga', 'noun', 'c'], ['stanotte', 'adverb', 'b'], ['stanza', 'noun', 'a'], ['star', 'noun', 'b'], ['stare', 'verb', 'a'], ['stasera', 'adverb', 'a'], ['statale', 'adjective', 'b'], ['statale', 'noun', 'b'], ['statistica', 'noun', 'b'], ['statistico', 'adjective', 'b'], ['statistico', 'noun', 'b'], ['stato', 'noun', 'a'], ['stato', 'noun', 'a'], ['statua', 'noun', 'b'], ['statunitense', 'adjective', 'b'], ['statunitense', 'noun', 'b'], ['status', 'noun', 'b'], ['stavolta', 'adverb', 'b'], ['stazione', 'noun', 'a'], ['stella', 'noun', 'a'], ['stellare', 'adjective', 'b'], ['stendere', 'verb', 'b'], ['stendibiancheria', 'noun', 'c'], ['stereo', 'adjective', 'c'], ['stereo', 'noun', 'c'], ['sterlina', 'noun', 'b'], ['sterzare', 'verb', 'c'], ['sterzo', 'noun', 'c'], ['stesso', 'adjective', 'a'], ['stesso', 'pronoun', 'a'], ['stile', 'noun', 'a'], ['stima', 'noun', 'b'], ['stimare', 'verb', 'b'], ['stimolare', 'verb', 'b'], ['stimolo', 'noun', 'b'], ['stinco', 'noun', 'c'], ['stipendiare', 'verb', 'c'], ['stipendio', 'noun', 'a'], ['stirare', 'verb', 'b'], ['stivaletto', 'noun', 'c'], ['stoffa', 'noun', 'b'], ['stomaco', 'noun', 'b'], ['stonare', 'verb', 'c'], ['stop', 'loc-comando', 'c'], ['stop', 'noun', 'c'], ['stoppa', 'noun', 'c'], ['storcere', 'verb', 'c'], ['storia', 'noun', 'a'], ['storico', 'adjective', 'a'], ['storico', 'noun', 'a'], ['stornello', 'noun', 'c'], ['storta', 'noun', 'c'], ['storto', 'past_part', 'b'], ['storto', 'adjective', 'b'], ['storto', 'adverb', 'b'], ['storto', 'noun', 'b'], ['stoviglia', 'noun', 'c'], ['stracchino', 'noun', 'c'], ['straccio', 'noun', 'b'], ['strada', 'noun', 'a'], ['stradale', 'adjective', 'b'], ['stradale', 'noun', 'b'], ['strage', 'noun', 'b'], ['strangolare', 'verb', 'c'], ['straniero', 'adjective', 'a'], ['straniero', 'noun', 'a'], ['strano', 'adjective', 'a'], ['straordinario', 'adjective', 'a'], ['straordinario', 'noun', 'a'], ['strappare', 'verb', 'b'], ['strategia', 'noun', 'a'], ['strategico', 'adjective', 'b'], ['strato', 'noun', 'b'], ['strega', 'noun', 'a'], ['stregare', 'verb', 'b'], ['stregone', 'noun', 'c'], ['stress', 'noun', 'b'], ['stretta', 'noun', 'b'], ['strettamente', 'adverb', 'b'], ['stretto', 'past_part', 'a'], ['stretto', 'adjective', 'a'], ['stretto', 'noun', 'a'], ['strillare', 'verb', 'b'], ['strillo', 'noun', 'c'], ['stringa', 'noun', 'c'], ['stringere', 'verb', 'a'], ['striscia', 'noun', 'b'], ['strisciare', 'verb', 'b'], ['strofinaccio', 'noun', 'c'], ['stronzata', 'noun', 'b'], ['stronzo', 'noun', 'a'], ['stronzo', 'adjective', 'a'], ['strumento', 'noun', 'a'], ['strutto', 'past_part', 'c'], ['strutto', 'adjective', 'c'], ['strutto', 'noun', 'c'], ['struttura', 'noun', 'a'], ['strutturale', 'adjective', 'b'], ['struzzo', 'noun', 'c'], ['studente', 'noun', 'a'], ['studiare', 'verb', 'a'], ['studio', 'noun', 'a'], ['studioso', 'adjective', 'b'], ['studioso', 'noun', 'b'], ['stufa', 'noun', 'c'], ['stuoia', 'noun', 'c'], ['stupefacente', 'pres_part', 'b'], ['stupefacente', 'adjective', 'b'], ['stupefacente', 'noun', 'b'], ['stupendo', 'adjective', 'b'], ['stupido', 'adjective', 'a'], ['stupido', 'noun', 'a'], ['stupire', 'verb', 'b'], ['stupito', 'past_part', 'b'], ['stupito', 'adjective', 'b'], ['stupore', 'noun', 'b'], ['stuzzicadenti', 'noun', 'c'], ['stuzzicare', 'verb', 'c'], ['style', 'noun', 'b'], ['su', 'preposition', 'a'], ['su', 'adverb', 'a'], ['su', 'exclamation', 'a'], ['su', 'noun', 'a'], ['subire', 'verb', 'a'], ['subito', 'adverb', 'a'], ['succedere', 'verb', 'a'], ['successione', 'noun', 'b'], ['successivamente', 'adverb', 'b'], ['successivo', 'adjective', 'a'], ['successo', 'noun', 'a'], ['succhiare', 'verb', 'b'], ['succo', 'noun', 'b'], ['sud', 'noun', 'a'], ['sud', 'adjective', 'a'], ['sudamericano', 'adjective', 'c'], ['sudamericano', 'noun', 'c'], ['sudare', 'verb', 'b'], ['sudato', 'past_part', 'c'], ['sudato', 'adjective', 'c'], ['suddito', 'noun', 'b'], ['suddito', 'adjective', 'b'], ['suddividere', 'verb', 'b'], ['sudicio', 'adjective', 'c'], ['sudicio', 'noun', 'c'], ['sudore', 'noun', 'b'], ['sudtirolese', 'adjective', 'c'], ['sudtirolese', 'noun', 'c'], ['sufficiente', 'adjective', 'a'], ['suggerimento', 'noun', 'b'], ['suggerire', 'verb', 'a'], ['suggestivo', 'adjective', 'b'], ['sughero', 'noun', 'c'], ['sugo', 'noun', 'b'], ['suicidio', 'noun', 'b'], ['suino', 'noun', 'c'], ['suino', 'adjective', 'c'], ['suo', 'adjective', 'a'], ['suo', 'pronoun', 'a'], ['suocera', 'noun', 'c'], ['suocero', 'noun', 'c'], ['suola', 'noun', 'c'], ['suolo', 'noun', 'b'], ['suonare', 'verb', 'a'], ['suono', 'noun', 'a'], ['suora', 'noun', 'a'], ['super', 'adjective', 'b'], ['super', 'noun', 'b'], ['superare', 'verb', 'a'], ['superbia', 'noun', 'c'], ['superficiale', 'adjective', 'b'], ['superficie', 'noun', 'a'], ['superiore', 'adjective', 'a'], ['superiore', 'noun', 'a'], ['supermercato', 'noun', 'b'], ['supporre', 'verb', 'b'], ['supportare', 'verb', 'b'], ['supporto', 'noun', 'a'], ['supremo', 'adjective', 'b'], ['surgelato', 'past_part', 'c'], ['surgelato', 'adjective', 'c'], ['surgelato', 'noun', 'c'], ['suscitare', 'verb', 'b'], ['susina', 'noun', 'c'], ['susino', 'noun', 'c'], ['susseguirsi', 'verb', 'c'], ['sussurrare', 'verb', 'b'], ['svanire', 'verb', 'b'], ['svedese', 'adjective', 'c'], ['svedese', 'noun', 'c'], ['sveglia', 'noun', 'c'], ['svegliare', 'verb', 'a'], ['svegliarsi', 'verb', 'c'], ['sveglio', 'past_part', 'b'], ['sveglio', 'adjective', 'b'], ['svelare', 'verb', 'b'], ['svelto', 'adjective', 'c'], ['svenire', 'verb', 'b'], ['sventola', 'noun', 'c'], ['sviluppare', 'verb', 'a'], ['sviluppato', 'past_part', 'b'], ['sviluppato', 'adjective', 'b'], ['sviluppo', 'noun', 'a'], ['svizzero', 'adjective', 'b'], ['svizzero', 'noun', 'b'], ['svolazzare', 'verb', 'c'], ['svolgere', 'verb', 'a'], ['svolgimento', 'noun', 'c'], ['svolta', 'noun', 'b'], ['svuotare', 'verb', 'b'], ['tabaccaio', 'noun', 'c'], ['tabella', 'noun', 'b'], ['tacca', 'noun', 'c'], ['tacchino', 'noun', 'c'], ['tacco', 'noun', 'b'], ['tacere', 'verb', 'a'], ['tacere', 'noun', 'a'], ['tag', 'noun', 'b'], ['taglia', 'noun', 'b'], ['tagliare', 'verb', 'a'], ['tagliatella', 'noun', 'c'], ['tagliato', 'past_part', 'b'], ['tagliato', 'adjective', 'b'], ['tagliere', 'noun', 'c'], ['taglio', 'noun', 'a'], ['tagliola', 'noun', 'c'], ['talco', 'noun', 'c'], ['tale', 'adjective', 'a'], ['tale', 'pronoun', 'a'], ['tale', 'adverb', 'a'], ['taleggio', 'noun', 'c'], ['talento', 'noun', 'b'], ['talmente', 'adverb', 'a'], ['talpa', 'noun', 'c'], ['talpa', 'adjective', 'c'], ['talpa', 'noun', 'c'], ['talvolta', 'adverb', 'b'], ['tamburo', 'noun', 'c'], ['tamponare', 'verb', 'c'], ['tangente', 'pres_part', 'b'], ['tangente', 'adjective', 'b'], ['tangente', 'noun', 'b'], ['tanto', 'adjective', 'a'], ['tanto', 'pronoun', 'a'], ['tanto', 'noun', 'a'], ['tanto', 'adverb', 'a'], ['tanto', 'conjunction', 'a'], ['tappa', 'noun', 'b'], ['tappare', 'verb', 'b'], ['tappetino', 'noun', 'c'], ['tappeto', 'noun', 'b'], ['tappezzare', 'verb', 'c'], ['tappo', 'noun', 'c'], ['tarallo', 'noun', 'c'], ['tarantella', 'noun', 'c'], ['tardi', 'adverb', 'a'], ['tardo', 'adjective', 'a'], ['tardo', 'adverb', 'a'], ['targa', 'noun', 'b'], ['tariffa', 'noun', 'b'], ['tarlo', 'noun', 'c'], ['tartaruga', 'noun', 'c'], ['tartufo', 'noun', 'c'], ['tasca', 'noun', 'a'], ['tassa', 'noun', 'a'], ['tassare', 'verb', 'c'], ['tassello', 'noun', 'c'], ['tasso', 'noun', 'b'], ['tastiera', 'noun', 'b'], ['tasto', 'noun', 'b'], ['tatto', 'noun', 'c'], ['tatuaggio', 'noun', 'b'], ['taverna', 'noun', 'c'], ['tavola', 'noun', 'a'], ['tavoletta', 'noun', 'c'], ['tavolino', 'noun', 'b'], ['tavolo', 'noun', 'a'], ['taxi', 'noun', 'b'], ['tazza', 'noun', 'b'], ['tè', 'noun', 'b'], ['te', 'pronoun', 'noun'], ['te', 'team', 'noun'], ['teatrale', 'adjective', 'b'], ['teatro', 'noun', 'a'], ['tecnica', 'noun', 'a'], ['tecnicamente', 'adverb', 'b'], ['tecnico', 'adjective', 'a'], ['tecnico', 'noun', 'a'], ['tecnologia', 'noun', 'a'], ['tecnologico', 'adjective', 'b'], ['tedesco', 'adjective', 'a'], ['tedesco', 'noun', 'a'], ['tegame', 'noun', 'c'], ['teglia', 'noun', 'c'], ['tegola', 'noun', 'c'], ['tela', 'noun', 'b'], ['telaio', 'noun', 'c'], ['telecamera', 'noun', 'b'], ['telecomandato', 'past_part', 'c'], ['telecomandato', 'adjective', 'c'], ['telecronaca', 'noun', 'c'], ['telecronista', 'noun', 'c'], ['telefilm', 'noun', 'b'], ['telefonare', 'verb', 'a'], ['telefonata', 'noun', 'a'], ['telefonico', 'adjective', 'a'], ['telefonino', 'noun', 'b'], ['telefono', 'noun', 'a'], ['telegiornale', 'noun', 'b'], ['telegrafico', 'adjective', 'c'], ['telegrafo', 'noun', 'c'], ['telegramma', 'noun', 'c'], ['telescopio', 'noun', 'b'], ['televisione', 'noun', 'a'], ['televisivo', 'adjective', 'a'], ['televisore', 'noun', 'b'], ['tema', 'noun', 'a'], ['temere', 'verb', 'a'], ['temperatura', 'noun', 'a'], ['tempesta', 'noun', 'b'], ['tempio', 'noun', 'b'], ['tempo', 'noun', 'a'], ['temporale', 'noun', 'b'], ['temporaneo', 'adjective', 'b'], ['tenaglia', 'noun', 'c'], ['tenda', 'noun', 'a'], ['tendenza', 'noun', 'a'], ['tendere', 'verb', 'a'], ['tenebra', 'noun', 'c'], ['tenente', 'noun', 'b'], ['tenere', 'verb', 'a'], ['tenerezza', 'noun', 'b'], ['tenero', 'adjective', 'b'], ['tenero', 'noun', 'b'], ['tennis', 'noun', 'b'], ['tensione', 'noun', 'a'], ['tentare', 'verb', 'a'], ['tentativo', 'noun', 'a'], ['tentazione', 'noun', 'b'], ['tenuta', 'noun', 'b'], ['teologia', 'noun', 'b'], ['teologo', 'noun', 'b'], ['teoria', 'noun', 'a'], ['teorico', 'adjective', 'b'], ['teorico', 'noun', 'b'], ['terapia', 'noun', 'a'], ['tergicristallo', 'noun', 'c'], ['terminale', 'adjective', 'b'], ['terminale', 'noun', 'b'], ['terminare', 'verb', 'a'], ['termine', 'noun', 'a'], ['termosifone', 'noun', 'c'], ['terra', 'noun', 'a'], ['terrazzo', 'noun', 'b'], ['terremoto', 'noun', 'b'], ['terreno', 'noun', 'a'], ['terrestre', 'adjective', 'b'], ['terrestre', 'noun', 'b'], ['terribile', 'adjective', 'a'], ['terriccio', 'noun', 'c'], ['territoriale', 'adjective', 'b'], ['territoriale', 'noun', 'b'], ['territorio', 'noun', 'a'], ['terrore', 'noun', 'b'], ['terrorismo', 'noun', 'b'], ['terrorista', 'adjective', 'b'], ['terrorista', 'noun', 'b'], ['terrorizzare', 'verb', 'b'], ['terzo', 'adjective', 'a'], ['terzo', 'noun', 'a'], ['teschio', 'noun', 'b'], ['tesi', 'noun', 'a'], ['teso', 'past_part', 'b'], ['teso', 'adjective', 'b'], ['tesoro', 'noun', 'a'], ['tessera', 'noun', 'b'], ['tessile', 'adjective', 'c'], ['tessile', 'noun', 'c'], ['tessuto', 'past_part', 'b'], ['tessuto', 'adjective', 'b'], ['tessuto', 'noun', 'b'], ['test', 'noun', 'a'], ['testa', 'noun', 'a'], ['testamento', 'noun', 'b'], ['testare', 'verb', 'b'], ['testimone', 'noun', 'a'], ['testimonianza', 'noun', 'b'], ['testimoniare', 'verb', 'b'], ['testo', 'noun', 'a'], ['tetta', 'noun', 'b'], ['tetto', 'noun', 'a'], ['tettoia', 'noun', 'c'], ['tg', 'sigla', 'b'], ['thermos', 'noun', 'c'], ['ti', 'noun', 'c'], ['ti', 'pronoun', 'a'], ['tic', 'noun', 'c'], ['ticchettio', 'noun', 'c'], ['tifare', 'verb', 'b'], ['tifo', 'noun', 'c'], ['tifoso', 'adjective', 'b'], ['tifoso', 'noun', 'b'], ['tigre', 'noun', 'b'], ['timbro', 'noun', 'c'], ['timidezza', 'noun', 'c'], ['timido', 'adjective', 'b'], ['timido', 'noun', 'b'], ['timone', 'noun', 'c'], ['timoniere', 'noun', 'c'], ['timore', 'noun', 'b'], ['tinello', 'noun', 'c'], ['tino', 'noun', 'c'], ['tipico', 'adjective', 'a'], ['tipo', 'noun', 'a'], ['tipologia', 'noun', 'b'], ['tiramisù', 'noun', 'c'], ['tiranno', 'noun', 'c'], ['tiranno', 'adjective', 'c'], ['tirare', 'verb', 'a'], ['tiro', 'noun', 'b'], ['tirocinio', 'noun', 'b'], ['tirrenico', 'adjective', 'c'], ['tisana', 'noun', 'c'], ['titolare', 'adjective', 'b'], ['titolare', 'noun', 'b'], ['titolo', 'noun', 'a'], ['tivù', 'noun', 'a'], ['tizio', 'noun', 'b'], ['toast', 'noun', 'c'], ['toccare', 'verb', 'a'], ['tocco', 'noun', 'b'], ['togliere', 'verb', 'a'], ['toilette', 'noun', 'c'], ['toletta', 'noun', 'c'], ['tolleranza', 'noun', 'b'], ['tollerare', 'verb', 'b'], ['tomba', 'noun', 'b'], ['tombola', 'noun', 'c'], ['tonaca', 'noun', 'c'], ['tondo', 'adjective', 'b'], ['tondo', 'noun', 'b'], ['tonnellata', 'noun', 'b'], ['tonno', 'noun', 'c'], ['tono', 'noun', 'a'], ['tonsilla', 'noun', 'c'], ['top', 'noun', 'b'], ['topo', 'noun', 'b'], ['topo', 'adjective', 'b'], ['toppa', 'noun', 'c'], ['torbido', 'adjective', 'c'], ['torbido', 'noun', 'c'], ['torcere', 'verb', 'b'], ['torcia', 'noun', 'c'], ['torcicollo', 'noun', 'c'], ['tordo', 'noun', 'c'], ['torero', 'noun', 'c'], ['torinese', 'adjective', 'c'], ['torinese', 'noun', 'c'], ['tormentare', 'verb', 'b'], ['tornaconto', 'noun', 'c'], ['tornare', 'verb', 'a'], ['torneo', 'noun', 'b'], ['tornio', 'noun', 'c'], ['toro', 'noun', 'b'], ['torre', 'noun', 'b'], ['torrone', 'noun', 'c'], ['torta', 'noun', 'b'], ['tortellino', 'noun', 'c'], ['torto', 'noun', 'b'], ['tortora', 'noun', 'c'], ['tortora', 'adjective', 'c'], ['tortora', 'noun', 'c'], ['tosare', 'verb', 'c'], ['toscano', 'adjective', 'b'], ['toscano', 'noun', 'b'], ['tosse', 'noun', 'b'], ['tossico', 'adjective', 'b'], ['tossico', 'noun', 'b'], ['tossire', 'verb', 'c'], ['tostapane', 'noun', 'c'], ['totale', 'adjective', 'a'], ['totale', 'noun', 'a'], ['totalmente', 'adverb', 'b'], ['tour', 'noun', 'b'], ['tovaglia', 'noun', 'b'], ['tovaglietta', 'noun', 'c'], ['tovagliolo', 'noun', 'c'], ['tra', 'preposition', 'a'], ['traballare', 'verb', 'c'], ['traboccare', 'verb', 'c'], ['trabocchetto', 'noun', 'c'], ['traccia', 'noun', 'a'], ['tracciare', 'verb', 'b'], ['tradimento', 'noun', 'b'], ['tradire', 'verb', 'b'], ['tradizionale', 'adjective', 'a'], ['tradizione', 'noun', 'a'], ['tradurre', 'verb', 'a'], ['traduzione', 'noun', 'a'], ['traffico', 'noun', 'a'], ['trafila', 'noun', 'c'], ['traforo', 'noun', 'c'], ['tragedia', 'noun', 'b'], ['traghetto', 'noun', 'c'], ['tragico', 'adjective', 'b'], ['tragico', 'noun', 'b'], ['trainare', 'verb', 'c'], ['trama', 'noun', 'b'], ['tramezzino', 'noun', 'c'], ['tramite', 'noun', 'preposition'], ['tramontare', 'verb', 'c'], ['tramonto', 'noun', 'b'], ['trampolino', 'noun', 'c'], ['trancio', 'noun', 'c'], ['tranne', 'preposition', 'a'], ['tranquillamente', 'adverb', 'b'], ['tranquillità', 'noun', 'b'], ['tranquillizzare', 'verb', 'c'], ['tranquillo', 'adjective', 'a'], ['tranquillo', 'adverb', 'a'], ['tranquillo', 'noun', 'a'], ['transito', 'noun', 'c'], ['trapano', 'noun', 'c'], ['trapezio', 'noun', 'c'], ['trapezio', 'adjective', 'c'], ['trapianto', 'noun', 'c'], ['trappola', 'noun', 'b'], ['trapunta', 'noun', 'c'], ['trarre', 'verb', 'a'], ['trascinare', 'verb', 'a'], ['trascorrere', 'verb', 'a'], ['trascrizione', 'noun', 'b'], ['trascurare', 'verb', 'b'], ['trasferimento', 'noun', 'b'], ['trasferire', 'verb', 'a'], ['trasformare', 'verb', 'a'], ['trasformazione', 'noun', 'b'], ['trasfusione', 'noun', 'c'], ['traslocare', 'verb', 'c'], ['trasloco', 'noun', 'c'], ['trasmettere', 'verb', 'a'], ['trasmissione', 'noun', 'a'], ['trasparente', 'adjective', 'b'], ['trasparente', 'noun', 'b'], ['trasparenza', 'noun', 'b'], ['trasportare', 'verb', 'b'], ['trasporto', 'noun', 'a'], ['trattamento', 'noun', 'a'], ['trattare', 'verb', 'a'], ['trattativa', 'noun', 'b'], ['trattato', 'noun', 'b'], ['trattenere', 'verb', 'a'], ['trattenuta', 'noun', 'c'], ['tratto', 'noun', 'a'], ['trattore', 'noun', 'c'], ['trauma', 'noun', 'b'], ['travasare', 'verb', 'c'], ['travestire', 'verb', 'c'], ['travolgere', 'verb', 'b'], ['tre', 'adjective', 'a'], ['tre', 'noun', 'a'], ['trebbiare', 'verb', 'c'], ['trecento', 'adjective', 'b'], ['trecento', 'noun', 'b'], ['tredici', 'adjective', 'b'], ['tredici', 'noun', 'b'], ['tremare', 'verb', 'b'], ['tremendo', 'adjective', 'b'], ['trend', 'noun', 'b'], ['treno', 'noun', 'a'], ['trenta', 'adjective', 'a'], ['trenta', 'noun', 'a'], ['trentino', 'adjective', 'c'], ['trentino', 'noun', 'c'], ['triangolo', 'noun', 'b'], ['tribù', 'noun', 'c'], ['tribunale', 'noun', 'a'], ['triestino', 'adjective', 'c'], ['triestino', 'noun', 'c'], ['trifoglio', 'noun', 'c'], ['trina', 'noun', 'c'], ['trincea', 'noun', 'c'], ['trionfo', 'noun', 'b'], ['triste', 'adjective', 'a'], ['tristezza', 'noun', 'b'], ['tritare', 'verb', 'c'], ['trofeo', 'noun', 'c'], ['tronco', 'noun', 'b'], ['trono', 'noun', 'b'], ['troppo', 'adjective', 'a'], ['troppo', 'pronoun', 'a'], ['troppo', 'adverb', 'a'], ['troppo', 'noun', 'a'], ['trota', 'noun', 'c'], ['trottare', 'verb', 'c'], ['trottola', 'noun', 'c'], ['trovare', 'verb', 'a'], ['truccare', 'verb', 'c'], ['trucco', 'noun', 'b'], ['trucco', 'noun', 'b'], ['truffa', 'noun', 'b'], ['truffare', 'verb', 'c'], ['truppa', 'noun', 'b'], ['t-shirt', 'noun', 'c'], ['tu', 'pronoun', 'a'], ['tubo', 'noun', 'b'], ['tuffare', 'verb', 'b'], ['tuffo', 'noun', 'c'], ['tulipano', 'noun', 'c'], ['tumore', 'noun', 'b'], ['tunica', 'noun', 'c'], ['tunisino', 'adjective', 'c'], ['tunisino', 'noun', 'c'], ['tunnel', 'noun', 'c'], ['tuo', 'adjective', 'a'], ['tuo', 'pronoun', 'a'], ['tuono', 'noun', 'c'], ['turbare', 'verb', 'b'], ['turco', 'adjective', 'b'], ['turco', 'noun', 'b'], ['turismo', 'noun', 'b'], ['turista', 'noun', 'b'], ['turistico', 'adjective', 'b'], ['turno', 'noun', 'a'], ['tuta', 'noun', 'b'], ['tutela', 'noun', 'b'], ['tutelare', 'verb', 'b'], ['tutore', 'noun', 'c'], ['tuttavia', 'conjunction', 'a'], ['tuttavia', 'adverb', 'a'], ['tutto', 'adjective', 'a'], ['tutto', 'pronoun', 'a'], ['tuttora', 'adverb', 'b'], ['u', 'noun', 'c'], ['ubriaco', 'adjective', 'b'], ['ubriaco', 'noun', 'b'], ['uccello', 'noun', 'a'], ['uccidere', 'verb', 'a'], ['ucraino', 'adjective', 'c'], ['ucraino', 'noun', 'c'], ['udienza', 'noun', 'b'], ['udinese', 'adjective', 'c'], ['udinese', 'noun', 'c'], ['udire', 'verb', 'b'], ['udire', 'noun', 'b'], ['ufficiale', 'noun', 'b'], ['ufficiale', 'adjective', 'a'], ['ufficialmente', 'adverb', 'b'], ['ufficio', 'noun', 'a'], ['uguale', 'adjective', 'a'], ['uguale', 'adverb', 'a'], ['uguale', 'noun', 'a'], ['ugualmente', 'adverb', 'b'], ['ulcera', 'noun', 'c'], ['ulteriore', 'adjective', 'a'], ['ulteriormente', 'adverb', 'b'], ['ultimamente', 'adverb', 'b'], ['ultimo', 'adjective', 'a'], ['ultimo', 'noun', 'a'], ['ultravioletto', 'noun', 'c'], ['ultravioletto', 'adjective', 'c'], ['umanità', 'noun', 'a'], ['umano', 'adjective', 'a'], ['umano', 'noun', 'a'], ['umbro', 'adjective', 'c'], ['umbro', 'noun', 'c'], ['umido', 'adjective', 'b'], ['umido', 'noun', 'b'], ['umile', 'adjective', 'b'], ['umile', 'noun', 'b'], ['umiliare', 'verb', 'b'], ['umore', 'noun', 'b'], ['umorismo', 'noun', 'c'], ['una', 'determiner', 'a'], ['una', 'pronoun', 'a'], ['undici', 'adjective', 'b'], ['undici', 'noun', 'b'], ['ungherese', 'adjective', 'c'], ['ungherese', 'noun', 'c'], ['unghia', 'noun', 'b'], ['unguento', 'noun', 'c'], ['unico', 'adjective', 'a'], ['unico', 'noun', 'a'], ['uniforme', 'adjective', 'b'], ['unione', 'noun', 'b'], ['unire', 'verb', 'a'], ['unità', 'noun', 'a'], ['unito', 'past_part', 'a'], ['unito', 'adjective', 'a'], ['unito', 'noun', 'a'], ['universale', 'adjective', 'b'], ['universale', 'noun', 'b'], ['università', 'noun', 'a'], ['universitario', 'adjective', 'b'], ['universitario', 'noun', 'b'], ['universo', 'noun', 'a'], ['uno', 'adjective', 'a'], ['uno', 'noun', 'a'], ['uno', 'determiner', 'a'], ['uno', 'pronoun', 'a'], ['uomo', 'noun', 'a'], ['uovo', 'noun', 'a'], ['uragano', 'noun', 'c'], ['urbanistico', 'adjective', 'b'], ['urbano', 'adjective', 'b'], ['urgente', 'adjective', 'b'], ['urgenza', 'noun', 'b'], ['urlare', 'verb', 'a'], ['urlo', 'noun', 'b'], ['urna', 'noun', 'c'], ['urtare', 'verb', 'b'], ['usare', 'verb', 'a'], ['usato', 'past_part', 'b'], ['usato', 'adjective', 'b'], ['usato', 'noun', 'b'], ['uscire', 'verb', 'a'], ['uscita', 'noun', 'a'], ['usignolo', 'noun', 'c'], ['uso', 'noun', 'a'], ['utensile', 'noun', 'c'], ['utente', 'noun', 'a'], ['utenza', 'noun', 'b'], ['utile', 'adjective', 'a'], ['utile', 'noun', 'a'], ['utilità', 'noun', 'b'], ['utilizzare', 'verb', 'a'], ['utilizzo', 'noun', 'b'], ['vabbè', 'exclamation', 'b'], ['vacanza', 'noun', 'a'], ['vacca', 'noun', 'b'], ['vaccino', 'noun', 'c'], ['vaffanculo', 'exclamation', 'b'], ['vagare', 'verb', 'b'], ['vagire', 'verb', 'c'], ['vago', 'adjective', 'b'], ['vago', 'noun', 'b'], ['valanga', 'noun', 'c'], ['valdostano', 'adjective', 'c'], ['valdostano', 'noun', 'c'], ['valere', 'verb', 'a'], ['valido', 'adjective', 'b'], ['valigia', 'noun', 'b'], ['valle', 'noun', 'b'], ['valore', 'noun', 'a'], ['valorizzare', 'verb', 'b'], ['valoroso', 'adjective', 'c'], ['valoroso', 'noun', 'c'], ['valutare', 'verb', 'a'], ['valutazione', 'noun', 'b'], ['valvola', 'noun', 'c'], ['vampata', 'noun', 'c'], ['vampiro', 'noun', 'b'], ['vandalo', 'adjective', 'c'], ['vandalo', 'noun', 'c'], ['vanga', 'noun', 'c'], ['vangelo', 'noun', 'b'], ['vanitoso', 'adjective', 'c'], ['vanitoso', 'noun', 'c'], ['vano', 'adjective', 'b'], ['vano', 'noun', 'b'], ['vantaggio', 'noun', 'a'], ['vantaggioso', 'adjective', 'c'], ['vantare', 'verb', 'b'], ['vanto', 'noun', 'c'], ['vapore', 'noun', 'b'], ['variabile', 'adjective', 'b'], ['variabile', 'noun', 'b'], ['variante', 'pres_part', 'b'], ['variante', 'adjective', 'b'], ['variante', 'noun', 'b'], ['variare', 'verb', 'b'], ['variazione', 'noun', 'b'], ['varietà', 'noun', 'b'], ['vario', 'adjective', 'a'], ['vario', 'adjective', 'a'], ['vario', 'pronoun', 'a'], ['variopinto', 'adjective', 'c'], ['vasca', 'noun', 'b'], ['vaso', 'noun', 'b'], ['vasto', 'adjective', 'b'], ['vasto', 'noun', 'b'], ['ve', 'pronoun', 'a'], ['ve', 'adverb', 'a'], ['vecchio', 'adjective', 'a'], ['vecchio', 'noun', 'a'], ['vedere', 'verb', 'a'], ['vedere', 'noun', 'a'], ['vedova', 'noun', 'b'], ['vegetale', 'adjective', 'b'], ['vegetale', 'noun', 'b'], ['veglia', 'noun', 'c'], ['veglione', 'noun', 'c'], ['veicolo', 'noun', 'b'], ['vela', 'noun', 'b'], ['veleno', 'noun', 'b'], ['velenoso', 'adjective', 'c'], ['vellutato', 'past_part', 'c'], ['vellutato', 'adjective', 'c'], ['velluto', 'noun', 'c'], ['velo', 'noun', 'b'], ['veloce', 'adjective', 'a'], ['veloce', 'adverb', 'a'], ['veloce', 'noun', 'a'], ['velocemente', 'adverb', 'b'], ['velocità', 'noun', 'a'], ['vena', 'noun', 'b'], ['vendemmiare', 'verb', 'c'], ['vendere', 'verb', 'a'], ['vendetta', 'noun', 'b'], ['vendicare', 'verb', 'b'], ['vendita', 'noun', 'a'], ['venditore', 'adjective', 'b'], ['venditore', 'noun', 'b'], ['venerdì', 'noun', 'a'], ['veneto', 'adjective', 'b'], ['veneto', 'noun', 'b'], ['veneziano', 'adjective', 'c'], ['veneziano', 'noun', 'c'], ['venire', 'verb', 'a'], ['ventaglio', 'noun', 'c'], ['ventata', 'noun', 'c'], ['venti', 'adjective', 'a'], ['venti', 'noun', 'a'], ['venticinque', 'adjective', 'b'], ['venticinque', 'noun', 'b'], ['ventilatore', 'adjective', 'c'], ['ventilatore', 'noun', 'c'], ['ventina', 'noun', 'b'], ['ventiquattro', 'adjective', 'b'], ['ventiquattro', 'noun', 'b'], ['vento', 'noun', 'a'], ['ventre', 'noun', 'b'], ['venuta', 'noun', 'c'], ['veramente', 'adverb', 'a'], ['verbale', 'adjective', 'a'], ['verbale', 'noun', 'a'], ['verbo', 'noun', 'b'], ['verde', 'adjective', 'a'], ['verde', 'noun', 'a'], ['verdura', 'noun', 'b'], ['vergine', 'adjective', 'b'], ['vergine', 'noun', 'b'], ['vergogna', 'noun', 'b'], ['vergognarsi', 'verb', 'b'], ['verifica', 'noun', 'b'], ['verificare', 'verb', 'a'], ['verità', 'noun', 'a'], ['verme', 'noun', 'b'], ['vernice', 'noun', 'b'], ['vero', 'adjective', 'a'], ['vero', 'noun', 'a'], ['versare', 'verb', 'a'], ['versione', 'noun', 'a'], ['verso', 'noun', 'a'], ['verso', 'preposition', 'a'], ['vertebra', 'noun', 'c'], ['verticale', 'adjective', 'b'], ['verticale', 'noun', 'b'], ['vertice', 'noun', 'b'], ['vertigine', 'noun', 'c'], ['vescovo', 'noun', 'b'], ['vescovo', 'adjective', 'b'], ['vespa', 'noun', 'c'], ['veste', 'noun', 'b'], ['vestire', 'verb', 'a'], ['vestito', 'noun', 'a'], ['vestito', 'past_part', 'b'], ['vestito', 'adjective', 'b'], ['veterinario', 'adjective', 'c'], ['veterinario', 'noun', 'c'], ['vetrina', 'noun', 'b'], ['vetro', 'noun', 'a'], ['vettura', 'noun', 'b'], ['vi', 'pronoun', 'a'], ['vi', 'adverb', 'a'], ['via', 'noun', 'a'], ['via', 'adverb', 'a'], ['via', 'exclamation', 'a'], ['via', 'noun', 'a'], ['viaggiare', 'verb', 'a'], ['viaggiatore', 'noun', 'b'], ['viaggiatrice', 'noun', 'c'], ['viaggio', 'noun', 'a'], ['viale', 'noun', 'b'], ['vibrare', 'verb', 'b'], ['vice', 'noun', 'b'], ['vicenda', 'noun', 'a'], ['viceversa', 'adverb', 'b'], ['vicinanza', 'noun', 'b'], ['vicino', 'adjective', 'a'], ['vicino', 'noun', 'a'], ['vicino', 'adverb', 'a'], ['vicolo', 'noun', 'b'], ['video', 'adjective', 'a'], ['video', 'noun', 'a'], ['videogioco', 'noun', 'b'], ['viennese', 'adjective', 'c'], ['viennese', 'noun', 'c'], ['vietare', 'verb', 'b'], ['vigile', 'adjective', 'b'], ['vigile', 'noun', 'b'], ['vigilia', 'noun', 'b'], ['vigna', 'noun', 'c'], ['vigore', 'noun', 'b'], ['villa', 'noun', 'a'], ['villaggio', 'noun', 'a'], ['vincente', 'pres_part', 'b'], ['vincente', 'adjective', 'b'], ['vincente', 'noun', 'b'], ['vincere', 'verb', 'a'], ['vincitore', 'adjective', 'b'], ['vincitore', 'noun', 'b'], ['vincolo', 'noun', 'b'], ['vino', 'noun', 'a'], ['vino', 'adjective', 'a'], ['viola', 'noun', 'b'], ['viola', 'adjective', 'b'], ['violare', 'verb', 'b'], ['violazione', 'noun', 'b'], ['violentare', 'verb', 'c'], ['violento', 'adjective', 'a'], ['violento', 'noun', 'a'], ['violenza', 'noun', 'a'], ['violetta', 'noun', 'c'], ['violetto', 'adjective', 'c'], ['violetto', 'noun', 'c'], ['violino', 'noun', 'b'], ['vipera', 'noun', 'c'], ['virgola', 'noun', 'b'], ['virtù', 'noun', 'b'], ['virtuale', 'adjective', 'b'], ['virus', 'noun', 'b'], ['visibile', 'adjective', 'b'], ['visibile', 'noun', 'b'], ['visione', 'noun', 'a'], ['visita', 'noun', 'a'], ['visitare', 'verb', 'a'], ['visitatore', 'noun', 'b'], ['visivo', 'adjective', 'b'], ['viso', 'noun', 'a'], ['vissuto', 'past_part', 'b'], ['vissuto', 'adjective', 'b'], ['vissuto', 'noun', 'b'], ['vista', 'noun', 'a'], ['vita', 'noun', 'a'], ['vitale', 'adjective', 'b'], ['vitale', 'noun', 'b'], ['vitamina', 'noun', 'c'], ['vite', 'noun', 'c'], ['vitello', 'noun', 'c'], ['vittima', 'noun', 'a'], ['vittoria', 'noun', 'a'], ['vivace', 'adjective', 'b'], ['vivace', 'adverb', 'b'], ['vivace', 'noun', 'b'], ['vivente', 'pres_part', 'b'], ['vivente', 'adjective', 'b'], ['vivente', 'noun', 'b'], ['vivere', 'verb', 'a'], ['vivere', 'noun', 'a'], ['vivo', 'adjective', 'a'], ['vivo', 'noun', 'a'], ['viziare', 'verb', 'c'], ['viziato', 'past_part', 'c'], ['viziato', 'adjective', 'c'], ['vizio', 'noun', 'b'], ['vocabolario', 'noun', 'b'], ['vocale', 'noun', 'b'], ['vocale', 'adjective', 'b'], ['vocazione', 'noun', 'b'], ['voce', 'noun', 'a'], ['vodka', 'noun', 'c'], ['voglia', 'noun', 'a'], ['voi', 'pronoun', 'a'], ['volantino', 'noun', 'c'], ['volare', 'verb', 'a'], ['volata', 'noun', 'c'], ['volenteroso', 'adjective', 'c'], ['volentieri', 'adverb', 'b'], ['volere', 'verb', 'a'], ['volgare', 'adjective', 'b'], ['volgare', 'noun', 'b'], ['volgere', 'verb', 'b'], ['volo', 'noun', 'a'], ['volontà', 'noun', 'a'], ['volontariato', 'noun', 'b'], ['volontario', 'adjective', 'b'], ['volontario', 'noun', 'b'], ['volta', 'noun', 'a'], ['voltare', 'verb', 'a'], ['volto', 'noun', 'a'], ['volume', 'noun', 'a'], ['vomitare', 'verb', 'b'], ['vomito', 'noun', 'c'], ['vongola', 'noun', 'c'], ['vostro', 'adjective', 'a'], ['vostro', 'pronoun', 'a'], ['votare', 'verb', 'a'], ['votazione', 'noun', 'c'], ['voto', 'noun', 'a'], ['vu', 'noun', 'c'], ['vuotare', 'verb', 'c'], ['vuoto', 'adjective', 'a'], ['vuoto', 'noun', 'a'], ['wafer', 'noun', 'c'], ['web', 'noun', 'a'], ['weekend', 'noun', 'b'], ['whisky', 'noun', 'c'], ['wurstel', 'noun', 'c'], ['yogurt', 'noun', 'c'], ['zaino', 'noun', 'b'], ['zampa', 'noun', 'b'], ['zampogna', 'noun', 'c'], ['zanna', 'noun', 'c'], ['zanzara', 'noun', 'c'], ['zattera', 'noun', 'c'], ['zebra', 'noun', 'c'], ['zero', 'adjective', 'a'], ['zero', 'noun', 'a'], ['zero', 'symbol', 'a'], ['zeta', 'noun', 'c'], ['zia', 'noun', 'a'], ['zingaro', 'adjective', 'c'], ['zingaro', 'noun', 'c'], ['zio', 'noun', 'a'], ['zitella', 'noun', 'c'], ['zitto', 'adjective', 'a'], ['zitto', 'noun', 'a'], ['zoccolo', 'noun', 'c'], ['zolla', 'noun', 'c'], ['zona', 'noun', 'a'], ['zoo', 'noun', 'c'], ['zoppicare', 'verb', 'c'], ['zoppo', 'adjective', 'c'], ['zoppo', 'noun', 'c'], ['zucca', 'noun', 'b'], ['zucchero', 'noun', 'b'], ['zucchina', 'noun', 'c'], ['zuffa', 'noun', 'c'], ['zuppa', 'noun', 'c'], ]
254,077
114,529
from __future__ import absolute_import, division, print_function, unicode_literals import os # this arbitrary-looking assortment of functionality is provided here # to have a central place for overrideable behavior. The motivating # use is the FB build environment, where this source file is replaced # by an equivalent. if os.path.basename(os.path.dirname(__file__)) == 'shared': torch_parent = os.path.dirname(os.path.dirname(os.path.dirname(__file__))) else: torch_parent = os.path.dirname(os.path.dirname(__file__)) def get_file_path(*path_components): return os.path.join(torch_parent, *path_components) def get_file_path_2(*path_components): return os.path.join(*path_components) def get_writable_path(path): return path def prepare_multiprocessing_environment(path): pass def resolve_library_path(path): return os.path.realpath(path) TEST_MASTER_ADDR = '127.0.0.1' TEST_MASTER_PORT = 29500
940
312
import json from collections import OrderedDict from typing import Union, List import clip import torch import torch.nn as nn import torch.nn.functional as F from libs.definitions import ROOT label_file = ROOT / 'imagenet_class_index.json' with open(label_file, 'r') as f: labels = json.load(f) _DEFAULT_CLASSNAMES = [value[1] for value in labels.values()] # templates are copied from https://github.com/openai/CLIP/blob/main/notebooks/Prompt_Engineering_for_ImageNet.ipynb _DEFAULT_TEMPLATES = [ 'a bad photo of a {}.', 'a photo of many {}.', 'a sculpture of a {}.', 'a photo of the hard to see {}.', 'a low resolution photo of the {}.', 'a rendering of a {}.', 'graffiti of a {}.', 'a bad photo of the {}.', 'a cropped photo of the {}.', 'a tattoo of a {}.', 'the embroidered {}.', 'a photo of a hard to see {}.', 'a bright photo of a {}.', 'a photo of a clean {}.', 'a photo of a dirty {}.', 'a dark photo of the {}.', 'a drawing of a {}.', 'a photo of my {}.', 'the plastic {}.', 'a photo of the cool {}.', 'a close-up photo of a {}.', 'a black and white photo of the {}.', 'a painting of the {}.', 'a painting of a {}.', 'a pixelated photo of the {}.', 'a sculpture of the {}.', 'a bright photo of the {}.', 'a cropped photo of a {}.', 'a plastic {}.', 'a photo of the dirty {}.', 'a jpeg corrupted photo of a {}.', 'a blurry photo of the {}.', 'a photo of the {}.', 'a good photo of the {}.', 'a rendering of the {}.', 'a {} in a video game.', 'a photo of one {}.', 'a doodle of a {}.', 'a close-up photo of the {}.', 'a photo of a {}.', 'the origami {}.', 'the {} in a video game.', 'a sketch of a {}.', 'a doodle of the {}.', 'a origami {}.', 'a low resolution photo of a {}.', 'the toy {}.', 'a rendition of the {}.', 'a photo of the clean {}.', 'a photo of a large {}.', 'a rendition of a {}.', 'a photo of a nice {}.', 'a photo of a weird {}.', 'a blurry photo of a {}.', 'a cartoon {}.', 'art of a {}.', 'a sketch of the {}.', 'a embroidered {}.', 'a pixelated photo of a {}.', 'itap of the {}.', 'a jpeg corrupted photo of the {}.', 'a good photo of a {}.', 'a plushie {}.', 'a photo of the nice {}.', 'a photo of the small {}.', 'a photo of the weird {}.', 'the cartoon {}.', 'art of the {}.', 'a drawing of the {}.', 'a photo of the large {}.', 'a black and white photo of a {}.', 'the plushie {}.', 'a dark photo of a {}.', 'itap of a {}.', 'graffiti of the {}.', 'a toy {}.', 'itap of my {}.', 'a photo of a cool {}.', 'a photo of a small {}.', 'a tattoo of the {}.', ] class DenseClip(nn.Module): _AVAILABLE_MODELS = ['RN50', 'RN50x16'] # refer to Table 3. in the paper def __init__(self, name: str, classnames: List[str] = None, templates: List[str] = None, device: Union[str, torch.device] = 'cuda' if torch.cuda.is_available() else 'cpu', jit: bool = False, download_root: str = None): super(DenseClip, self).__init__() self.clip_model, self.preprocess = clip.load(name, device, jit, download_root) if classnames is None: classnames = _DEFAULT_CLASSNAMES if templates is None: templates = _DEFAULT_TEMPLATES self._init_visual(device) self._init_zeroshot_classifier(classnames, templates, device) def _init_visual(self, device): self.visual = self.clip_model.visual self.conv1 = nn.Conv2d(self.visual.attnpool.v_proj.in_features, self.visual.attnpool.v_proj.out_features, kernel_size=(1, 1)).to(device).to(self.dtype) self.conv2 = nn.Conv2d(self.visual.attnpool.c_proj.in_features, self.visual.attnpool.c_proj.out_features, kernel_size=(1, 1)).to(device).to(self.dtype) conv1_weight_shape = (*self.visual.attnpool.v_proj.weight.shape, 1, 1) conv2_weight_shape = (*self.visual.attnpool.c_proj.weight.shape, 1, 1) self.conv1.load_state_dict( OrderedDict(weight=self.visual.attnpool.v_proj.weight.reshape(conv1_weight_shape), bias=self.visual.attnpool.v_proj.bias)) self.conv2.load_state_dict( OrderedDict(weight=self.visual.attnpool.c_proj.weight.reshape(conv2_weight_shape), bias=self.visual.attnpool.c_proj.bias)) @torch.no_grad() def _init_zeroshot_classifier(self, classnames, templates, device): # refer to: https://github.com/openai/CLIP/blob/main/notebooks/Prompt_Engineering_for_ImageNet.ipynb zeroshot_weights = [] for classname in classnames: texts = [template.format(classname) for template in templates] # format with class texts = clip.tokenize(texts).to(device) # tokenize class_embeddings = self.clip_model.encode_text(texts) # embed with text encoder class_embeddings /= class_embeddings.norm(dim=-1, keepdim=True) class_embedding = class_embeddings.mean(dim=0) class_embedding /= class_embedding.norm() zeroshot_weights.append(class_embedding) # shape: [E, C] # where E is the dimension of an embedding and C is the number of classes. self.zeroshot_weights = torch.stack(zeroshot_weights, dim=1).to(device) @property def dtype(self): return self.visual.conv1.weight.dtype def _stem(self, x): for conv, bn in [(self.visual.conv1, self.visual.bn1), (self.visual.conv2, self.visual.bn2), (self.visual.conv3, self.visual.bn3)]: x = self.visual.relu(bn(conv(x))) x = self.visual.avgpool(x) return x def encode_image(self, image): image = image.type(self.dtype) feature = self._stem(image) feature = self.visual.layer1(feature) feature = self.visual.layer2(feature) feature = self.visual.layer3(feature) feature = self.visual.layer4(feature) # removed attnpool feature = self.conv1(feature) feature = self.conv2(feature) return feature def forward(self, images): # [B, E, h, w] features = self.encode_image(images) # [B, w, h, E] features_t = features.transpose(1, 3) # [B, w, h, C] output_t = features_t @ self.zeroshot_weights # [B, C, h, w] output = output_t.transpose(1, 3) output = F.interpolate(output, size=images.shape[-2:], mode='bilinear') return output @staticmethod def available_models(): return DenseClip._AVAILABLE_MODELS class Clip(nn.Module): _AVAILABLE_MODELS = ['RN50', 'RN50x16'] # refer to Table 3. in the paper def __init__(self, name: str, classnames: List[str] = None, templates: List[str] = None, device: Union[str, torch.device] = 'cuda' if torch.cuda.is_available() else 'cpu', jit: bool = False, download_root: str = None): super(Clip, self).__init__() self.clip_model, self.preprocess = clip.load(name, device, jit, download_root) if classnames is None: classnames = _DEFAULT_CLASSNAMES if templates is None: templates = _DEFAULT_TEMPLATES self._init_zeroshot_classifier(classnames, templates, device) @torch.no_grad() def _init_zeroshot_classifier(self, classnames, templates, device): # refer to: https://github.com/openai/CLIP/blob/main/notebooks/Prompt_Engineering_for_ImageNet.ipynb zeroshot_weights = [] for classname in classnames: texts = [template.format(classname) for template in templates] # format with class texts = clip.tokenize(texts).to(device) # tokenize class_embeddings = self.clip_model.encode_text(texts) # embed with text encoder class_embeddings /= class_embeddings.norm(dim=-1, keepdim=True) class_embedding = class_embeddings.mean(dim=0) class_embedding /= class_embedding.norm() zeroshot_weights.append(class_embedding) # shape: [E, C] # where E is the dimension of an embedding and C is the number of classes. self.zeroshot_weights = torch.stack(zeroshot_weights, dim=1).to(device) def encode_image(self, image): feature = self.clip_model.encode_image(image) feature /= feature.norm(dim=-1, keepdim=True) return feature def forward(self, images): features = self.encode_image(images) output = features @ self.zeroshot_weights return F.softmax(output, dim=-1) @staticmethod def available_models(): return Clip._AVAILABLE_MODELS
9,099
3,063
#!/usr/bin/env python from __future__ import print_function from optparse import OptionParser import os import random import subprocess import matplotlib matplotlib.use('Agg') import numpy as np import matplotlib.pyplot as plt import pysam from scipy.stats import binom from scipy.stats.mstats import mquantiles import seaborn as sns import stats ################################################################################ # basset_sick_loss.py # # Shuffle SNPs that overlap DNase sites within their sites and compare the SAD # distributions. # # Todo: # -Control for GC% changes introduced by mutation shuffles. # -Control for positional changes within the DHS regions. # -Properly handle indels. ################################################################################ ################################################################################ # main ################################################################################ def main(): usage = 'usage: %prog [options] <vcf_file> <sample_beds_file> <model_file>' parser = OptionParser(usage) parser.add_option('-f', dest='genome_fasta', default='%s/assembly/hg19.fa'%os.environ['HG19'], help='Genome FASTA [Default: %default]') parser.add_option('-g', dest='gpu', default=False, action='store_true', help='Run on GPU [Default: %default]') parser.add_option('-l', dest='seq_len', type='int', default=600, help='Sequence length provided to the model [Default: %default]') parser.add_option('-o', dest='out_dir', default='sad_shuffle', help='Output directory') parser.add_option('-r', dest='replot', default=False, action='store_true', help='Re-plot only, without re-computing [Default: %default]') parser.add_option('-s', dest='num_shuffles', default=1, type='int', help='Number of SNP shuffles [Default: %default]') parser.add_option('-t', dest='sad_table_file', help='Pre-computed SAD scores for the SNPs') (options,args) = parser.parse_args() if len(args) != 3: parser.error('Must provide VCF file, sample BEDs file, and model file') else: vcf_file = args[0] sample_beds_file = args[1] model_file = args[2] if not os.path.isdir(options.out_dir): os.mkdir(options.out_dir) # open reference genome genome = pysam.Fastafile(options.genome_fasta) # open binomial stats file binom_out = open('%s/binom.txt' % options.out_dir, 'w') # open mann-whitney stats file mw_out = open('%s/mannwhitney.txt' % options.out_dir, 'w') # plot defaults sns.set(font_scale=1.5, style='ticks') si = 0 for line in open(sample_beds_file): sample, bed_file = line.split() print(sample) ######################################### # compute SAD ######################################### # filter VCF to overlapping SNPs print(" intersecting SNPs") sample_vcf_file = '%s/%s.vcf' % (options.out_dir,sample) if not options.replot: filter_vcf(vcf_file, bed_file, sample_vcf_file) # compute SAD scores for this sample's SNPs print(" computing SAD") if options.sad_table_file: true_sad = retrieve_sad(sample_vcf_file, options.sad_table_file, si) else: true_sad = compute_sad(sample_vcf_file, model_file, si, '%s/%s_sad'%(options.out_dir,sample), options.seq_len, options.gpu, options.replot) ######################################### # compute shuffled SAD ######################################### shuffle_sad = np.zeros((true_sad.shape[0],options.num_shuffles)) for ni in range(options.num_shuffles): # shuffle the SNPs within their overlapping DHS print(" shuffle %d" % ni) sample_vcf_shuf_file = '%s/%s_shuf%d.vcf' % (options.out_dir, sample, ni) shuffle_snps(sample_vcf_file, sample_vcf_shuf_file, genome) # compute SAD scores for shuffled SNPs print(" computing shuffle SAD") shuffle_sad[:,ni] = compute_sad(sample_vcf_shuf_file, model_file, si, '%s/%s_shuf%d_sad'%(options.out_dir,sample,ni), options.seq_len, options.gpu, options.replot) ######################################### # simple stats ######################################### # compute shuffle means shuffle_sad_mean = shuffle_sad.mean(axis=1) # print sample table sample_sad_out = open('%s/%s_table.txt' % (options.out_dir,sample), 'w') for vi in range(len(true_sad)): print('%f\t%f' % (true_sad[vi], shuffle_sad_mean[vi]), file=sample_sad_out) sample_sad_out.close() # scatter plot # plt.figure() # plt.scatter(true_sad, shuffle_sad_mean, color='black', alpha=0.7) # plt.gca().grid(True, linestyle=':') # plt.savefig('%s/%s_scatter.pdf' % (options.out_dir,sample)) # plt.close() # plot CDFs sns_colors = sns.color_palette('deep') plt.figure() plt.hist(true_sad, 1000, normed=1, histtype='step', cumulative=True, color=sns_colors[0], linewidth=1, label='SNPs') plt.hist(shuffle_sad.flatten(), 1000, normed=1, histtype='step', cumulative=True, color=sns_colors[2], linewidth=1, label='Shuffle') ax = plt.gca() ax.grid(True, linestyle=':') ax.set_xlim(-.2, .2) plt.legend() plt.savefig('%s/%s_cdf.pdf' % (options.out_dir,sample)) plt.close() # plot Q-Q true_q = mquantiles(true_sad, np.linspace(0,1,min(10000,true_sad.shape[0]))) shuf_q = mquantiles(shuffle_sad_mean, np.linspace(0,1,min(10000,true_sad.shape[0]))) plt.figure() plt.scatter(true_q, shuf_q, color=sns_colors[0]) pmin = 1.05*min(true_q[0], shuf_q[0]) pmax = 1.05*max(true_q[-1], shuf_q[-1]) plt.plot([pmin,pmax], [pmin,pmax], color='black', linewidth=1) ax = plt.gca() ax.set_xlim(pmin,pmax) ax.set_ylim(pmin,pmax) ax.set_xlabel('True SAD') ax.set_ylabel('Shuffled SAD') ax.grid(True, linestyle=':') plt.savefig('%s/%s_qq.pdf' % (options.out_dir,sample)) plt.close() ######################################### # statistical tests ######################################### # compute matched binomial test true_great = sum((true_sad-shuffle_sad_mean) > 0) true_lo = np.log2(true_great) - np.log2(len(true_sad)-true_great) if true_lo > 0: binom_p = 1.0 - binom.cdf(true_great-1, n=len(true_sad), p=0.5) else: binom_p = binom.cdf(true_great, n=len(true_sad), p=0.5) # print significance stats cols = (sample, len(true_sad), true_great, true_lo, binom_p) print('%-20s %5d %5d %6.2f %6.1e' % cols, file=binom_out) # compute Mann-Whitney mw_z, mw_p = stats.mannwhitneyu(true_sad, shuffle_sad.flatten()) cols = (sample, len(true_sad), true_sad.mean(), shuffle_sad.mean(), mw_z, mw_p) print('%-20s %5d %6.3f %6.3f %6.2f %6.1e' % cols, file=mw_out) # update sample index si += 1 binom_out.close() mw_out.close() genome.close() def compute_sad(sample_vcf_file, model_file, si, out_dir, seq_len, gpu, replot): ''' Run basset_sad.py to compute scores. ''' cuda_str = '' if gpu: cuda_str = '--cudnn' cmd = 'basset_sad.py %s -l %d -o %s %s %s' % (cuda_str, seq_len, out_dir, model_file, sample_vcf_file) if not replot: subprocess.call(cmd, shell=True) sad = [] for line in open('%s/sad_table.txt' % out_dir): a = line.split() if a[3] == 't%d'%si: sad.append(float(a[-1])) return np.array(sad) def filter_vcf(vcf_file, bed_file, sample_vcf_file): ''' Filter the VCF file for SNPs that overlap the BED file, removing indels. ''' # open filtered file sample_vcf_out = open(sample_vcf_file, 'w') # intersect p = subprocess.Popen('bedtools intersect -wo -a %s -b %s' % (vcf_file, bed_file), stdout=subprocess.PIPE, shell=True) for line in p.stdout: a = line.split() if len(a[3]) == len(a[4]) == 1: print(line, file=sample_vcf_out, end='') sample_vcf_out.close() def retrieve_sad(sample_vcf_file, sad_table_file, si): ''' Retrieve SAD scores from a pre-computed table. Note that I'm assuming here the table has all SAD scores in one row for each SNP so I can pull out the score I want as column si+1. ''' snp_indexes = {} vi = 0 for line in open(sample_vcf_file): a = line.split() snp_indexes[a[2]] = vi vi += 1 sad = np.zeros(len(snp_indexes)) for line in open(sad_table_file): a = line.split() print(a) if a[0] in snp_indexes: sad[snp_indexes[a[0]]] = float(a[si+1]) return sad def shuffle_snps(in_vcf_file, out_vcf_file, genome): ''' Shuffle the SNPs within their overlapping DHS. ''' out_vcf_open = open(out_vcf_file, 'w') for line in open(in_vcf_file): a = line.split() # read SNP info snp_chrom = a[0] snp_pos = int(a[1]) snp_nt = a[3] # determine BED start bi = 5 while a[bi] != snp_chrom: bi += 1 # read BED info bed_chrom = a[bi] bed_start = int(a[bi+1]) bed_end = int(a[bi+2]) # get sequence bed_seq = genome.fetch(bed_chrom, bed_start, bed_end) # determine matching positions bed_nt_matches = [i for i in range(len(bed_seq)) if bed_seq[i] == snp_nt] while len(bed_nt_matches) == 0: # expand segment by 10 nt bed_start = max(0, bed_start-10) bed_end += 10 bed_seq = genome.fetch(bed_chrom, bed_start, bed_end) # sample new SNP position shuf_pos = bed_start + 1 + random.choice(bed_nt_matches) # write into columns a[1] = str(shuf_pos) print('\t'.join(a), file=out_vcf_open) out_vcf_open.close() def shuffle_snps_old(in_vcf_file, out_vcf_file, genome): ''' Shuffle the SNPs within their overlapping DHS. ''' out_vcf_open = open(out_vcf_file, 'w') for line in open(in_vcf_file): a = line.split() # read SNP info snp_chrom = a[0] snp_pos = int(a[1]) # determine BED start bi = 5 while a[bi] != snp_chrom: bi += 1 # read BED info bed_chrom = a[bi] bed_start = int(a[bi+1]) bed_end = int(a[bi+2]) # sample new SNP position shuf_pos = random.randint(bed_start, bed_end-1) while shuf_pos == snp_pos: shuf_pos = random.randint(bed_start, bed_end-1) # set reference allele ref_nt = genome.fetch(snp_chrom, shuf_pos-1, shuf_pos) # sample alternate allele alt_nt = random.choice('ACGT') while alt_nt == ref_nt: alt_nt = random.choice('ACGT') # write into columns a[1] = str(shuf_pos) a[3] = ref_nt a[4] = alt_nt print('\t'.join(a), file=out_vcf_open) ################################################################################ # __main__ ################################################################################ if __name__ == '__main__': main()
11,463
4,091
#!/usr/bin/env python3 import codecs import re def trans_addr(addr): """Traduit une position de fichier en une adresse du programme""" if addr < 0x1000: return 0 if 0x0000000000001000 <= addr < 0x0000000000001000 + 0x00001ef000000000: return 0x00002b0000000000 + addr - 0x0000000000001000 if 0x00002afffffe1000 <= addr < 0x00002afffffe1000 + 0x0000161000000000: return 0x000049f000000000 + addr - 0x00002afffffe1000 if 0x000049effffe1000 <= addr < 0x000049effffe1000 + 0x00002afffffe0000: return 0x0000000000020000 + addr - 0x000049effffe1000 raise Exception("Invalid addr {:#x}".format(addr)) blobs = {} with open('strace_tar_output.log', 'r') as f: curseek = 0 for line in f: m = re.match(r'lseek\(4, ([^,]*), SEEK_SET\)', line) if m is not None: curseek = int(m.group(1)) continue if line.startswith('write(4, "'): m = re.match(r'write\(4, "(.*)", ([0-9]*)\) = ([0-9]*)', line) assert m is not None: rawdata, count1, count2 = m.groups() assert count1 == count2 addr = curseek curseek += int(count1) data = codecs.escape_decode(rawdata.encode('ascii'))[0] # Trouve le premier octet non-nul dans le bloc de données i = 0 while i < len(data) and not data[i]: i += 1 if i >= len(data): continue addr = trans_addr(addr + i) data = data[i:].rstrip(b'\0') with open('out/blob-{:016x}.bin'.format(addr), 'wb') as f: f.write(data)
1,657
717
import pygame_gui import pygame class colors: black = "#000000" dimgrey = "#696969" # lmao darkslategray = "#2F4F4F" TITLE = "TUNAPRO1234" BACKGROUND = colors.darkslategray WIDTH, HEIGHT = 1920, 1080 """ Hızlıca bir plan yapacağım Neural ağları kontrol edebileceğimiz küçük bir framework Ağların gelişimini görebileceğiz değiştirebileceğiz ve kaydedebileceğiz Bunun için -select box yapısı -başlatmak için buton -kaydetme olayları için üstteki şeyden Pencereler -tıkladığımız nöronun bilgilerini gösteren ve değiştirebilen bir pencere -tıkladığımız weightin değişmesini sağlayan bir pencere Norön, katman ve ağ için pygame wrapperları yazacağım Weigth için de bir class olur Kaydetme olayına daha var """ # elements: dict: {"buttons": butonlar, "entries", entryler} class Window: def __init__(self, screen, buttons={}, entries={}): self.buttons = buttons self.entries = entries self.screen = screen def main(): pygame.init() pygame.display.set_caption(TITLE) window_surface = pygame.display.set_mode((WIDTH, HEIGHT), pygame.FULLSCREEN) background = pygame.Surface((WIDTH, HEIGHT)) background.fill(pygame.Color(BACKGROUND)) manager = pygame_gui.UIManager((WIDTH, HEIGHT)) buttons = {} entries = {} selects = {} sliders = {} windows = {} # labels = {} dropdowns = {} #yapf: disable entries["Hello"] = pygame_gui.elements.UITextEntryLine(relative_rect=pygame.Rect((400, 500), (200, 50)), manager=manager) buttons["Hello"] = pygame_gui.elements.UIButton(relative_rect=pygame.Rect((605, 500), (95, 29)), text='ok', manager=manager) sliders["Hello"] = pygame_gui.elements.UIHorizontalSlider(relative_rect=pygame.Rect((400, 534), (300, 20)), start_value=0, value_range=(-20.0, 20.0), manager=manager) dropdowns["Hello"] = pygame_gui.elements.UIDropDownMenu(relative_rect=pygame.Rect((500, 100), (100, 20)), options_list=["1", "2", "3", "4"], starting_option="select", manager=manager) selects["Hello"] = pygame_gui.elements.UISelectionList(relative_rect=pygame.Rect((100, 500), (100, 100)), item_list=["1", "2", "3", "4"], manager=manager) # links["Hello"] = pygame_gui.elements.UITextBox(relative_rect=pygame.Rect((100, 500), (100, 50)), text="LABEL TUNAPRO", manager=manager) windows["Hello"] = pygame_gui.elements.UIWindow(rect=pygame.Rect((100, 100), (200, 200)), manager=manager, window_display_title="test", resizable=True) buttonRect = pygame.Rect(0, 0, 100, 20) buttonRect.bottomright = (-30, -20) anchors = { 'left': 'right', 'right': 'right', 'top': 'bottom', 'bottom': 'bottom' } pygame_gui.elements.UIButton(relative_rect=buttonRect, text='Hello', manager=manager, container=windows["Hello"], anchors=anchors) # yapf: enable # activate: text_box.set_active_effect(pygame_gui.TEXT_EFFECT_TYPING_APPEAR) # activate: text_box.set_active_effect(pygame_gui.TEXT_EFFECT_FADE_OUT) # activate: text_box.set_active_effect(pygame_gui.TEXT_EFFECT_FADE_IN) # deactivate: text_box.set_active_effect(None) clock = pygame.time.Clock() isRunning = True while isRunning: time_delta = clock.tick(60) / 1000.0 for event in pygame.event.get(): if event.type == pygame.QUIT: isRunning = False if event.type == pygame.USEREVENT: if event.ui_element == buttons["Hello"]: if event.user_type == pygame_gui.UI_BUTTON_PRESSED: print('Hello World!') if event.ui_element == dropdowns["Hello"]: if event.user_type == pygame_gui.UI_DROP_DOWN_MENU_CHANGED: print("Selected option:", event.text) if event.ui_element == entries["Hello"]: if event.user_type == pygame_gui.UI_TEXT_ENTRY_FINISHED: print("Entered text:", event.text) # if event.user_type == pygame_gui.UI_TEXT_ENTRY_CHANGED: # print("Changed text:", event.text) if event.ui_element == sliders["Hello"]: if event.user_type == pygame_gui.UI_HORIZONTAL_SLIDER_MOVED: print('current slider value:', event.value) if event.ui_element == selects["Hello"]: if event.user_type == pygame_gui.UI_SELECTION_LIST_NEW_SELECTION: print("Selected item:", event.text) manager.process_events(event) manager.update(time_delta) window_surface.blit(background, (0, 0)) manager.draw_ui(window_surface) pygame.display.update() if __name__ == "__main__": main()
4,826
1,753
import time import pytest from flask import g from flask import session import paho.mqtt.client as paho from SmartSleep.db import get_db from flask import json import runpy msg_nr = 0 messages = [""] broker = 'broker.emqx.io' port = 1883 def update_contor(): global msg_nr msg_nr += 1 def on_message(client, userdata, message): received = json.loads(message.payload) if "status" in received: assert received['status'] == messages[msg_nr] update_contor() elif "db" in received: assert received["db"] == messages[msg_nr] update_contor() def test_cooling_system(client, auth): global msg_nr msg_nr = 0 global messages messages = ['16', "Setting the temperature system level to 1.0", "New temperature system level set to 1.0", '16', "Setting the temperature system level to 2.0", "New temperature system level set to 2.0", '16', "Setting the temperature system level to 3.0", "New temperature system level set to 3.0", '16', "Setting the temperature system level to 4.0", "New temperature system level set to 4.0", '19', "Setting the temperature system level to 3.0", "New temperature system level set to 3.0", '16', "Setting the temperature system level to 4.0", "New temperature system level set to 4.0", "18" ] time.sleep(2) client_mqtt = paho.Client("client-test-snoring") client_mqtt.on_message = on_message client_mqtt.connect(broker) client_mqtt.loop_start() client_mqtt.subscribe("SmartSleep/SoundSensor") auth.login() response = client.post(f"/config/start_to_sleep?sleep_now={True}") assert response.status_code == 200 response = client.post("/config/temp?temperature=18") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=16") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=16") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=16") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=16") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=19") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=16") assert response.status_code == 200 time.sleep(1.5) response = client.post("/config/current_temp?sensor=18") assert response.status_code == 200 time.sleep(1.5)
2,814
913
import theano.tensor as tt def explin(x): return tt.where(x >= 0, 1 + x, tt.exp(x)) def log_exp1p(x): return tt.log1p(tt.exp(x))
136
73
from pyEtherCAT import MasterEtherCAT #ライブラリの読出し nic = "eth0" # ネットワークカードのアドレスを記載 cat = MasterEtherCAT.MasterEtherCAT(nic) ADP = 0x0000 #1台目 ADDR = 0x0E00 #コアレジスタのアドレス cat.APRD(IDX=0x00, ADP=ADP, ADO=ADDR, DATA=[0,0,0,0,0,0,0,0]) #DATAは0を8個(64bit分)の枠を指示 (DATA, WKC) = cat.socket_read() #結果を読出し print("[0x{:04X}]= 0x{:02x}{:02x},0x{:02x}{:02x},0x{:02x}{:02x},0x{:02x}{:02x}".format(ADDR, DATA[7],DATA[6],DATA[5],DATA[4],DATA[3],DATA[2],DATA[1],DATA[0])) #読み出したデータを表示する
470
319
#!/usr/bin/env python # -*- coding: UTF-8 -*- import pprint from base import BaseObject from base import FileIO class PythonLOCParser(BaseObject): """ Parse T/LOC from a Python File """ def __init__(self, file_path: str, is_debug: bool = False): """ Created: 24-Dec-2019 craig.trim@ibm.com * https://github.ibm.com/GTS-CDO/unstructured-analytics/issues/1637#issuecomment-16802191 :param file_path: link to a python file :param is_debug: """ BaseObject.__init__(self, __name__) self._is_debug = is_debug self._file_path = file_path def _lines(self) -> list: lines = FileIO.file_to_lines(self._file_path, use_sort=False) return lines def process(self) -> dict: lines = self._lines() loc = len(lines) tloc = len([line for line in lines if line and len(line.strip())]) d_result = { "Provenance": str(self.__class__.__name__), "FilePath": self._file_path, "LOC": str(loc), "TLOC": str(tloc)} if self._is_debug: self.logger.debug('\n'.join([ "LOC Parsing Complete", pprint.pformat(d_result, indent=4)])) return d_result
1,353
431
import pandas as pd import time df = pd.read_csv("data/sample.csv") for num in range(1000): argx = str(df["x"][num:num+1].get_values()) argy = str(df["y"][num:num+1].get_values()) print("x:{0} / y:{1}".format(argx,argy)) time.sleep(0.1)
255
113
from django.urls import path from . import views app_name = "graphs" urlpatterns = [ path("graphs/", views.GraphListView.as_view(), name="graph.list"), path("graphs/create/", views.GraphCreateView.as_view(), name="graph.create"), path( "graphs/<str:pk>/", views.GraphDetailView.as_view(), name="graph.detail", ), path( "graphs/<str:pk>/update/", views.GraphUpdateView.as_view(), name="graph.update", ), path( "graphs/<str:pk>/delete/", views.GraphDeleteView.as_view(), name="graph.delete", ), ]
601
202
# -------------------------------------------------------- # Fast R-CNN # Copyright (c) 2015 Microsoft # Licensed under The MIT License [see LICENSE for details] # Written by Ross Girshick - extended by Lukas Tuggener # -------------------------------------------------------- """Blob helper functions.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np import cv2 import random def im_list_to_blob(ims): """Convert a list of images into a network input. Assumes images are already prepared (means subtracted, BGR order, ...). """ max_shape = np.array([im.shape for im in ims]).max(axis=0) num_images = len(ims) blob = np.zeros((num_images, max_shape[0], max_shape[1], 3), dtype=np.float32) for i in range(num_images): im = ims[i] blob[i, 0:im.shape[0], 0:im.shape[1], :] = im return blob def prep_im_for_blob(im, pixel_means, global_scale, args): """Mean subtract and scale an image for use in a blob.""" im = im.astype(np.float32, copy=False) # substract mean if args.substract_mean == "True": im -= pixel_means # do global scaling im = cv2.resize(im, None, None, fx=global_scale, fy=global_scale, interpolation=cv2.INTER_LINEAR) im_size_max = np.max(im.shape[0:2]) # Prevent the biggest axis from being more than MAX_SIZE if im_size_max > args.max_edge: if not args.crop == "True": # scale down if bigger than max size re_scale = (float(args.max_edge) / float(im_size_max)) im = cv2.resize(im, None, None, fx=re_scale, fy=re_scale, interpolation=cv2.INTER_LINEAR) global_scale = global_scale*re_scale crop_box = [0,0,im.shape[0],im.shape[1]] else: # Crop image topleft = random.uniform(0,1)<args.crop_top_left_bias # crop to max size if necessary if im.shape[0] <= args.max_edge or topleft: crop_0 = 0 else: crop_0 = random.randint(0,im.shape[0]-args.max_edge) if im.shape[1] <= args.max_edge or topleft: crop_1 = 0 else: crop_1 = random.randint(0,im.shape[1]-args.max_edge) crop_box = [crop_0, crop_1, min(crop_0+args.max_edge,im.shape[0]), min(crop_1+args.max_edge,im.shape[1])] im = im[crop_box[0]:crop_box[2],crop_box[1]:crop_box[3]] else: crop_box = [0, 0, im.shape[0], im.shape[1]] if not args.pad_to == 0: # pad to fit RefineNet #TODO fix refinenet padding problem y_mulity = int(np.ceil(im.shape[0] / float(args.pad_to))) x_mulity = int(np.ceil(im.shape[1] / float(args.pad_to))) canv = np.ones([y_mulity * args.pad_to, x_mulity * args.pad_to,3], dtype=np.uint8) * 255 canv[0:im.shape[0], 0:im.shape[1]] = im im = canv return im, global_scale, crop_box
2,823
1,081
# coding=utf-8 """ Gathers the media folders """ import elib from ._context import Context def get_media_folders(ctx: Context): """ Gathers the media folders """ ctx.info('gathering media folders') media_folders = [] this_folder = ctx.source_folder while True: ctx.debug(f'traversing: "{this_folder}"') media_folder_candidate = elib.path.ensure_path(this_folder, 'media', must_exist=False).absolute() if media_folder_candidate.exists() and media_folder_candidate.is_dir(): ctx.debug(f'media folder found: "{media_folder_candidate}"') media_folders.append(media_folder_candidate) if len(this_folder.parents) is 1: ctx.debug(f'reach mount point at: "{this_folder}"') break this_folder = this_folder.parent # if not media_folders: # raise ConvertError('no media folder found', ctx) ctx.info(f'media folders:\n{elib.pretty_format(media_folders)}') ctx.media_folders = media_folders
1,022
333
from inspect import signature import twister2.tset.TSet as ts from twister2.utils import function_wrapper class TLink: def __init__(self, java_ref, env): self.__java_ref = java_ref self.__env = env def map(self, lam): map_wrapper = function_wrapper(lam) map_func_java_ref = self.__env.functions.map.build(map_wrapper) map_t_set_java_ref = self.__java_ref.map(map_func_java_ref) return ts.TSet(map_t_set_java_ref, self.__env) def flat_map(self, lam): flat_map_wrapper = function_wrapper(lam) flat_map_func_java_ref = self.__env.functions.flat_map.build(flat_map_wrapper) flat_map_t_set_java_ref = self.__java_ref.flatmap(flat_map_func_java_ref) return ts.TSet(flat_map_t_set_java_ref, self.__env) def sink(self, sink_func): sink_wrapper = function_wrapper(sink_func) sink_func_java_ref = self.__env.functions.sink.build(sink_wrapper) self.__java_ref.sink(sink_func_java_ref) def compute(self, compute_func): compute_wrapper = function_wrapper(compute_func) # if function has two arguments, user is expecting the collector version of compute if len(signature(compute_func).parameters) is 3: compute_collector_func_java_ref = self.__env.functions \ .compute_with_collector.build(compute_wrapper) return ts.TSet(self.__java_ref.compute(compute_collector_func_java_ref), self.__env) else: compute_func_java_ref = self.__env.functions.compute.build(compute_wrapper) return ts.TSet(self.__java_ref.compute(compute_func_java_ref), self.__env) def for_each(self, foreach_func): foreach_wrapper = function_wrapper(foreach_func) foreach_func_java_ref = self.__env.functions.apply.build(foreach_wrapper) return ts.TSet(self.__java_ref.forEach(foreach_func_java_ref), self.__env)
1,924
619
from e2cnn.gspaces import * from e2cnn.nn import FieldType from e2cnn.nn import GeometricTensor from e2cnn.group import Representation from e2cnn.group.representation import build_from_discrete_group_representation from ..equivariant_module import EquivariantModule import torch from typing import List, Tuple, Any import numpy as np import math __all__ = ["ConcatenatedNonLinearity"] class ConcatenatedNonLinearity(EquivariantModule): def __init__(self, in_type, function = "c_relu"): r""" Concatenated non-linearities. For each input channel, the module applies the specified activation function both to its value and its opposite (the value multiplied by -1). The number of channels is, therefore, doubled. Notice that not all the representations support this kind of non-linearity. Indeed, only representations with the same pattern of permutation matrices and containing only values in :math:`\{0, 1, -1\}` support it. Args: in_type (FieldType): the input field type function (str): the identifier of the non-linearity. It is used to specify which function to apply. By default (``'c_relu'``), ReLU is used. """ assert isinstance(in_type.gspace, GeneralOnR2) for r in in_type.representations: assert "concatenated" in r.supported_nonlinearities, ( 'Error! Representation "{}" does not support "concatenated"' " non-linearity".format(r.name) ) super(ConcatenatedNonLinearity, self).__init__() self.space = in_type.gspace self.in_type = in_type # compute the output representation given the input one self.out_type = ConcatenatedNonLinearity._transform_fiber_representation( in_type ) # retrieve the activation function to apply if function == "c_relu": self._function = torch.relu elif function == "c_sigmoid": self._function = torch.sigmoid elif function == "c_tanh": self._function = torch.tanh else: raise ValueError('Function "{}" not recognized!'.format(function)) def forward(self, input): assert input.type == self.in_type b, c, w, h = input.tensor.shape # build the output tensor output = torch.empty( b, 2 * c, w, h, dtype=torch.float, device=input.tensor.device ) # each channels is transformed to 2 channels: # first, apply the non-linearity to its value output[:, ::2, ...] = self._function(input.tensor) # then, apply the non-linearity to its values with the sign inverted output[:, 1::2, ...] = self._function(-1 * input.tensor) # wrap the result in a GeometricTensor return GeometricTensor(output, self.out_type) def evaluate_output_shape(self, input_shape): assert len(input_shape) == 4 assert input_shape[1] == self.in_type.size b, c, hi, wi = input_shape return b, self.out_type.size, hi, wi def check_equivariance(self, atol = 1e-6, rtol = 1e-5): c = self.in_type.size x = torch.randn(3, c, 10, 10) x = GeometricTensor(x, self.in_type) errors = [] for el in self.space.testing_elements: out1 = self(x).transform_fibers(el) out2 = self(x.transform_fibers(el)) errs = (out1.tensor - out2.tensor).detach().numpy() errs = np.abs(errs).reshape(-1) print(el, errs.max(), errs.mean(), errs.var()) assert torch.allclose(out1.tensor, out2.tensor, atol=atol, rtol=rtol), ( 'The error found during equivariance check with element "{}" is too' " high: max = {}, mean = {} var ={}".format( el, errs.max(), errs.mean(), errs.var() ) ) errors.append((el, errs.mean())) return errors @staticmethod def _transform_fiber_representation(in_type): r""" Compute the output representation from the input one after applying the concatenated non-linearity. Args: in_type (FieldType): the input field type Returns: (FieldType): the new output field type """ transformed = {} # transform each different input Representation for repr in in_type._unique_representations: transformed[repr] = ConcatenatedNonLinearity._transform_representation(repr) new_representations = [] # concatenate the new representations for repr in in_type.representations: new_representations.append(transformed[repr]) return FieldType(in_type.gspace, new_representations) @staticmethod def _transform_representation(representation): r""" Transform an input :class:`~e2cnn.group.Representation` according to the concatenated non-linearity. The input representation needs to have the pattern of a permutation matrix, with values -1 or 1. The output representation has double the size of the input one and is built by substituting the ``1`` s with 2x2 identity matrices and the ``-1`` s with 2x2 antidiagonal matrix containing ``1`` s. Args: representation (Representation): the input representation Returns: (Representation): the new output representation """ group = representation.group assert not group.continuous # the name of the new representation name = "concatenated_{}".format(representation.name) if name in group.representations: # if the representation has already been built, return it r = group.representations[name] else: # otherwise, build the new representation s = representation.size rep = {} # build the representation for each element for element in group.elements: # retrieve the input representation of the current element r = representation(element) # build the matrix for the output representation of the current element rep[element] = np.zeros((2 * s, 2 * s)) # check if the input matrix has the pattern of a permutation matrix e = [-1] * s for i in range(s): for j in range(s): if not math.isclose(r[i, j], 0, abs_tol=1e-9): if e[i] < 0: e[i] = j else: raise ValueError( '''Error! the representation should have the pattern of a permutation matrix but 2 values have been found in a row for element "{}"'''.format( element ) ) if len(set(e)) != len(e): raise ValueError( '''Error! the representation should have the pattern of a permutation matrix but 2 values have been found in a column for element "{}"'''.format( element ) ) # parse the input representation matrix and fill the output representation accordingly for i in range(s): for j in range(s): if math.isclose(r[i, j], 1, abs_tol=1e-9): # if the current cell contains 1, fill the output submatrix with the 2x2 identity rep[element][2 * i : 2 * i + 2, 2 * j : 2 * j + 2] = np.eye( 2 ) elif math.isclose(r[i, j], -1, abs_tol=1e-9): # if the current cell contains -1, fill the output submatrix with the 2x2 antidigonal matrix rep[element][ 2 * i : 2 * i + 2, 2 * j : 2 * j + 2 ] = np.flipud(np.eye(2)) elif not math.isclose(r[i, j], 0, abs_tol=1e-9): # otherwise the cell has to contain a 0 raise ValueError( '''Error! The representation should be a signed permutation matrix and, therefore, contain only -1, 1 or 0 values but {} found in position({}, {}) for element "{}"'''.format( r[i, j], i, j, element ) ) # the resulting representation is a quotient repreentation and, therefore, # it also supports pointwise non-linearities nonlinearities = representation.supported_nonlinearities.union( ["pointwise"] ) # build the output representation r = build_from_discrete_group_representation( rep, name, group, supported_nonlinearities=nonlinearities ) return r
9,373
2,467
import requests import re import socket from common.colors import bad,que, info, good,run,W,end from common.uriParser import parsing_url as hostd class GatherHost(): def __init__(self,url,headers=None): self.url = url self.headers = headers def match_info(self,regex,data): match = re.search(regex, data) if match: return dict( data=match.group(1) ) def match_printer(self,to_match,match): if match['data']: print(' {0} {1} : {2}'.format(good,to_match,match['data'])) def os_server(self): response = requests.get(self.url, headers=self.headers).headers try: regx = re.compile(r"(.+) \((.+)\)") data = regx.search(response["server"]) try: print(' {0} {1}Server :{2} {3}' .format(good, W, end, data.group(1))) print(' {0} {1}OS :{2} {3}' .format(good, W, end, data.group(2))) except AttributeError: print(' {0} Cannot Find OS & HostingServer ' .format(bad)) except KeyError: print(' {0} Cannot Find the server headers ' .format(bad)) def web_host(self): urldate = "https://input.payapi.io/v1/api/fraud/domain/age/" + hostd(self.url) getinfo = requests.get(urldate, self.headers).text regex_date = r'Date: (.+?)-(.+?)' regex_date = re.compile(regex_date) matches = re.search(regex_date, getinfo) try: if matches: print(' {0} Domain Created on : {1}'.format(good, matches.group(1))) ip = socket.gethostbyname(hostd(self.url)) print(' {0} CloudFlare IP : {1}'.format(good, ip)) ipinfo = "http://ipinfo.io/" + ip + "/json" gather = requests.get(ipinfo, self.headers).text self.match_printer('Country',self.match_info(r'country\": \"(.+?)\"',gather)) self.match_printer('Region',self.match_info(r'region\": \"(.+?)\"',gather)) self.match_printer('Timezone',self.match_info(r'timezone\": \"(.+?)\"',gather)) self.match_printer('Postal',self.match_info(r'postal\": \"(.+?)\"',gather)) self.match_printer('Org',self.match_info(r'org\": \"(.+?)\"',gather)) self.match_printer('Location',self.match_info(r'loc\": \"(.+?)\"',gather)) except Exception as err: print(' {0} Parse Error : {1}' .format(bad,err))
2,551
811
from key import Key from query import Cursor class Datastore(object): '''A Datastore represents storage for any key-value pair. Datastores are general enough to be backed by all kinds of different storage: in-memory caches, databases, a remote datastore, flat files on disk, etc. The general idea is to wrap a more complicated storage facility in a simple, uniform interface, keeping the freedom of using the right tools for the job. In particular, a Datastore can aggregate other datastores in interesting ways, like sharded (to distribute load) or tiered access (caches before databases). While Datastores should be written general enough to accept all sorts of values, some implementations will undoubtedly have to be specific (e.g. SQL databases where fields should be decomposed into columns), particularly to support queries efficiently. ''' # Main API. Datastore mplementations MUST implement these methods. def get(self, key): '''Return the object named by key or None if it does not exist. None takes the role of default value, so no KeyError exception is raised. Args: key: Key naming the object to retrieve Returns: object or None ''' raise NotImplementedError def put(self, key, value): '''Stores the object `value` named by `key`. How to serialize and store objects is up to the underlying datastore. It is recommended to use simple objects (strings, numbers, lists, dicts). Args: key: Key naming `value` value: the object to store. ''' raise NotImplementedError def delete(self, key): '''Removes the object named by `key`. Args: key: Key naming the object to remove. ''' raise NotImplementedError def query(self, query): '''Returns an iterable of objects matching criteria expressed in `query` Implementations of query will be the largest differentiating factor amongst datastores. All datastores **must** implement query, even using query's worst case scenario, see :ref:class:`Query` for details. Args: query: Query object describing the objects to return. Raturns: iterable cursor with all objects matching criteria ''' raise NotImplementedError # Secondary API. Datastores MAY provide optimized implementations. def contains(self, key): '''Returns whether the object named by `key` exists. The default implementation pays the cost of a get. Some datastore implementations may optimize this. Args: key: Key naming the object to check. Returns: boalean whether the object exists ''' return self.get(key) is not None class NullDatastore(Datastore): '''Stores nothing, but conforms to the API. Useful to test with.''' def get(self, key): '''Return the object named by key or None if it does not exist (None).''' return None def put(self, key, value): '''Store the object `value` named by `key` (does nothing).''' pass def delete(self, key): '''Remove the object named by `key` (does nothing).''' pass def query(self, query): '''Returns an iterable of objects matching criteria in `query` (empty).''' return query([]) class DictDatastore(Datastore): '''Simple straw-man in-memory datastore backed by nested dicts.''' def __init__(self): self._items = dict() def _collection(self, key): '''Returns the namespace collection for `key`.''' collection = str(key.path) if not collection in self._items: self._items[collection] = dict() return self._items[collection] def get(self, key): '''Return the object named by `key` or None. Retrieves the object from the collection corresponding to ``key.path``. Args: key: Key naming the object to retrieve. Returns: object or None ''' try: return self._collection(key)[key] except KeyError, e: return None def put(self, key, value): '''Stores the object `value` named by `key`. Stores the object in the collection corresponding to ``key.path``. Args: key: Key naming `value` value: the object to store. ''' if value is None: self.delete(key) else: self._collection(key)[key] = value def delete(self, key): '''Removes the object named by `key`. Removes the object from the collection corresponding to ``key.path``. Args: key: Key naming the object to remove. ''' try: del self._collection(key)[key] if len(self._collection(key)) == 0: del self._items[str(key.path)] except KeyError, e: pass def contains(self, key): '''Returns whether the object named by `key` exists. Checks for the object in the collection corresponding to ``key.path``. Args: key: Key naming the object to check. Returns: boalean whether the object exists ''' return key in self._collection(key) def query(self, query): '''Returns an iterable of objects matching criteria expressed in `query` Naively applies the query operations on the objects within the namespaced collection corresponding to ``query.key.path``. Args: query: Query object describing the objects to return. Raturns: iterable cursor with all objects matching criteria ''' # entire dataset already in memory, so ok to apply query naively if str(query.key) in self._items: return query(self._items[str(query.key)].values()) else: return query([]) def __len__(self): return sum(map(len, self._items.values())) class InterfaceMappingDatastore(Datastore): '''Represents simple wrapper datastore around an object that, though not a Datastore, implements data storage through a similar interface. For example, memcached and redis both implement a `get`, `set`, `delete` interface. ''' def __init__(self, service, get='get', put='put', delete='delete', key=str): '''Initialize the datastore with given `service`. Args: service: A service that provides data storage through a similar interface to Datastore. Using the service should only require a simple mapping of methods, such as {put : set}. get: The attribute name of the `service` method implementing get put: The attribute name of the `service` method implementing put delete: The attribute name of the `service` method implementing delete key: A function converting a Datastore key (of type Key) into a `service` key. The conversion will often be as simple as `str`. ''' self._service = service self._service_key = key self._service_ops = {} self._service_ops['get'] = getattr(service, get) self._service_ops['put'] = getattr(service, put) self._service_ops['delete'] = getattr(service, delete) # AttributeError will be raised if service does not implement the interface def get(self, key): '''Return the object in `service` named by `key` or None. Args: key: Key naming the object to retrieve. Returns: object or None ''' key = self._service_key(key) return self._service_ops['get'](key) def put(self, key, value): '''Stores the object `value` named by `key` in `service`. Args: key: Key naming `value`. value: the object to store. ''' key = self._service_key(key) self._service_ops['put'](key, value) def delete(self, key): '''Removes the object named by `key` in `service`. Args: key: Key naming the object to remove. ''' key = self._service_key(key) self._service_ops['delete'](key) class ShimDatastore(Datastore): '''Represents a non-concrete datastore that adds functionality between the client and a lower level datastore. Shim datastores do not actually store data themselves; instead, they delegate storage to an underlying child datastore. The default implementation just passes all calls to the child. ''' def __init__(self, datastore): '''Initializes this ShimDatastore with child `datastore`.''' if not isinstance(datastore, Datastore): errstr = 'datastore must be of type %s. Got %s.' raise TypeError(errstr % (Datastore, datastore)) self.child_datastore = datastore # default implementation just passes all calls to child def get(self, key): '''Return the object named by key or None if it does not exist. Default shim implementation simply returns ``child_datastore.get(key)`` Override to provide different functionality, for example:: def get(self, key): value = self.child_datastore.get(key) return json.loads(value) Args: key: Key naming the object to retrieve Returns: object or None ''' return self.child_datastore.get(key) def put(self, key, value): '''Stores the object `value` named by `key`. Default shim implementation simply calls ``child_datastore.put(key, value)`` Override to provide different functionality, for example:: def put(self, key, value): value = json.dumps(value) self.child_datastore.put(key, value) Args: key: Key naming `value`. value: the object to store. ''' self.child_datastore.put(key, value) def delete(self, key): '''Removes the object named by `key`. Default shim implementation simply calls ``child_datastore.delete(key)`` Override to provide different functionality. Args: key: Key naming the object to remove. ''' self.child_datastore.delete(key) def query(self, query): '''Returns an iterable of objects matching criteria expressed in `query`. Default shim implementation simply returns ``child_datastore.query(query)`` Override to provide different functionality, for example:: def query(self, query): cursor = self.child_datastore.query(query) cursor._iterable = deserialized(cursor._iterable) return cursor Args: query: Query object describing the objects to return. Raturns: iterable cursor with all objects matching criteria ''' return self.child_datastore.query(query) class CacheShimDatastore(ShimDatastore): '''Wraps a datastore with a caching shim optimizes some calls.''' def __init__(self, *args, **kwargs): self.cache_datastore = kwargs.pop('cache') if not isinstance(self.cache_datastore, Datastore): errstr = 'datastore must be of type %s. Got %s.' raise TypeError(errstr % (Datastore, self.cache_datastore)) super(CacheShimDatastore, self).__init__(*args, **kwargs) def get(self, key): '''Return the object named by key or None if it does not exist. CacheShimDatastore first checks its ``cache_datastore``. ''' value = self.cache_datastore.get(key) return value if value is not None else self.child_datastore.get(key) def put(self, key, value): '''Stores the object `value` named by `key`self. Writes to both ``cache_datastore`` and ``child_datastore``. ''' self.cache_datastore.put(key, value) self.child_datastore.put(key, value) def delete(self, key): '''Removes the object named by `key`. Writes to both ``cache_datastore`` and ``child_datastore``. ''' self.cache_datastore.delete(key) self.child_datastore.delete(key) def contains(self, key): '''Returns whether the object named by `key` exists. First checks ``cache_datastore``. ''' return self.cache_datastore.contains(key) \ or self.child_datastore.contains(key) class LoggingDatastore(ShimDatastore): '''Wraps a datastore with a logging shim.''' def __init__(self, child_datastore, logger=None): if not logger: import logging logger = logging self.logger = logger super(LoggingDatastore, self).__init__(child_datastore) def get(self, key): '''Return the object named by key or None if it does not exist. LoggingDatastore logs the access. ''' self.logger.info('%s: get %s' % (self, key)) value = super(LoggingDatastore, self).get(key) self.logger.debug('%s: %s' % (self, value)) return value def put(self, key, value): '''Stores the object `value` named by `key`self. LoggingDatastore logs the access. ''' self.logger.info('%s: put %s' % (self, key)) self.logger.debug('%s: %s' % (self, value)) super(LoggingDatastore, self).put(key, value) def delete(self, key): '''Removes the object named by `key`. LoggingDatastore logs the access. ''' self.logger.info('%s: delete %s' % (self, key)) super(LoggingDatastore, self).delete(key) def contains(self, key): '''Returns whether the object named by `key` exists. LoggingDatastore logs the access. ''' self.logger.info('%s: contains %s' % (self, key)) return super(LoggingDatastore, self).contains(key) def query(self, query): '''Returns an iterable of objects matching criteria expressed in `query`. LoggingDatastore logs the access. ''' self.logger.info('%s: query %s' % (self, query)) return super(LoggingDatastore, self).query(query) class KeyTransformDatastore(ShimDatastore): '''Represents a simple ShimDatastore that applies a transform on all incoming keys. For example: >>> import datastore.core >>> def transform(key): ... return key.reverse ... >>> ds = datastore.DictDatastore() >>> kt = datastore.KeyTransformDatastore(ds, keytransform=transform) None >>> ds.put(datastore.Key('/a/b/c'), 'abc') >>> ds.get(datastore.Key('/a/b/c')) 'abc' >>> kt.get(datastore.Key('/a/b/c')) None >>> kt.get(datastore.Key('/c/b/a')) 'abc' >>> ds.get(datastore.Key('/c/b/a')) None ''' def __init__(self, *args, **kwargs): '''Initializes KeyTransformDatastore with `keytransform` function.''' self.keytransform = kwargs.pop('keytransform', None) super(KeyTransformDatastore, self).__init__(*args, **kwargs) def get(self, key): '''Return the object named by keytransform(key).''' return self.child_datastore.get(self._transform(key)) def put(self, key, value): '''Stores the object names by keytransform(key).''' return self.child_datastore.put(self._transform(key), value) def delete(self, key): '''Removes the object named by keytransform(key).''' return self.child_datastore.delete(self._transform(key)) def contains(self, key): '''Returns whether the object named by key is in this datastore.''' return self.child_datastore.contains(self._transform(key)) def query(self, query): '''Returns a sequence of objects matching criteria expressed in `query`''' query = query.copy() query.key = self._transform(query.key) return self.child_datastore.query(query) def _transform(self, key): '''Returns a `key` transformed by `self.keytransform`.''' return self.keytransform(key) if self.keytransform else key class LowercaseKeyDatastore(KeyTransformDatastore): '''Represents a simple ShimDatastore that lowercases all incoming keys. For example: >>> import datastore.core >>> ds = datastore.DictDatastore() >>> ds.put(datastore.Key('hello'), 'world') >>> ds.put(datastore.Key('HELLO'), 'WORLD') >>> ds.get(datastore.Key('hello')) 'world' >>> ds.get(datastore.Key('HELLO')) 'WORLD' >>> ds.get(datastore.Key('HeLlO')) None >>> lds = datastore.LowercaseKeyDatastore(ds) >>> lds.get(datastore.Key('HeLlO')) 'world' >>> lds.get(datastore.Key('HeLlO')) 'world' >>> lds.get(datastore.Key('HeLlO')) 'world' ''' def __init__(self, *args, **kwargs): '''Initializes KeyTransformDatastore with keytransform function.''' super(LowercaseKeyDatastore, self).__init__(*args, **kwargs) self.keytransform = self.lowercaseKey @classmethod def lowercaseKey(cls, key): '''Returns a lowercased `key`.''' return Key(str(key).lower()) class NamespaceDatastore(KeyTransformDatastore): '''Represents a simple ShimDatastore that namespaces all incoming keys. For example: >>> import datastore.core >>> >>> ds = datastore.DictDatastore() >>> ds.put(datastore.Key('/a/b'), 'ab') >>> ds.put(datastore.Key('/c/d'), 'cd') >>> ds.put(datastore.Key('/a/b/c/d'), 'abcd') >>> >>> nd = datastore.NamespaceDatastore('/a/b', ds) >>> nd.get(datastore.Key('/a/b')) None >>> nd.get(datastore.Key('/c/d')) 'abcd' >>> nd.get(datastore.Key('/a/b/c/d')) None >>> nd.put(datastore.Key('/c/d'), 'cd') >>> ds.get(datastore.Key('/a/b/c/d')) 'cd' ''' def __init__(self, namespace, *args, **kwargs): '''Initializes NamespaceDatastore with `key` namespace.''' super(NamespaceDatastore, self).__init__(*args, **kwargs) self.keytransform = self.namespaceKey self.namespace = Key(namespace) def namespaceKey(self, key): '''Returns a namespaced `key`: namespace.child(key).''' return self.namespace.child(key) class NestedPathDatastore(KeyTransformDatastore): '''Represents a simple ShimDatastore that shards/namespaces incoming keys. Incoming keys are sharded into nested namespaces. The idea is to use the key name to separate into nested namespaces. This is akin to the directory structure that ``git`` uses for objects. For example: >>> import datastore.core >>> >>> ds = datastore.DictDatastore() >>> np = datastore.NestedPathDatastore(ds, depth=3, length=2) >>> >>> np.put(datastore.Key('/abcdefghijk'), 1) >>> np.get(datastore.Key('/abcdefghijk')) 1 >>> ds.get(datastore.Key('/abcdefghijk')) None >>> ds.get(datastore.Key('/ab/cd/ef/abcdefghijk')) 1 >>> np.put(datastore.Key('abc'), 2) >>> np.get(datastore.Key('abc')) 2 >>> ds.get(datastore.Key('/ab/ca/bc/abc')) 2 ''' _default_depth = 3 _default_length = 2 _default_keyfn = lambda key: key.name _default_keyfn = staticmethod(_default_keyfn) def __init__(self, *args, **kwargs): '''Initializes KeyTransformDatastore with keytransform function. kwargs: depth: the nesting level depth (e.g. 3 => /1/2/3/123) default: 3 length: the nesting level length (e.g. 2 => /12/123456) default: 2 ''' # assign the nesting variables self.nest_depth = kwargs.pop('depth', self._default_depth) self.nest_length = kwargs.pop('length', self._default_length) self.nest_keyfn = kwargs.pop('keyfn', self._default_keyfn) super(NestedPathDatastore, self).__init__(*args, **kwargs) self.keytransform = self.nestKey def query(self, query): # Requires supporting * operator on queries. raise NotImplementedError def nestKey(self, key): '''Returns a nested `key`.''' nest = self.nest_keyfn(key) # if depth * length > len(key.name), we need to pad. mult = 1 + int(self.nest_depth * self.nest_length / len(nest)) nest = nest * mult pref = Key(self.nestedPath(nest, self.nest_depth, self.nest_length)) return pref.child(key) @staticmethod def nestedPath(path, depth, length): '''returns a nested version of `basename`, using the starting characters. For example: >>> NestedPathDatastore.nested_path('abcdefghijk', 3, 2) 'ab/cd/ef' >>> NestedPathDatastore.nested_path('abcdefghijk', 4, 2) 'ab/cd/ef/gh' >>> NestedPathDatastore.nested_path('abcdefghijk', 3, 4) 'abcd/efgh/ijk' >>> NestedPathDatastore.nested_path('abcdefghijk', 1, 4) 'abcd' >>> NestedPathDatastore.nested_path('abcdefghijk', 3, 10) 'abcdefghij/k' ''' components = [path[n:n+length] for n in xrange(0, len(path), length)] components = components[:depth] return '/'.join(components) class SymlinkDatastore(ShimDatastore): '''Datastore that creates filesystem-like symbolic link keys. A symbolic link key is a way of naming the same value with multiple keys. For example: >>> import datastore.core >>> >>> dds = datastore.DictDatastore() >>> sds = datastore.SymlinkDatastore(dds) >>> >>> a = datastore.Key('/A') >>> b = datastore.Key('/B') >>> >>> sds.put(a, 1) >>> sds.get(a) 1 >>> sds.link(a, b) >>> sds.get(b) 1 >>> sds.put(b, 2) >>> sds.get(b) 2 >>> sds.get(a) 2 >>> sds.delete(a) >>> sds.get(a) None >>> sds.get(b) None >>> sds.put(a, 3) >>> sds.get(a) 3 >>> sds.get(b) 3 >>> sds.delete(b) >>> sds.get(b) None >>> sds.get(a) 3 ''' sentinel = 'datastore_link' def _link_value_for_key(self, source_key): '''Returns the link value for given `key`.''' return str(source_key.child(self.sentinel)) def _link_for_value(self, value): '''Returns the linked key if `value` is a link, or None.''' try: key = Key(value) if key.name == self.sentinel: return key.parent except: pass return None def _follow_link(self, value): '''Returns given `value` or, if it is a symlink, the `value` it names.''' seen_keys = set() while True: link_key = self._link_for_value(value) if not link_key: return value assert link_key not in seen_keys, 'circular symlink reference' seen_keys.add(link_key) value = super(SymlinkDatastore, self).get(link_key) def _follow_link_gen(self, iterable): '''A generator that follows links in values encountered.''' for item in iterable: yield self._follow_link(item) def link(self, source_key, target_key): '''Creates a symbolic link key pointing from `target_key` to `source_key`''' link_value = self._link_value_for_key(source_key) # put straight into the child, to avoid following previous links. self.child_datastore.put(target_key, link_value) # exercise the link. ensure there are no cycles. self.get(target_key) def get(self, key): '''Return the object named by `key. Follows links.''' value = super(SymlinkDatastore, self).get(key) return self._follow_link(value) def put(self, key, value): '''Stores the object named by `key`. Follows links.''' # if value is a link, don't follow links if self._link_for_value(value): super(SymlinkDatastore, self).put(key, value) return # if `key` points to a symlink, need to follow it. current_value = super(SymlinkDatastore, self).get(key) link_key = self._link_for_value(current_value) if link_key: self.put(link_key, value) # self.put: could be another link. else: super(SymlinkDatastore, self).put(key, value) def query(self, query): '''Returns objects matching criteria expressed in `query`. Follows links.''' results = super(SymlinkDatastore, self).query(query) return self._follow_link_gen(results) class DirectoryDatastore(ShimDatastore): '''Datastore that allows manual tracking of directory entries. For example: >>> ds = DirectoryDatastore(ds) >>> >>> # initialize directory at /foo >>> ds.directory(Key('/foo')) >>> >>> # adding directory entries >>> ds.directoryAdd(Key('/foo'), Key('/foo/bar')) >>> ds.directoryAdd(Key('/foo'), Key('/foo/baz')) >>> >>> # value is a generator returning all the keys in this dir >>> for key in ds.directoryRead(Key('/foo')): ... print key Key('/foo/bar') Key('/foo/baz') >>> >>> # querying for a collection works >>> for item in ds.query(Query(Key('/foo'))): ... print item 'bar' 'baz' ''' def directory(self, dir_key): '''Initializes directory at dir_key.''' dir_items = self.get(dir_key) if not isinstance(dir_items, list): self.put(dir_key, []) def directoryRead(self, dir_key): '''Returns a generator that iterates over all keys in the directory referenced by `dir_key` Returns None if the directory `dir_key` does not exist ''' return self.directory_entries_generator(dir_key) def directoryAdd(self, dir_key, key): '''Adds directory entry `key` to directory at `dir_key`. If the directory `dir_key` does not exist, it is created. ''' key = str(key) dir_items = self.get(dir_key) or [] if key not in dir_items: dir_items.append(key) self.put(dir_key, dir_items) def directoryRemove(self, dir_key, key): '''Removes directory entry `key` from directory at `dir_key`. If either the directory `dir_key` or the directory entry `key` don't exist, this method is a no-op. ''' key = str(key) dir_items = self.get(dir_key) or [] if key in dir_items: dir_items = [k for k in dir_items if k != key] self.put(dir_key, dir_items) def directory_entries_generator(self, dir_key): dir_items = self.get(dir_key) or [] for item in dir_items: yield Key(item) class DirectoryTreeDatastore(ShimDatastore): '''Datastore that tracks directory entries, like in a filesystem. All key changes cause changes in a collection-like directory. For example: >>> import datastore.core >>> >>> dds = datastore.DictDatastore() >>> rds = datastore.DirectoryTreeDatastore(dds) >>> >>> a = datastore.Key('/A') >>> b = datastore.Key('/A/B') >>> c = datastore.Key('/A/C') >>> >>> rds.get(a) [] >>> rds.put(b, 1) >>> rds.get(b) 1 >>> rds.get(a) ['/A/B'] >>> rds.put(c, 1) >>> rds.get(c) 1 >>> rds.get(a) ['/A/B', '/A/C'] >>> rds.delete(b) >>> rds.get(a) ['/A/C'] >>> rds.delete(c) >>> rds.get(a) [] ''' def put(self, key, value): '''Stores the object `value` named by `key`self. DirectoryTreeDatastore stores a directory entry. ''' super(DirectoryTreeDatastore, self).put(key, value) str_key = str(key) # ignore root if str_key == '/': return # retrieve directory, to add entry dir_key = key.parent.instance('directory') directory = self.directory(dir_key) # ensure key is in directory if str_key not in directory: directory.append(str_key) super(DirectoryTreeDatastore, self).put(dir_key, directory) def delete(self, key): '''Removes the object named by `key`. DirectoryTreeDatastore removes the directory entry. ''' super(DirectoryTreeDatastore, self).delete(key) str_key = str(key) # ignore root if str_key == '/': return # retrieve directory, to remove entry dir_key = key.parent.instance('directory') directory = self.directory(dir_key) # ensure key is not in directory if directory and str_key in directory: directory.remove(str_key) if len(directory) > 0: super(DirectoryTreeDatastore, self).put(dir_key, directory) else: super(DirectoryTreeDatastore, self).delete(dir_key) def query(self, query): '''Returns objects matching criteria expressed in `query`. DirectoryTreeDatastore uses directory entries. ''' return query(self.directory_values_generator(query.key)) def directory(self, key): '''Retrieves directory entries for given key.''' if key.name != 'directory': key = key.instance('directory') return self.get(key) or [] def directory_values_generator(self, key): '''Retrieve directory values for given key.''' directory = self.directory(key) for key in directory: yield self.get(Key(key)) class DatastoreCollection(ShimDatastore): '''Represents a collection of datastores.''' def __init__(self, stores=[]): '''Initialize the datastore with any provided datastores.''' if not isinstance(stores, list): stores = list(stores) for store in stores: if not isinstance(store, Datastore): raise TypeError("all stores must be of type %s" % Datastore) self._stores = stores def datastore(self, index): '''Returns the datastore at `index`.''' return self._stores[index] def appendDatastore(self, store): '''Appends datastore `store` to this collection.''' if not isinstance(store, Datastore): raise TypeError("stores must be of type %s" % Datastore) self._stores.append(store) def removeDatastore(self, store): '''Removes datastore `store` from this collection.''' self._stores.remove(store) def insertDatastore(self, index, store): '''Inserts datastore `store` into this collection at `index`.''' if not isinstance(store, Datastore): raise TypeError("stores must be of type %s" % Datastore) self._stores.insert(index, store) class TieredDatastore(DatastoreCollection): '''Represents a hierarchical collection of datastores. Each datastore is queried in order. This is helpful to organize access order in terms of speed (i.e. read caches first). Datastores should be arranged in order of completeness, with the most complete datastore last, as it will handle query calls. Semantics: * get : returns first found value * put : writes through to all * delete : deletes through to all * contains : returns first found value * query : queries bottom (most complete) datastore ''' def get(self, key): '''Return the object named by key. Checks each datastore in order.''' value = None for store in self._stores: value = store.get(key) if value is not None: break # add model to lower stores only if value is not None: for store2 in self._stores: if store == store2: break store2.put(key, value) return value def put(self, key, value): '''Stores the object in all underlying datastores.''' for store in self._stores: store.put(key, value) def delete(self, key): '''Removes the object from all underlying datastores.''' for store in self._stores: store.delete(key) def query(self, query): '''Returns a sequence of objects matching criteria expressed in `query`. The last datastore will handle all query calls, as it has a (if not the only) complete record of all objects. ''' # queries hit the last (most complete) datastore return self._stores[-1].query(query) def contains(self, key): '''Returns whether the object is in this datastore.''' for store in self._stores: if store.contains(key): return True return False class ShardedDatastore(DatastoreCollection): '''Represents a collection of datastore shards. A datastore is selected based on a sharding function. Sharding functions should take a Key and return an integer. WARNING: adding or removing datastores while mid-use may severely affect consistency. Also ensure the order is correct upon initialization. While this is not as important for caches, it is crucial for persistent datastores. ''' def __init__(self, stores=[], shardingfn=hash): '''Initialize the datastore with any provided datastore.''' if not callable(shardingfn): raise TypeError('shardingfn (type %s) is not callable' % type(shardingfn)) super(ShardedDatastore, self).__init__(stores) self._shardingfn = shardingfn def shard(self, key): '''Returns the shard index to handle `key`, according to sharding fn.''' return self._shardingfn(key) % len(self._stores) def shardDatastore(self, key): '''Returns the shard to handle `key`.''' return self.datastore(self.shard(key)) def get(self, key): '''Return the object named by key from the corresponding datastore.''' return self.shardDatastore(key).get(key) def put(self, key, value): '''Stores the object to the corresponding datastore.''' self.shardDatastore(key).put(key, value) def delete(self, key): '''Removes the object from the corresponding datastore.''' self.shardDatastore(key).delete(key) def contains(self, key): '''Returns whether the object is in this datastore.''' return self.shardDatastore(key).contains(key) def query(self, query): '''Returns a sequence of objects matching criteria expressed in `query`''' cursor = Cursor(query, self.shard_query_generator(query)) cursor.apply_order() # ordering sharded queries is expensive (no generator) return cursor def shard_query_generator(self, query): '''A generator that queries each shard in sequence.''' shard_query = query.copy() for shard in self._stores: # yield all items matching within this shard cursor = shard.query(shard_query) for item in cursor: yield item # update query with results of first query shard_query.offset = max(shard_query.offset - cursor.skipped, 0) if shard_query.limit: shard_query.limit = max(shard_query.limit - cursor.returned, 0) if shard_query.limit <= 0: break # we're already done! ''' Hello Tiered Access >>> import pymongo >>> import datastore.core >>> >>> from datastore.impl.mongo import MongoDatastore >>> from datastore.impl.lrucache import LRUCache >>> from datastore.impl.filesystem import FileSystemDatastore >>> >>> conn = pymongo.Connection() >>> mongo = MongoDatastore(conn.test_db) >>> >>> cache = LRUCache(1000) >>> fs = FileSystemDatastore('/tmp/.test_db') >>> >>> ds = datastore.TieredDatastore([cache, mongo, fs]) >>> >>> hello = datastore.Key('hello') >>> ds.put(hello, 'world') >>> ds.contains(hello) True >>> ds.get(hello) 'world' >>> ds.delete(hello) >>> ds.get(hello) None Hello Sharding >>> import datastore.core >>> >>> shards = [datastore.DictDatastore() for i in range(0, 10)] >>> >>> ds = datastore.ShardedDatastore(shards) >>> >>> hello = datastore.Key('hello') >>> ds.put(hello, 'world') >>> ds.contains(hello) True >>> ds.get(hello) 'world' >>> ds.delete(hello) >>> ds.get(hello) None '''
34,379
10,916
from django.contrib import admin from django.contrib.admin import widgets from .models import Person, Publication, Position, Talk, Project, Poster, Keyword, News, Banner, Video, Project_header, Photo, Project_umbrella, Project_Role, Sponsor from website.admin_list_filters import PositionRoleListFilter, PositionTitleListFilter, PubVenueTypeListFilter, PubVenueListFilter from sortedm2m_filter_horizontal_widget.forms import SortedFilteredSelectMultiple import django from django import forms from django.http import HttpResponse from datetime import datetime from django.template import loader from django.template import RequestContext from django.shortcuts import redirect from django import forms import urllib import bibtexparser from image_cropping import ImageCroppingMixin class BannerAdmin(ImageCroppingMixin, admin.ModelAdmin): fieldsets = [ (None, {'fields': ["page", "title", "caption", "alt_text", "link", "favorite", "project"]}), # ('Image', {'fields': ["image", "image_preview"]}) ('Image', {'fields': ["image", "cropping"]}) ] # The list display lets us control what is shown in the default persons table at Home > Website > Banners # info on displaying multiple entries comes from http://stackoverflow.com/questions/9164610/custom-columns-using-django-admin list_display = ('title', 'project', 'page', 'favorite', 'image') # readonly_fields = ["image_preview"] class PositionInline(admin.StackedInline): model = Position # This specifies that the Inline is linked to the main owner of the position rather than any of the advisor roles. fk_name = "person" # This specifies that the field appears only once (by default) extra = 0 def formfield_for_foreignkey(self, db_field, request, **kwargs): print("PositionInline.formfield_for_foreignkey: db_field: {} db_field.name {} request: {}".format(db_field, db_field.name, request)) if db_field.name == "advisor" or db_field.name == "co_advisor": # Filters advisors to professors and sorts by first name # Based on: http://stackoverflow.com/a/30627555 professor_ids = [person.id for person in Person.objects.all() if person.is_professor()] filtered_persons = Person.objects.filter(id__in=professor_ids).order_by('first_name') print(filtered_persons) kwargs["queryset"] = filtered_persons elif db_field.name == "grad_mentor": # Filters grad mentor list to current grad students (either member or collaborator) grad_ids = [person.id for person in Person.objects.all() if person.is_grad_student() and (person.is_current_member() or person.is_current_collaborator())] filtered_persons = Person.objects.filter(id__in=grad_ids).order_by('first_name') print(filtered_persons) kwargs["queryset"] = filtered_persons return super(PositionInline, self).formfield_for_foreignkey(db_field, request, **kwargs) class ProjectRoleInline(admin.StackedInline): model = Project_Role extra = 0 class ProjectHeaderInline(ImageCroppingMixin, admin.StackedInline): model = Project_header extra = 0 # Uses format as per https://github.com/jonasundderwolf/django-image-cropping to add cropping to the admin page class NewsAdmin(ImageCroppingMixin, admin.ModelAdmin): # Filters authors only to current members and sorts by firstname # Based on: http://stackoverflow.com/a/30627555 def formfield_for_foreignkey(self, db_field, request, **kwargs): # print("NewsAdmin.formfield_for_foreignkey: db_field: {} db_field.name {} request: {}".format(db_field, db_field.name, request)) if db_field.name == "author": current_member_ids = [person.id for person in Person.objects.all() if person.is_current_member()] filtered_persons = Person.objects.filter(id__in=current_member_ids).order_by('first_name') print(filtered_persons) kwargs["queryset"] = filtered_persons return super(NewsAdmin, self).formfield_for_foreignkey(db_field, request, **kwargs) def formfield_for_manytomany(self, db_field, request=None, **kwargs): if db_field.name == "project": kwargs["widget"] = widgets.FilteredSelectMultiple("project", is_stacked=False) return super(NewsAdmin, self).formfield_for_manytomany(db_field, request, **kwargs) class PhotoAdmin(ImageCroppingMixin, admin.ModelAdmin): list_display = ('__str__', 'admin_thumbnail') class ProjectAdmin(ImageCroppingMixin, admin.ModelAdmin): inlines = [ProjectHeaderInline] # The list display lets us control what is shown in the Project table at Home > Website > Project # info on displaying multiple entries comes from http://stackoverflow.com/questions/9164610/custom-columns-using-django-admin list_display = ('name', 'start_date', 'end_date', 'has_ended', 'get_people_count', 'get_current_member_count', 'get_past_member_count', 'get_most_recent_artifact_date', 'get_most_recent_artifact_type', 'get_publication_count', 'get_video_count', 'get_talk_count') def formfield_for_manytomany(self, db_field, request=None, **kwargs): if db_field.name == "sponsors": kwargs["widget"] = widgets.FilteredSelectMultiple("sponsors", is_stacked=False) if db_field.name == "keywords": kwargs["widget"] = widgets.FilteredSelectMultiple("keywords", is_stacked=False) if db_field.name == "project_umbrellas": kwargs["widget"] = widgets.FilteredSelectMultiple("project umbrellas", is_stacked=False) return super(ProjectAdmin, self).formfield_for_manytomany(db_field, request, **kwargs) class PersonAdmin(ImageCroppingMixin, admin.ModelAdmin): # inlines allow us to edit models on the same page as a parent model # see: https://docs.djangoproject.com/en/1.11/ref/contrib/admin/#inlinemodeladmin-objects inlines = [PositionInline, ProjectRoleInline] # The list display lets us control what is shown in the default persons table at Home > Website > People # info on displaying multiple entries comes from http://stackoverflow.com/questions/9164610/custom-columns-using-django-admin list_display = ('get_full_name', 'get_current_title', 'get_current_role', 'is_active', 'get_start_date', 'get_end_date', 'get_time_in_current_position', 'get_total_time_as_member') #TODO setup filter here that has diff categories (like active members, past, etc.): #https://www.elements.nl/2015/03/16/getting-the-most-out-of-django-admin-filters/ #related to: https://github.com/jonfroehlich/makeabilitylabwebsite/issues/238 list_filter = (PositionRoleListFilter, PositionTitleListFilter) class VideoAdmin(admin.ModelAdmin): # The list display lets us control what is shown in the default persons table at Home > Website > Videos # info on displaying multiple entries comes from http://stackoverflow.com/questions/9164610/custom-columns-using-django-admin list_display = ('title', 'date', 'caption', 'project') # search_fields are used for auto-complete, see: # https://docs.djangoproject.com/en/3.0/ref/contrib/admin/#django.contrib.admin.ModelAdmin.autocomplete_fields search_fields = ['title', 'get_video_host_str', 'date'] # default the sort order in table to descending order by date ordering = ('-date',) class TalkAdmin(admin.ModelAdmin): # The list display lets us control what is shown in the default talk table at Home > Website > Talk # See: https://docs.djangoproject.com/en/dev/ref/contrib/admin/#django.contrib.admin.ModelAdmin.list_display list_display = ('title', 'date', 'get_speakers_as_csv', 'forum_name', 'location', 'talk_type') # search_fields are used for auto-complete, see: # https://docs.djangoproject.com/en/3.0/ref/contrib/admin/#django.contrib.admin.ModelAdmin.autocomplete_fields # for example, the PublicationAdmin uses auto-complete select2 for talks search_fields = ['title', 'forum_name'] # Filters speakers only to current members and collaborators and sorts by first name # Based on: https://stackoverflow.com/a/17457828 # Update: we no longer do this because sometimes we want to add a talk by a former member or collaborator def formfield_for_manytomany(self, db_field, request, **kwargs): print("TalkAdmin.formfield_for_manytomany: db_field: {} db_field.name {} request: {}".format(db_field, db_field.name, request)) if db_field.name == "projects": kwargs["widget"] = widgets.FilteredSelectMultiple("projects", is_stacked=False) if db_field.name == "project_umbrellas": kwargs["widget"] = widgets.FilteredSelectMultiple("project umbrellas", is_stacked=False, ) if db_field.name == "speakers": # Uncomment the following block of code to limit the speakers field in the admin UI only to current lab members # Note: we don't actually want to do this (see https://github.com/jonfroehlich/makeabilitylabwebsite/issues/534) # but keeping it here because code may be useful in the future for other areas of admin interface # current_member_and_collab_ids = [person.id for person in Person.objects.all() if person.is_current_member()] # filtered_speakers = Person.objects.filter(id__in=current_member_and_collab_ids).order_by('first_name') # kwargs["queryset"] = filtered_speakers kwargs["widget"] = widgets.FilteredSelectMultiple("speakers", is_stacked=False) if db_field.name == "keywords": kwargs["widget"] = widgets.FilteredSelectMultiple("keywords", is_stacked=False) return super(TalkAdmin, self).formfield_for_manytomany(db_field, request, **kwargs) class PosterAdmin(admin.ModelAdmin): # search_fields are used for auto-complete, see: # https://docs.djangoproject.com/en/3.0/ref/contrib/admin/#django.contrib.admin.ModelAdmin.autocomplete_fields search_fields = ['title', 'date'] def formfield_for_manytomany(self, db_field, request, **kwargs): print("PosterAdmin.formfield_for_manytomany: db_field: {} db_field.name {} request: {}".format(db_field, db_field.name, request)) if db_field.name == "projects": kwargs["widget"] = widgets.FilteredSelectMultiple("projects", is_stacked=False) if db_field.name == "authors": kwargs["widget"] = widgets.FilteredSelectMultiple("authors", is_stacked=False) if db_field.name == "keywords": kwargs["widget"] = widgets.FilteredSelectMultiple("keywords", is_stacked=False) return super(PosterAdmin, self).formfield_for_manytomany(db_field, request, **kwargs) class ProjectUmbrellaAdmin(admin.ModelAdmin): def formfield_for_manytomany(self, db_field, request=None, **kwargs): if db_field.name == "keywords": kwargs["widget"] = widgets.FilteredSelectMultiple("keywords", is_stacked=False) return super(ProjectUmbrellaAdmin, self).formfield_for_manytomany(db_field, request, **kwargs) #from https://stackoverflow.com/questions/9602217/define-an-order-for-manytomanyfield-with-django #display items inline class PublicationAuthorInline(admin.TabularInline): model = Publication.authors.through verbose_name = "Author" verbose_name_plural = "Author Order" class PublicationAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title', 'authors', 'date']}), ('Files', {'fields': ['pdf_file']}), ('Pub Venue information', {'fields': ['pub_venue_url','pub_venue_type', 'book_title', 'book_title_short', 'geo_location', 'total_papers_submitted', 'total_papers_accepted']}), ('Archival Info', {'fields': ['official_url', 'extended_abstract', 'peer_reviewed', 'award' ]}), ('Page Info', {'fields': ['num_pages', 'page_num_start', 'page_num_end']}), ('Supplementary Artifacts', {'fields': ['poster', 'video', 'talk', 'code_repo_url']}), ('Project Info', {'fields': ['projects', 'project_umbrellas']}), ('Keyword Info', {'fields': ['keywords']}), ] list_display = ('title', 'book_title_short', 'date') # default the sort order in table to descending order by date ordering = ('-date',) list_filter = (PubVenueTypeListFilter, PubVenueListFilter) # add in auto-complete fields for talks, see: # https://docs.djangoproject.com/en/3.0/ref/contrib/admin/#django.contrib.admin.ModelAdmin.autocomplete_fields # this addresses: https://github.com/jonfroehlich/makeabilitylabwebsite/issues/553 # You must also update the search_fields in the respective admins like PosterAdmin, VideoAdmin, and TalkAdmin # these search fields become what the auto-complete function searches for filtering autocomplete_fields = ['poster', 'video', 'talk'] def get_form(self, request, obj=None, **kwargs): """We custom style some of the admin UI, including expanding the width of the talk select interface""" form = super(PublicationAdmin, self).get_form(request, obj, **kwargs) # we style the talks select2 widget so that it's wider, see: # https://docs.djangoproject.com/en/2.2/ref/forms/widgets/#customizing-widget-instances # see also: # https://stackoverflow.com/questions/10588275/django-change-field-size-of-modelmultiplechoicefield # https://stackoverflow.com/questions/110378/change-the-width-of-form-elements-created-with-modelform-in-django # and finally, this is what worked for me: # https://stackoverflow.com/q/35211809 # to address: https://github.com/jonfroehlich/makeabilitylabwebsite/issues/851 text_min_width = 750 form.base_fields['title'].widget.attrs['style'] = 'min-width: {}px;'.format(text_min_width) form.base_fields['book_title'].widget.attrs['style'] = 'min-width: {}px;'.format(text_min_width) form.base_fields['book_title_short'].widget.attrs['style'] = 'min-width: {}px;'.format(500) select_min_width = 600 form.base_fields['poster'].widget.attrs['style'] = 'min-width: {}px;'.format(select_min_width) form.base_fields['video'].widget.attrs['style'] = 'min-width: {}px;'.format(select_min_width) form.base_fields['talk'].widget.attrs['style'] = 'min-width: {}px;'.format(select_min_width) return form def formfield_for_manytomany(self, db_field, request=None, **kwargs): if db_field.name == "authors": kwargs['widget'] = SortedFilteredSelectMultiple() elif db_field.name == "projects": kwargs["widget"] = widgets.FilteredSelectMultiple("projects", is_stacked=False) elif db_field.name == "project_umbrellas": kwargs["widget"] = widgets.FilteredSelectMultiple("project umbrellas", is_stacked=False) elif db_field.name == "keywords": kwargs["widget"] = widgets.FilteredSelectMultiple("keywords", is_stacked=False) return super(PublicationAdmin, self).formfield_for_manytomany(db_field, request, **kwargs) admin.site.register(Person, PersonAdmin) admin.site.register(Publication, PublicationAdmin) admin.site.register(Talk, TalkAdmin) admin.site.register(Project, ProjectAdmin) admin.site.register(Poster, PosterAdmin) admin.site.register(Keyword) admin.site.register(News, NewsAdmin) admin.site.register(Banner, BannerAdmin) admin.site.register(Video, VideoAdmin) admin.site.register(Photo, PhotoAdmin) admin.site.register(Project_umbrella, ProjectUmbrellaAdmin) admin.site.register(Sponsor) # For modifying more on the front admin landing page, see https://medium.com/django-musings/customizing-the-django-admin-site-b82c7d325510 admin.site.index_title = "Makeability Lab Admin. Django version: " + django.get_version() + " ML Version: 0.5.7a"
15,988
4,873
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jun 26 21:39:04 2019 @author: fmk """ import argparse,subprocess,string,random import pandas as pd ''' positional and optional argument parser''' parser = argparse.ArgumentParser(formatter_class=argparse.RawDescriptionHelpFormatter, description='''\ Infer orthologs across two or more prokka-based annotations, and returns overview table for all genes. Homology is inferred using CD-HIT and annotations need to be in fasta format (nucleotide (*.ffn) or amino acid (*.faa)) CD-HIT: %identity optional. Fixed: -s 0.9, ie. shorter sequences need to be at least 90% length of the representative of the cluster. ''', epilog="Questions or comments? --> fkey@mit.edu") parser.add_argument("-f", dest="file_sample_annotation", help="2-col TSV file with subject-identifier and annotation file path.",type=argparse.FileType('r'),required=True) parser.add_argument('-p', dest="percentIdentity", action="store", default='0.98', help="Percent identity cd-hit. Default: 0.98") parser.add_argument('-o', dest="outpath", action="store", help="Output path.",required=True) parser.add_argument("-c", dest="cdhit", help="Path to CD-HIT executable", action="store",default="cd-hit") args = parser.parse_args() ''' FUNCTIONS''' def fix_path(path): # make sure path has trailing "/" if path[-1] != "/": path = path + "/" return path else: return path def read_merge_sample_annotation_file(file_sample_annotation): # get list of annotation file paths, tuple/dict of subjectID and prokka-assigned gene-tag (use first line) subj_tag_dict = {} subprocess.run(['mkdir','-p',outpath],check=True) with open(outpath+"merged_annotation.fa", 'w') as outfile: # with open(file_sample_annotation,'r') as infile: for line in file_sample_annotation: line = line.strip().split('\t') print('1') print(line) # read annotation file: extract prokka-gene-tag (from 1st line) and merge all annotation files into one with open(line[1]) as annofile: # link subjectID with prokka gene tag first_line = annofile.readline() if first_line.startswith('>'): prokka_gene_tag = first_line.strip().split(' ')[0].split('>')[1].split('_')[0] subj_tag_dict[ line[0] ] = prokka_gene_tag else: raise ValueError('Annotation file does not start with ">": '+first_line+' in '+line[1]) # write annotation files into merged outfile outfile.write(first_line) # necessary bcs already read outfile.write(annofile.read()) file_sample_annotation.close() return subj_tag_dict def read_cdhit_cluster(cdhit_clstr,percentIdentity,prokka_tag_list): # read cdhit results and build for each cluster entry in dict ctr = 1 ctr_cluster_within_subject = 0 rdm_tag = ''.join(random.choice(string.ascii_lowercase) for x in range(4)) # added to cluster-tags to avoid confusion w/ different runs saab_cluster_genes = {} # value_default = ['NA' for i in prokka_tag_list] with open(cdhit_clstr) as infile: for line in infile: if line.startswith('>'): cluster_tag = "SAAB_" + "%05d" % ctr + "_pid" + percentIdentity + "_" + rdm_tag saab_cluster_genes[cluster_tag] = ['NA' for i in prokka_tag_list] ctr += 1 else: line = line.strip().split('\t')[1].split(" ")[1].split('.')[0].split('>')[1] # remodel string so it matches prokka-gene_id eg. "0 10155aa, >JKPBNODD_00001... *" subject_identifier_prokka = line.split('_')[0] subject_index = prokka_tag_list.index( subject_identifier_prokka ) if saab_cluster_genes[cluster_tag][subject_index] == 'NA': saab_cluster_genes[cluster_tag][subject_index] = line else: saab_cluster_genes[cluster_tag][subject_index] = saab_cluster_genes[cluster_tag][subject_index] + "," + line ctr_cluster_within_subject += 1 if ctr_cluster_within_subject > 0: print('NOTE: ' + str(ctr_cluster_within_subject) + " occasions where a gene cluster had >1 gene from the same individual assigned.") return saab_cluster_genes ''' MAIN ''' # TEST Vars #file_sample_annotation = "/Users/fmk/Documents/mit/stapAD/tmp/pycode/prokka_ffn/subject_4_9_16.list" ##annopath = "/Users/fmk/Documents/mit/stapAD/mlst" ##filetype = "txt" #outpath = "/Users/fmk/Documents/mit/stapAD/tmp/pycode" #percentIdentity=0.95 #cdhit_executable = '/usr/local/bin/cd-hit' if __name__ == "__main__": # assign argparse arguments file_sample_annotation = args.file_sample_annotation # annopath = fix_path(args.annopath) # fix path to annotation has trailing "/" outpath = fix_path(args.outpath) # filetype = args.filetype cdhit_executable = args.cdhit percentIdentity = args.percentIdentity # get concatenated annotation file (output: merged_annotation.fa) and dict[subject]=prokka-tag subj_tag_dict = read_merge_sample_annotation_file(file_sample_annotation) subject_list_ord = list(subj_tag_dict.keys()) prokkaTag_list_ord = [ subj_tag_dict[k] for k in subject_list_ord ] # cd-hit command_cdhit = cdhit_executable + " -s 0.9 -c " + percentIdentity + " -i " + outpath + "merged_annotation.fa" + " -o " + outpath+"cdhit_results" subprocess.run(command_cdhit,shell=True) # read-in cdhit results: dict[SAAB_XXXXX_pidZZZ_YYY]=[geneX,geneY,geneZ] cdhit_res_dict = read_cdhit_cluster(outpath+"cdhit_results.clstr",percentIdentity,prokkaTag_list_ord) # build table of gene annotation cdhit_res_df = pd.DataFrame.from_dict(cdhit_res_dict,orient='index',columns=subject_list_ord) # write cdhit res cdhit_res_df.to_csv(outpath+'annotation_orthologs.tsv',sep="\t")
6,191
2,008
import os import angr test_location = os.path.join(os.path.dirname(os.path.realpath(__file__)), '..', '..', 'binaries', 'tests') def test_vtable_extraction_x86_64(): p = angr.Project(os.path.join(test_location, "x86_64", "cpp_classes"), auto_load_libs=False) vtables_sizes = {0x403cb0: 24, 0x403cd8: 16, 0x403cf8: 16, 0x403d18: 16} vtable_analysis = p.analyses.VtableFinder() vtables = vtable_analysis.vtables_list assert len(vtables) == 4 for vtable in vtables: assert vtable.vaddr in [0x403cb0, 0x403cd8, 0x403cf8, 0x403d18] assert vtables_sizes[vtable.vaddr] == vtable.size if __name__ == "__main__": test_vtable_extraction_x86_64()
708
336
from zeep import Client def test(): client = Client('http://www.dneonline.com/calculator.asmx?wsdl') result = client.service.Add(100,220) print(result) def test2(): client = Client('http://172.16.237.11:8080/SendNotify_Interface_Gmail/services/Notify?wsdl') result = client.service.sendEmail("loitd@vnptepay.com.vn", "deglkxtfyjpnjqtq", "Ahihi", "Xin chao", "loitd@vnptepay.com.vn", 1) print(result) if __name__ == "__main__": test2()
465
198
'''The HTTP agent - provides observation using http push to remote agent and expects action in the reply''' import json import time import os import threading import requests from . import BaseAgent from .. import utility from .. import characters class HttpAgent(BaseAgent): """The HTTP Agent that connects to a port with a remote agent where the character runs. It uses the same interface as the docker agent and is useful for debugging.""" def __init__(self, port=8080, host='localhost', timeout=120, character=characters.Bomber): self._port = port self._host = host self._timeout = timeout super(HttpAgent, self).__init__(character) self._wait_for_remote() def _wait_for_remote(self): """Wait for network service to appear. A timeout of 0 waits forever.""" timeout = self._timeout backoff = .25 max_backoff = min(timeout, 16) if timeout: # time module is needed to calc timeout shared between two exceptions end = time.time() + timeout while True: try: now = time.time() if timeout and end < now: print("Timed out - %s:%s" % (self._host, self._port)) raise request_url = 'http://%s:%s/ping' % (self._host, self._port) req = requests.get(request_url) self._acknowledged = True return True except requests.exceptions.ConnectionError as e: print("ConnectionError: ", e) backoff = min(max_backoff, backoff * 2) time.sleep(backoff) except requests.exceptions.HTTPError as e: print("HTTPError: ", e) backoff = min(max_backoff, backoff * 2) time.sleep(backoff) def act(self, obs, action_space): obs_serialized = json.dumps(obs, cls=utility.PommermanJSONEncoder) request_url = "http://{}:{}/action".format(self._host, self._port) try: req = requests.post( request_url, timeout=0.15, json={ "obs": obs_serialized, "action_space": json.dumps(action_space, cls=utility.PommermanJSONEncoder) }) action = req.json()['action'] except requests.exceptions.Timeout as e: print('Timeout!') # TODO: Fix this. It's ugly. action = [0] * len(action_space.shape) if len(action) == 1: action = action[0] return action
2,760
726
# -*- coding: utf-8 -*- from rest_framework import serializers from django_countries.serializer_fields import CountryField from .models import Event, CountryEvent class CountryEventSerializer(serializers.ModelSerializer): code = serializers.ReadOnlyField(source='country.code') name = serializers.SerializerMethodField() class Meta: model = CountryEvent fields = ('code', 'name') def get_name(self, obj): return obj.country.name class EventsSerializer(serializers.ModelSerializer): events_country = CountryEventSerializer(many=True, read_only=True) class Meta: model = Event fields = ( 'name', 'description', 'start', 'end', 'events_country' )
737
202
#!/usr/bin/env python """ Prints a map of the entire world. """ import os, sys import math from struct import pack # local module try: import nbt except ImportError: # nbt not in search path. Let's see if it can be found in the parent folder extrasearchpath = os.path.realpath(os.path.join(__file__,os.pardir,os.pardir)) if not os.path.exists(os.path.join(extrasearchpath,'nbt')): raise sys.path.append(extrasearchpath) from nbt.region import RegionFile from nbt.chunk import Chunk from nbt.world import WorldFolder,McRegionWorldFolder # PIL module (not build-in) try: from PIL import Image except ImportError: # PIL not in search path. Let's see if it can be found in the parent folder sys.stderr.write("Module PIL/Image not found. Pillow (a PIL fork) can be found at http://python-imaging.github.io/\n") # Note: it may also be possible that PIL is installed, but JPEG support is disabled or broken sys.exit(70) # EX_SOFTWARE def get_heightmap_image(chunk, buffer=False, gmin=False, gmax=False): points = chunk.blocks.generate_heightmap(buffer, True) # Normalize the points hmin = min(points) if (gmin == False) else gmin # Allow setting the min/max explicitly, in case this is part of a bigger map hmax = max(points) if (gmax == False) else gmax hdelta = hmax-hmin+0.0 pixels = "" for y in range(16): for x in range(16): # pix X => mc -Z # pix Y => mc X offset = (15-x)*16+y height = int((points[offset]-hmin)/hdelta*255) if (height < 0): height = 0 if (height > 255): height = 255 pixels += pack(">B", height) im = Image.fromstring('L', (16,16), pixels) return im # List of blocks to ignore # Uncomment all the lines to show underground structures # TODO: move this list into a separate config file block_ignore = [ 'air', # At least this one # 'cave_air', 'water', 'lava', 'snow', 'ice', # 'grass', 'tall_grass', 'dead_bush', # 'seagrass', 'tall_seagrass', 'kelp', 'kelp_plant', # 'dandelion', 'poppy', 'oxeye_daisy', 'white_tulip', # 'azure_bluet', 'lilac', 'rose_bush', 'peony', 'blue_orchid', # 'lily_pad', 'sugar_cane', 'vine', 'pumpkin', 'cactus', # 'wheat', 'potatoes', 'beetroots', 'carrots', # 'oak_leaves', 'dark_oak_leaves', 'birch_leaves', # 'acacia_leaves', 'spruce_leaves', # 'oak_log', 'dark_oak_log', 'birch_log', # 'acacia_log', 'spruce_log', # 'brown_mushroom', 'red_mushroom', # 'brown_mushroom_block', 'red_mushroom_block', 'mushroom_stem', # 'grass_block', 'grass_path', 'farmland', 'dirt', # 'stone', 'sand', 'gravel', 'clay', # 'sandstone', 'diorite', 'andesite', 'granite', 'obsidian', # 'coal_ore', 'iron_ore', 'gold_ore', 'diamond_ore', # 'redstone_ore', 'lapis_ore', 'emerald_ore', # 'cobweb', ] # Map of block colors from names # Legacy block numeric identifiers are now hidden by Block class # and mapped to alpha identifiers in best effort # TODO: move this map into a separate config file block_colors = { 'acacia_leaves': {'h':114, 's':64, 'l':22 }, 'acacia_log': {'h':35, 's':93, 'l':30 }, 'air': {'h':0, 's':0, 'l':0 }, 'andesite': {'h':0, 's':0, 'l':32 }, 'azure_bluet': {'h':0, 's':0, 'l':100}, 'bedrock': {'h':0, 's':0, 'l':10 }, 'birch_leaves': {'h':114, 's':64, 'l':22 }, 'birch_log': {'h':35, 's':93, 'l':30 }, 'blue_orchid': {'h':0, 's':0, 'l':100}, 'bookshelf': {'h':0, 's':0, 'l':100}, 'brown_mushroom': {'h':0, 's':0, 'l':100}, 'brown_mushroom_block': {'h':0, 's':0, 'l':100}, 'cactus': {'h':126, 's':61, 'l':20 }, 'cave_air': {'h':0, 's':0, 'l':0 }, 'chest': {'h':0, 's':100, 'l':50 }, 'clay': {'h':7, 's':62, 'l':23 }, 'coal_ore': {'h':0, 's':0, 'l':10 }, 'cobblestone': {'h':0, 's':0, 'l':25 }, 'cobblestone_stairs': {'h':0, 's':0, 'l':25 }, 'crafting_table': {'h':0, 's':0, 'l':100}, 'dandelion': {'h':60, 's':100, 'l':60 }, 'dark_oak_leaves': {'h':114, 's':64, 'l':22 }, 'dark_oak_log': {'h':35, 's':93, 'l':30 }, 'dark_oak_planks': {'h':35, 's':93, 'l':30 }, 'dead_bush': {'h':0, 's':0, 'l':100}, 'diorite': {'h':0, 's':0, 'l':32 }, 'dirt': {'h':27, 's':51, 'l':15 }, 'end_portal_frame': {'h':0, 's':100, 'l':50 }, 'farmland': {'h':35, 's':93, 'l':15 }, 'fire': {'h':55, 's':100, 'l':50 }, 'flowing_lava': {'h':16, 's':100, 'l':48 }, 'flowing_water': {'h':228, 's':50, 'l':23 }, 'glass_pane': {'h':0, 's':0, 'l':100}, 'granite': {'h':0, 's':0, 'l':32 }, 'grass': {'h':94, 's':42, 'l':25 }, 'grass_block': {'h':94, 's':42, 'l':32 }, 'gravel': {'h':21, 's':18, 'l':20 }, 'ice': {'h':240, 's':10, 'l':95 }, 'infested_stone': {'h':320, 's':100, 'l':50 }, 'iron_ore': {'h':22, 's':65, 'l':61 }, 'iron_bars': {'h':22, 's':65, 'l':61 }, 'ladder': {'h':35, 's':93, 'l':30 }, 'lava': {'h':16, 's':100, 'l':48 }, 'lilac': {'h':0, 's':0, 'l':100}, 'lily_pad': {'h':114, 's':64, 'l':18 }, 'lit_pumpkin': {'h':24, 's':100, 'l':45 }, 'mossy_cobblestone': {'h':115, 's':30, 'l':50 }, 'mushroom_stem': {'h':0, 's':0, 'l':100}, 'oak_door': {'h':35, 's':93, 'l':30 }, 'oak_fence': {'h':35, 's':93, 'l':30 }, 'oak_fence_gate': {'h':35, 's':93, 'l':30 }, 'oak_leaves': {'h':114, 's':64, 'l':22 }, 'oak_log': {'h':35, 's':93, 'l':30 }, 'oak_planks': {'h':35, 's':93, 'l':30 }, 'oak_pressure_plate': {'h':35, 's':93, 'l':30 }, 'oak_stairs': {'h':114, 's':64, 'l':22 }, 'peony': {'h':0, 's':0, 'l':100}, 'pink_tulip': {'h':0, 's':0, 'l':0 }, 'poppy': {'h':0, 's':100, 'l':50 }, 'pumpkin': {'h':24, 's':100, 'l':45 }, 'rail': {'h':33, 's':81, 'l':50 }, 'red_mushroom': {'h':0, 's':50, 'l':20 }, 'red_mushroom_block': {'h':0, 's':50, 'l':20 }, 'rose_bush': {'h':0, 's':0, 'l':100}, 'sugar_cane': {'h':123, 's':70, 'l':50 }, 'sand': {'h':53, 's':22, 'l':58 }, 'sandstone': {'h':48, 's':31, 'l':40 }, 'seagrass': {'h':94, 's':42, 'l':25 }, 'sign': {'h':114, 's':64, 'l':22 }, 'spruce_leaves': {'h':114, 's':64, 'l':22 }, 'spruce_log': {'h':35, 's':93, 'l':30 }, 'stone': {'h':0, 's':0, 'l':32 }, 'stone_slab': {'h':0, 's':0, 'l':32 }, 'tall_grass': {'h':94, 's':42, 'l':25 }, 'tall_seagrass': {'h':94, 's':42, 'l':25 }, 'torch': {'h':60, 's':100, 'l':50 }, 'snow': {'h':240, 's':10, 'l':85 }, 'spawner': {'h':180, 's':100, 'l':50 }, 'vine': {'h':114, 's':64, 'l':18 }, 'wall_torch': {'h':60, 's':100, 'l':50 }, 'water': {'h':228, 's':50, 'l':23 }, 'wheat': {'h':123, 's':60, 'l':50 }, 'white_wool': {'h':0, 's':0, 'l':100}, } def get_map(chunk): # Show an image of the chunk from above pixels = b"" for z in range(16): for x in range(16): # Find the highest block in this column max_height = chunk.get_max_height() ground_height = max_height tints = [] for y in range(max_height,-1,-1): block_id = chunk.get_block(x, y, z) if block_id != None: #block_data = 0 # TODO: use block properties #if (block_id == 'water' or block_id == 'water'): #tints.append({'h':228, 's':50, 'l':23}) # Water #elif (block_id == 'leaves'): # TODO: old id - update #if (block_data == 1): #tints.append({'h':114, 's':64, 'l':22}) # Redwood Leaves #elif (block_data == 2): #tints.append({'h':93, 's':39, 'l':10}) # Birch Leaves #else: #tints.append({'h':114, 's':64, 'l':22}) # Normal Leaves #elif (block_id == 'ice'): #tints.append({'h':240, 's':5, 'l':95}) # Ice #elif (block_id == 'fire'): #tints.append({'h':55, 's':100, 'l':50}) # Fire #elif (block_id != 'air' or block_id != 'cave_air' or y == 0): if (block_id not in block_ignore or y == 0): # Here is ground level ground_height = y break if block_id != None: if block_id in block_colors: color = block_colors[block_id] else: color = {'h':0, 's':0, 'l':100} print("warning: unknown color for block id: %s" % block_id) print("hint: add that block to the 'block_colors' map") else: color = {'h':0, 's':0, 'l':0} height_shift = 0 #(ground_height-64)*0.25 final_color = {'h':color['h'], 's':color['s'], 'l':color['l'] + height_shift} if final_color['l'] > 100: final_color['l'] = 100 if final_color['l'] < 0: final_color['l'] = 0 # Apply tints from translucent blocks for tint in reversed(tints): final_color = hsl_slide(final_color, tint, 0.4) rgb = hsl2rgb(final_color['h'], final_color['s'], final_color['l']) pixels += pack("BBB", rgb[0], rgb[1], rgb[2]) im = Image.frombytes('RGB', (16,16), pixels) return im ## Color functions for map generation ## # Hue given in degrees, # saturation and lightness given either in range 0-1 or 0-100 and returned in kind def hsl_slide(hsl1, hsl2, ratio): if (abs(hsl2['h'] - hsl1['h']) > 180): if (hsl1['h'] > hsl2['h']): hsl1['h'] -= 360 else: hsl1['h'] += 360 # Find location of two colors on the H/S color circle p1x = math.cos(math.radians(hsl1['h']))*hsl1['s'] p1y = math.sin(math.radians(hsl1['h']))*hsl1['s'] p2x = math.cos(math.radians(hsl2['h']))*hsl2['s'] p2y = math.sin(math.radians(hsl2['h']))*hsl2['s'] # Slide part of the way from tint to base color avg_x = p1x + ratio*(p2x-p1x) avg_y = p1y + ratio*(p2y-p1y) avg_h = math.atan(avg_y/avg_x) avg_s = avg_y/math.sin(avg_h) avg_l = hsl1['l'] + ratio*(hsl2['l']-hsl1['l']) avg_h = math.degrees(avg_h) #print('tint: %s base: %s avg: %s %s %s' % (tint,final_color,avg_h,avg_s,avg_l)) return {'h':avg_h, 's':avg_s, 'l':avg_l} # From http://www.easyrgb.com/index.php?X=MATH&H=19#text19 def hsl2rgb(H,S,L): H = H/360.0 S = S/100.0 # Turn into a percentage L = L/100.0 if (S == 0): return (int(L*255), int(L*255), int(L*255)) var_2 = L * (1+S) if (L < 0.5) else (L+S) - (S*L) var_1 = 2*L - var_2 def hue2rgb(v1, v2, vH): if (vH < 0): vH += 1 if (vH > 1): vH -= 1 if ((6*vH)<1): return v1 + (v2-v1)*6*vH if ((2*vH)<1): return v2 if ((3*vH)<2): return v1 + (v2-v1)*(2/3.0-vH)*6 return v1 R = int(255*hue2rgb(var_1, var_2, H + (1.0/3))) G = int(255*hue2rgb(var_1, var_2, H)) B = int(255*hue2rgb(var_1, var_2, H - (1.0/3))) return (R,G,B) def main(world_folder, show=True): world = WorldFolder(world_folder) bb = world.get_boundingbox() world_map = Image.new('RGB', (16*bb.lenx(),16*bb.lenz())) t = world.chunk_count() try: i =0.0 for chunk in world.iter_chunks(): if i % 50 ==0: sys.stdout.write("Rendering image") elif i % 2 == 0: sys.stdout.write(".") sys.stdout.flush() elif i % 50 == 49: sys.stdout.write("%5.1f%%\n" % (100*i/t)) i +=1 chunkmap = get_map(chunk) x,z = chunk.get_coords() world_map.paste(chunkmap, (16*(x-bb.minx),16*(z-bb.minz))) print(" done\n") filename = os.path.basename(world_folder)+".png" world_map.save(filename,"PNG") print("Saved map as %s" % filename) except KeyboardInterrupt: print(" aborted\n") filename = os.path.basename(world_folder)+".partial.png" world_map.save(filename,"PNG") print("Saved map as %s" % filename) return 75 # EX_TEMPFAIL if show: world_map.show() return 0 # NOERR if __name__ == '__main__': if (len(sys.argv) == 1): print("No world folder specified!") sys.exit(64) # EX_USAGE if sys.argv[1] == '--noshow' and len(sys.argv) > 2: show = False world_folder = sys.argv[2] else: show = True world_folder = sys.argv[1] # clean path name, eliminate trailing slashes. required for os.path.basename() world_folder = os.path.normpath(world_folder) if (not os.path.exists(world_folder)): print("No such folder as "+world_folder) sys.exit(72) # EX_IOERR sys.exit(main(world_folder, show))
13,998
5,844
from flaskcbv.url import Url, make_urls from .views import mainView namespases = make_urls( Url('', mainView(), name="main"), )
136
50
# Rest framework from rest_framework import permissions class IsAuthorOrReadOnly(permissions.BasePermission): """ Object level permission. Check if the requesting user is the author or not. If he/she the author then we will give him/her a read and write permission otherwise ready only """ def has_object_permission(self, request, view, obj): # Check if he requesting for only a get, etc if request.method in permissions.SAFE_METHODS: return True return obj.author == request.user
536
137
""" femagtools.vtu ~~~~~~~~~~~~~~ Read FEMAG vtu files """ import vtk import pathlib import numpy as np class Reader(object): '''Class to read vtu-files''' def __init__(self, pathname): '''Read the vtu-files Parameters ---------- pathname : str Directory of result files (vtu-files) or a single vtu file ''' self.data = {} self.reader = vtk.vtkXMLUnstructuredGridReader() self.output = self.reader.GetOutput() self.field_data_names = [] self.point_data_names = [] self.cell_data_names = [] if pathlib.Path(pathname).suffix == '.vtu': self.filenames = [pathlib.Path(pathname)] else: self.filenames = sorted(pathlib.Path(pathname).glob( "*.vtu")) self.reader.SetFileName(str(self.filenames[0])) self.reader.Update() self.field_data_names = [ self.output.GetFieldData().GetAbstractArray(i).GetName() for i in range(self.output.GetFieldData().GetNumberOfArrays())] self.point_data_names = [ self.output.GetPointData().GetAbstractArray(i).GetName() for i in range(self.output.GetPointData().GetNumberOfArrays())] self.cell_data_names = [ self.output.GetCellData().GetAbstractArray(i).GetName() for i in range(self.output.GetCellData().GetNumberOfArrays())] self.set_time_window(0.0, 0.0) def get_data_names(self): '''Read the list of values stored in the vtu files Parameters ---------- None Returns ------- List of values stored in the vtu files ''' return (self.field_data_names + self.point_data_names + self.cell_data_names) def read_data(self, data_list): '''Extracts data from the vtu files Parameters ---------- data_list : fist of str List of values to extract from vtu_files ''' for data_name in data_list: if data_name in self.field_data_names: self.data[data_name] = [] elif data_name in self.point_data_names: self.data[data_name] = [] elif data_name in self.cell_data_names: self.data[data_name] = [] else: raise Exception('unknown data name "' + data_name+'"') for filename in self.filenames: self.reader.SetFileName(str(filename)) self.reader.Update() for data_name in data_list: if data_name in self.field_data_names: self.data[data_name].append( self.output.GetFieldData().GetAbstractArray(data_name)) if data_name in self.point_data_names: self.data[data_name].append( self.output.GetPointData().GetAbstractArray(data_name)) if data_name in self.cell_data_names: self.data[data_name].append( self.output.GetCellData().GetAbstractArray(data_name)) return "done" def set_time_window(self, start, end): '''Set time window Parameters ---------- start: float Start of the time window end: float End of the time window Only values within the time window are output by the functions get_field_vector get_point_vector get_cell_vector get_data_vector At start = 0.0 the values are read out starting from the first value At end = 0.0 the values are read out up to the last value ''' try: if "time [s]" not in self.data: self.read_data(['time [s]']) if start == 0 or start <= self.data['time [s]'][0].GetValue(0): self.istart = 0 else: self.istart = 0 for i in range(len(self.data['time [s]'])): if start >= self.data['time [s]'][i].GetValue(0): self.istart = i+1 if end == 0 or end >= self.data['time [s]'][-1].GetValue(0): self.iend = len(self.data['time [s]']) else: self.iend = 0 for i in range(len(self.data['time [s]'])): if end <= self.data['time [s]'][i].GetValue(0): self.iend = i except: self.istart = None self.iend = None def get_field_vector(self, field_data): '''Read field data Parameters ---------- field_data : str Name of field to read Returns ------- field_vec : list of float List of field values within the time window ''' if field_data not in self.data: self.read_data([field_data]) if self.istart: start = self.istart else: start = 0 if self.iend: end = self.iend else: end = len(self.data[field_data]) field_vec = [] # for i in range(self.istart,self.iend): for i in range(start, end): field_vec.append(self.data[field_data][i].GetValue(0)) return field_vec # pnt = node-key, >0 def get_point_vector(self, pnt_data, pnt): '''Read point data Parameters ---------- point_data : str Name of field to read pnt : int Key of point Returns ------- point_vec : list of float List of point values within the time window ''' if pnt_data not in self.data: self.read_data([pnt_data]) if self.istart: start = self.istart else: start = 0 if self.iend: end = self.iend else: end = len(self.data[pnt_data]) point_vec = [] for i in range(start, end): point_vec.append(self.data[pnt_data][i].GetValue(pnt-1)) return point_vec def get_cell_vector(self, cell_data, cell): '''Read cell data Parameters ---------- cell_data : str Name of field to read cell : int Key of cell Returns ------- cell_vec : list of float List of cell values within the time window ''' if cell_data not in self.data: self.read_data([cell_data]) i = self.cell_data_names.index(cell_data) noc = self.output.GetCellData().GetAbstractArray(i).GetNumberOfComponents() if noc == 1: cell_vec = [] else: cell_vec_x = [] cell_vec_y = [] cell_vec_z = [] if self.istart: start = self.istart else: start = 0 if self.iend: end = self.iend else: end = int(len(self.data[cell_data])) for i in range(start, end): if noc == 1: cell_vec.append(self.data[cell_data][i].GetValue(cell-1)) else: cell_vec_x.append( self.data[cell_data][i].GetValue(noc*(cell-1))) cell_vec_y.append( self.data[cell_data][i].GetValue(noc*(cell-1)+1)) cell_vec_z.append( self.data[cell_data][i].GetValue(noc*(cell-1)+2)) if noc == 1: return cell_vec else: return [cell_vec_x, cell_vec_y, cell_vec_z] def get_data_vector(self, data_name, key=0): '''Read data of fiels, point or cell Parameters ---------- data_name : str Name of data to read hey : int (optional) Key of point or cell Returns ------- data_vec : list of float List of values within the time window ''' if data_name in self.field_data_names: return self.get_field_vector(data_name) if data_name in self.point_data_names: return self.get_point_vector(data_name, key) if data_name in self.cell_data_names: return self.get_cell_vector(data_name, key) return [] def read(filename): """ Read vtu file and return Reader object. Arguments: filename: name of vtu file to be read """ return Reader(filename)
8,595
2,494
#!/usr/bin/python #-------------------------------IMPORT--------------------------------# from lib import * #-------------------------------EXPORT--------------------------------# __all__ = ['<#PREFIX#>_app','<#PREFIX#>_index']
232
62
from collections import Counter import numpy as np def keep_word(word): return word.is_alpha def unique_words(problems): return set([word.lemma_ for problem in problems for word in problem.tokens() if keep_word(word)]) def create_word2idx(vocab): return {word: idx for idx, word in enumerate(vocab)} class BagOfWordsFeature(): def __init__(self, corpus): self.vocab = list(unique_words(corpus)) # Mapping from words to their index in the feature vector. self.word2idx = create_word2idx(self.vocab) def process(self, problem): features = np.zeros(len(self.vocab)) words = [word.lemma_ for word in problem.tokens() if keep_word(word)] freqs = Counter(words) for word in freqs: # Skip unknown words. if word in self.word2idx: features[self.word2idx[word]] = freqs[word] return features
916
288
#!/usr/bin/python ''' Tests for MPS and MPO ''' from numpy import * import matplotlib.pyplot as plt from numpy.testing import dec, assert_, assert_raises, assert_almost_equal, assert_allclose import pdb from ..sweep import * def test_iterator(): start = (1, '->', 2) stop = (3, '<-', 1) print('Testing iterator start = %s, stop= %s' % (start, stop)) iterator = get_sweeper(start=start, stop=stop, nsite=4 - 2, iprint=2) order = [(1, '->', 2), (1, '<-', 1), (1, '<-', 0), (2, '->', 1), (2, '->', 2), (2, '<-', 1), (2, '<-', 0), (3, '->', 1), (3, '->', 2), (3, '<-', 1), ] plt.ion() visualize_sweeper(iterator, nsite=3) for od, it in zip(order, iterator): assert_(od == it) print('Testing 2-site iterator.') start = (1, '->', 0) stop = (3, '->', 0) order = [(1, '->', 0), (2, '->', 0), (3, '->', 0)] iterator = get_sweeper(start=start, stop=stop, nsite=2 - 2) for od, it in zip(order, iterator): assert_(od == it) print('Testing periodic case.') iterator = get_psweeper(start=(1, 2), stop=(3, 1), nsite=4, iprint=2) order = [(1, 2), (1, 3), (2, 0), (2, 1), (2, 2), (2, 3), (3, 0), (3, 1)] for od, it in zip(order, iterator): assert_(od == it) iterator = get_psweeper(start=(1, 0), stop=(3, 0), nsite=2, iprint=2) order = [(1, 0), (1, 1), (2, 0)] for od, it in zip(order, iterator): assert_(od == it) if __name__ == '__main__': test_iterator()
1,507
636
from threading import Lock from time import time from ui import Menu from ui.utils import clamp, check_value_lock, to_be_foreground class NumberedMenu(Menu): """ This Menu allows the user to jump to entries using the numpad. If the menu is 10 entries or less the navigation is instant. Otherwise, it lets the user type multiple digits to navigate to entries beyond 10th. The `input_delay` parameter controls how long, and if, the menu waits before considering an input as definitive. If `input_delay` is 0, then only the 10 first entries can be navigated to using the keypad. The `prepend_numbers` parameters controls whether the entries should be prefixed by their number. (default: `True`) """ def __init__(self, *args, **kwargs): self.prepend_numbers = kwargs.pop('prepend_numbers', True) self.input_delay = kwargs.pop('input_delay', 1) Menu.__init__(self, *args, **kwargs) self.__locked_name__ = None self.value_lock = Lock() self.numeric_keymap = {"KEY_{}".format(i): i for i in range(10)} self.last_input_time = 0 self.current_input = None @property def entry_count(self): return len(self.contents) def before_activate(self): Menu.before_activate(self) self.last_input_time = -self.input_delay def idle_loop(self): Menu.idle_loop(self) self.check_character_state() def set_keymap(self): Menu.set_keymap(self) self.i.set_streaming(self.on_key_pressed) def deactivate(self): Menu.deactivate(self) self.i.remove_streaming() @to_be_foreground def on_key_pressed(self, key): if key == "KEY_RIGHT" and self.is_multi_digit(): self.confirm_current_input() if key not in self.numeric_keymap: return if self.is_multi_digit(): self.process_multi_digit_input(key) else: self.process_single_digit_input(key) self.view.refresh() def process_single_digit_input(self, key): self.move_to_entry(self.numeric_keymap[key]) def process_multi_digit_input(self, key): self.last_input_time = time() if not self.current_input: self.current_input = str(self.numeric_keymap[key]) else: self.current_input += str(self.numeric_keymap[key]) def move_to_entry(self, index): if self.pointer == index: # Moving to the same item that's already selected # let's interpret this as KEY_ENTER self.current_input = None self.select_entry() return self.pointer = clamp(index, 0, len(self.contents) - 1) self.current_input = None self.view.refresh() def process_contents(self): Menu.process_contents(self) if self.prepend_numbers: self.prepend_entry_text() def prepend_entry_text(self): # prepend numbers to each entry name if self.is_multi_digit(): self.contents = [["{} {}".format(i, entry[0]), entry[1]] for i, entry in enumerate(self.contents)] else: for i, entry in enumerate(self.contents[:10]): entry[0] = "{} {}".format(i, entry[0]) @check_value_lock def check_character_state(self): if self.is_current_input_finished(): self.move_to_entry(int(self.current_input)) def is_multi_digit(self): return self.input_delay > 0 def is_current_input_finished(self): # nothing in the buffer if not self.current_input: return False # no need to let the user input '100' if we have 20 entries if len(str(self.current_input)) == len(str(self.entry_count)): return True # user typed 2 and we have 19 entries, going to the most likely option if int(self.current_input) * 10 > self.entry_count: return True # user typed 17 and we have 12 entries if int(self.current_input) >= self.entry_count: return True now = time() elapsed = now - self.last_input_time if self.is_multi_digit() and elapsed >= self.input_delay: # delay wait is over return True return False def confirm_current_input(self): if self.current_input is None: return self.move_to_entry(int(self.current_input))
4,471
1,353
# coding=utf-8 """Russian Peasant Multiplication algorithm Python implementation.""" def russ_peasant(a, b): res = 0 while b > 0: if b & 1: res += a a <<= 1 b >>= 1 return res if __name__ == '__main__': for x in range(10): for y in range(10): print(x, y, x * y, russ_peasant(x, y))
358
135
from urllib.parse import urlparse, parse_qs from io import BytesIO import json class MockRequest: def getmethod(self, path): apiend = "v2/" start = path.find(apiend) + len(apiend) return path[start:] def urlopen(self, url): output = BytesIO() parsed = urlparse(url) params = parse_qs(parsed.query) method = self.getmethod(parsed.path) assert("key" in params) s = getattr(self, method)(**params) output.write(s.encode("UTF8")) output.seek(0) return output def gettime(self, **kwargs): return json.dumps( {"bustime-response": {"tm": "20141012 10:21:04"}} ) def getdirections(self, **kwargs): assert("rt" in kwargs.keys()) return json.dumps( {"bustime-response": {"directions": [{"dir":"INBOUND"}, {"dir":"OUTBOUND"}]} } ) def getstops(self, **kwargs): assert("rt" in kwargs.keys()) assert("dir" in kwargs.keys()) return json.dumps( { "bustime-response": {"stops": [ {'stpid': '2564', 'stpnm': '5th Ave at Meyran Ave', 'lon': -79.959239533731, 'lat': 40.441172012068}]} }) def getpredictions(self, **kwargs): assert("stpid" in kwargs) assert(type(kwargs.get("rt", [])) == list) return json.dumps({ "bustime-response":{ "prd": [{'rt': '71C', 'typ': 'A', 'prdctdn': '5', 'prdtm': '20141022 12:37', 'rtdir': 'INBOUND', 'zone': '', 'des': 'Downtown', 'dly': False, 'dstp': 4198, 'stpnm': '5th Ave at Chesterfield Rd', 'stpid': '38', 'tatripid': '159261', 'tmstmp': '20141022 12:31', 'tablockid': '071C-150', 'vid': '5678'}]} }) def getvehicles(self, **kwargs): assert(len(kwargs.get("vid", [])) > 0 or len(kwargs.get("rt", [])) > 0) return json.dumps({ "bustime-response":{ "vehicle":[{'tablockid': '071C-148', 'pid': 2363, 'tatripid': '159264', 'zone': '', 'des': 'Downtown', 'dly': False, 'spd': 0, 'pdist': 17607, 'hdg': '299', 'tmstmp': '20141022 12:52', 'lat': '40.46042251586914', 'rt': '71C', 'vid': '5669', 'lon': '-79.92157814719461'}] } }) def getroutes(self, **kwargs): return json.dumps({ "bustime-response": { "routes": [{'rt': '12', 'rtnm': 'MCKNIGHT', 'rtclr': '#cc00cc'}] } }) def getpatterns(self, **kwargs): assert(len(kwargs.get("pid", [])) > 0 or len(kwargs.get("rt", [])) > 0) return json.dumps({ "bustime-response": { #yes, it's a fucking long line. I wanted to have a solid pattern for testing. "ptr": [{'ln': 48062.0, 'rtdir': 'INBOUND', 'pid': 2363, 'pt': [{'typ': 'S', 'stpid': '19988', 'seq': 1, 'lon': -79.89193733333298, 'stpnm': 'LAYOVER BRUSHTON TERMINAL', 'pdist': 0.0, 'lat': 40.449901185011}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.449815459172, 'lon': -79.89199634193102, 'seq': 2}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.449160346541, 'lon': -79.89106988890501, 'seq': 3}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.44854800997, 'lon': -79.89031887038101, 'seq': 4}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.447739717154, 'lon': -79.88950347883997, 'seq': 5}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.447155944074, 'lon': -79.88896703703699, 'seq': 6}, {'typ': 'S', 'stpid': '16111', 'seq': 7, 'lon': -79.88872563822599, 'stpnm': 'Wilkinsburg Station stop B', 'pdist': 1664.0, 'lat': 40.446845684989}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.445680436902, 'lon': -79.88800934987802, 'seq': 8}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.4445128499, 'lon': -79.88743535714798, 'seq': 9}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.444227073768, 'lon': -79.88744608598398, 'seq': 10}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.44292065312, 'lon': -79.88707057672201, 'seq': 11}, {'typ': 'S', 'stpid': '8153', 'seq': 12, 'lon': -79.88699011045202, 'stpnm': 'Hay St ramp outbound shelter', 'pdist': 3184.0, 'lat': 40.442822670548}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442713549274, 'lon': -79.886595606477, 'seq': 13}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442472674453, 'lon': -79.886718988092, 'seq': 14}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442713549274, 'lon': -79.887507557543, 'seq': 15}, {'typ': 'S', 'stpid': '19871', 'seq': 16, 'lon': -79.887233972223, 'stpnm': 'Pennwood Ave at Franklin', 'pdist': 3809.0, 'lat': 40.441803119568}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.44066404441, 'lon': -79.886885285051, 'seq': 17}, {'typ': 'S', 'stpid': '7084', 'seq': 18, 'lon': -79.887757249342, 'stpnm': 'Rowland Connector opp Highrise', 'pdist': 4362.0, 'lat': 40.440518615096}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.4404410428, 'lon': -79.887928910719, 'seq': 19}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.440228739215, 'lon': -79.888191767202, 'seq': 20}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.440196077066, 'lon': -79.888325877653, 'seq': 21}, {'typ': 'S', 'stpid': '7085', 'seq': 22, 'lon': -79.889442955026, 'stpnm': 'Kelly Ave at Pitt St', 'pdist': 4987.0, 'lat': 40.440550595192}, {'typ': 'S', 'stpid': '7086', 'seq': 23, 'lon': -79.891097240081, 'stpnm': 'Kelly Ave at West St', 'pdist': 5545.0, 'lat': 40.441113108446}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.441213613573, 'lon': -79.891389862428, 'seq': 24}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.44125444064, 'lon': -79.891786829362, 'seq': 25}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.441266688755, 'lon': -79.892339364419, 'seq': 26}, {'typ': 'S', 'stpid': '8224', 'seq': 27, 'lon': -79.892355457673, 'stpnm': 'Trenton Ave opp Rebecca (nearside)', 'pdist': 6113.0, 'lat': 40.441887256997}, {'typ': 'S', 'stpid': '8225', 'seq': 28, 'lon': -79.892379833333, 'stpnm': 'Trenton Ave at Franklin', 'pdist': 6606.0, 'lat': 40.443183767554}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.443363401196, 'lon': -79.892342282407, 'seq': 29}, {'typ': 'S', 'stpid': '8226', 'seq': 30, 'lon': -79.892166166667, 'stpnm': 'Trenton Ave at South', 'pdist': 6878.0, 'lat': 40.443920100848}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.445097455012, 'lon': -79.891708178573, 'seq': 31}, {'typ': 'S', 'stpid': '8227', 'seq': 32, 'lon': -79.891327304892, 'stpnm': 'Trenton Ave at Penn', 'pdist': 7526.0, 'lat': 40.445815966242}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.445893532335, 'lon': -79.891284389548, 'seq': 33}, {'typ': 'S', 'stpid': '18527', 'seq': 34, 'lon': -79.891631411372, 'stpnm': 'Penn Ave past Trenton', 'pdist': 7803.0, 'lat': 40.44601741031}, {'typ': 'S', 'stpid': '7628', 'seq': 35, 'lon': -79.89356461905, 'stpnm': 'Penn Ave at Brushton', 'pdist': 8206.0, 'lat': 40.446668068321}, {'typ': 'S', 'stpid': '7629', 'seq': 36, 'lon': -79.895090308202, 'stpnm': 'Penn Ave at Braddock', 'pdist': 8685.0, 'lat': 40.44715925159}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.447673625313, 'lon': -79.896608438504, 'seq': 37}, {'typ': 'S', 'stpid': '7630', 'seq': 38, 'lon': -79.898152870373, 'stpnm': 'Penn Ave opp Carnegie', 'pdist': 9608.0, 'lat': 40.44816623597}, {'typ': 'S', 'stpid': '7631', 'seq': 39, 'lon': -79.899677632935, 'stpnm': 'Penn Ave at Lexington', 'pdist': 10093.0, 'lat': 40.448684819403}, {'typ': 'S', 'stpid': '7632', 'seq': 40, 'lon': -79.901225994708, 'stpnm': 'Penn Ave at Homewood', 'pdist': 10522.0, 'lat': 40.449202240648}, {'typ': 'S', 'stpid': '7633', 'seq': 41, 'lon': -79.903692774473, 'stpnm': 'Penn Ave at Lang', 'pdist': 11274.0, 'lat': 40.450018821594}, {'typ': 'S', 'stpid': '18528', 'seq': 42, 'lon': -79.906128449073, 'stpnm': 'Penn Ave past Murtland', 'pdist': 12021.0, 'lat': 40.450823187894}, {'typ': 'S', 'stpid': '20762', 'seq': 43, 'lon': -79.908102313491, 'stpnm': 'Penn Ave at Dallas ', 'pdist': 12609.0, 'lat': 40.451481247155}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.451775156577, 'lon': -79.908847967597, 'seq': 44}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.452232346455, 'lon': -79.909700910064, 'seq': 45}, {'typ': 'S', 'stpid': '8229', 'seq': 46, 'lon': -79.910224384258, 'stpnm': 'Penn Ave at Linden', 'pdist': 13331.0, 'lat': 40.452589523632}, {'typ': 'S', 'stpid': '8230', 'seq': 47, 'lon': -79.912039435186, 'stpnm': 'Penn Ave at 5th Ave', 'pdist': 14004.0, 'lat': 40.453822374}, {'typ': 'S', 'stpid': '18942', 'seq': 48, 'lon': -79.915107988093, 'stpnm': 'Penn Ave at Bakery Square', 'pdist': 15237.0, 'lat': 40.455933324408}, {'typ': 'S', 'stpid': '20011', 'seq': 49, 'lon': -79.917314336642, 'stpnm': 'Penn Ave at East Liberty Blvd', 'pdist': 15990.0, 'lat': 40.457427481869}, {'typ': 'S', 'stpid': '8233', 'seq': 50, 'lon': -79.918370864419, 'stpnm': 'Penn Ave at Village of Eastside Shpg Ctr', 'pdist': 16430.0, 'lat': 40.458190021715}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.459063494845, 'lon': -79.919679782418, 'seq': 51}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.45921568281, 'lon': -79.91981071693, 'seq': 52}, {'typ': 'S', 'stpid': '8234', 'seq': 53, 'lon': -79.920277421298, 'stpnm': 'Penn Ave opp Shady Ave (nearside)', 'pdist': 17088.0, 'lat': 40.459550374354}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.459827922173, 'lon': -79.92062610847, 'seq': 54}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460195263114, 'lon': -79.921221558871, 'seq': 55}, {'typ': 'S', 'stpid': '19573', 'seq': 56, 'lon': -79.921578570767, 'stpnm': 'Penn Ave opp Eastside III Dr (Target)', 'pdist': 17587.0, 'lat': 40.460350125013}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460386858912, 'lon': -79.92178778307, 'seq': 57}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460676647854, 'lon': -79.922694369718, 'seq': 58}, {'typ': 'S', 'stpid': '8881', 'seq': 59, 'lon': -79.923569593255, 'stpnm': 'Penn Ave at Sheridan Ave', 'pdist': 18160.0, 'lat': 40.461008500156}, {'typ': 'S', 'stpid': '8882', 'seq': 60, 'lon': -79.92460004233, 'stpnm': 'Penn Ave at Highland Ave', 'pdist': 18488.0, 'lat': 40.46138158881}, {'typ': 'S', 'stpid': '8883', 'seq': 61, 'lon': -79.925521833333, 'stpnm': 'Penn Ave at Whitfield St', 'pdist': 18764.0, 'lat': 40.461677433225}, {'typ': 'S', 'stpid': '9124', 'seq': 62, 'lon': -79.928575802248, 'stpnm': 'Penn Ave at St Clair', 'pdist': 19679.0, 'lat': 40.462834921535}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.463556607817, 'lon': -79.930664010581, 'seq': 63}, {'typ': 'S', 'stpid': '19060', 'seq': 64, 'lon': -79.930969782409, 'stpnm': 'Negley Ave past Penn', 'pdist': 20408.0, 'lat': 40.463050518435}, {'typ': 'S', 'stpid': '8188', 'seq': 65, 'lon': -79.93160924008, 'stpnm': 'Negley Ave at Coral St', 'pdist': 20857.0, 'lat': 40.462015444544}, {'typ': 'S', 'stpid': '8189', 'seq': 66, 'lon': -79.932365500001, 'stpnm': 'Negley Ave at Friendship Ave', 'pdist': 21357.0, 'lat': 40.460807766604}, {'typ': 'S', 'stpid': '8190', 'seq': 67, 'lon': -79.933393646824, 'stpnm': 'Negley Ave at #370', 'pdist': 22054.0, 'lat': 40.459079351742}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.457797710009, 'lon': -79.93416612302, 'seq': 68}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.45757729782, 'lon': -79.9342197672, 'seq': 69}, {'typ': 'S', 'stpid': '8191', 'seq': 70, 'lon': -79.934196499999, 'stpnm': 'Negley Ave at Centre Ave', 'pdist': 22689.0, 'lat': 40.457446933451}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.457279582648, 'lon': -79.934116033729, 'seq': 71}, {'typ': 'S', 'stpid': '8192', 'seq': 72, 'lon': -79.936404395502, 'stpnm': 'Centre Ave at Graham St', 'pdist': 23393.0, 'lat': 40.456588788684}, {'typ': 'S', 'stpid': '8193', 'seq': 73, 'lon': -79.938546226853, 'stpnm': 'Centre Ave at Aiken Ave', 'pdist': 23939.0, 'lat': 40.45594147008}, {'typ': 'S', 'stpid': '8194', 'seq': 74, 'lon': -79.940120982802, 'stpnm': 'Centre Ave opp Shadyside Hospital', 'pdist': 24570.0, 'lat': 40.455471658241}, {'typ': 'S', 'stpid': '8195', 'seq': 75, 'lon': -79.941432683864, 'stpnm': 'Centre Ave at Cypress St', 'pdist': 24887.0, 'lat': 40.455069439451}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.454661251586, 'lon': -79.942848890224, 'seq': 76}, {'typ': 'S', 'stpid': '8196', 'seq': 77, 'lon': -79.94441278836, 'stpnm': 'Centre Ave at Morewood Ave', 'pdist': 25777.0, 'lat': 40.454203439285}, {'typ': 'S', 'stpid': '8197', 'seq': 78, 'lon': -79.947320525796, 'stpnm': 'Centre Ave at Millvale Ave', 'pdist': 26631.0, 'lat': 40.453325441844}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.452851703376, 'lon': -79.94905713492, 'seq': 79}, {'typ': 'S', 'stpid': '2631', 'seq': 80, 'lon': -79.949652585321, 'stpnm': 'Centre Ave opp Neville St', 'pdist': 27299.0, 'lat': 40.452651685031}, {'typ': 'S', 'stpid': '2632', 'seq': 81, 'lon': -79.950813788362, 'stpnm': 'Centre Ave at Melwood Ave', 'pdist': 27681.0, 'lat': 40.452286780054}, {'typ': 'S', 'stpid': '2633', 'seq': 82, 'lon': -79.952102351194, 'stpnm': 'Centre Ave at Craig St (nearside)', 'pdist': 28159.0, 'lat': 40.45194294909}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.451898046457, 'lon': -79.952070164685, 'seq': 83}, {'typ': 'S', 'stpid': '2634', 'seq': 84, 'lon': -79.950836166666, 'stpnm': 'Craig St at Bayard St', 'pdist': 28824.0, 'lat': 40.449890100525}, {'typ': 'S', 'stpid': '2635', 'seq': 85, 'lon': -79.949188166667, 'stpnm': 'Craig St at 5th Ave', 'pdist': 29896.0, 'lat': 40.447257934}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446951757745, 'lon': -79.949064785052, 'seq': 86}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446853781049, 'lon': -79.95005183797, 'seq': 87}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446796627911, 'lon': -79.951156908084, 'seq': 88}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446714980486, 'lon': -79.951516324092, 'seq': 89}, {'typ': 'S', 'stpid': '1171', 'seq': 90, 'lon': -79.951805621695, 'stpnm': '5th Ave at Bellefield', 'pdist': 30703.0, 'lat': 40.446598625546}, {'typ': 'S', 'stpid': '33', 'seq': 91, 'lon': -79.953128988756, 'stpnm': '5th Ave at Tennyson Ave', 'pdist': 31230.0, 'lat': 40.445674281789}, {'typ': 'S', 'stpid': '34', 'seq': 92, 'lon': -79.954654511245, 'stpnm': '5th Ave at Bigelow Blvd', 'pdist': 31839.0, 'lat': 40.444563354959}, {'typ': 'S', 'stpid': '35', 'seq': 93, 'lon': -79.956426053571, 'stpnm': '5th Ave at Thackeray Ave', 'pdist': 32525.0, 'lat': 40.443287941739}, {'typ': 'S', 'stpid': '36', 'seq': 94, 'lon': -79.958504737436, 'stpnm': '5th Ave opp Atwood St', 'pdist': 33276.0, 'lat': 40.441797374239}, {'typ': 'S', 'stpid': '38', 'seq': 95, 'lon': -79.961631833334, 'stpnm': '5th Ave at Chesterfield Rd', 'pdist': 34541.0, 'lat': 40.439498767751}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438439620008, 'lon': -79.963037673274, 'seq': 96}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438194646982, 'lon': -79.963327351848, 'seq': 97}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.43796192178, 'lon': -79.9637296832, 'seq': 98}, {'typ': 'S', 'stpid': '2639', 'seq': 99, 'lon': -79.964008632937, 'stpnm': '5th Ave opp Craft Ave', 'pdist': 35395.0, 'lat': 40.43786393198}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437761859121, 'lon': -79.964722100536, 'seq': 100}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437786356621, 'lon': -79.965043965617, 'seq': 101}, {'typ': 'S', 'stpid': '2640', 'seq': 102, 'lon': -79.965337675265, 'stpnm': '5th Ave at Robinson St', 'pdist': 35782.0, 'lat': 40.437783509968}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437501788161, 'lon': -79.965552251986, 'seq': 103}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437269060561, 'lon': -79.965616625003, 'seq': 104}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437056746959, 'lon': -79.965766828707, 'seq': 105}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436913843196, 'lon': -79.965874117068, 'seq': 106}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436762773174, 'lon': -79.966045778445, 'seq': 107}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436615785799, 'lon': -79.966260355166, 'seq': 108}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.43652187703, 'lon': -79.966560762576, 'seq': 109}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436472881098, 'lon': -79.966989916018, 'seq': 110}, {'typ': 'S', 'stpid': '18165', 'seq': 111, 'lon': -79.968430881616, 'stpnm': '5th Ave past Brenham St', 'pdist': 36887.0, 'lat': 40.436465920651}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436457754658, 'lon': -79.968661551591, 'seq': 112}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436484215649, 'lon': -79.968868819433, 'seq': 113}, {'typ': 'S', 'stpid': '2643', 'seq': 114, 'lon': -79.970167008596, 'stpnm': '5th Ave opp #2410', 'pdist': 37351.0, 'lat': 40.436912928767}, {'typ': 'S', 'stpid': '2644', 'seq': 115, 'lon': -79.970976146824, 'stpnm': '5th Ave opp #2358', 'pdist': 37622.0, 'lat': 40.437166181894}, {'typ': 'S', 'stpid': '18164', 'seq': 116, 'lon': -79.973009164022, 'stpnm': '5th Ave past Kirkpatrick St', 'pdist': 38216.0, 'lat': 40.437822924516}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437953577631, 'lon': -79.973400766539, 'seq': 117}, {'typ': 'S', 'stpid': '3147', 'seq': 118, 'lon': -79.974486406746, 'stpnm': '5th Ave past Moultrie', 'pdist': 38618.0, 'lat': 40.438030101164}, {'typ': 'S', 'stpid': '3148', 'seq': 119, 'lon': -79.975773604499, 'stpnm': '5th Ave opp Seneca St', 'pdist': 38981.0, 'lat': 40.438166929226}, {'typ': 'S', 'stpid': '3149', 'seq': 120, 'lon': -79.977394310846, 'stpnm': '5th Ave at Wyandotte St', 'pdist': 39420.0, 'lat': 40.438247594947}, {'typ': 'S', 'stpid': '3150', 'seq': 121, 'lon': -79.97944159061, 'stpnm': '5th Ave opp Gist St', 'pdist': 39979.0, 'lat': 40.438339593968}, {'typ': 'S', 'stpid': '3151', 'seq': 122, 'lon': -79.980569064815, 'stpnm': '5th Ave at Dinwiddie St', 'pdist': 40343.0, 'lat': 40.438395013737}, {'typ': 'S', 'stpid': '3152', 'seq': 123, 'lon': -79.982623861111, 'stpnm': '5th Ave opp Van Braam St', 'pdist': 40957.0, 'lat': 40.438484094732}, {'typ': 'S', 'stpid': '3154', 'seq': 124, 'lon': -79.984903019843, 'stpnm': '5th Ave at Pride St', 'pdist': 41518.0, 'lat': 40.438582096493}, {'typ': 'S', 'stpid': '3155', 'seq': 125, 'lon': -79.986305788361, 'stpnm': '5th Ave at Stevenson St', 'pdist': 41893.0, 'lat': 40.438632511731}, {'typ': 'S', 'stpid': '3156', 'seq': 126, 'lon': -79.988097240079, 'stpnm': '5th Ave at Magee St', 'pdist': 42426.0, 'lat': 40.438710016375}, {'typ': 'S', 'stpid': '3157', 'seq': 127, 'lon': -79.990663968915, 'stpnm': '5th Ave at Washington Pl', 'pdist': 43145.0, 'lat': 40.43883051923}, {'typ': 'S', 'stpid': '18161', 'seq': 128, 'lon': -79.992593541, 'stpnm': '5th Ave opp Diamond St', 'pdist': 43844.0, 'lat': 40.43889764138}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.43895888408, 'lon': -79.99469102845, 'seq': 129}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.439742785712, 'lon': -79.994202866409, 'seq': 130}, {'typ': 'S', 'stpid': '20292', 'seq': 131, 'lon': -79.994164892194, 'stpnm': '6th Ave at Center Ave', 'pdist': 44671.0, 'lat': 40.439906541484}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.44002494221, 'lon': -79.994191714284, 'seq': 132}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.440151508273, 'lon': -79.994266816137, 'seq': 133}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.440265825802, 'lon': -79.994379468915, 'seq': 134}, {'typ': 'S', 'stpid': '1509', 'seq': 135, 'lon': -79.994631135581, 'stpnm': '6th Ave at Bigelow Sq', 'pdist': 44962.0, 'lat': 40.440377273463}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.440657024606, 'lon': -79.995568611756, 'seq': 136}, {'typ': 'S', 'stpid': '3158', 'seq': 137, 'lon': -79.997612455025, 'stpnm': '6th Ave at Smithfield St', 'pdist': 45844.0, 'lat': 40.441420492964}, {'typ': 'S', 'stpid': '3159', 'seq': 138, 'lon': -79.999358695108, 'stpnm': '6th Ave at Wood St', 'pdist': 46378.0, 'lat': 40.442103148627}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442299115726, 'lon': -79.999809306222, 'seq': 139}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442462421206, 'lon': -79.999916594583, 'seq': 140}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.44171937807, 'lon': -80.002727549631, 'seq': 141}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.441507078522, 'lon': -80.00256661709, 'seq': 142}, {'typ': 'S', 'stpid': '20690', 'seq': 143, 'lon': -80.00084666204202, 'stpnm': '5th Ave at Wood St', 'pdist': 48062.0, 'lat': 40.44088273536}]}, {'ln': 43730.0, 'rtdir': 'OUTBOUND', 'pid': 2364, 'pt': [{'typ': 'S', 'stpid': '20690', 'seq': 1, 'lon': -80.00084666204202, 'stpnm': '5th Ave at Wood St', 'pdist': 0.0, 'lat': 40.44088273536}, {'typ': 'S', 'stpid': '20691', 'seq': 2, 'lon': -79.99866483333403, 'stpnm': '5th Ave at Smithfield St', 'pdist': 542.0, 'lat': 40.440055767723}, {'typ': 'S', 'stpid': '20293', 'seq': 3, 'lon': -79.995996042329, 'stpnm': '5th Ave at Ross St', 'pdist': 1454.0, 'lat': 40.439031583534}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438937678139, 'lon': -79.995684906083, 'seq': 4}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438896849666, 'lon': -79.994783683854, 'seq': 5}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438758032671, 'lon': -79.994783683854, 'seq': 6}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.43801494859, 'lon': -79.994343801576, 'seq': 7}, {'typ': 'S', 'stpid': '3241', 'seq': 8, 'lon': -79.993441129629, 'stpnm': 'Forbes Ave at Boyd St', 'pdist': 2309.0, 'lat': 40.437999108078}, {'typ': 'S', 'stpid': '3242', 'seq': 9, 'lon': -79.991346486111, 'stpnm': 'Forbes Ave at McAnulty Dr', 'pdist': 2920.0, 'lat': 40.437917437913}, {'typ': 'S', 'stpid': '3243', 'seq': 10, 'lon': -79.9887468168, 'stpnm': 'Forbes Ave at Magee St', 'pdist': 3680.0, 'lat': 40.437810426188}, {'typ': 'S', 'stpid': '3244', 'seq': 11, 'lon': -79.986794570767, 'stpnm': 'Forbes Ave at Stevenson St', 'pdist': 4191.0, 'lat': 40.437725178342}, {'typ': 'S', 'stpid': '3245', 'seq': 12, 'lon': -79.98534843783, 'stpnm': 'Forbes Ave at Pride St', 'pdist': 4573.0, 'lat': 40.437646432071}, {'typ': 'S', 'stpid': '3246', 'seq': 13, 'lon': -79.982940666667, 'stpnm': 'Forbes Ave at Van Braam St', 'pdist': 5263.0, 'lat': 40.437541934523}, {'typ': 'S', 'stpid': '3247', 'seq': 14, 'lon': -79.981758166667, 'stpnm': 'Forbes Ave at Miltenberger St', 'pdist': 5584.0, 'lat': 40.437488434527}, {'typ': 'S', 'stpid': '3248', 'seq': 15, 'lon': -79.979675333334, 'stpnm': 'Forbes Ave at Gist St', 'pdist': 6165.0, 'lat': 40.43739501512}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437313356274, 'lon': -79.977685134245, 'seq': 16}, {'typ': 'S', 'stpid': '3249', 'seq': 17, 'lon': -79.977664844577, 'stpnm': 'Jumonville St at Watson St', 'pdist': 6823.0, 'lat': 40.437740517434}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438185923097, 'lon': -79.977653302922, 'seq': 18}, {'typ': 'S', 'stpid': '3250', 'seq': 19, 'lon': -79.975984968915, 'stpnm': '5th Ave at Seneca St', 'pdist': 7437.0, 'lat': 40.438112431006}, {'typ': 'S', 'stpid': '3251', 'seq': 20, 'lon': -79.974548949073, 'stpnm': '5th Ave at Moulrie St', 'pdist': 7832.0, 'lat': 40.43797818236}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437889166532, 'lon': -79.97340458863, 'seq': 21}, {'typ': 'S', 'stpid': '3252', 'seq': 22, 'lon': -79.972862782409, 'stpnm': '5th Ave at Birmingham Bridge', 'pdist': 8327.0, 'lat': 40.437725849943}, {'typ': 'S', 'stpid': '2556', 'seq': 23, 'lon': -79.971299708996, 'stpnm': '5th Ave at #2358', 'pdist': 8811.0, 'lat': 40.437214433778}, {'typ': 'S', 'stpid': '2557', 'seq': 24, 'lon': -79.9702225, 'stpnm': '5th Ave at #2410', 'pdist': 9150.0, 'lat': 40.43686293456}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436434221124, 'lon': -79.968918946419, 'seq': 25}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436409723131, 'lon': -79.968758013878, 'seq': 26}, {'typ': 'S', 'stpid': '2558', 'seq': 27, 'lon': -79.968582166668, 'stpnm': '5th Ave opp Brenham St', 'pdist': 9629.0, 'lat': 40.436405101253}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436421433251, 'lon': -79.968040360447, 'seq': 28}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436425516249, 'lon': -79.966962112423, 'seq': 29}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436478595211, 'lon': -79.966511501309, 'seq': 30}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436580670019, 'lon': -79.966243280407, 'seq': 31}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436727657471, 'lon': -79.965996517178, 'seq': 32}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436850146768, 'lon': -79.965846313473, 'seq': 33}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.436988967702, 'lon': -79.965728296276, 'seq': 34}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437119622437, 'lon': -79.965658558842, 'seq': 35}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437262131201, 'lon': -79.965561438496, 'seq': 36}, {'typ': 'S', 'stpid': '19850', 'seq': 37, 'lon': -79.965362955029, 'stpnm': '5th Ave at Maurice St', 'pdist': 10655.0, 'lat': 40.437429531855}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437552019874, 'lon': -79.965164471562, 'seq': 38}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437641844279, 'lon': -79.964917708333, 'seq': 39}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437739834402, 'lon': -79.96443491071, 'seq': 40}, {'typ': 'S', 'stpid': '2561', 'seq': 41, 'lon': -79.964154864419, 'stpnm': '5th Ave at Craft Ave', 'pdist': 11043.0, 'lat': 40.437787848263}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437836843236, 'lon': -79.96390810119, 'seq': 42}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.437951164701, 'lon': -79.963613058198, 'seq': 43}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438171641264, 'lon': -79.963285828698, 'seq': 44}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438318625238, 'lon': -79.963103438485, 'seq': 45}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.438592177888, 'lon': -79.962744022477, 'seq': 46}, {'typ': 'S', 'stpid': '2562', 'seq': 47, 'lon': -79.962188666667, 'stpnm': '5th Ave at Halket St', 'pdist': 11768.0, 'lat': 40.439010601112}, {'typ': 'S', 'stpid': '17671', 'seq': 48, 'lon': -79.959905053573, 'stpnm': '5th Ave past McKee Pl', 'pdist': 12632.0, 'lat': 40.440680195665}, {'typ': 'S', 'stpid': '2564', 'seq': 49, 'lon': -79.959239533731, 'stpnm': '5th Ave at Meyran Ave', 'pdist': 12879.0, 'lat': 40.441172012068}, {'typ': 'S', 'stpid': '2565', 'seq': 50, 'lon': -79.957992499998, 'stpnm': '5th Ave at Oakland Ave', 'pdist': 13394.0, 'lat': 40.442085267612}, {'typ': 'S', 'stpid': '2566', 'seq': 51, 'lon': -79.956747457672, 'stpnm': '5th Ave opp Thackeray Ave', 'pdist': 13872.0, 'lat': 40.442990015366}, {'typ': 'S', 'stpid': '2567', 'seq': 52, 'lon': -79.954881824734, 'stpnm': '5th Ave at Bigelow Blvd', 'pdist': 14532.0, 'lat': 40.444335993965}, {'typ': 'S', 'stpid': '3253', 'seq': 53, 'lon': -79.953202708996, 'stpnm': '5th Ave opp Tennyson Ave', 'pdist': 15193.0, 'lat': 40.445546749707}, {'typ': 'S', 'stpid': '3254', 'seq': 54, 'lon': -79.952111093253, 'stpnm': '5th Ave at Bellefield Ave', 'pdist': 15623.0, 'lat': 40.44630592845}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446623373397, 'lon': -79.951609216288, 'seq': 55}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446721350429, 'lon': -79.951314173296, 'seq': 56}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446770338891, 'lon': -79.950579248026, 'seq': 57}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446835656786, 'lon': -79.949640474871, 'seq': 58}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446913221703, 'lon': -79.948980651453, 'seq': 59}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.447374526992, 'lon': -79.94917913492, 'seq': 60}, {'typ': 'S', 'stpid': '2571', 'seq': 61, 'lon': -79.949420533732, 'stpnm': 'Craig St at Park Plaza', 'pdist': 16851.0, 'lat': 40.447762347572}, {'typ': 'S', 'stpid': '2572', 'seq': 62, 'lon': -79.950570087302, 'stpnm': 'Craig St at Bayard St', 'pdist': 17630.0, 'lat': 40.449607534262}, {'typ': 'S', 'stpid': '2573', 'seq': 63, 'lon': -79.951959353173, 'stpnm': 'Craig St at Centre Ave', 'pdist': 18457.0, 'lat': 40.451872385571}, {'typ': 'S', 'stpid': '2574', 'seq': 64, 'lon': -79.951001553571, 'stpnm': 'Centre Ave at Melwood Ave', 'pdist': 18766.0, 'lat': 40.452173174704}, {'typ': 'S', 'stpid': '2575', 'seq': 65, 'lon': -79.949852615741, 'stpnm': 'Centre Ave at Neville St', 'pdist': 19116.0, 'lat': 40.452535756554}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.452783201296, 'lon': -79.94906657475, 'seq': 66}, {'typ': 'S', 'stpid': '8198', 'seq': 67, 'lon': -79.947097833333, 'stpnm': 'Centre Ave opp Millvale Ave', 'pdist': 19930.0, 'lat': 40.453342433673}, {'typ': 'S', 'stpid': '8199', 'seq': 68, 'lon': -79.944541689153, 'stpnm': 'Centre Ave at Morewood Ave', 'pdist': 20688.0, 'lat': 40.454097103115}, {'typ': 'S', 'stpid': '8201', 'seq': 69, 'lon': -79.941522513889, 'stpnm': 'Centre Ave opp Cypress St', 'pdist': 21562.0, 'lat': 40.45498568797}, {'typ': 'S', 'stpid': '8202', 'seq': 70, 'lon': -79.939839104496, 'stpnm': 'Centre Ave at Shadyside Hosptital', 'pdist': 22083.0, 'lat': 40.455500688037}, {'typ': 'S', 'stpid': '8203', 'seq': 71, 'lon': -79.938622420634, 'stpnm': 'Centre Ave at Aiken St', 'pdist': 22408.0, 'lat': 40.455849760816}, {'typ': 'S', 'stpid': '8204', 'seq': 72, 'lon': -79.936516138226, 'stpnm': 'Centre Ave at Graham St', 'pdist': 23051.0, 'lat': 40.456481021498}, {'typ': 'S', 'stpid': '8205', 'seq': 73, 'lon': -79.934715333332, 'stpnm': 'Centre Ave at Negley Ave', 'pdist': 23617.0, 'lat': 40.457034933475}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.457255347443, 'lon': -79.934001865734, 'seq': 74}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.457496169285, 'lon': -79.934119882931, 'seq': 75}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.457647192712, 'lon': -79.934130611767, 'seq': 76}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.457851277885, 'lon': -79.934082332004, 'seq': 77}, {'typ': 'S', 'stpid': '8206', 'seq': 78, 'lon': -79.933821595899, 'stpnm': 'Centre Ave at Baum Blvd', 'pdist': 24178.0, 'lat': 40.458263705021}, {'typ': 'S', 'stpid': '8207', 'seq': 79, 'lon': -79.932419593254, 'stpnm': 'Negley Ave at Friendship Ave', 'pdist': 25101.0, 'lat': 40.460559766618}, {'typ': 'S', 'stpid': '8208', 'seq': 80, 'lon': -79.931591155423, 'stpnm': 'Negley Ave opp Coral St', 'pdist': 25663.0, 'lat': 40.461936270313}, {'typ': 'S', 'stpid': '8209', 'seq': 81, 'lon': -79.930680564813, 'stpnm': 'Negley Ave at Penn Ave', 'pdist': 26175.0, 'lat': 40.463411168093}, {'typ': 'S', 'stpid': '9128', 'seq': 82, 'lon': -79.927021771164, 'stpnm': 'Penn Ave at Beatty St', 'pdist': 27341.0, 'lat': 40.462186421807}, {'typ': 'S', 'stpid': '9129', 'seq': 83, 'lon': -79.925920042329, 'stpnm': 'Penn Ave at Whitfield St', 'pdist': 27695.0, 'lat': 40.461761433221}, {'typ': 'S', 'stpid': '3133', 'seq': 84, 'lon': -79.924964833335, 'stpnm': 'Penn Ave at Highland Ave', 'pdist': 27978.0, 'lat': 40.461429599904}, {'typ': 'S', 'stpid': '18975', 'seq': 85, 'lon': -79.923960135581, 'stpnm': 'Penn Ave at Sheridan Sq', 'pdist': 28284.0, 'lat': 40.461080255601}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460778223843, 'lon': -79.923107193114, 'seq': 86}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460631288983, 'lon': -79.922769234778, 'seq': 87}, {'typ': 'S', 'stpid': '19574', 'seq': 88, 'lon': -79.921938462961, 'stpnm': 'Penn Ave at Eastside III Dr', 'pdist': 28861.0, 'lat': 40.460351862483}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460323291659, 'lon': -79.921852632273, 'seq': 89}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.460237579116, 'lon': -79.921788259256, 'seq': 90}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.45995186985, 'lon': -79.921112342585, 'seq': 91}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.4596865673, 'lon': -79.920774384249, 'seq': 92}, {'typ': 'S', 'stpid': '8276', 'seq': 93, 'lon': -79.920097333333, 'stpnm': 'Penn Ave past Shady Ave', 'pdist': 29549.0, 'lat': 40.459199365493}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.459080998544, 'lon': -79.919963222882, 'seq': 94}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.459015692551, 'lon': -79.919754010579, 'seq': 95}, {'typ': 'S', 'stpid': '8277', 'seq': 96, 'lon': -79.918572725527, 'stpnm': 'Penn Ave opp Village of Eastside Shpg Ctr', 'pdist': 30029.0, 'lat': 40.458219989936}, {'typ': 'S', 'stpid': '19381', 'seq': 97, 'lon': -79.9176930959, 'stpnm': 'Penn Ave opp East Liberty Blvd (farside)', 'pdist': 30400.0, 'lat': 40.457613502192}, {'typ': 'S', 'stpid': '8279', 'seq': 98, 'lon': -79.915559208994, 'stpnm': 'Penn Ave opp Bakery Square', 'pdist': 30705.0, 'lat': 40.456158097123}, {'typ': 'S', 'stpid': '8280', 'seq': 99, 'lon': -79.912942375661, 'stpnm': 'Penn Ave at 5th Ave', 'pdist': 31694.0, 'lat': 40.454361927924}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.453982309293, 'lon': -79.912395205022, 'seq': 100}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.453194491742, 'lon': -79.911279406072, 'seq': 101}, {'typ': 'S', 'stpid': '8281', 'seq': 102, 'lon': -79.910441722881, 'stpnm': 'Penn Ave at Linden', 'pdist': 32619.0, 'lat': 40.452643023645}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.452187911929, 'lon': -79.909776638236, 'seq': 103}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.451938906948, 'lon': -79.909347484794, 'seq': 104}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.451751132091, 'lon': -79.908955882277, 'seq': 105}, {'typ': 'S', 'stpid': '20763', 'seq': 106, 'lon': -79.908237050262, 'stpnm': 'Penn Ave at Dallas Ave', 'pdist': 33320.0, 'lat': 40.451444976306}, {'typ': 'S', 'stpid': '7692', 'seq': 107, 'lon': -79.906119717594, 'stpnm': 'Penn Ave opp Murtland', 'pdist': 33990.0, 'lat': 40.450752764737}, {'typ': 'S', 'stpid': '7693', 'seq': 108, 'lon': -79.903907833332, 'stpnm': 'Penn Ave at Lang', 'pdist': 34657.0, 'lat': 40.450015933852}, {'typ': 'S', 'stpid': '7694', 'seq': 109, 'lon': -79.901542666665, 'stpnm': 'Penn Ave at Homewood', 'pdist': 35380.0, 'lat': 40.449207267228}, {'typ': 'S', 'stpid': '7695', 'seq': 110, 'lon': -79.899738177912, 'stpnm': 'Penn Ave at Lexington', 'pdist': 35928.0, 'lat': 40.44861834093}, {'typ': 'S', 'stpid': '18542', 'seq': 111, 'lon': -79.898263060188, 'stpnm': 'Penn Ave past Carnegie', 'pdist': 36455.0, 'lat': 40.448122273561}, {'typ': 'S', 'stpid': '7697', 'seq': 112, 'lon': -79.895344437829, 'stpnm': 'Penn Ave at Braddock', 'pdist': 37204.0, 'lat': 40.447155681048}, {'typ': 'S', 'stpid': '7698', 'seq': 113, 'lon': -79.894080240081, 'stpnm': 'Penn Ave at East End', 'pdist': 37577.0, 'lat': 40.44674726544}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.445849137648, 'lon': -79.891376573394, 'seq': 114}, {'typ': 'S', 'stpid': '8283', 'seq': 115, 'lon': -79.891540499999, 'stpnm': 'Trenton Ave past Penn', 'pdist': 38534.0, 'lat': 40.445526100761}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.445060696134, 'lon': -79.891780072747, 'seq': 116}, {'typ': 'S', 'stpid': '8284', 'seq': 117, 'lon': -79.892091208993, 'stpnm': 'Trenton Ave at South', 'pdist': 38980.0, 'lat': 40.44430135164}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.444219701183, 'lon': -79.892101937829, 'seq': 118}, {'typ': 'S', 'stpid': '7329', 'seq': 119, 'lon': -79.890609666666, 'stpnm': 'South Ave at West', 'pdist': 39454.0, 'lat': 40.443714100858}, {'typ': 'S', 'stpid': '7330', 'seq': 120, 'lon': -79.888255084655, 'stpnm': 'South Ave at Pitt', 'pdist': 40151.0, 'lat': 40.442899450303}, {'typ': 'S', 'stpid': '20013', 'seq': 121, 'lon': -79.887481977515, 'stpnm': 'South Ave at Pennwood', 'pdist': 40395.0, 'lat': 40.442649715114}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442433335871, 'lon': -79.886747052245, 'seq': 122}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442743615323, 'lon': -79.886575390868, 'seq': 123}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442772193622, 'lon': -79.886698772483, 'seq': 124}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.442845680619, 'lon': -79.886800696425, 'seq': 125}, {'typ': 'S', 'stpid': '16014', 'seq': 126, 'lon': -79.88690183333199, 'stpnm': 'Hay St Ramp inbound shelter', 'pdist': 40996.0, 'lat': 40.443279227407}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.443626246017, 'lon': -79.88700912169202, 'seq': 127}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.443961015213, 'lon': -79.887186147487, 'seq': 128}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.444859168235, 'lon': -79.887545563495, 'seq': 129}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.445712402489, 'lon': -79.887985445774, 'seq': 130}, {'typ': 'S', 'stpid': '16118', 'seq': 131, 'lon': -79.88844805952203, 'stpnm': 'Wilkinsburg Station stop D', 'pdist': 42200.0, 'lat': 40.446499209682}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.446980929603, 'lon': -79.88878065344, 'seq': 132}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.447519798645, 'lon': -79.88925272222701, 'seq': 133}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.448026005021, 'lon': -79.88972479101301, 'seq': 134}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.448548536635, 'lon': -79.89027732607002, 'seq': 135}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.448956761632, 'lon': -79.89076548811101, 'seq': 136}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.449258846534, 'lon': -79.89114099737299, 'seq': 137}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.449581341079, 'lon': -79.89159160848703, 'seq': 138}, {'typ': 'W', 'pdist': 0.0, 'lat': 40.449760957623, 'lon': -79.89176326986399, 'seq': 139}, {'typ': 'S', 'stpid': '19988', 'seq': 140, 'lon': -79.89193733333298, 'stpnm': 'LAYOVER BRUSHTON TERMINAL', 'pdist': 43730.0, 'lat': 40.449901185011}]}] } })
36,420
23,109
import math import torch from bisect import bisect_right class _LRScheduler: def __init__(self, optimizer, last_epoch=-1): self.optimizer = optimizer self.base_lr = optimizer.lr self.last_epoch = last_epoch def step(self): self.last_epoch += 1 self.optimizer.lr = self.get_lr() class StepLR(_LRScheduler): def __init__(self, optimizer, step_size, gamma=0.1, last_epoch=-1): super().__init__(optimizer, last_epoch) self.step_size = step_size self.gamma = gamma def get_lr(self): return self.base_lr * self.gamma ** (self.last_epoch // self.step_size) class MultiStepLR(_LRScheduler): def __init__(self, optimizer, milestones, gamma=0.1, last_epoch=-1): super().__init__(optimizer, last_epoch) self.milestones = milestones self.gamma = gamma def get_lr(self): return self.base_lr * self.gamma ** bisect_right(self.milestones, self.last_epoch) class ExponentialLR(_LRScheduler): def __init__(self, optimizer, gamma, last_epoch=-1): super().__init__(optimizer, last_epoch) self.gamma = gamma def get_lr(self): return self.base_lr * self.gamma ** self.last_epoch class CosineAnnealingLR(_LRScheduler): def __init__(self, optimizer, T_max, eta_min=0, last_epoch=-1): super().__init__(optimizer, last_epoch) self.T_max = T_max self.eta_min = eta_min def get_lr(self): return self.eta_min + (self.base_lr - self.eta_min) * (1 + math.cos(math.pi * self.last_epoch / self.T_max)) / 2
1,589
567
# -*- coding: utf-8 -*- from .._globals import IDENTITY from ..connection import ConnectionPool from .base import BaseAdapter class TeradataAdapter(BaseAdapter): drivers = ('pyodbc',) types = { 'boolean': 'CHAR(1)', 'string': 'VARCHAR(%(length)s)', 'text': 'VARCHAR(2000)', 'json': 'VARCHAR(4000)', 'password': 'VARCHAR(%(length)s)', 'blob': 'BLOB', 'upload': 'VARCHAR(%(length)s)', 'integer': 'INT', 'bigint': 'BIGINT', 'float': 'REAL', 'double': 'DOUBLE', 'decimal': 'NUMERIC(%(precision)s,%(scale)s)', 'date': 'DATE', 'time': 'TIME', 'datetime': 'TIMESTAMP', # Modified Constraint syntax for Teradata. # Teradata does not support ON DELETE. 'id': 'INT GENERATED ALWAYS AS IDENTITY', # Teradata Specific 'reference': 'INT', 'list:integer': 'VARCHAR(4000)', 'list:string': 'VARCHAR(4000)', 'list:reference': 'VARCHAR(4000)', 'geometry': 'ST_GEOMETRY', # http://www.info.teradata.com/HTMLPubs/DB_TTU_14_00/index.html#page/Database_Management/B035_1094_111A/ch14.055.160.html 'big-id': 'BIGINT GENERATED ALWAYS AS IDENTITY', # Teradata Specific 'big-reference': 'BIGINT', 'reference FK': ' REFERENCES %(foreign_key)s', 'reference TFK': ' FOREIGN KEY (%(field_name)s) REFERENCES %(foreign_table)s (%(foreign_key)s)', } def __init__(self,db,uri,pool_size=0,folder=None,db_codec ='UTF-8', credential_decoder=IDENTITY, driver_args={}, adapter_args={}, do_connect=True, after_connection=None): self.db = db self.dbengine = "teradata" self.uri = uri if do_connect: self.find_driver(adapter_args,uri) self.pool_size = pool_size self.folder = folder self.db_codec = db_codec self._after_connection = after_connection self.find_or_make_work_folder() ruri = uri.split('://', 1)[1] def connector(cnxn=ruri,driver_args=driver_args): return self.driver.connect(cnxn,**driver_args) self.connector = connector if do_connect: self.reconnect() def close(self,action='commit',really=True): # Teradata does not implicitly close off the cursor # leading to SQL_ACTIVE_STATEMENTS limit errors self.cursor.close() ConnectionPool.close(self, action, really) def LEFT_JOIN(self): return 'LEFT OUTER JOIN' # Similar to MSSQL, Teradata can't specify a range (for Pageby) def select_limitby(self, sql_s, sql_f, sql_t, sql_w, sql_o, limitby): if limitby: (lmin, lmax) = limitby sql_s += ' TOP %i' % lmax return 'SELECT %s %s FROM %s%s%s;' % (sql_s, sql_f, sql_t, sql_w, sql_o) def _truncate(self, table, mode=''): tablename = table._tablename return ['DELETE FROM %s ALL;' % (tablename)]
2,972
1,051
import os import platform import unittest import nose from conans import tools from conans.errors import ConanException from conans.model.version import Version from conans import __version__ as client_version from conans.model import settings from conans.test.utils.tools import TestClient from conans.test.assets.visual_project_files import get_vs_project_files class vswhereTest(unittest.TestCase): # Environment supossed: # - BuildTools 14 (2015) # - VS Community 14 (2015) # # - BuildTools 15 (2017) OR VS Community 15 (2017) modern_products = 1 # 2017 or higher versions without BuildTools -> vswhere() all_modern_products = 2 # 2017 or higher versions with BuildTools -> vswhere(products=["*"]) modern_and_legacy_products = 2 # 2017 and lower versions (without BuildTools) -> vswhere(legacy=True) only_legacy_products = 1 all_products = 3 def setUp(self): if platform.system() != "Windows": raise nose.SkipTest("Only Windows test") if Version(client_version) < Version("1.1.0-dev"): raise nose.SkipTest("Only >= 1.1.0-dev version") def vs_comntools_test(self): # Fake path with tools.environment_append({"VS150COMNTOOLS": "fake/path/here"}): path = tools.vs_comntools("15") self.assertEqual(path, "fake/path/here") # VS 14 path path = tools.vs_comntools("14") self.assertEqual(path, "C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\") # VS 15 path (shouldn't be found as VS150COMNTOOLS is not set by default) path = tools.vs_comntools("15") self.assertEqual(path, None) def vswhere_test(self): # products and legacy not allowed self.assertRaises(ConanException, tools.vswhere, products=["*"], legacy=True) # Detect only one product (VS Community 15) as vswhere default detection nproducts = len(tools.vswhere()) self.assertEqual(nproducts, self.modern_products) # Detect only modern products (VS Community 15 & BuildTools 15) products = tools.vswhere(products=["*"]) nproducts = len(products) self.assertEqual(nproducts, self.all_modern_products) installation_paths = [product["installationPath"] for product in products] self.assertTrue(any("Community" in install_path for install_path in installation_paths)) self.assertTrue(any("BuildTools" in install_path for install_path in installation_paths)) # Detect also legacy products but no modern BuildTools products = tools.vswhere(legacy=True) nproducts = len(products) self.assertEqual(nproducts, self.modern_and_legacy_products) installation_paths = [product["installationPath"] for product in products] self.assertTrue(any("Community" in install_path for install_path in installation_paths)) self.assertTrue(any("Microsoft Visual Studio 14.0" in install_path for install_path in installation_paths)) # Detect all installed products products = tools.vswhere(products=["*"]) products += tools.vswhere(legacy=["*"]) seen_products = [] for product in products: if product not in seen_products: seen_products.append(product) products = seen_products nproducts = len(products) self.assertEqual(nproducts, self.all_products) installation_paths = [product["installationPath"] for product in products] self.assertTrue(any("Community" in install_path for install_path in installation_paths)) self.assertTrue(any("BuildTools" in install_path for install_path in installation_paths)) self.assertTrue(any("Microsoft Visual Studio 14.0" in install_path for install_path in installation_paths)) def vs_installation_path_test(self): # Default behaviour install_path = tools.vs_installation_path("15") self.assertIn("Community", install_path) install_path = tools.vs_installation_path("14") self.assertIn("Microsoft Visual Studio 14.0", install_path) # only BuildTools detection install_path = tools.vs_installation_path("15", preference=["BuildTools"]) self.assertIn("BuildTools", install_path) install_path = tools.vs_installation_path("14", preference=["BuildTools"]) self.assertIn("Microsoft Visual Studio 14.0", install_path) # Ask for not installed versions install_path = tools.vs_installation_path("15", preference=["Enterprise"]) self.assertIsNone(install_path) install_path = tools.vs_installation_path("15", preference=["Professional"]) self.assertIsNone(install_path) # Change preference order install_path = tools.vs_installation_path("15", preference=["BuildTools", "Community", "Professional", "Enterprise"]) self.assertIn("BuildTools", install_path) install_path = tools.vs_installation_path("15", preference=["Professional", "Enterprise", "Community"]) self.assertIn("Community", install_path) # Preference order by env var with(tools.environment_append({"CONAN_VS_INSTALLATION_PREFERENCE":"BuildTools, Community,Professional, Enterprise"})): install_path = tools.vs_installation_path("15") self.assertIn("BuildTools", install_path) with(tools.environment_append({"CONAN_VS_INSTALLATION_PREFERENCE":"Professional, Enterprise,Community"})): install_path = tools.vs_installation_path("15") self.assertIn("Community", install_path) def vvcars_command_test(self): fake_settings = settings.Settings({"os":"Windows", "arch": "x86_64"}) # preference order with VS 15 with(tools.environment_append({"CONAN_VS_INSTALLATION_PREFERENCE":"BuildTools, Community,Professional, Enterprise"})): command = tools.vcvars_command(settings=fake_settings, compiler_version="15") self.assertNotIn("Community", command) self.assertIn("VC/Auxiliary/Build/vcvarsall.bat", command) self.assertIn("Microsoft Visual Studio\\2017\\BuildTools", command) self.assertIn("VSCMD_START_DIR", command) with(tools.environment_append({"CONAN_VS_INSTALLATION_PREFERENCE":"Professional, Enterprise,Community"})): command = tools.vcvars_command(settings=fake_settings, compiler_version="15") self.assertNotIn("BuildTools", command) self.assertIn("VC/Auxiliary/Build/vcvarsall.bat", command) self.assertIn("Microsoft Visual Studio\\2017\\Community", command) self.assertIn("VSCMD_START_DIR", command) # With VS 14 order of preference does not apply command = tools.vcvars_command(settings=fake_settings, compiler_version="14") self.assertNotIn("VSCMD_START_DIR", command) self.assertIn("VC/vcvarsall.bat", command) self.assertIn("Microsoft Visual Studio 14.0\\", command) def build_test(self): conan_build_vs = """ from conans import ConanFile, MSBuild, tools class HelloConan(ConanFile): name = "Hello" version = "1.2.1" settings = "os", "build_type", "arch", "compiler" export_source = "*" def build(self): msbuild = MSBuild(self) msbuild.build("MyProject.sln", upgrade_project=False) """ client = TestClient() files = get_vs_project_files() files["conanfile.py"] = conan_build_vs client.save(files) with(tools.environment_append({"CONAN_PRINT_RUN_COMMANDS": "1"})): with(tools.environment_append({"CONAN_VS_INSTALLATION_PREFERENCE": "BuildTools"})): client.run("install .") client.run("build .") self.assertIn("BuildTools", client.out) conan_build_vs = conan_build_vs.replace("upgrade_project=False", "upgrade_project=True") files["conanfile.py"] = conan_build_vs client.save(files) with(tools.environment_append({"CONAN_VS_INSTALLATION_PREFERENCE":"BuildTools", "CONAN_SKIP_VS_PROJECTS_UPGRADE":"True"})): client.run("install .") client.run("build .") self.assertIn("BuildTools", client.out)
8,367
2,492
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import argparse import json import os from pinliner import __version__ import sys TEMPLATE_FILE = 'importer.template' TEMPLATE_PATTERN = '${CONTENTS}' def output(cfg, what, newline=True): # We need indentation for PEP8 cfg.outfile.write(what) if newline: cfg.outfile.write(os.linesep) def process_file(cfg, base_dir, package_path): if cfg.tagging: output(cfg, '<tag:' + package_path + '>') path = os.path.splitext(package_path)[0].replace(os.path.sep, '.') package_start = cfg.outfile.tell() full_path = os.path.join(base_dir, package_path) with open(full_path, 'r') as f: # Read the whole file code = f.read() # Insert escape character before ''' since we'll be using ''' to insert # the code as a string output(cfg, code.replace("'''", r"\'''"), newline=cfg.tagging) package_end = cfg.outfile.tell() is_package = 1 if path.endswith('__init__') else 0 if is_package: path = path[:-9] # Get file timestamp timestamp = int(os.path.getmtime(full_path)) return path, is_package, package_start, package_end, timestamp def template(cfg): template_path = os.path.join(os.path.dirname(__file__), TEMPLATE_FILE) with open(template_path) as f: template = f.read() prefix_end = template.index(TEMPLATE_PATTERN) prefix_data = template[:prefix_end].replace('%{FORCE_EXC_HOOK}', str(cfg.set_hook)) prefix_data = prefix_data.replace('%{DEFAULT_PACKAGE}', cfg.default_package) cfg.outfile.write(prefix_data) postfix_begin = prefix_end + len(TEMPLATE_PATTERN) return template[postfix_begin:] def process_directory(cfg, base_dir, package_path): files = [] contents = os.listdir(os.path.join(base_dir, package_path)) for content in contents: next_path = os.path.join(package_path, content) path = os.path.join(base_dir, next_path) if is_module(path): files.append(process_file(cfg, base_dir, next_path)) elif is_package(path): files.extend(process_directory(cfg, base_dir, next_path)) return files def process_files(cfg): # template would look better as a context manager postfix = template(cfg) files = [] output(cfg, "'''") for package_path in cfg.packages: base_dir, module_name = os.path.split(package_path) files.extend(process_directory(cfg, base_dir, module_name)) output(cfg, "'''") # Transform the list into a dictionary inliner_packages = {data[0]: data[1:] for data in files} # Generate the references to the positions of the different packages and # modules inside the main file. # We don't use indent to decrease the number of bytes in the file data = json.dumps(inliner_packages) output(cfg, 2 * os.linesep + 'inliner_packages = ', newline=False) data = data.replace('],', '],' + os.linesep + ' ') data = data.replace('[', '[' + os.linesep + 8 * ' ') data = '%s%s %s%s%s' % (data[0], os.linesep, data[1:-1], os.linesep, data[-1]) output(cfg, data) # No newline on last line, as we want output file to be PEP8 compliant. output(cfg, postfix, newline=False) cfg.outfile.close() def parse_args(): class MyParser(argparse.ArgumentParser): """Class to print verbose help on error.""" def error(self, message): self.print_help() sys.stderr.write('\nERROR: %s\n' % message) sys.exit(2) general_description = """Pinliner - Python Inliner (Version %s) This tool allows you to merge all files that comprise a Python package into a single file and be able to use this single file as if it were a package. Imports will work as usual so if you have a package structure like: . └── [my_package] ├── file_a.py ├── [sub_package] │   ├── file_b.py │   └── __init__.py ├── __init__.py And you execute: $ mkdir test $ pinliner my_package test/my_package.py $ cd test $ python You'll be able to use this file as if it were the real package: >>> import my_package >>> from my_package import file_a as a_file >>> from my_package.sub_package import file_b And __init__.py contents will be executed as expected when importing my_package and you'll be able to access its contents like you would with your normal package. Modules will also behave as usual. By default there is no visible separation between the different modules' source code, but one can be enabled for clarity with option --tag, which will include a newline and a <tag:file_path> tag before each of the source files. """ % __version__ general_epilog = None parser = MyParser(description=general_description, epilog=general_epilog, argument_default='', formatter_class=argparse.RawTextHelpFormatter) parser.add_argument('packages', nargs='+', help='Packages to inline.') parser.add_argument('--version', action='version', version=__version__) parser.add_argument('-o', '--outfile', nargs='?', type=argparse.FileType('w'), default=sys.stdout, help='Output file.') parser.add_argument('--set-except', default=None, dest='set_hook', action='store_true', help='Force setting handler for uncaught exceptions.') parser.add_argument('--no-except', default=None, dest='set_hook', action='store_false', help="Don't set handler for uncaught exceptions.") parser.add_argument('--tag', default=False, dest='tagging', action='store_true', help="Mark with <tag:file_path> each added file.") parser.add_argument('-d', '--default-pkg', default=None, dest='default_package', help='Define the default package when multiple ' 'packages are inlined.') cfg = parser.parse_args() # If user didn't pass a default package determine one ourselves. if cfg.default_package is None: # For single package file default is the package, for multiple packaged # files default is none (act as a bundle). def_file = cfg.packages[0] if len(cfg.packages) == 1 else '' cfg.default_package = def_file return cfg def is_module(module): # This validation is poor, but good enough for now return os.path.isfile(module) and module.endswith('.py') def is_package(package): init_file = os.path.join(package, '__init__.py') return os.path.isdir(package) and os.path.isfile(init_file) def validate_args(cfg): missing = False # This is weird now, but in the future we'll allow to inline multiple # packages for package in cfg.packages: if not is_package(package): sys.stderr.write('ERROR: %s is not a python package' % package) missing = True if missing: sys.exit(1) if cfg.default_package: if cfg.default_package not in cfg.packages: sys.stderr.write('ERROR: %s is not a valid default package' % cfg.default_pkg) sys.exit(2) # Convert the default package from path to package cfg.default_package = os.path.split(cfg.default_package)[1] def main(): cfg = parse_args() validate_args(cfg) process_files(cfg) if __name__ == '__main__': main()
7,797
2,352
from dotenv import load_dotenv load_dotenv() import os import boto3 #s3 = boto3.resource('s3') s3 = boto3.resource('s3', aws_access_key_id=os.environ.get("AWS_KEY_ID"), aws_secret_access_key=os.environ.get("AWS_SECRET_KEY")) for bucket in s3.buckets.all(): print(bucket.name)
305
121
import os from sonar.testbench import Testbench, Module, TestVector, Thread from sonar.interfaces import AXIS from sonar_strToInt import strToInt hold_buffer = Testbench.default('hold_buffer') filepath = os.path.join(os.path.dirname(__file__), 'build/hold_buffer/') dut = Module.default("DUT") dut.add_clock_port('ap_clk', '20ns') dut.add_reset_port('ap_rst_n') dut.add_port('dataRelease_V', 'input', 16) axis_input = AXIS('axis_input', 'slave', 'ap_clk', c_struct='axis_word', c_stream='uaxis_l') axis_input.port.init_channels('tkeep', 64, True) dut.add_interface(axis_input) axis_output = AXIS('axis_output', 'master', 'ap_clk', c_struct='axis_word', c_stream='uaxis_l') axis_output.port.init_channels('tkeep', 64, True) dut.add_interface(axis_output) hold_buffer.add_module(dut) ################################################################################ # Test Vectors ################################################################################ # Initialization thread (added to each test vector to reset everything) initT = Thread() initT.init_signals() initT.wait_negedge('ap_clk') initT.add_delay('40ns') initT.set_signal('ap_rst_n', 1) initT.set_signal('axis_output_tready', 1) #------------------------------------------------------------------------------- # #------------------------------------------------------------------------------- Release_A = TestVector() Release_A.add_thread(initT) rA_t1 = Thread() rA_t1.add_delay('100ns') rA_t1.init_timer() rA_t1.set_signal('dataRelease_V', 1) axis_input.writes(rA_t1, [ {"tdata": 0xDEF, "callTB": 1}, {"tdata": 0xFED, "callTB": 1}, ]) Release_A.add_thread(rA_t1) rA_t2 = Thread() axis_output.read(rA_t2, 0xDEF) axis_output.read(rA_t2, 0xFED) rA_t2.print_elapsed_time("Release_A") rA_t2.end_vector() Release_A.add_thread(rA_t2) #------------------------------------------------------------------------------- # Medium Message A # # #------------------------------------------------------------------------------- Release_B = TestVector() Release_B.add_thread(initT) rB_t1 = Thread() rB_t1.add_delay('100ns') rB_t1.init_timer() axis_input.writes(rB_t1, [ {"tdata": 0xDEF, "callTB": 1}, {"tdata": 0xFED, "callTB": 1}, ]) rB_t1.set_signal('dataRelease_V', 1) Release_B.add_thread(rB_t1) rB_t2 = Thread() axis_output.read(rB_t2, 0xDEF) axis_output.read(rB_t2, 0xFED) rB_t2.print_elapsed_time("Release_B") rB_t2.end_vector() Release_B.add_thread(rB_t2) hold_buffer.add_test_vector(Release_A) hold_buffer.add_test_vector(Release_B) hold_buffer.generateTB(filepath, 'all')
2,570
995
#!/usr/bin/env python # filename: pair.py # # Copyright (c) 2015 Bryan Briney # License: The MIT license (http://opensource.org/licenses/MIT) # # Permission is hereby granted, free of charge, to any person obtaining a copy of this software # and associated documentation files (the "Software"), to deal in the Software without restriction, # including without limitation the rights to use, copy, modify, merge, publish, distribute, # sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all copies or # substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING # BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, # DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # import copy import sys import traceback from Bio.Seq import Seq from Bio.Alphabet import generic_dna from abtools import germlines from abtools.alignment import global_alignment from abtools.sequence import Sequence class Pair(object): ''' Holds a pair of sequences, corresponding to HC and LC of a single mAb. Input is a list of dicts, with each dict containing sequence information from a single chain, formatted as would be returned from a query on a MongoDB database containing AbStar output. ''' def __init__(self, seqs, name=None, h_selection_func=None, l_selection_func=None): self._seqs = seqs self._heavy = None self._light = None self._heavies = [s for s in seqs if s['chain'] == 'heavy'] self._lights = [s for s in seqs if s['chain'] in ['kappa', 'lambda']] self._name = name self._fasta = None self._sample = None self._subject = None self._group = None self._experiment = None self._timepoint = None self._is_pair = None self._vrc01_like = None self._lineage = None self._select_heavy = h_selection_func self._select_light = l_selection_func def __eq__(self, other): return (self.heavy, self.light) == (other.heavy, other.light) def __ne__(self, other): return not self == other def __hash(self): return hash((self.heavy, self.light)) @property def heavy(self): if self._heavy is None: # self._heavies = [s for s in self._seqs if s['chain'] == 'heavy'] if len(self._heavies) > 0: if self._select_heavy is not None: self._heavy = Sequence(self._select_heavy(self._heavies)) else: self._heavy = Sequence(self._heavies[0]) else: self._heavy = None return self._heavy @heavy.setter def heavy(self, heavy): self._heavy = heavy @property def light(self): if self._light is None: # self._lights = [s for s in self._seqs if s['chain'] in ['kappa', 'lambda']] if len(self._lights) > 0: if self._select_light is not None: self._light = Sequence(self._select_light(self._lights)) else: self._light = Sequence(self._lights[0]) else: self._light = None return self._light @light.setter def light(self, light): self._light = light @property def is_pair(self): if all([self.heavy is not None, self.light is not None]): return True return False @property def lineage(self): if self._lineage is None: self._lineage = self.heavy['clonify']['id'] return self._lineage @property def vrc01_like(self): if self._vrc01_like is None: if any([self.heavy is None, self.light is None]): self._vrc01_like = False else: self._vrc01_like = all([self.heavy['v_gene']['gene'] == 'IGHV1-2', self.light['cdr3_len'] == 5]) return self._vrc01_like @property def name(self): if self._name is None: if self.heavy is not None: self._name = self.heavy['seq_id'] elif self.light is not None: self._name = self.light['seq_id'] return self._name @name.setter def name(self, name): self._name = name @property def sample(self): if self._sample is None: slist = [] if self.experiment is not None: slist.append(str(self.experiment)) if self.group is not None: slist.append(str(self.group)) if self.subject is not None: slist.append(str(self.subject)) if self.timepoint is not None: slist.append(str(self.timepoint)) if slist: self._sample = '|'.join(slist) return self._sample @property def subject(self): if self._subject is None: if self.heavy is not None and 'subject' in list(self.heavy.keys()): self._subject = self.heavy['subject'] elif self.light is not None and 'subject' in list(self.light.keys()): self._subject = self.light['subject'] return self._subject @subject.setter def subject(self, subject): self._subject = subject @property def group(self): if self._group is None: if self.heavy is not None and 'group' in list(self.heavy.keys()): self._group = self.heavy['group'] elif self.light is not None and 'group' in list(self.light.keys()): self._group = self.light['group'] return self._group @group.setter def group(self, group): self._group = group @property def experiment(self): if self._experiment is None: if self.heavy is not None and 'experiment' in list(self.heavy.keys()): self._experiment = self.heavy['experiment'] elif self.light is not None and 'experiment' in list(self.light.keys()): self._experiment = self.light['experiment'] return self._experiment @experiment.setter def experiment(self, experiment): self._experiment = experiment @property def timepoint(self): if self._timepoint is None: if self.heavy is not None and 'timepoint' in list(self.heavy.keys()): self._timepoint = self.heavy['timepoint'] elif self.light is not None and 'timepoint' in list(self.light.keys()): self._timepoint = self.light['timepoint'] return self._timepoint @timepoint.setter def timepoint(self, timepoint): self._timepoint = timepoint def refine(self, heavy=True, light=True, species='human'): for seq in [s for s in [self.heavy, self.light] if s is not None]: try: self.remove_ambigs(seq) self._refine_v(seq, species) self._refine_j(seq, species) self._retranslate(seq) except: print('REFINEMENT FAILED: {}, {} chain'.format(s['seq_id'], s['chain'])) print(traceback.format_exception_only(sys.exc_info()[0], sys.exc_info()[1])) @staticmethod def remove_ambigs(seq): # fix Ns in the nucleotide sequence vdj = '' for s, g in zip(seq['vdj_nt'], seq['vdj_germ_nt']): if s.upper() == 'N': vdj += g else: vdj += s seq['vdj_nt'] = vdj # fix Xs in the amino acid sequence vdj = '' for s, g in zip(seq['vdj_aa'], seq['vdj_germ_aa']): if s.upper() == 'X': vdj += g else: vdj += s seq['vdj_aa'] = vdj @staticmethod def _refine_v(seq, species): ''' Completes the 5' end of a a truncated sequence with germline nucleotides. Input is a MongoDB dict (seq) and the species. ''' vgerm = germlines.get_germline(seq['v_gene']['full'], species) aln = global_alignment(seq['vdj_nt'], vgerm) prepend = '' for s, g in zip(aln.aligned_query, aln.aligned_target): if s != '-': break else: prepend += g seq['vdj_nt'] = prepend + seq['vdj_nt'] @staticmethod def _refine_j(seq, species): ''' Completes the 3' end of a a truncated sequence with germline nucleotides. Input is a MongoDB dict (seq) and the species. ''' jgerm = germlines.get_germline(seq['j_gene']['full'], species) aln = global_alignment(seq['vdj_nt'], jgerm) append = '' for s, g in zip(aln.aligned_query[::-1], aln.aligned_target[::-1]): if s != '-': break else: append += g seq['vdj_nt'] = seq['vdj_nt'] + append[::-1] @staticmethod def _retranslate(seq): ''' Retranslates a nucleotide sequence following refinement. Input is a Pair sequence (basically a dict of MongoDB output). ''' if len(seq['vdj_nt']) % 3 != 0: trunc = len(seq['vdj_nt']) % 3 seq['vdj_nt'] = seq['vdj_nt'][:-trunc] seq['vdj_aa'] = Seq(seq['vdj_nt'], generic_dna).translate() def fasta(self, key='vdj_nt', append_chain=True): ''' Returns the sequence pair as a fasta string. If the Pair object contains both heavy and light chain sequences, both will be returned as a single string. By default, the fasta string contains the 'vdj_nt' sequence for each chain. To change, use the <key> option to select an alternate sequence. By default, the chain (heavy or light) will be appended to the sequence name: >MySequence_heavy To just use the pair name (which will result in duplicate sequence names for Pair objects with both heavy and light chains), set <append_chain> to False. ''' fastas = [] for s, chain in [(self.heavy, 'heavy'), (self.light, 'light')]: if s is not None: c = '_{}'.format(chain) if append_chain else '' fastas.append('>{}{}\n{}'.format(s['seq_id'], c, s[key])) return '\n'.join(fastas) def get_pairs(db, collection, experiment=None, subject=None, group=None, name='seq_id', delim=None, delim_occurance=1, pairs_only=False): ''' Gets sequences and assigns them to the appropriate mAb pair, based on the sequence name. Inputs: ::db:: is a pymongo database connection object ::collection:: is the collection name, as a string If ::subject:: is provided, only sequences with a 'subject' field matching ::subject:: will be included. ::subject:: can be either a single subject (as a string) or an iterable (list or tuple) of subject strings. If ::group:: is provided, only sequences with a 'group' field matching ::group:: will be included. ::group:: can be either a single group (as a string) or an iterable (list or tuple) of group strings. ::name:: is the dict key of the field to be used to group the sequences into pairs. Default is 'seq_id' ::delim:: is an optional delimiter used to truncate the contents of the ::name:: field. Default is None, which results in no name truncation. ::delim_occurance:: is the occurance of the delimiter at which to trim. Trimming is performed as delim.join(name.split(delim)[:delim_occurance]), so setting delim_occurance to -1 will trucate after the last occurance of delim. Default is 1. ::pairs_only:: setting to True results in only truly paired sequences (pair.is_pair == True) will be returned. Default is False. Returns a list of Pair objects, one for each mAb pair. ''' match = {} if subject is not None: if type(subject) in (list, tuple): match['subject'] = {'$in': subject} elif type(subject) in (str, str): match['subject'] = subject if group is not None: if type(group) in (list, tuple): match['group'] = {'$in': group} elif type(group) in (str, str): match['group'] = group if experiment is not None: if type(experiment) in (list, tuple): match['experiment'] = {'$in': experiment} elif type(experiment) in (str, str): match['experiment'] = experiment seqs = list(db[collection].find(match)) return assign_pairs(seqs, name=name, delim=delim, delim_occurance=delim_occurance, pairs_only=pairs_only) def assign_pairs(seqs, name='seq_id', delim=None, delim_occurance=1, pairs_only=False): ''' Assigns sequences to the appropriate mAb pair, based on the sequence name. Inputs: ::seqs:: is a list of dicts, of the format returned by querying a MongoDB containing Abstar output. ::name:: is the dict key of the field to be used to group the sequences into pairs. Default is 'seq_id' ::delim:: is an optional delimiter used to truncate the contents of the ::name:: field. Default is None, which results in no name truncation. ::delim_occurance:: is the occurance of the delimiter at which to trim. Trimming is performed as delim.join(name.split(delim)[:delim_occurance]), so setting delim_occurance to -1 will trucate after the last occurance of delim. Default is 1. ::pairs_only:: setting to True results in only truly paired sequences (pair.is_pair == True) will be returned. Default is False. Returns a list of Pair objects, one for each mAb pair. ''' pdict = {} for s in seqs: if delim is not None: pname = delim.join(s[name].split(delim)[:delim_occurance]) else: pname = s[name] if pname not in pdict: pdict[pname] = [s, ] else: pdict[pname].append(s) pairs = [Pair(pdict[n], name=n) for n in list(pdict.keys())] if pairs_only: pairs = [p for p in pairs if p.is_pair] return pairs def deduplicate(pairs, aa=False, ignore_primer_regions=False): ''' Removes duplicate sequences from a list of Pair objects. If a Pair has heavy and light chains, both chains must identically match heavy and light chains from another Pair to be considered a duplicate. If a Pair has only a single chain, identical matches to that chain will cause the single chain Pair to be considered a duplicate, even if the comparison Pair has both chains. Note that identical sequences are identified by simple string comparison, so sequences of different length that are identical over the entirety of the shorter sequence are not considered duplicates. By default, comparison is made on the nucleotide sequence. To use the amino acid sequence instead, set aa=True. ''' nr_pairs = [] just_pairs = [p for p in pairs if p.is_pair] single_chains = [p for p in pairs if not p.is_pair] _pairs = just_pairs + single_chains for p in _pairs: duplicates = [] for nr in nr_pairs: identical = True vdj = 'vdj_aa' if aa else 'vdj_nt' offset = 4 if aa else 12 if p.heavy is not None: if nr.heavy is None: identical = False else: heavy = p.heavy[vdj][offset:-offset] if ignore_primer_regions else p.heavy[vdj] nr_heavy = nr.heavy[vdj][offset:-offset] if ignore_primer_regions else nr.heavy[vdj] if heavy != nr_heavy: identical = False if p.light is not None: if nr.light is None: identical = False else: light = p.light[vdj][offset:-offset] if ignore_primer_regions else p.light[vdj] nr_light = nr.light[vdj][offset:-offset] if ignore_primer_regions else nr.light[vdj] if light != nr_light: identical = False duplicates.append(identical) if any(duplicates): continue else: nr_pairs.append(p) return nr_pairs def refine(pairs, heavy=True, light=True, species='human'): refined_pairs = copy.deepcopy(pairs) for p in refined_pairs: p.refine(heavy, light, species) return refined_pairs
16,957
5,065
import tensorflow as tf from t3f.tensor_train import TensorTrain from t3f.tensor_train_batch import TensorTrainBatch from t3f import shapes from t3f import decompositions def project_sum(what, where, weights=None): """Project sum of `what` TTs on the tangent space of `where` TT. project_sum(what, x) = P_x(what) project_sum(batch_what, x) = P_x(\sum_i batch_what[i]) project_sum(batch_what, x, weights) = P_x(\sum_j weights[j] * batch_what[j]) This function implements the algorithm from the paper [1], theorem 3.1. [1] C. Lubich, I. Oseledets and B. Vandereycken, Time integration of Tensor Trains. Args: what: TensorTrain or TensorTrainBatch. In the case of batch returns projection of the sum of elements in the batch. where: TensorTrain, TT-tensor or TT-matrix on which tangent space to project weights: python list or tf.Tensor of numbers or None, weights of the sum Returns: a TensorTrain with the TT-ranks equal 2 * tangent_space_tens.get_tt_ranks() Complexity: O(d r_where^3 m) for orthogonalizing the TT-cores of where +O(batch_size d r_what r_where n (r_what + r_where)) d is the number of TT-cores (what.ndims()); r_what is the largest TT-rank of what max(what.get_tt_rank()) r_where is the largest TT-rank of where n is the size of the axis dimension of what and where e.g. for a tensor of size 4 x 4 x 4, n is 4; for a 9 x 64 matrix of raw shape (3, 3, 3) x (4, 4, 4) n is 12 """ # Always work with batch of TT objects for simplicity. what = shapes.expand_batch_dim(what) if weights is not None: weights = tf.convert_to_tensor(weights, dtype=where.dtype) if not isinstance(where, TensorTrain): raise ValueError('The first argument should be a TensorTrain object, got ' '"%s".' % where) if where.get_raw_shape() != what.get_raw_shape(): raise ValueError('The shapes of the tensor we want to project and of the ' 'tensor on which tangent space we want to project should ' 'match, got %s and %s.' % (where.get_raw_shape(), what.get_raw_shape())) dtypes_compatible = (where.dtype.is_compatible_with(what.dtype) or what.dtype.is_compatible_with(where.dtype)) if not dtypes_compatible: raise ValueError('Dtypes of the arguments should coincide, got %s and %s.' % (where.dtype, what.dtype)) left_tangent_space_tens = decompositions.orthogonalize_tt_cores( where) right_tangent_space_tens = decompositions.orthogonalize_tt_cores( left_tangent_space_tens, left_to_right=False) ndims = where.ndims() dtype = where.dtype raw_shape = shapes.lazy_raw_shape(where) batch_size = shapes.lazy_batch_size(what) right_tangent_tt_ranks = shapes.lazy_tt_ranks(right_tangent_space_tens) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left_tangent_space_tens) # For einsum notation. mode_str = 'ij' if where.is_tt_matrix() else 'i' right_rank_dim = where.right_tt_rank_dim left_rank_dim = where.left_tt_rank_dim if weights is not None: weights_shape = weights.get_shape() output_is_batch = len(weights_shape) > 1 and weights_shape[1] > 1 else: output_is_batch = False output_batch_str = 'o' if output_is_batch else '' if output_is_batch: right_rank_dim += 1 left_rank_dim += 1 output_batch_size = weights.get_shape()[1].value # Prepare rhs vectors. # rhs[core_idx] is of size # batch_size x tensor_tt_ranks[core_idx] x tangent_tt_ranks[core_idx] rhs = [None] * (ndims + 1) rhs[ndims] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1, 0, -1): tens_core = what.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sa{0}b,sbd,c{0}d->sac'.format(mode_str) rhs[core_idx] = tf.einsum(einsum_str, tens_core, rhs[core_idx + 1], right_tang_core) # Prepare lhs vectors. # lhs[core_idx] is of size # batch_size x tangent_tt_ranks[core_idx] x tensor_tt_ranks[core_idx] lhs = [None] * (ndims + 1) lhs[0] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sab,a{0}c,sb{0}d->scd'.format(mode_str) lhs[core_idx + 1] = tf.einsum(einsum_str, lhs[core_idx], left_tang_core, tens_core) # Left to right sweep. res_cores_list = [] for core_idx in range(ndims): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] if core_idx < ndims - 1: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) einsum_str = 'a{0}b,sbc->sa{0}c'.format(mode_str) proj_core -= tf.einsum(einsum_str, left_tang_core, lhs[core_idx + 1]) if weights is None: einsum_str = 'sa{0}b,sbc->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, proj_core, rhs[core_idx + 1]) else: einsum_str = 'sa{0}b,sbc->sa{0}c'.format(mode_str, output_batch_str) proj_core_s = tf.einsum(einsum_str, proj_core, rhs[core_idx + 1]) einsum_str = 's{1},sa{0}c->{1}a{0}c'.format(mode_str, output_batch_str) proj_core = tf.einsum(einsum_str, weights, proj_core_s) if core_idx == ndims - 1: if weights is None: einsum_str = 'sab,sb{0}c->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) else: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str, output_batch_str) proj_core_s = tf.einsum(einsum_str, lhs[core_idx], tens_core) einsum_str = 's{1},sa{0}c->{1}a{0}c'.format(mode_str, output_batch_str) proj_core = tf.einsum(einsum_str, weights, proj_core_s) if output_is_batch: # Add batch dimension of size output_batch_size to left_tang_core and # right_tang_core extended_left_tang_core = tf.expand_dims(left_tang_core, 0) extended_right_tang_core = tf.expand_dims(right_tang_core, 0) if where.is_tt_matrix(): extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1, 1]) else: extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1]) else: extended_left_tang_core = left_tang_core extended_right_tang_core = right_tang_core if core_idx == 0: res_core = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) elif core_idx == ndims - 1: res_core = tf.concat((extended_right_tang_core, proj_core), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[core_idx] rank_2 = left_tangent_tt_ranks[core_idx + 1] if where.is_tt_matrix(): mode_size_n = raw_shape[0][core_idx] mode_size_m = raw_shape[1][core_idx] shape = [rank_1, mode_size_n, mode_size_m, rank_2] else: mode_size = raw_shape[0][core_idx] shape = [rank_1, mode_size, rank_2] if output_is_batch: shape = [output_batch_size] + shape zeros = tf.zeros(shape, dtype) upper = tf.concat((extended_right_tang_core, zeros), axis=right_rank_dim) lower = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) res_core = tf.concat((upper, lower), axis=left_rank_dim) res_cores_list.append(res_core) # TODO: TT-ranks. if output_is_batch: res = TensorTrainBatch(res_cores_list, where.get_raw_shape(), batch_size=output_batch_size) else: res = TensorTrain(res_cores_list, where.get_raw_shape()) res.projection_on = where return res def project(what, where): """Project `what` TTs on the tangent space of `where` TT. project(what, x) = P_x(what) project(batch_what, x) = batch(P_x(batch_what[0]), ..., P_x(batch_what[N])) This function implements the algorithm from the paper [1], theorem 3.1. [1] C. Lubich, I. Oseledets and B. Vandereycken, Time integration of Tensor Trains. Args: what: TensorTrain or TensorTrainBatch. In the case of batch returns batch with projection of each individual tensor. where: TensorTrain, TT-tensor or TT-matrix on which tangent space to project Returns: a TensorTrain with the TT-ranks equal 2 * tangent_space_tens.get_tt_ranks() Complexity: O(d r_where^3 m) for orthogonalizing the TT-cores of where +O(batch_size d r_what r_where n (r_what + r_where)) d is the number of TT-cores (what.ndims()); r_what is the largest TT-rank of what max(what.get_tt_rank()) r_where is the largest TT-rank of where n is the size of the axis dimension of what and where e.g. for a tensor of size 4 x 4 x 4, n is 4; for a 9 x 64 matrix of raw shape (3, 3, 3) x (4, 4, 4) n is 12 """ if not isinstance(where, TensorTrain): raise ValueError('The first argument should be a TensorTrain object, got ' '"%s".' % where) if where.get_raw_shape() != what.get_raw_shape(): raise ValueError('The shapes of the tensor we want to project and of the ' 'tensor on which tangent space we want to project should ' 'match, got %s and %s.' % (where.get_raw_shape(), what.get_raw_shape())) dtypes_compatible = (where.dtype.is_compatible_with(what.dtype) or what.dtype.is_compatible_with(where.dtype)) if not dtypes_compatible: raise ValueError('Dtypes of the arguments should coincide, got %s and %s.' % (where.dtype, what.dtype)) left_tangent_space_tens = decompositions.orthogonalize_tt_cores( where) right_tangent_space_tens = decompositions.orthogonalize_tt_cores( left_tangent_space_tens, left_to_right=False) ndims = where.ndims() dtype = where.dtype raw_shape = shapes.lazy_raw_shape(where) right_tangent_tt_ranks = shapes.lazy_tt_ranks(right_tangent_space_tens) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left_tangent_space_tens) # For einsum notation. mode_str = 'ij' if where.is_tt_matrix() else 'i' right_rank_dim = what.right_tt_rank_dim left_rank_dim = what.left_tt_rank_dim output_is_batch = isinstance(what, TensorTrainBatch) if output_is_batch: output_batch_size = what.batch_size # Always work with batch of TT objects for simplicity. what = shapes.expand_batch_dim(what) batch_size = shapes.lazy_batch_size(what) # Prepare rhs vectors. # rhs[core_idx] is of size # batch_size x tensor_tt_ranks[core_idx] x tangent_tt_ranks[core_idx] rhs = [None] * (ndims + 1) rhs[ndims] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1, 0, -1): tens_core = what.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sa{0}b,sbd,c{0}d->sac'.format(mode_str) rhs[core_idx] = tf.einsum(einsum_str, tens_core, rhs[core_idx + 1], right_tang_core) # Prepare lhs vectors. # lhs[core_idx] is of size # batch_size x tangent_tt_ranks[core_idx] x tensor_tt_ranks[core_idx] lhs = [None] * (ndims + 1) lhs[0] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sab,a{0}c,sb{0}d->scd'.format(mode_str) lhs[core_idx + 1] = tf.einsum(einsum_str, lhs[core_idx], left_tang_core, tens_core) # Left to right sweep. res_cores_list = [] for core_idx in range(ndims): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] if core_idx < ndims - 1: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) einsum_str = 'a{0}b,sbc->sa{0}c'.format(mode_str) proj_core -= tf.einsum(einsum_str, left_tang_core, lhs[core_idx + 1]) if output_is_batch: einsum_str = 'sa{0}b,sbc->sa{0}c'.format(mode_str) else: einsum_str = 'sa{0}b,sbc->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, proj_core, rhs[core_idx + 1]) if core_idx == ndims - 1: if output_is_batch: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str) else: einsum_str = 'sab,sb{0}c->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) if output_is_batch: # Add batch dimension of size output_batch_size to left_tang_core and # right_tang_core extended_left_tang_core = tf.expand_dims(left_tang_core, 0) extended_right_tang_core = tf.expand_dims(right_tang_core, 0) if where.is_tt_matrix(): extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1, 1]) else: extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1]) else: extended_left_tang_core = left_tang_core extended_right_tang_core = right_tang_core if core_idx == 0: res_core = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) elif core_idx == ndims - 1: res_core = tf.concat((extended_right_tang_core, proj_core), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[core_idx] rank_2 = left_tangent_tt_ranks[core_idx + 1] if where.is_tt_matrix(): mode_size_n = raw_shape[0][core_idx] mode_size_m = raw_shape[1][core_idx] shape = [rank_1, mode_size_n, mode_size_m, rank_2] else: mode_size = raw_shape[0][core_idx] shape = [rank_1, mode_size, rank_2] if output_is_batch: shape = [output_batch_size] + shape zeros = tf.zeros(shape, dtype) upper = tf.concat((extended_right_tang_core, zeros), axis=right_rank_dim) lower = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) res_core = tf.concat((upper, lower), axis=left_rank_dim) res_cores_list.append(res_core) # TODO: TT-ranks. if output_is_batch: res = TensorTrainBatch(res_cores_list, where.get_raw_shape(), batch_size=output_batch_size) else: res = TensorTrain(res_cores_list, where.get_raw_shape()) res.projection_on = where return res def project_matmul(what, where, matrix): """Project `matrix` * `what` TTs on the tangent space of `where` TT. project(what, x) = P_x(what) project(batch_what, x) = batch(P_x(batch_what[0]), ..., P_x(batch_what[N])) This function implements the algorithm from the paper [1], theorem 3.1. [1] C. Lubich, I. Oseledets and B. Vandereycken, Time integration of Tensor Trains. Args: what: TensorTrain or TensorTrainBatch. In the case of batch returns batch with projection of each individual tensor. where: TensorTrain, TT-tensor or TT-matrix on which tangent space to project matrix: TensorTrain, TT-matrix to multiply by what Returns: a TensorTrain with the TT-ranks equal 2 * tangent_space_tens.get_tt_ranks() Complexity: O(d r_where^3 m) for orthogonalizing the TT-cores of where +O(batch_size d R r_what r_where (n r_what + n m R + m r_where)) d is the number of TT-cores (what.ndims()); r_what is the largest TT-rank of what max(what.get_tt_rank()) r_where is the largest TT-rank of where matrix is of TT-rank R and of raw-shape (m, m, ..., m) x (n, n, ..., n). """ if not isinstance(where, TensorTrain): raise ValueError('The first argument should be a TensorTrain object, got ' '"%s".' % where) if where.get_raw_shape() != what.get_raw_shape(): raise ValueError('The shapes of the tensor we want to project and of the ' 'tensor on which tangent space we want to project should ' 'match, got %s and %s.' % (where.get_raw_shape(), what.get_raw_shape())) dtypes_compatible = (where.dtype.is_compatible_with(what.dtype) or what.dtype.is_compatible_with(where.dtype)) if not dtypes_compatible: raise ValueError('Dtypes of the arguments should coincide, got %s and %s.' % (where.dtype, what.dtype)) left_tangent_space_tens = decompositions.orthogonalize_tt_cores( where) right_tangent_space_tens = decompositions.orthogonalize_tt_cores( left_tangent_space_tens, left_to_right=False) ndims = where.ndims() dtype = where.dtype raw_shape = shapes.lazy_raw_shape(where) batch_size = shapes.lazy_batch_size(what) right_tangent_tt_ranks = shapes.lazy_tt_ranks(right_tangent_space_tens) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left_tangent_space_tens) # For einsum notation. right_rank_dim = what.right_tt_rank_dim left_rank_dim = what.left_tt_rank_dim output_is_batch = isinstance(what, TensorTrainBatch) if output_is_batch: output_batch_size = what.batch_size # Always work with batch of TT objects for simplicity. what = shapes.expand_batch_dim(what) # Prepare rhs vectors. # rhs[core_idx] is of size # batch_size x tensor_tt_ranks[core_idx] x matrix_tt_ranks[core_idx] x tangent_tt_ranks[core_idx] rhs = [None] * (ndims + 1) rhs[ndims] = tf.ones((batch_size, 1, 1, 1), dtype=dtype) for core_idx in range(ndims - 1, 0, -1): tens_core = what.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] matrix_core = matrix.tt_cores[core_idx] rhs[core_idx] = tf.einsum('bije,cikf,sdef,sajkd->sabc', matrix_core, right_tang_core, rhs[core_idx + 1], tens_core) # Prepare lhs vectors. # lhs[core_idx] is of size # batch_size x tangent_tt_ranks[core_idx] x matrix_tt_ranks[core_idx] x tensor_tt_ranks[core_idx] lhs = [None] * (ndims + 1) lhs[0] = tf.ones((batch_size, 1, 1, 1), dtype=dtype) for core_idx in range(ndims - 1): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] matrix_core = matrix.tt_cores[core_idx] # TODO: brutforce order of indices in lhs?? lhs[core_idx + 1] = tf.einsum('bije,aikd,sabc,scjkf->sdef', matrix_core, left_tang_core, lhs[core_idx], tens_core) # Left to right sweep. res_cores_list = [] for core_idx in range(ndims): tens_core = what.tt_cores[core_idx] matrix_core = matrix.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] if core_idx < ndims - 1: proj_core = tf.einsum('scjke,sabc,bijd->saikde', tens_core, lhs[core_idx], matrix_core) proj_core -= tf.einsum('aikb,sbcd->saikcd', left_tang_core, lhs[core_idx + 1]) proj_core = tf.einsum('saikcb,sbcd->saikd', proj_core, rhs[core_idx + 1]) if core_idx == ndims - 1: # d and e dimensions take 1 value, since its the last rank. # To make the result shape (?, ?, ?, 1), we are summing d and leaving e, # but we could have done the opposite -- sum e and leave d. proj_core = tf.einsum('sabc,bijd,scjke->saike', lhs[core_idx], matrix_core, tens_core) if output_is_batch: # Add batch dimension of size output_batch_size to left_tang_core and # right_tang_core extended_left_tang_core = tf.expand_dims(left_tang_core, 0) extended_right_tang_core = tf.expand_dims(right_tang_core, 0) extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1, 1]) else: extended_left_tang_core = left_tang_core extended_right_tang_core = right_tang_core if core_idx == 0: res_core = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) elif core_idx == ndims - 1: res_core = tf.concat((extended_right_tang_core, proj_core), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[core_idx] rank_2 = left_tangent_tt_ranks[core_idx + 1] mode_size_n = raw_shape[0][core_idx] mode_size_m = raw_shape[1][core_idx] shape = [rank_1, mode_size_n, mode_size_m, rank_2] if output_is_batch: shape = [output_batch_size] + shape zeros = tf.zeros(shape, dtype) upper = tf.concat((extended_right_tang_core, zeros), axis=right_rank_dim) lower = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) res_core = tf.concat((upper, lower), axis=left_rank_dim) res_cores_list.append(res_core) # TODO: TT-ranks. if output_is_batch: res = TensorTrainBatch(res_cores_list, where.get_raw_shape(), batch_size=output_batch_size) else: res = TensorTrain(res_cores_list, where.get_raw_shape()) res.projection_on = where return res def pairwise_flat_inner_projected(projected_tt_vectors_1, projected_tt_vectors_2): """Scalar products between two batches of TTs from the same tangent space. res[i, j] = t3f.flat_inner(projected_tt_vectors_1[i], projected_tt_vectors_1[j]). pairwise_flat_inner_projected(projected_tt_vectors_1, projected_tt_vectors_2) is equivalent to pairwise_flat_inner(projected_tt_vectors_1, projected_tt_vectors_2) , but works only on objects from the same tangent space and is much faster than general pairwise_flat_inner. Args: projected_tt_vectors_1: TensorTrainBatch of tensors projected on the same tangent space as projected_tt_vectors_2. projected_tt_vectors_2: TensorTrainBatch. Returns: tf.tensor with the scalar product matrix. Complexity: O(batch_size^2 d r^2 n), where d is the number of TT-cores (projected_tt_vectors_1.ndims()); r is the largest TT-rank max(projected_tt_vectors_1.get_tt_rank()) (i.e. 2 * {the TT-rank of the object we projected vectors onto}. and n is the size of the axis dimension, e.g. for a tensor of size 4 x 4 x 4, n is 4; for a 9 x 64 matrix of raw shape (3, 3, 3) x (4, 4, 4) n is 12. """ if not hasattr(projected_tt_vectors_1, 'projection_on') or \ not hasattr(projected_tt_vectors_2, 'projection_on'): raise ValueError('Both arguments should be projections on the tangent ' 'space of some other TT-object. All projection* functions ' 'leave .projection_on field in the resulting TT-object ' 'which is not present in the arguments you\'ve provided') if projected_tt_vectors_1.projection_on != projected_tt_vectors_2.projection_on: raise ValueError('Both arguments should be projections on the tangent ' 'space of the same TT-object. The provided arguments are ' 'projections on different TT-objects (%s and %s). Or at ' 'least the pointers are different.' % (projected_tt_vectors_1.projection_on, projected_tt_vectors_2.projection_on)) # Always work with batches of objects for simplicity. projected_tt_vectors_1 = shapes.expand_batch_dim(projected_tt_vectors_1) projected_tt_vectors_2 = shapes.expand_batch_dim(projected_tt_vectors_2) ndims = projected_tt_vectors_1.ndims() tt_ranks = shapes.lazy_tt_ranks(projected_tt_vectors_1) if projected_tt_vectors_1.is_tt_matrix(): right_size = tt_ranks[1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[0] curr_core_2 = projected_tt_vectors_2.tt_cores[0] curr_du_1 = curr_core_1[:, :, :, :, :right_size] curr_du_2 = curr_core_2[:, :, :, :, :right_size] res = tf.einsum('paijb,qaijb->pq', curr_du_1, curr_du_2) for core_idx in range(1, ndims): left_size = tt_ranks[core_idx] // 2 right_size = tt_ranks[core_idx + 1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[core_idx] curr_core_2 = projected_tt_vectors_2.tt_cores[core_idx] curr_du_1 = curr_core_1[:, left_size:, :, :, :right_size] curr_du_2 = curr_core_2[:, left_size:, :, :, :right_size] res += tf.einsum('paijb,qaijb->pq', curr_du_1, curr_du_2) left_size = tt_ranks[-2] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[-1] curr_core_2 = projected_tt_vectors_2.tt_cores[-1] curr_du_1 = curr_core_1[:, left_size:, :, :, :] curr_du_2 = curr_core_2[:, left_size:, :, :, :] res += tf.einsum('paijb,qaijb->pq', curr_du_1, curr_du_2) else: # Working with TT-tensor, not TT-matrix. right_size = tt_ranks[1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[0] curr_core_2 = projected_tt_vectors_2.tt_cores[0] curr_du_1 = curr_core_1[:, :, :, :right_size] curr_du_2 = curr_core_2[:, :, :, :right_size] res = tf.einsum('paib,qaib->pq', curr_du_1, curr_du_2) for core_idx in range(1, ndims): left_size = tt_ranks[core_idx] // 2 right_size = tt_ranks[core_idx + 1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[core_idx] curr_core_2 = projected_tt_vectors_2.tt_cores[core_idx] curr_du_1 = curr_core_1[:, left_size:, :, :right_size] curr_du_2 = curr_core_2[:, left_size:, :, :right_size] res += tf.einsum('paib,qaib->pq', curr_du_1, curr_du_2) left_size = tt_ranks[-2] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[-1] curr_core_2 = projected_tt_vectors_2.tt_cores[-1] curr_du_1 = curr_core_1[:, left_size:, :, :] curr_du_2 = curr_core_2[:, left_size:, :, :] res += tf.einsum('paib,qaib->pq', curr_du_1, curr_du_2) return res def add_n_projected(tt_objects, coef=None): """Adds all input TT-objects that are projections on the same tangent space. add_projected((a, b)) is equivalent add(a, b) for a and b that are from the same tangent space, but doesn't increase the TT-ranks. Args: tt_objects: a list of TT-objects that are projections on the same tangent space. coef: a list of numbers or anything else convertable to tf.Tensor. If provided, computes weighted sum. The size of this array should be len(tt_objects) x tt_objects[0].batch_size Returns: TT-objects representing the sum of the tt_objects (weighted sum if coef is provided). The TT-rank of the result equals to the TT-ranks of the arguments. """ for tt in tt_objects: if not hasattr(tt, 'projection_on'): raise ValueError('Both arguments should be projections on the tangent ' 'space of some other TT-object. All projection* functions ' 'leave .projection_on field in the resulting TT-object ' 'which is not present in the argument you\'ve provided.') projection_on = tt_objects[0].projection_on for tt in tt_objects[1:]: if tt.projection_on != projection_on: raise ValueError('All tt_objects should be projections on the tangent ' 'space of the same TT-object. The provided arguments are ' 'projections on different TT-objects (%s and %s). Or at ' 'least the pointers are different.' % (tt.projection_on, projection_on)) if coef is not None: coef = tf.convert_to_tensor(coef, dtype=tt_objects[0].dtype) if coef.get_shape().ndims > 1: # In batch case we will need to multiply each core by this coefficients # along the first axis. To do it need to reshape the coefs to match # the TT-cores number of dimensions. some_core = tt_objects[0].tt_cores[0] dim_array = [1] * (some_core.get_shape().ndims + 1) dim_array[0] = coef.get_shape()[0].value dim_array[1] = coef.get_shape()[1].value coef = tf.reshape(coef, dim_array) ndims = tt_objects[0].ndims() tt_ranks = shapes.lazy_tt_ranks(tt_objects[0]) left_rank_dim = tt_objects[0].left_tt_rank_dim right_rank_dim = tt_objects[0].right_tt_rank_dim res_cores = [] def slice_tt_core(tt_core, left_idx, right_idx): num_tt_core_dims = len(tt_core.get_shape()) idx = [slice(None)] * num_tt_core_dims idx[left_rank_dim] = left_idx idx[right_rank_dim] = right_idx return tt_core[idx] right_half_rank = tt_ranks[1] // 2 left_chunks = [] for obj_idx, tt in enumerate(tt_objects): curr_core = slice_tt_core(tt.tt_cores[0], slice(None), slice(0, right_half_rank)) if coef is not None: curr_core *= coef[obj_idx] left_chunks.append(curr_core) left_part = tf.add_n(left_chunks) first_obj_core = tt_objects[0].tt_cores[0] right_part = slice_tt_core(first_obj_core, slice(None), slice(right_half_rank, None)) first_core = tf.concat((left_part, right_part), axis=right_rank_dim) res_cores.append(first_core) for core_idx in range(1, ndims - 1): first_obj_core = tt_objects[0].tt_cores[core_idx] left_half_rank = tt_ranks[core_idx] // 2 right_half_rank = tt_ranks[core_idx + 1] // 2 upper_part = slice_tt_core(tt.tt_cores[core_idx], slice(0, left_half_rank), slice(None)) lower_right_part = slice_tt_core(first_obj_core, slice(left_half_rank, None), slice(right_half_rank, None)) lower_left_chunks = [] for obj_idx, tt in enumerate(tt_objects): curr_core = slice_tt_core(tt.tt_cores[core_idx], slice(left_half_rank, None), slice(0, right_half_rank)) if coef is not None: curr_core *= coef[obj_idx] lower_left_chunks.append(curr_core) lower_left_part = tf.add_n(lower_left_chunks) lower_part = tf.concat((lower_left_part, lower_right_part), axis=right_rank_dim) curr_core = tf.concat((upper_part, lower_part), axis=left_rank_dim) res_cores.append(curr_core) left_half_rank = tt_ranks[ndims - 1] // 2 upper_part = slice_tt_core(tt.tt_cores[-1], slice(0, left_half_rank), slice(None)) lower_chunks = [] for obj_idx, tt in enumerate(tt_objects): curr_core = slice_tt_core(tt.tt_cores[-1], slice(left_half_rank, None), slice(None)) if coef is not None: curr_core *= coef[obj_idx] lower_chunks.append(curr_core) lower_part = tf.add_n(lower_chunks) last_core = tf.concat((upper_part, lower_part), axis=left_rank_dim) res_cores.append(last_core) raw_shape = tt_objects[0].get_raw_shape() static_tt_ranks = tt_objects[0].get_tt_ranks() if isinstance(tt_objects[0], TensorTrain): res = TensorTrain(res_cores, raw_shape, static_tt_ranks) elif isinstance(tt_objects[0], TensorTrainBatch): res = TensorTrainBatch(res_cores, raw_shape, static_tt_ranks, tt_objects[0].batch_size) # Maintain the projection_on property. res.projection_on = tt_objects[0].projection_on return res def tangent_space_to_deltas(tt, name='t3f_tangent_space_to_deltas'): """Convert an element of the tangent space to deltas representation. Tangent space elements (outputs of t3f.project) look like: dP1 V2 ... Vd + U1 dP2 V3 ... Vd + ... + U1 ... Ud-1 dPd. This function takes as input an element of the tangent space and converts it to the list of deltas [dP1, ..., dPd]. Args: tt: `TensorTrain` or `TensorTrainBatch` that is a result of t3f.project, t3f.project_matmul, or other similar functions. name: string, name of the Op. Returns: A list of delta-cores (tf.Tensors). """ if not hasattr(tt, 'projection_on') or tt.projection_on is None: raise ValueError('tt argument is supposed to be a projection, but it ' 'lacks projection_on field') num_dims = tt.ndims() left_tt_rank_dim = tt.left_tt_rank_dim right_tt_rank_dim = tt.right_tt_rank_dim deltas = [None] * num_dims tt_ranks = shapes.lazy_tt_ranks(tt) for i in range(1, num_dims - 1): if int(tt_ranks[i] / 2) != tt_ranks[i] / 2: raise ValueError('tt argument is supposed to be a projection, but its ' 'ranks are not even.') with tf.compat.v1.name_scope(name, values=tt.tt_cores): for i in range(1, num_dims - 1): r1, r2 = tt_ranks[i], tt_ranks[i + 1] curr_core = tt.tt_cores[i] slc = [slice(None)] * len(curr_core.shape) slc[left_tt_rank_dim] = slice(int(r1 / 2), None) slc[right_tt_rank_dim] = slice(0, int(r2 / 2)) deltas[i] = curr_core[slc] slc = [slice(None)] * len(tt.tt_cores[0].shape) slc[right_tt_rank_dim] = slice(0, int(tt_ranks[1] / 2)) deltas[0] = tt.tt_cores[0][slc] slc = [slice(None)] * len(tt.tt_cores[0].shape) slc[left_tt_rank_dim] = slice(int(tt_ranks[-2] / 2), None) deltas[num_dims - 1] = tt.tt_cores[num_dims - 1][slc] return deltas def deltas_to_tangent_space(deltas, tt, left=None, right=None, name='t3f_deltas_to_tangent_space'): """Converts deltas representation of tangent space vector to TT object. Takes as input a list of [dP1, ..., dPd] and returns dP1 V2 ... Vd + U1 dP2 V3 ... Vd + ... + U1 ... Ud-1 dPd. This function is hard to use correctly because deltas should abey the so called gauge conditions. If the don't, the function will silently return incorrect result. This is why this function is not imported in __init__. Args: deltas: a list of deltas (essentially TT-cores) obeying the gauge conditions. tt: `TensorTrain` object on which the tangent space tensor represented by delta is projected. left: t3f.orthogonilize_tt_cores(tt). If you have it already compute, you may pass it as argument to avoid recomputing. right: t3f.orthogonilize_tt_cores(left, left_to_right=False). If you have it already compute, you may pass it as argument to avoid recomputing. name: string, name of the Op. Returns: `TensorTrain` object constructed from deltas, that is from the tangent space at point `tt`. """ cores = [] dtype = tt.dtype num_dims = tt.ndims() # TODO: add cache instead of mannually pasisng precomputed stuff? input_tensors = list(tt.tt_cores) + list(deltas) if left is not None: input_tensors += list(left.tt_cores) if right is not None: input_tensors += list(right.tt_cores) with tf.compat.v1.name_scope(name, values=input_tensors): if left is None: left = decompositions.orthogonalize_tt_cores(tt) if right is None: right = decompositions.orthogonalize_tt_cores(left, left_to_right=False) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left) right_tangent_tt_ranks = shapes.lazy_tt_ranks(left) raw_shape = shapes.lazy_raw_shape(left) right_rank_dim = left.right_tt_rank_dim left_rank_dim = left.left_tt_rank_dim is_batch_case = len(deltas[0].shape) > len(tt.tt_cores[0].shape) if is_batch_case: right_rank_dim += 1 left_rank_dim += 1 batch_size = deltas[0].shape.as_list()[0] for i in range(num_dims): left_tt_core = left.tt_cores[i] right_tt_core = right.tt_cores[i] if is_batch_case: tile = [1] * len(left_tt_core.shape) tile = [batch_size] + tile left_tt_core = tf.tile(left_tt_core[None, ...], tile) right_tt_core = tf.tile(right_tt_core[None, ...], tile) if i == 0: tangent_core = tf.concat((deltas[i], left_tt_core), axis=right_rank_dim) elif i == num_dims - 1: tangent_core = tf.concat((right_tt_core, deltas[i]), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[i] rank_2 = left_tangent_tt_ranks[i + 1] if tt.is_tt_matrix(): mode_size_n = raw_shape[0][i] mode_size_m = raw_shape[1][i] shape = [rank_1, mode_size_n, mode_size_m, rank_2] else: mode_size_n = raw_shape[0][i] shape = [rank_1, mode_size_n, rank_2] if is_batch_case: shape = [batch_size] + shape zeros = tf.zeros(shape, dtype=dtype) upper = tf.concat((right_tt_core, zeros), axis=right_rank_dim) lower = tf.concat((deltas[i], left_tt_core), axis=right_rank_dim) tangent_core = tf.concat((upper, lower), axis=left_rank_dim) cores.append(tangent_core) if is_batch_case: tangent = TensorTrainBatch(cores, batch_size=batch_size) else: tangent = TensorTrain(cores) tangent.projection_on = tt return tangent
38,104
14,392
from django.contrib import admin from . import models @admin.register(models.Username) class UsernameAdmin(admin.ModelAdmin): list_display = ('user', 'name', 'is_mod') readonly_fields = ('normalized_name',) search_fields = ('user__email', 'name')
262
81
import logging from iptv_proxy.providers.beast.constants import BeastConstants from iptv_proxy.providers.iptv_provider.json_api import ProviderConfigurationJSONAPI logger = logging.getLogger(__name__) class BeastConfigurationJSONAPI(ProviderConfigurationJSONAPI): __slots__ = [] _provider_name = BeastConstants.PROVIDER_NAME.lower()
346
105
#!/usr/bin/env python # Copyright 2015 Wieger Wesselink. # Distributed under the Boost Software License, Version 1.0. # (See accompanying file LICENSE_1_0.txt or http://www.boost.org/LICENSE_1_0.txt) import os import os.path import random import re import sys import traceback sys.path += [os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'python'))] import random_state_formula_generator from random_bes_generator import make_bes from random_pbes_generator import make_pbes import random_process_expression from testing import YmlTest from text_utility import write_text MCRL2_ROOT = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', '..')) MCRL2_INSTALL_DIR = os.path.join(MCRL2_ROOT, 'install', 'bin') def ymlfile(file): return '{}/tests/specifications/{}.yml'.format(MCRL2_ROOT, file) def mcrl2file(file): return os.path.join(MCRL2_ROOT, file) class RandomTest(YmlTest): def __init__(self, name, ymlfile, settings): super(RandomTest, self).__init__(name, ymlfile, [], settings) # create input files for the random test, and add the filenames to self.inputfiles def create_inputfiles(self, runpath = '.'): raise NotImplementedError # removes input files that are in the runpath directory def remove_inputfiles(self, runpath = '.'): for filename in self.inputfiles: if os.path.abspath(runpath) == os.path.abspath(os.path.dirname(filename)): os.remove(filename) def execute(self, runpath = '.'): self.create_inputfiles(runpath) super(RandomTest, self).execute(runpath) self.remove_inputfiles(runpath) class ProcessTest(RandomTest): def __init__(self, name, ymlfile, settings): super(ProcessTest, self).__init__(name, ymlfile, settings) self.actions = ['a', 'b', 'c', 'd'] self.process_identifiers = ['P', 'Q', 'R'] self.process_size = 13 self.parallel_operator_generators = random_process_expression.default_parallel_operator_generators self.process_expression_generators = random_process_expression.default_process_expression_generators self.init = None self.generate_process_parameters = False def create_inputfiles(self, runpath = '.'): filename = '{0}.mcrl2'.format(self.name, self.settings) p = random_process_expression.make_process_specification(self.parallel_operator_generators, self.process_expression_generators, self.actions, self.process_identifiers, self.process_size, self.init, self.generate_process_parameters) write_text(filename, str(p)) self.inputfiles += [filename] # generates stochastic random processes class StochasticProcessTest(ProcessTest): def __init__(self, name, ymlfile, settings): super(StochasticProcessTest, self).__init__(name, ymlfile, settings) self.process_expression_generators = { random_process_expression.make_action : 8, random_process_expression.make_delta : 1, random_process_expression.make_tau : 1, random_process_expression.make_process_instance: 2, random_process_expression.make_sum : 2, random_process_expression.make_if_then : 2, random_process_expression.make_if_then_else : 2, random_process_expression.make_choice : 5, random_process_expression.make_seq : 5, random_process_expression.make_multi_action : 1, random_process_expression.make_dist : 3, } # generates random process with higher probability of tau transitions class ProcessTauTest(ProcessTest): def __init__(self, name, testfile, settings): super(ProcessTauTest, self).__init__(name, testfile, settings) self.actions = ['a', 'b', 'c'] self.init = 'hide({a}, allow({a, b, c}, P || Q || R))' self.process_expression_generators = { random_process_expression.make_action: 8, random_process_expression.make_delta: 1, random_process_expression.make_tau: 4, random_process_expression.make_process_instance: 1, random_process_expression.make_sum: 0, random_process_expression.make_if_then: 0, random_process_expression.make_if_then_else: 0, random_process_expression.make_choice: 5, random_process_expression.make_seq: 5, random_process_expression.make_multi_action: 1, random_process_expression.make_dist: 0, } class AlphabetReduceTest(ProcessTest): def __init__(self, name, settings): super(AlphabetReduceTest, self).__init__(name, ymlfile('alphabet-reduce'), settings) self.actions = ['a', 'b', 'c', 'd', 'e'] class LpsSuminstTest(ProcessTest): def __init__(self, name, settings): super(LpsSuminstTest, self).__init__(name, ymlfile('lpssuminst'), settings) class LpsSumelmTest(ProcessTest): def __init__(self, name, settings): super(LpsSumelmTest, self).__init__(name, ymlfile('lpssumelm'), settings) class LpsParelmTest(ProcessTest): def __init__(self, name, settings): super(LpsParelmTest, self).__init__(name, ymlfile('lpsparelm'), settings) self.generate_process_parameters = True class LpsOnePointRuleRewriteTest(ProcessTest): def __init__(self, name, settings): super(LpsOnePointRuleRewriteTest, self).__init__(name, ymlfile('lpstransform'), settings) self.add_command_line_options('t2', ['-alps-one-point-rule-rewriter']) class LpsConfcheckTest(ProcessTauTest): def __init__(self, name, confluence_type, settings): self.option_map = { 'commutative' : 'C', 'commutative-disjoint' : 'c', 'disjoint' : 'd', 'triangular' : 'T', 'trivial' : 'Z' } assert confluence_type in self.option_map super(LpsConfcheckTest, self).__init__(name, ymlfile('lpsconfcheck'), settings) self.add_command_line_options('t2', ['-x' + self.option_map[confluence_type]]) class LtscompareTest(ProcessTauTest): def __init__(self, name, equivalence_type, settings): assert equivalence_type in ['bisim', 'bisim-gv', 'bisim-gjkw', 'branching-bisim', 'branching-bisim-gv', 'branching-bisim-gjkw', 'dpbranching-bisim', 'dpbranching-bisim-gv', 'dpbranching-bisim-gjkw', 'weak-bisim', 'dpweak-bisim', 'sim', 'ready-sim' , 'trace', 'weak-trace'] super(LtscompareTest, self).__init__(name, ymlfile('ltscompare'), settings) self.add_command_line_options('t3', ['-e' + equivalence_type]) self.add_command_line_options('t4', ['-e' + equivalence_type]) class StochasticLtscompareTest(StochasticProcessTest): def __init__(self, name, settings): super(StochasticLtscompareTest, self).__init__(name, ymlfile('stochastic-ltscompare'), settings) class BisimulationTest(ProcessTauTest): def __init__(self, name, equivalence_type, settings): assert equivalence_type in ['bisim', 'bisim-gv', 'bisim-gjkw', 'branching-bisim', 'branching-bisim-gv', 'branching-bisim-gjkw', 'weak-bisim'] super(BisimulationTest, self).__init__(name, ymlfile('bisimulation'), settings) self.add_command_line_options('t3', ['-e' + equivalence_type]) self.add_command_line_options('t4', ['-e' + equivalence_type]) if equivalence_type in ['branching-bisim-gv', 'branching-bisim-gjkw']: self.add_command_line_options('t7', ['-bbranching-bisim']) elif equivalence_type in ['bisim', 'bisim-gv', 'bisim-gjkw']: self.add_command_line_options('t7', ['-bstrong-bisim']) else: self.add_command_line_options('t7', ['-b' + equivalence_type]) class Lps2ltsAlgorithmsTest(ProcessTauTest): def __init__(self, name, settings): super(Lps2ltsAlgorithmsTest, self).__init__(name, ymlfile('lps2lts-algorithms'), settings) # randomly choose an algorithm actions = random.choice(['a', 'a,b', 'a,b,c']) options = [random.choice(['--deadlock', '--divergence', '--nondeterminism', '--action={}'.format(actions)])] options = [random.choice(['--deadlock', '--nondeterminism', '--action={}'.format(actions)])] if 'divergence' in options[0]: tau_actions = random.choice(['', '', 'b', 'b,c']) if tau_actions: options.append('--tau={}'.format(tau_actions)) self.add_command_line_options('t2', options) self.add_command_line_options('t3', options) class LpsConstelmTest(ProcessTest): def __init__(self, name, settings): super(LpsConstelmTest, self).__init__(name, ymlfile('lpsconstelm'), settings) self.generate_process_parameters = True class LpsBinaryTest(ProcessTest): def __init__(self, name, settings): super(LpsBinaryTest, self).__init__(name, ymlfile('lpsbinary'), settings) self.generate_process_parameters = True class LpsstategraphTest(ProcessTest): def __init__(self, name, settings): super(LpsstategraphTest, self).__init__(name, ymlfile('lpsstategraph'), settings) self.generate_process_parameters = True class Lps2pbesTest(ProcessTest): def __init__(self, name, settings): super(Lps2pbesTest, self).__init__(name, ymlfile('lps2pbes'), settings) def create_inputfiles(self, runpath = '.'): super(Lps2pbesTest, self).create_inputfiles(runpath) self.inputfiles.append(mcrl2file('examples/modal-formulas/nodeadlock.mcf')) class Lts2pbesTest(ProcessTest): def __init__(self, name, settings): super(Lts2pbesTest, self).__init__(name, ymlfile('lts2pbes'), settings) def create_inputfiles(self, runpath = '.'): super(Lts2pbesTest, self).create_inputfiles(runpath) self.inputfiles.append(mcrl2file('examples/modal-formulas/nodeadlock.mcf')) class PbesTest(RandomTest): def __init__(self, name, ymlfile, settings): super(PbesTest, self).__init__(name, ymlfile, settings) self.equation_count = 4 self.atom_count = 4 self.propvar_count = 3 self.use_quantifiers = True self.use_integers = True def create_inputfiles(self, runpath = '.'): filename = '{0}.txt'.format(self.name) p = make_pbes(self.equation_count, self.atom_count, self.propvar_count, self.use_quantifiers, use_integers=self.use_integers) write_text(filename, str(p)) self.inputfiles += [filename] # N.B. does not work yet due to unusable abstraction map class PbesabsintheTest(PbesTest): def __init__(self, name, settings): super(PbesabsintheTest, self).__init__(name, ymlfile('pbesabsinthe'), settings) # N.B. This test has been disabled, since the tool has been deprecated. class PbesabstractTest(PbesTest): def __init__(self, name, settings): super(PbesabstractTest, self).__init__(name, ymlfile('pbesabstract'), settings) class PbesbddsolveTest(PbesTest): def __init__(self, name, settings): super(PbesbddsolveTest, self).__init__(name, ymlfile('pbesbddsolve'), settings) self.use_integers = False self.use_quantifiers = False class PbesconstelmTest(PbesTest): def __init__(self, name, settings): super(PbesconstelmTest, self).__init__(name, ymlfile('pbesconstelm'), settings) class PbesparelmTest(PbesTest): def __init__(self, name, settings): super(PbesparelmTest, self).__init__(name, ymlfile('pbesparelm'), settings) class PbespareqelmTest(PbesTest): def __init__(self, name, settings): super(PbespareqelmTest, self).__init__(name, ymlfile('pbespareqelm'), settings) class Pbespor1Test(PbesTest): def __init__(self, name, settings): super(Pbespor1Test, self).__init__(name, ymlfile('pbespor1'), settings) class Pbespor2Test(ProcessTest): def __init__(self, name, settings): super(Pbespor2Test, self).__init__(name, ymlfile('pbespor2'), settings) def create_inputfiles(self, runpath = '.'): super(Pbespor2Test, self).create_inputfiles(runpath) filename = '{0}.mcf'.format(self.name, self.settings) formula = random_state_formula_generator.make_modal_formula() write_text(filename, str(formula)) self.inputfiles += [filename] class PbesrewrTest(PbesTest): def __init__(self, name, rewriter, settings): super(PbesrewrTest, self).__init__(name, ymlfile('pbesrewr'), settings) self.add_command_line_options('t2', ['-p' + rewriter]) class PbestransformTest(PbesTest): def __init__(self, name, rewriter, settings): super(PbestransformTest, self).__init__(name, ymlfile('pbestransform'), settings) self.add_command_line_options('t2', ['-a' + rewriter]) class PbesinstTest(PbesTest): def __init__(self, name, options, settings): super(PbesinstTest, self).__init__(name, ymlfile('pbesinst'), settings) self.add_command_line_options('t2', options) class PbespgsolveTest(PbesTest): def __init__(self, name, settings): super(PbespgsolveTest, self).__init__(name, ymlfile('pbespgsolve'), settings) class PbesstategraphTest(PbesTest): def __init__(self, name, settings): super(PbesstategraphTest, self).__init__(name, ymlfile('pbesstategraph'), settings) class PbessymbolicbisimTest(PbesTest): def __init__(self, name, settings): super(PbessymbolicbisimTest, self).__init__(name, ymlfile('pbessymbolicbisim'), settings) class PbessolvesymbolicTest(PbesTest): def __init__(self, name, settings): super(PbessolvesymbolicTest, self).__init__(name, ymlfile('pbessolvesymbolic'), settings) class Pbes2boolTest(PbesTest): def __init__(self, name, settings): super(Pbes2boolTest, self).__init__(name, ymlfile('pbessolve'), settings) class Pbes2boolDepthFirstTest(PbesTest): def __init__(self, name, settings): super(Pbes2boolDepthFirstTest, self).__init__(name, ymlfile('pbessolve'), settings) self.add_command_line_options('t2', ['-zdepth-first']) self.add_command_line_options('t3', ['-zdepth-first']) self.add_command_line_options('t4', ['-zdepth-first']) self.add_command_line_options('t5', ['-zdepth-first']) self.add_command_line_options('t6', ['-zdepth-first']) self.add_command_line_options('t7', ['-zdepth-first']) self.add_command_line_options('t8', ['-zdepth-first']) class Pbes2bool_counter_exampleTest(ProcessTest): def __init__(self, name, optimization, settings): super(Pbes2bool_counter_exampleTest, self).__init__(name, ymlfile('pbessolve-counter-example'), settings) if optimization in [4, 5]: self.add_command_line_options('t3', ['-l{}'.format(optimization), '--aggressive', '--prune-todo-list']) else: self.add_command_line_options('t3', ['-l{}'.format(optimization), '--prune-todo-list']) def create_inputfiles(self, runpath = '.'): super(Pbes2bool_counter_exampleTest, self).create_inputfiles(runpath) filename = '{0}.mcf'.format(self.name, self.settings) formula = random_state_formula_generator.make_modal_formula() write_text(filename, str(formula)) self.inputfiles += [filename] class Pbes_unify_parametersTest(PbesTest): def __init__(self, name, settings): super(Pbes_unify_parametersTest, self).__init__(name, ymlfile('pbes-unify-parameters'), settings) class Pbes_srfTest(PbesTest): def __init__(self, name, settings): super(Pbes_srfTest, self).__init__(name, ymlfile('pbes-srf'), settings) # N.B does not work due to unknown expressions (F_or) class SymbolicExplorationTest(PbesTest): def __init__(self, name, settings): super(SymbolicExplorationTest, self).__init__(name, ymlfile('symbolic_exploration'), settings) class BesTest(RandomTest): def __init__(self, name, ymlfile, settings): super(BesTest, self).__init__(name, ymlfile, settings) self.equation_count = 4 self.term_size = 3 def create_inputfiles(self, runpath = '.'): filename = '{0}.txt'.format(self.name, self.settings) p = make_bes(self.equation_count, self.term_size) write_text(filename, str(p)) self.inputfiles += [filename] class BessolveTest(BesTest): def __init__(self, name, settings): super(BessolveTest, self).__init__(name, ymlfile('bessolve'), settings) available_tests = { 'alphabet-reduce' : lambda name, settings: AlphabetReduceTest(name, settings) , 'lpssuminst' : lambda name, settings: LpsSuminstTest(name, settings) , 'lpssumelm' : lambda name, settings: LpsSumelmTest(name, settings) , 'lpsparelm' : lambda name, settings: LpsParelmTest(name, settings) , 'lps-quantifier-one-point' : lambda name, settings: LpsOnePointRuleRewriteTest(name, settings) , 'lpsconfcheck-commutative' : lambda name, settings: LpsConfcheckTest(name, 'commutative', settings) , 'lpsconfcheck-commutative-disjoint' : lambda name, settings: LpsConfcheckTest(name, 'commutative-disjoint', settings) , 'lpsconfcheck-disjoint' : lambda name, settings: LpsConfcheckTest(name, 'disjoint', settings) , 'lpsconfcheck-triangular' : lambda name, settings: LpsConfcheckTest(name, 'triangular', settings) , 'lpsconfcheck-trivial' : lambda name, settings: LpsConfcheckTest(name, 'trivial', settings) , 'lpsconstelm' : lambda name, settings: LpsConstelmTest(name, settings) , 'lpsbinary' : lambda name, settings: LpsBinaryTest(name, settings) , 'lps2lts-algorithms' : lambda name, settings: Lps2ltsAlgorithmsTest(name, settings) , 'lps2pbes' : lambda name, settings: Lps2pbesTest(name, settings) , 'lpsstategraph' : lambda name, settings: LpsstategraphTest(name, settings) , 'lts2pbes' : lambda name, settings: Lts2pbesTest(name, settings) , 'ltscompare-bisim' : lambda name, settings: LtscompareTest(name, 'bisim', settings) , 'ltscompare-bisim-gv' : lambda name, settings: LtscompareTest(name, 'bisim-gv', settings) , 'ltscompare-bisim-gjkw' : lambda name, settings: LtscompareTest(name, 'bisim-gjkw', settings) , 'ltscompare-branching-bisim' : lambda name, settings: LtscompareTest(name, 'branching-bisim', settings) , 'ltscompare-branching-bisim-gv' : lambda name, settings: LtscompareTest(name, 'branching-bisim-gv', settings) , 'ltscompare-branching-bisim-gjkw' : lambda name, settings: LtscompareTest(name, 'branching-bisim-gjkw', settings) , 'ltscompare-dpbranching-bisim' : lambda name, settings: LtscompareTest(name, 'dpbranching-bisim', settings) , 'ltscompare-dpbranching-bisim-gv' : lambda name, settings: LtscompareTest(name, 'dpbranching-bisim-gv', settings) , 'ltscompare-dpbranching-bisim-gjkw' : lambda name, settings: LtscompareTest(name, 'dpbranching-bisim-gjkw', settings) , 'ltscompare-weak-bisim' : lambda name, settings: LtscompareTest(name, 'weak-bisim', settings) , 'ltscompare-dpweak-bisim' : lambda name, settings: LtscompareTest(name, 'dpweak-bisim', settings) , 'ltscompare-sim' : lambda name, settings: LtscompareTest(name, 'sim', settings) , 'ltscompare-ready-sim' : lambda name, settings: LtscompareTest(name, 'ready-sim', settings) , 'ltscompare-trace' : lambda name, settings: LtscompareTest(name, 'trace', settings) , 'ltscompare-weak-trace' : lambda name, settings: LtscompareTest(name, 'weak-trace', settings) , 'bisimulation-bisim' : lambda name, settings: BisimulationTest(name, 'bisim', settings) , 'bisimulation-bisim-gv' : lambda name, settings: BisimulationTest(name, 'bisim-gv', settings) , 'bisimulation-bisim-gjkw' : lambda name, settings: BisimulationTest(name, 'bisim-gjkw', settings) , 'bisimulation-branching-bisim' : lambda name, settings: BisimulationTest(name, 'branching-bisim', settings) , 'bisimulation-branching-bisim-gv' : lambda name, settings: BisimulationTest(name, 'branching-bisim-gv', settings) , 'bisimulation-branching-bisim-gjkw' : lambda name, settings: BisimulationTest(name, 'branching-bisim-gjkw', settings) , 'bisimulation-weak-bisim' : lambda name, settings: BisimulationTest(name, 'weak-bisim', settings) , 'pbesconstelm' : lambda name, settings: PbesconstelmTest(name, settings) , 'pbesparelm' : lambda name, settings: PbesparelmTest(name, settings) , 'pbespareqelm' : lambda name, settings: PbespareqelmTest(name, settings) , 'pbespor2' : lambda name, settings: Pbespor2Test(name, settings) , 'pbesrewr-simplify' : lambda name, settings: PbesrewrTest(name, 'simplify', settings) , 'pbesrewr-pfnf' : lambda name, settings: PbesrewrTest(name, 'pfnf', settings) , 'pbesrewr-quantifier-all' : lambda name, settings: PbesrewrTest(name, 'quantifier-all', settings) , 'pbesrewr-quantifier-finite' : lambda name, settings: PbesrewrTest(name, 'quantifier-finite', settings) , 'pbesrewr-quantifier-inside' : lambda name, settings: PbesrewrTest(name, 'quantifier-inside', settings) , 'pbesrewr-quantifier-one-point' : lambda name, settings: PbesrewrTest(name, 'quantifier-one-point', settings) , 'pbesrewr-data-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-data-rewriter', settings) , 'pbesrewr-simplify-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-rewriter', settings) , 'pbesrewr-simplify-data-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-data-rewriter', settings) , 'pbesrewr-simplify-quantifiers-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-quantifiers-rewriter', settings) , 'pbesrewr-simplify-quantifiers-data-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-quantifiers-data-rewriter', settings), 'pbesinst-lazy' : lambda name, settings: PbesinstTest(name, ['-slazy'], settings) , 'pbesinst-alternative_lazy' : lambda name, settings: PbesinstTest(name, ['-salternative-lazy'], settings) , 'pbesinst-finite' : lambda name, settings: PbesinstTest(name, ['-sfinite', '-f*(*:Bool)'], settings) , 'pbespgsolve' : lambda name, settings: PbespgsolveTest(name, settings) , 'pbessolve' : lambda name, settings: Pbes2boolTest(name, settings) , 'pbessolve-depth-first' : lambda name, settings: Pbes2boolDepthFirstTest(name, settings) , 'pbessolve-counter-example-optimization-0' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 0, settings) , 'pbessolve-counter-example-optimization-1' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 1, settings) , 'pbessolve-counter-example-optimization-2' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 2, settings) , 'pbessolve-counter-example-optimization-3' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 3, settings) , 'pbessolve-counter-example-optimization-4' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 4, settings) , 'pbessolve-counter-example-optimization-5' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 5, settings) , 'pbessolve-counter-example-optimization-6' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 6, settings) , 'pbessolve-counter-example-optimization-7' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 7, settings) , 'pbesstategraph' : lambda name, settings: PbesstategraphTest(name, settings) , 'pbes-unify-parameters' : lambda name, settings: Pbes_unify_parametersTest(name, settings) , 'pbes-srf' : lambda name, settings: Pbes_srfTest(name, settings) , # 'pbessymbolicbisim' : lambda name, settings: PbessymbolicbisimTest(name, settings) , # excluded from the tests because of Z3 dependency 'bessolve' : lambda name, settings: BessolveTest(name, settings) , #'stochastic-ltscompare' : lambda name, settings: StochasticLtscompareTest(name, settings) , } # These test do not work on Windows due to dependencies. if os.name != 'nt': available_tests.update({'pbessolvesymbolic' : lambda name, settings: PbessolvesymbolicTest(name, settings) }) # available_tests.update({ 'pbesbddsolve' : lambda name, settings: PbesbddsolveTest(name, settings) }) def print_names(tests): for name in sorted(tests): print(name) # Return all tests that match with pattern. In case of an exact match, only this exact match is returned. def matching_tests(tests, pattern): matches = [name for name in sorted(tests) if re.search(pattern, name)] if pattern in matches: return [pattern] return matches def main(tests): import argparse cmdline_parser = argparse.ArgumentParser() cmdline_parser.add_argument('-t', '--toolpath', dest='toolpath', help='The path where the mCRL2 tools are installed') cmdline_parser.add_argument('-r', '--repetitions', dest='repetitions', metavar='N', default='10', help='Perform N repetitions of each test') cmdline_parser.add_argument('-v', '--verbose', dest='verbose', action='store_true', help='Display additional progress messages.') cmdline_parser.add_argument('-k', '--keep-files', dest='keep_files', action='store_true', help='Keep the files produced by the test') cmdline_parser.add_argument('-n', '--names', dest='names', action='store_true', help='Print the names of the available tests') cmdline_parser.add_argument('-p', '--pattern', dest='pattern', metavar='P', default='.', action='store', help='Run the tests that match with pattern P') cmdline_parser.add_argument('-o', '--output', dest='output', metavar='o', action='store', help='Run the tests in the given directory') args = cmdline_parser.parse_args() if args.names: print_names(tests) return toolpath = args.toolpath if not toolpath: toolpath = MCRL2_INSTALL_DIR settings = {'toolpath': toolpath, 'verbose': args.verbose, 'cleanup_files': not args.keep_files, 'allow-non-zero-return-values': True} I = range(int(args.repetitions)) if args.output: if not os.path.exists(args.output): os.mkdir(args.output) os.chdir(args.output) test_failed = False for name in matching_tests(tests, args.pattern): try: for i in I: test = tests[name]('{}_{}'.format(name, i), settings) test.execute_in_sandbox() except Exception as e: print('An exception occurred:', e.__class__, e) traceback.print_exc() test_failed = True if (test_failed): sys.exit(-1) if __name__ == '__main__': main(available_tests)
30,912
9,017
""" Main program to launch proc/hdfs.py """ import argparse import logging from pars import addargs import sys import logging logging.basicConfig(format='%(levelname)s:%(message)s', level=logging.INFO) from proc.hdfs import DIRHDFS def gettestargs(parser) : i = "/home/sbartkowski/work/webhdfsdirectory/testdata/inputhdfs.txt" return parser.parse_args([i,"inimical1","14000","sb","/user/sb","dir1","/tmp/download","--dryrun"]) def getargs(parser) : return parser.parse_args(sys.argv[1:]) def readargs(): parser = argparse.ArgumentParser( description='Download HDFS using WEB REST/API') addargs(parser) # return gettestargs(parser) return getargs(parser) def main(): args = readargs() T = DIRHDFS(args.host[0], args.port[0], args.user[0],args.regexp,args.dryrun) T.downloadhdfsdir(args.userdir[0], args.usersubdir[0], args.localdir[0]) if __name__ == "__main__": # execute only if run as a script main()
969
350
"""JSON (de)serialization framework. The framework presented here is somewhat based on `Go's "json" package`_ (especially the ``omitempty`` functionality). .. _`Go's "json" package`: http://golang.org/pkg/encoding/json/ """ import abc import binascii import logging import OpenSSL import six from josepy import b64, errors, interfaces, util logger = logging.getLogger(__name__) class Field(object): """JSON object field. :class:`Field` is meant to be used together with :class:`JSONObjectWithFields`. ``encoder`` (``decoder``) is a callable that accepts a single parameter, i.e. a value to be encoded (decoded), and returns the serialized (deserialized) value. In case of errors it should raise :class:`~josepy.errors.SerializationError` (:class:`~josepy.errors.DeserializationError`). Note, that ``decoder`` should perform partial serialization only. :ivar str json_name: Name of the field when encoded to JSON. :ivar default: Default value (used when not present in JSON object). :ivar bool omitempty: If ``True`` and the field value is empty, then it will not be included in the serialized JSON object, and ``default`` will be used for deserialization. Otherwise, if ``False``, field is considered as required, value will always be included in the serialized JSON objected, and it must also be present when deserializing. """ __slots__ = ('json_name', 'default', 'omitempty', 'fdec', 'fenc') def __init__(self, json_name, default=None, omitempty=False, decoder=None, encoder=None): # pylint: disable=too-many-arguments self.json_name = json_name self.default = default self.omitempty = omitempty self.fdec = self.default_decoder if decoder is None else decoder self.fenc = self.default_encoder if encoder is None else encoder @classmethod def _empty(cls, value): """Is the provided value considered "empty" for this field? This is useful for subclasses that might want to override the definition of being empty, e.g. for some more exotic data types. """ return not isinstance(value, bool) and not value def omit(self, value): """Omit the value in output?""" return self._empty(value) and self.omitempty def _update_params(self, **kwargs): current = dict(json_name=self.json_name, default=self.default, omitempty=self.omitempty, decoder=self.fdec, encoder=self.fenc) current.update(kwargs) return type(self)(**current) # pylint: disable=star-args def decoder(self, fdec): """Descriptor to change the decoder on JSON object field.""" return self._update_params(decoder=fdec) def encoder(self, fenc): """Descriptor to change the encoder on JSON object field.""" return self._update_params(encoder=fenc) def decode(self, value): """Decode a value, optionally with context JSON object.""" return self.fdec(value) def encode(self, value): """Encode a value, optionally with context JSON object.""" return self.fenc(value) @classmethod def default_decoder(cls, value): """Default decoder. Recursively deserialize into immutable types ( :class:`josepy.util.frozendict` instead of :func:`dict`, :func:`tuple` instead of :func:`list`). """ # bases cases for different types returned by json.loads if isinstance(value, list): return tuple(cls.default_decoder(subvalue) for subvalue in value) elif isinstance(value, dict): return util.frozendict( dict((cls.default_decoder(key), cls.default_decoder(value)) for key, value in six.iteritems(value))) else: # integer or string return value @classmethod def default_encoder(cls, value): """Default (passthrough) encoder.""" # field.to_partial_json() is no good as encoder has to do partial # serialization only return value class JSONObjectWithFieldsMeta(abc.ABCMeta): """Metaclass for :class:`JSONObjectWithFields` and its subclasses. It makes sure that, for any class ``cls`` with ``__metaclass__`` set to ``JSONObjectWithFieldsMeta``: 1. All fields (attributes of type :class:`Field`) in the class definition are moved to the ``cls._fields`` dictionary, where keys are field attribute names and values are fields themselves. 2. ``cls.__slots__`` is extended by all field attribute names (i.e. not :attr:`Field.json_name`). Original ``cls.__slots__`` are stored in ``cls._orig_slots``. In a consequence, for a field attribute name ``some_field``, ``cls.some_field`` will be a slot descriptor and not an instance of :class:`Field`. For example:: some_field = Field('someField', default=()) class Foo(object): __metaclass__ = JSONObjectWithFieldsMeta __slots__ = ('baz',) some_field = some_field assert Foo.__slots__ == ('some_field', 'baz') assert Foo._orig_slots == () assert Foo.some_field is not Field assert Foo._fields.keys() == ['some_field'] assert Foo._fields['some_field'] is some_field As an implementation note, this metaclass inherits from :class:`abc.ABCMeta` (and not the usual :class:`type`) to mitigate the metaclass conflict (:class:`ImmutableMap` and :class:`JSONDeSerializable`, parents of :class:`JSONObjectWithFields`, use :class:`abc.ABCMeta` as its metaclass). """ def __new__(mcs, name, bases, dikt): fields = {} for base in bases: fields.update(getattr(base, '_fields', {})) # Do not reorder, this class might override fields from base classes! for key, value in tuple(six.iteritems(dikt)): # not six.iterkeys() (in-place edit!) if isinstance(value, Field): fields[key] = dikt.pop(key) dikt['_orig_slots'] = dikt.get('__slots__', ()) dikt['__slots__'] = tuple( list(dikt['_orig_slots']) + list(six.iterkeys(fields))) dikt['_fields'] = fields return abc.ABCMeta.__new__(mcs, name, bases, dikt) @six.add_metaclass(JSONObjectWithFieldsMeta) class JSONObjectWithFields(util.ImmutableMap, interfaces.JSONDeSerializable): # pylint: disable=too-few-public-methods """JSON object with fields. Example:: class Foo(JSONObjectWithFields): bar = Field('Bar') empty = Field('Empty', omitempty=True) @bar.encoder def bar(value): return value + 'bar' @bar.decoder def bar(value): if not value.endswith('bar'): raise errors.DeserializationError('No bar suffix!') return value[:-3] assert Foo(bar='baz').to_partial_json() == {'Bar': 'bazbar'} assert Foo.from_json({'Bar': 'bazbar'}) == Foo(bar='baz') assert (Foo.from_json({'Bar': 'bazbar', 'Empty': '!'}) == Foo(bar='baz', empty='!')) assert Foo(bar='baz').bar == 'baz' """ @classmethod def _defaults(cls): """Get default fields values.""" return dict([(slot, field.default) for slot, field in six.iteritems(cls._fields)]) def __init__(self, **kwargs): # pylint: disable=star-args super(JSONObjectWithFields, self).__init__( **(dict(self._defaults(), **kwargs))) def encode(self, name): """Encode a single field. :param str name: Name of the field to be encoded. :raises errors.SerializationError: if field cannot be serialized :raises errors.Error: if field could not be found """ try: field = self._fields[name] except KeyError: raise errors.Error("Field not found: {0}".format(name)) return field.encode(getattr(self, name)) def fields_to_partial_json(self): """Serialize fields to JSON.""" jobj = {} omitted = set() for slot, field in six.iteritems(self._fields): value = getattr(self, slot) if field.omit(value): omitted.add((slot, value)) else: try: jobj[field.json_name] = field.encode(value) except errors.SerializationError as error: raise errors.SerializationError( 'Could not encode {0} ({1}): {2}'.format( slot, value, error)) return jobj def to_partial_json(self): return self.fields_to_partial_json() @classmethod def _check_required(cls, jobj): missing = set() for _, field in six.iteritems(cls._fields): if not field.omitempty and field.json_name not in jobj: missing.add(field.json_name) if missing: raise errors.DeserializationError( 'The following fields are required: {0}'.format( ','.join(missing))) @classmethod def fields_from_json(cls, jobj): """Deserialize fields from JSON.""" cls._check_required(jobj) fields = {} for slot, field in six.iteritems(cls._fields): if field.json_name not in jobj and field.omitempty: fields[slot] = field.default else: value = jobj[field.json_name] try: fields[slot] = field.decode(value) except errors.DeserializationError as error: raise errors.DeserializationError( 'Could not decode {0!r} ({1!r}): {2}'.format( slot, value, error)) return fields @classmethod def from_json(cls, jobj): return cls(**cls.fields_from_json(jobj)) def encode_b64jose(data): """Encode JOSE Base-64 field. :param bytes data: :rtype: `unicode` """ # b64encode produces ASCII characters only return b64.b64encode(data).decode('ascii') def decode_b64jose(data, size=None, minimum=False): """Decode JOSE Base-64 field. :param unicode data: :param int size: Required length (after decoding). :param bool minimum: If ``True``, then `size` will be treated as minimum required length, as opposed to exact equality. :rtype: bytes """ error_cls = TypeError if six.PY2 else binascii.Error try: decoded = b64.b64decode(data.encode()) except error_cls as error: raise errors.DeserializationError(error) if size is not None and ((not minimum and len(decoded) != size) or (minimum and len(decoded) < size)): raise errors.DeserializationError( "Expected at least or exactly {0} bytes".format(size)) return decoded def encode_hex16(value): """Hexlify. :param bytes value: :rtype: unicode """ return binascii.hexlify(value).decode() def decode_hex16(value, size=None, minimum=False): """Decode hexlified field. :param unicode value: :param int size: Required length (after decoding). :param bool minimum: If ``True``, then `size` will be treated as minimum required length, as opposed to exact equality. :rtype: bytes """ value = value.encode() if size is not None and ((not minimum and len(value) != size * 2) or (minimum and len(value) < size * 2)): raise errors.DeserializationError() error_cls = TypeError if six.PY2 else binascii.Error try: return binascii.unhexlify(value) except error_cls as error: raise errors.DeserializationError(error) def encode_cert(cert): """Encode certificate as JOSE Base-64 DER. :type cert: `OpenSSL.crypto.X509` wrapped in `.ComparableX509` :rtype: unicode """ return encode_b64jose(OpenSSL.crypto.dump_certificate( OpenSSL.crypto.FILETYPE_ASN1, cert.wrapped)) def decode_cert(b64der): """Decode JOSE Base-64 DER-encoded certificate. :param unicode b64der: :rtype: `OpenSSL.crypto.X509` wrapped in `.ComparableX509` """ try: return util.ComparableX509(OpenSSL.crypto.load_certificate( OpenSSL.crypto.FILETYPE_ASN1, decode_b64jose(b64der))) except OpenSSL.crypto.Error as error: raise errors.DeserializationError(error) def encode_csr(csr): """Encode CSR as JOSE Base-64 DER. :type csr: `OpenSSL.crypto.X509Req` wrapped in `.ComparableX509` :rtype: unicode """ return encode_b64jose(OpenSSL.crypto.dump_certificate_request( OpenSSL.crypto.FILETYPE_ASN1, csr.wrapped)) def decode_csr(b64der): """Decode JOSE Base-64 DER-encoded CSR. :param unicode b64der: :rtype: `OpenSSL.crypto.X509Req` wrapped in `.ComparableX509` """ try: return util.ComparableX509(OpenSSL.crypto.load_certificate_request( OpenSSL.crypto.FILETYPE_ASN1, decode_b64jose(b64der))) except OpenSSL.crypto.Error as error: raise errors.DeserializationError(error) class TypedJSONObjectWithFields(JSONObjectWithFields): """JSON object with type.""" typ = NotImplemented """Type of the object. Subclasses must override.""" type_field_name = "type" """Field name used to distinguish different object types. Subclasses will probably have to override this. """ TYPES = NotImplemented """Types registered for JSON deserialization""" @classmethod def register(cls, type_cls, typ=None): """Register class for JSON deserialization.""" typ = type_cls.typ if typ is None else typ cls.TYPES[typ] = type_cls return type_cls @classmethod def get_type_cls(cls, jobj): """Get the registered class for ``jobj``.""" if cls in six.itervalues(cls.TYPES): if cls.type_field_name not in jobj: raise errors.DeserializationError( "Missing type field ({0})".format(cls.type_field_name)) # cls is already registered type_cls, force to use it # so that, e.g Revocation.from_json(jobj) fails if # jobj["type"] != "revocation". return cls if not isinstance(jobj, dict): raise errors.DeserializationError( "{0} is not a dictionary object".format(jobj)) try: typ = jobj[cls.type_field_name] except KeyError: raise errors.DeserializationError("missing type field") try: return cls.TYPES[typ] except KeyError: raise errors.UnrecognizedTypeError(typ, jobj) def to_partial_json(self): """Get JSON serializable object. :returns: Serializable JSON object representing ACME typed object. :meth:`validate` will almost certainly not work, due to reasons explained in :class:`josepy.interfaces.IJSONSerializable`. :rtype: dict """ jobj = self.fields_to_partial_json() jobj[self.type_field_name] = self.typ return jobj @classmethod def from_json(cls, jobj): """Deserialize ACME object from valid JSON object. :raises josepy.errors.UnrecognizedTypeError: if type of the ACME object has not been registered. """ # make sure subclasses don't cause infinite recursive from_json calls type_cls = cls.get_type_cls(jobj) return type_cls(**type_cls.fields_from_json(jobj))
15,748
4,722
from norm.models.norm import Status, Lambda from norm.executable import NormExecutable from typing import Union, List import logging logger = logging.getLogger(__name__) class VariableName(NormExecutable): def __init__(self, scope, name): """ The variable and its scope :param scope: the scope of the variable :type scope: Union[VariableName, EvaluationExpr] :param name: the name of the variable :type name: str """ super().__init__() from norm.executable.expression.evaluation import EvaluationExpr self.scope: Union[VariableName, EvaluationExpr] = scope self.name: str = name self.output_projection: str = None @property def eval_lam(self): return self.lam def __str__(self): if self.scope is not None: return '{}{}{}'.format(self.scope.name, self.VARIABLE_SEPARATOR, self.name) else: return self.name def variable_type(self): return self.lam def compile(self, context): if self.name == context.THAT_VARIABLE_NAME: self.lam = context.that return self session = context.session if self.scope is None: name = self.name scope = context.get_scope(name) if scope is not None: return ColumnVariable(scope, name).compile(context) else: lam = self.try_retrieve_type(session, context.context_namespace, name) if lam is None: lam = self.try_retrieve_type(session, context.search_namespaces, name, status=Status.READY) self.lam = lam return self else: if isinstance(self.scope, ColumnVariable) and str(self) in self.scope.lam: # Already joined self.scope.name = str(self) return self.scope lam = self.scope.variable_type() if self.name in lam: if isinstance(self.scope, (ColumnVariable, JoinVariable)): # Join lam to the scope for the column return JoinVariable(self.scope, self.name, lam).compile(context) else: # A column of the scope variable return ColumnVariable(self.scope, self.name).compile(context) else: # An evaluation whose first argument is the scope lam = self.try_retrieve_type(session, context.context_namespace, self.name) if lam is None: lam = self.try_retrieve_type(session, context.search_namespaces, self.name, status=Status.READY) assert(lam is not None) self.lam = lam from norm.executable.expression.argument import ArgumentExpr arg = ArgumentExpr(expr=self.scope) self.scope = None from norm.executable.expression.evaluation import EvaluationExpr return EvaluationExpr([arg], self) class UnquoteVariable(VariableName): def __init__(self, name, unquoted_variables): """ The variable and its scope :param name: the name of the variable :type name: str :param unquoted_variables: a list of variables to unquote :type unquoted_variables: List[VariableName] """ super().__init__(None, name) self.unquoted_variables: List[VariableName] = unquoted_variables def __str__(self): return self.name def variable_type(self): raise NotImplementedError def compile(self, context): assert(len(self.unquoted_variables) > 0) assert(all([isinstance(v, ColumnVariable) for v in self.unquoted_variables])) lam = self.unquoted_variables[0].lam assert(all([v.lam is lam for v in self.unquoted_variables])) self.lam = lam return self def execute(self, context): # TODO: hacky assert(len(self.unquoted_variables) == 1) vname = self.unquoted_variables[0].name data = self.lam.data from pandas import DataFrame def _execute(x): try: result = context.execute(x[vname].values[0]) if isinstance(result, DataFrame): return result else: return None except: return None results = data.groupby(vname).apply(_execute).reset_index() if self.output_projection is not None: cols = dict((col, self.VARIABLE_SEPARATOR.join([self.output_projection, col])) for col in results.columns if col != vname and col != self.lam.VAR_OID) if self.lam.VAR_OID in results.columns: cols[self.lam.VAR_OID] = self.output_projection results = results.rename(columns=cols) return data.merge(results, on=vname) class ColumnVariable(VariableName): def __init__(self, scope, name): super().__init__(scope, name) def __str__(self): return self.name def variable_type(self): return self.lam.get_type(self.name) def compile(self, context): from norm.engine import QuantifiedLambda if self.scope is None: assert(context.scope is not None) self.lam = context.scope elif isinstance(self.scope, Lambda): self.lam = self.scope elif isinstance(self.scope, QuantifiedLambda): self.lam = self.scope else: self.lam = self.scope.lam return self def execute(self, context): result = self.lam.data[self.name] if self.output_projection is not None: return result.rename(self.output_projection) else: return result class JoinVariable(VariableName): def __init__(self, scope, name, joiner): super().__init__(scope, name) self.lam = joiner def variable_type(self): return self.lam.get_type(self.name) def compile(self, context): return self def execute(self, context): lam = self.scope.lam joiner = self.lam if str(self) not in lam.data.columns: to_join = joiner.data[[self.name]].rename(columns={self.name: str(self)}) lam.data = lam.data.join(to_join, on=str(self.scope)) return lam.data
6,463
1,782
from __future__ import print_function import os import sys import json import tika from tqdm import tqdm from utils import LogUtil from parser import Parser from ioutils import read_lines from tika import parser as tk_parser class TikaParser(Parser): """ The TikaParser class is wrapper of the Apache TIKA parse """ def __init__(self, tika_server_url): super(TikaParser, self).__init__('tika_parser') if tika_server_url: os.environ['TIKA_CLIENT_ONLY'] = 'True' os.environ['TIKA_SERVER_ENDPOINT'] = tika_server_url print("Tika Server Endpoint %s" % os.environ['TIKA_SERVER_ENDPOINT']) tika.initVM() def parse(self, file_path): """ Parse one PDF file using Apache TIKA parser Args: file_path (str): Path to a PDF file Return: parsed content stored in a dictionary """ if not os.path.exists(file_path): raise RuntimeError('%s error. File not found: %s' % (self.parse_name, os.path.abspath(file_path))) try: tika_dict = tk_parser.from_file(file_path) except Exception: raise RuntimeError('Internal TIKA error occurred while parsing the ' 'file: %s' % os.path.abspath(file_path)) tika_dict['file'] = os.path.abspath(file_path) return tika_dict def process(in_file, in_list, out_file, log_file, tika_server_url): # Log input parameters logger = LogUtil('lpsc-parser', log_file) logger.info('Input parameters') logger.info('in_file: %s' % in_file) logger.info('in_list: %s' % in_list) logger.info('out_file: %s' % out_file) logger.info('tika_server_url: %s' % tika_server_url) if in_file and in_list: print('[ERROR] in_file and in_list cannot be provided simultaneously') sys.exit(1) tika_parser = TikaParser(tika_server_url) if in_file: files = [in_file] else: files = read_lines(in_list) out_f = open(out_file, 'wb', 1) for f in tqdm(files): try: tika_dict = tika_parser.parse(f) out_f.write(json.dumps(tika_dict)) out_f.write('\n') except Exception as e: logger.info('TIKA parser failed: %s' % os.path.abspath(f)) logger.error(e) out_f.close() def main(): import argparse parser = argparse.ArgumentParser() input_parser = parser.add_mutually_exclusive_group(required=True) input_parser.add_argument('-i', '--in_file', help='Path to input file') input_parser.add_argument('-li', '--in_list', help='Path to input list') parser.add_argument('-o', '--out_file', required=True, help='Path to output JSON file') parser.add_argument('-l', '--log_file', default='./tika-parser-log.txt', help='Log file that contains processing information. ' 'It is default to ./tika-parser-log.txt unless ' 'otherwise specified.') parser.add_argument('-p', '--tika_server_url', required=False, help='Tika server URL') args = parser.parse_args() process(**vars(args)) if __name__ == '__main__': main()
3,318
1,063
from checkov.cloudformation.checks.resource.base_resource_check import BaseResourceCheck from checkov.common.parsers.node import DictNode from checkov.common.models.enums import CheckResult, CheckCategories class DocDBAuditLogs(BaseResourceCheck): def __init__(self) -> None: name = "Ensure DocDB has audit logs enabled" id = "CKV_AWS_104" supported_resources = ["AWS::DocDB::DBClusterParameterGroup"] categories = [CheckCategories.LOGGING] super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources) def scan_resource_conf(self, conf: DictNode) -> CheckResult: params = conf.get("Properties", {}).get("Parameters", {}) if params.get("audit_logs") == "enabled": return CheckResult.PASSED return CheckResult.FAILED check = DocDBAuditLogs()
869
263
# # Copyright (c) 2016-2022 Deephaven Data Labs and Patent Pending # import jpy _JCallbackAdapter = jpy.get_type('io.deephaven.server.plugin.python.CallbackAdapter') def initialize_all_and_register_into(callback: _JCallbackAdapter): try: from . import register except ModuleNotFoundError as e: # deephaven.plugin is an optional dependency, so if it can't be found, there are no Deephaven python plugins # to register if e.name == 'deephaven.plugin': return raise e register.initialize_all_and_register_into(callback)
584
183
"""Tests for parse tree pretty printing that preserves formatting Test case descriptions are in file test/data/output.test. """ import os.path import re from typing import Undefined, Any from mypy import build from mypy.myunit import Suite, run_test from mypy.test.helpers import assert_string_arrays_equal from mypy.test.data import parse_test_cases from mypy.test.config import test_data_prefix, test_temp_dir from mypy.parse import parse from mypy.output import OutputVisitor from mypy.errors import CompileError # Files which contain test case descriptions. output_files = ['output.test'] class OutputSuite(Suite): def cases(self): c = [] for f in output_files: c += parse_test_cases(os.path.join(test_data_prefix, f), test_output, test_temp_dir, True) return c def test_output(testcase): """Perform an identity source code transformation test case.""" expected = testcase.output if expected == []: expected = testcase.input try: src = '\n'.join(testcase.input) # Parse and semantically analyze the source program. # Test case names with a special suffix get semantically analyzed. This # lets us test that semantic analysis does not break source code pretty # printing. if testcase.name.endswith('_SemanticAnalyzer'): result = build.build('main', target=build.SEMANTIC_ANALYSIS, program_text=src, flags=[build.TEST_BUILTINS], alt_lib_path=test_temp_dir) files = result.files else: files = {'main': parse(src, 'main')} a = [] first = True # Produce an output containing the pretty-printed forms (with original # formatting) of all the relevant source files. for fnam in sorted(files.keys()): f = files[fnam] # Omit the builtins and files marked for omission. if (not f.path.endswith(os.sep + 'builtins.py') and '-skip.' not in f.path): # Add file name + colon for files other than the first. if not first: a.append('{}:'.format(fix_path(remove_prefix( f.path, test_temp_dir)))) v = OutputVisitor() f.accept(v) s = v.output() if s != '': a += s.split('\n') first = False except CompileError as e: a = e.messages assert_string_arrays_equal( expected, a, 'Invalid source code output ({}, line {})'.format( testcase.file, testcase.line)) def remove_prefix(path, prefix): regexp = '^' + prefix.replace('\\', '\\\\') np = re.sub(regexp, '', path) if np.startswith(os.sep): np = np[1:] return np def fix_path(path): return path.replace('\\', '/') if __name__ == '__main__': import sys run_test(OutputSuite(), sys.argv[1:])
3,133
877
# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # -------------------------------------------------------------------------- import unittest from sklearn.tree import DecisionTreeClassifier from sklearn.tree import DecisionTreeRegressor from skl2onnx.common.data_types import onnx_built_with_ml from test_utils import ( dump_one_class_classification, dump_binary_classification, dump_multiple_classification, ) from test_utils import dump_multiple_regression, dump_single_regression class TestSklearnDecisionTreeModels(unittest.TestCase): @unittest.skipIf(not onnx_built_with_ml(), reason="Requires ONNX-ML extension.") def test_decision_tree_classifier(self): model = DecisionTreeClassifier() dump_one_class_classification( model, # Operator cast-1 is not implemented in onnxruntime allow_failure="StrictVersion(onnx.__version__)" " < StrictVersion('1.3') or " "StrictVersion(onnxruntime.__version__)" " <= StrictVersion('0.2.1')", ) dump_binary_classification( model, allow_failure="StrictVersion(onnx.__version__)" " < StrictVersion('1.3') or " "StrictVersion(onnxruntime.__version__)" " <= StrictVersion('0.2.1')", ) dump_multiple_classification( model, allow_failure="StrictVersion(onnx.__version__)" " < StrictVersion('1.3') or " "StrictVersion(onnxruntime.__version__)" " <= StrictVersion('0.2.1')", ) def test_decision_tree_regressor(self): model = DecisionTreeRegressor() dump_single_regression( model, allow_failure="StrictVersion(onnx.__version__)" " < StrictVersion('1.2')", ) dump_multiple_regression( model, allow_failure="StrictVersion(onnx.__version__)" " < StrictVersion('1.2')", ) if __name__ == "__main__": unittest.main()
2,395
638
import logging import os import tempfile import PIL.Image import numpy import tensorflow import microscopeimagequality.constants import microscopeimagequality.data_provider import microscopeimagequality.evaluation import microscopeimagequality.prediction class Inference(tensorflow.test.TestCase): def setUp(self): self.input_directory = os.path.join(os.path.dirname(os.path.abspath(__file__)) , "data") self.test_data_directory = os.path.join(os.path.dirname(os.path.abspath(__file__)) , "data") self.test_dir = tempfile.mkdtemp() self.glob_images = os.path.join(self.input_directory, 'images_for_glob_test/*') self.patch_width = 84 self.num_classes = 11 def testPatchValuesToMask(self): values = numpy.round( numpy.array([[0.2, 0.4, 0.5], [1.0, 0.0, 0.3]]) * numpy.iinfo(numpy.uint16).max).astype(numpy.uint16) mask = microscopeimagequality.prediction.patch_values_to_mask(values, self.patch_width) self.assertEquals((168, 252), mask.shape) self.assertEquals(numpy.iinfo(numpy.uint16).max, numpy.max(mask)) def testSaveMasksAndAnnotatedVisualization(self): test_filename = 'BBBC006_z_aligned__a01__s1__w1_10.png' orig_name = os.path.join(self.test_data_directory, test_filename) prediction = 1 certainties = {name: 0.3 for name in microscopeimagequality.evaluation.CERTAINTY_NAMES} num_patches = 4 np_images = numpy.ones((num_patches, self.patch_width, self.patch_width, 1)) np_probabilities = numpy.ones( (num_patches, self.num_classes)) / self.num_classes np_probabilities[0, :] = 0 np_probabilities[0, 1] = 1.0 np_probabilities[1, :] = 0 np_probabilities[1, 2] = 0.4 np_probabilities[1, -1] = 0.6 np_labels = 2 * numpy.ones(num_patches) image_height = int(numpy.sqrt(num_patches)) * self.patch_width image_width = image_height microscopeimagequality.prediction.save_masks_and_annotated_visualization( orig_name, self.test_dir, prediction, certainties, np_images, np_probabilities, np_labels, self.patch_width, image_height, image_width) # Check that output has been generated and is the correct shape. expected_size = PIL.Image.open(orig_name, 'r').size expected_visualization_path = os.path.join( self.test_dir, 'actual2_pred1_mean_certainty=0.300orig_name=%s' % test_filename) expected_predictions_path = os.path.join(self.test_dir, microscopeimagequality.constants.PREDICTIONS_MASK_FORMAT % test_filename) expected_certainties_path = os.path.join(self.test_dir, microscopeimagequality.constants.CERTAINTY_MASK_FORMAT % test_filename) expected_valid_path = os.path.join(self.test_dir, microscopeimagequality.constants.VALID_MASK_FORMAT % test_filename) img = PIL.Image.open(expected_visualization_path, 'r') self.assertEquals(expected_size, img.size) img = PIL.Image.open(expected_predictions_path, 'r') self.assertEquals(expected_size, img.size) img = PIL.Image.open(expected_certainties_path, 'r') self.assertEquals(expected_size, img.size) img = PIL.Image.open(expected_valid_path, 'r') self.assertEquals(expected_size, img.size) def testSaveMasksAndAnnotatedVisualizationTif(self): test_filename = ('00_mcf-z-stacks-03212011_k06_s2_w12667264a' '-6432-4f7e-bf58-625a1319a1c9.tif') orig_name = os.path.join(self.test_data_directory, test_filename) prediction = 1 certainties = {name: 0.3 for name in microscopeimagequality.evaluation.CERTAINTY_NAMES} num_patches = 4 np_images = numpy.ones((num_patches, self.patch_width, self.patch_width, 1)) np_probabilities = numpy.ones( (num_patches, self.num_classes)) / self.num_classes image_height = int(numpy.sqrt(num_patches)) * self.patch_width image_width = image_height np_labels = 2 * numpy.ones(num_patches) microscopeimagequality.prediction.save_masks_and_annotated_visualization( orig_name, self.test_dir, prediction, certainties, np_images, np_probabilities, np_labels, self.patch_width, image_height, image_width) mask_formats = [ microscopeimagequality.constants.CERTAINTY_MASK_FORMAT, microscopeimagequality.constants.PREDICTIONS_MASK_FORMAT, microscopeimagequality.constants.VALID_MASK_FORMAT ] for mask_format in mask_formats: orig_name_png = os.path.splitext(os.path.basename(orig_name))[0] + '.png' expected_file = os.path.join(self.test_dir, mask_format % orig_name_png) self.assertTrue(os.path.isfile(expected_file)) def testRunModelInferenceFirstHalfRuns(self): batch_size = 1 num_classes = 11 model_patch_width = 84 image_width = 84 image_height = 84 tfexamples_tfrecord = microscopeimagequality.prediction.build_tfrecord_from_pngs( [self.glob_images], use_unlabeled_data=True, num_classes=num_classes, eval_directory=self.test_dir, image_background_value=0, image_brightness_scale=1, shard_num=0, num_shards=1, image_width=image_width, image_height=image_height) num_samples = microscopeimagequality.data_provider.get_num_records(tfexamples_tfrecord % microscopeimagequality.prediction._SPLIT_NAME) logging.info('TFRecord has %g samples.', num_samples) g = tensorflow.Graph() with g.as_default(): images, one_hot_labels, _, _ = microscopeimagequality.data_provider.provide_data( tfexamples_tfrecord, split_name=microscopeimagequality.prediction._SPLIT_NAME, batch_size=batch_size, num_classes=num_classes, image_width=84, image_height=84, patch_width=model_patch_width, randomize=False, num_threads=1) labels = microscopeimagequality.evaluation.get_model_and_metrics( images, num_classes=num_classes, one_hot_labels=one_hot_labels, is_training=False).labels self.assertEquals(batch_size, labels.get_shape())
7,005
2,181
""" QuickBot wiring config. Specifies which pins are used for motor control, IR sensors and wheel encoders. """ # Motor pins: (dir1_pin, dir2_pin, pwd_pin) RIGHT_MOTOR_PINS = 'P8_12', 'P8_10', 'P9_14' LEFT_MOTOR_PINS = 'P8_14', 'P8_16', 'P9_16' # IR sensors (clock-wise, starting with the rear left sensor): # rear-left, front-left, front, front-right, rear-right IR_PINS = ('P9_38', 'P9_40', 'P9_36', 'P9_35', 'P9_33') # Wheel encoder sensors: (left, right) ENC_PINS = ('P9_39', 'P9_37')
498
246
#!/usr/bin/env python # Copyright (c) 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. # Self-test for skimage. import filecmp import os import subprocess import sys import tempfile class BinaryNotFoundException(Exception): def __str__ (self): return ("Could not find binary!\n" "Did you forget to build the tools project?\n" "Self tests failed") # Find a path to the binary to use. Iterates through a list of possible # locations the binary may be. def PickBinaryPath(base_dir): POSSIBLE_BINARY_PATHS = [ 'out/Debug/skimage', 'out/Release/skimage', 'xcodebuild/Debug/skimage', 'xcodebuild/Release/skimage', ] for binary in POSSIBLE_BINARY_PATHS: binary_full_path = os.path.join(base_dir, binary) if (os.path.exists(binary_full_path)): return binary_full_path raise BinaryNotFoundException # Quit early if two files have different content. def DieIfFilesMismatch(expected, actual): if not filecmp.cmp(expected, actual): print 'Error: file mismatch! expected=%s , actual=%s' % ( expected, actual) exit(1) def test_invalid_file(file_dir, skimage_binary): """ Test the return value of skimage when an invalid file is decoded. If there is no expectation file, or the file expects a particular result, skimage should return nonzero indicating failure. If the file has no expectation, or ignore-failure is set to true, skimage should return zero indicating success. """ invalid_file = os.path.join(file_dir, "skimage", "input", "bad-images", "invalid.png") # No expectations file: args = [skimage_binary, "--readPath", invalid_file] result = subprocess.call(args) if 0 == result: print "'%s' should have reported failure!" % " ".join(args) exit(1) # Directory holding all expectations files expectations_dir = os.path.join(file_dir, "skimage", "input", "bad-images") # Expectations file expecting a valid decode: incorrect_expectations = os.path.join(expectations_dir, "incorrect-results.json") args = [skimage_binary, "--readPath", invalid_file, "--readExpectationsPath", incorrect_expectations] result = subprocess.call(args) if 0 == result: print "'%s' should have reported failure!" % " ".join(args) exit(1) # Empty expectations: empty_expectations = os.path.join(expectations_dir, "empty-results.json") output = subprocess.check_output([skimage_binary, "--readPath", invalid_file, "--readExpectationsPath", empty_expectations], stderr=subprocess.STDOUT) if not "Missing" in output: # Another test (in main()) tests to ensure that "Missing" does not appear # in the output. That test could be passed if the output changed so # "Missing" never appears. This ensures that an error is not missed if # that happens. print "skimage output changed! This may cause other self tests to fail!" exit(1) # Ignore failure: ignore_expectations = os.path.join(expectations_dir, "ignore-results.json") output = subprocess.check_output([skimage_binary, "--readPath", invalid_file, "--readExpectationsPath", ignore_expectations], stderr=subprocess.STDOUT) if not "failures" in output: # Another test (in main()) tests to ensure that "failures" does not # appear in the output. That test could be passed if the output changed # so "failures" never appears. This ensures that an error is not missed # if that happens. print "skimage output changed! This may cause other self tests to fail!" exit(1) def test_incorrect_expectations(file_dir, skimage_binary): """ Test that comparing to incorrect expectations fails, unless ignore-failures is set to true. """ valid_file = os.path.join(file_dir, "skimage", "input", "images-with-known-hashes", "1209453360120438698.png") expectations_dir = os.path.join(file_dir, "skimage", "input", "images-with-known-hashes") incorrect_results = os.path.join(expectations_dir, "incorrect-results.json") args = [skimage_binary, "--readPath", valid_file, "--readExpectationsPath", incorrect_results] result = subprocess.call(args) if 0 == result: print "'%s' should have reported failure!" % " ".join(args) exit(1) ignore_results = os.path.join(expectations_dir, "ignore-failures.json") subprocess.check_call([skimage_binary, "--readPath", valid_file, "--readExpectationsPath", ignore_results]) def main(): # Use the directory of this file as the out directory file_dir = os.path.abspath(os.path.dirname(__file__)) trunk_dir = os.path.normpath(os.path.join(file_dir, os.pardir, os.pardir)) # Find the binary skimage_binary = PickBinaryPath(trunk_dir) print "Running " + skimage_binary # Generate an expectations file from known images. images_dir = os.path.join(file_dir, "skimage", "input", "images-with-known-hashes") expectations_path = os.path.join(file_dir, "skimage", "output-actual", "create-expectations", "expectations.json") subprocess.check_call([skimage_binary, "--readPath", images_dir, "--createExpectationsPath", expectations_path]) # Make sure the expectations file was generated correctly. golden_expectations = os.path.join(file_dir, "skimage", "output-expected", "create-expectations", "expectations.json") DieIfFilesMismatch(expected=golden_expectations, actual=expectations_path) # Tell skimage to read back the expectations file it just wrote, and # confirm that the images in images_dir match it. output = subprocess.check_output([skimage_binary, "--readPath", images_dir, "--readExpectationsPath", expectations_path], stderr=subprocess.STDOUT) # Although skimage succeeded, it would have reported success if the file # was missing from the expectations file. Consider this a failure, since # the expectations file was created from this same image. (It will print # "Missing" in this case before listing the missing expectations). if "Missing" in output: print "Expectations file was missing expectations!" print output exit(1) # Again, skimage would succeed if there were known failures (and print # "failures"), but there should be no failures, since the file just # created did not include failures to ignore. if "failures" in output: print "Image failed!" print output exit(1) test_incorrect_expectations(file_dir=file_dir, skimage_binary=skimage_binary) # Generate an expectations file from an empty directory. empty_dir = tempfile.mkdtemp() expectations_path = os.path.join(file_dir, "skimage", "output-actual", "empty-dir", "expectations.json") subprocess.check_call([skimage_binary, "--readPath", empty_dir, "--createExpectationsPath", expectations_path]) golden_expectations = os.path.join(file_dir, "skimage", "output-expected", "empty-dir", "expectations.json") DieIfFilesMismatch(expected=golden_expectations, actual=expectations_path) os.rmdir(empty_dir) # Generate an expectations file from a nonexistent directory. expectations_path = os.path.join(file_dir, "skimage", "output-actual", "nonexistent-dir", "expectations.json") subprocess.check_call([skimage_binary, "--readPath", "/nonexistent/dir", "--createExpectationsPath", expectations_path]) golden_expectations = os.path.join(file_dir, "skimage", "output-expected", "nonexistent-dir", "expectations.json") DieIfFilesMismatch(expected=golden_expectations, actual=expectations_path) test_invalid_file(file_dir=file_dir, skimage_binary=skimage_binary) # Done with all tests. print "Self tests succeeded!" if __name__ == "__main__": main()
8,912
2,489
from collections import namedtuple Condition = namedtuple( 'Condition', ['current_state', 'input_character'], )
121
33
"""Asset definitions for the simple_lakehouse example.""" import pandas as pd from lakehouse import Column, computed_table, source_table from pyarrow import date32, float64, string sfo_q2_weather_sample_table = source_table( path="data", columns=[Column("tmpf", float64()), Column("valid_date", string())], ) @computed_table( input_assets=[sfo_q2_weather_sample_table], columns=[Column("valid_date", date32()), Column("max_tmpf", float64())], ) def daily_temperature_highs_table(sfo_q2_weather_sample: pd.DataFrame) -> pd.DataFrame: """Computes the temperature high for each day""" sfo_q2_weather_sample["valid_date"] = pd.to_datetime(sfo_q2_weather_sample["valid"]) return sfo_q2_weather_sample.groupby("valid_date").max().rename(columns={"tmpf": "max_tmpf"})
788
271
#!/usr/bin/env python # -*- coding: utf-8 -*- # #@created: 07.01.2018 #@author: Aleksey Komissarov #@contact: ad3002@gmail.com from aindex import * settings = { "index_prefix": "tests/kmers.23", "aindex_prefix": "tests/kmers.23", "reads_file": "tests/reads.reads", } index = load_aindex(settings) k = 23 sequence = "TAAGTTATTATTTAGTTAATACTTTTAACAATATTATTAAGGTATTTAAAAAATACTATTATAGTATTTAACATAGTTAAATACCTTCCTTAATACTGTTAAATTATATTCAATCAATACATATATAATATTATTAAAATACTTGATAAGTATTATTTAGATATTAGACAAATACTAATTTTATATTGCTTTAATACTTAATAAATACTACTTATGTATTAAGTAAATATTACTGTAATACTAATAACAATATTATTACAATATGCTAGAATAATATTGCTAGTATCAATAATTACTAATATAGTATTAGGAAAATACCATAATAATATTTCTACATAATACTAAGTTAATACTATGTGTAGAATAATAAATAATCAGATTAAAAAAATTTTATTTATCTGAAACATATTTAATCAATTGAACTGATTATTTTCAGCAGTAATAATTACATATGTACATAGTACATATGTAAAATATCATTAATTTCTGTTATATATAATAGTATCTATTTTAGAGAGTATTAATTATTACTATAATTAAGCATTTATGCTTAATTATAAGCTTTTTATGAACAAAATTATAGACATTTTAGTTCTTATAATAAATAATAGATATTAAAGAAAATAAAAAAATAGAAATAAATATCATAACCCTTGATAACCCAGAAATTAATACTTAATCAAAAATGAAAATATTAATTAATAAAAGTGAATTGAATAAAATTTTGAAAAAAATGAATAACGTTATTATTTCCAATAACAAAATAAAACCACATCATTCATATTTTTTAATAGAGGCAAAAGAAAAAGAAATAAACTTTTATGCTAACAATGAATACTTTTCTGTCAAATGTAATTTAAATAAAAATATTGATATTCTTGAACAAGGCTCCTTAATTGTTAAAGGAAAAATTTTTAACGATCTTATTAATGGCATAAAAGAAGAGATTATTACTATTCAAGAAAAAGATCAAACACTTTTGGTTAAAACAAAAAAAACAAGTATTAATTTAAACACAATTAATGTGAATGAATTTCCAAGAATAAGGTTTAATGAAAAAAACGATTTAAGTGAATTTAATCAATTCAAAATAAATTATTCACTTTTAGTAAAAGGCATTAAAAAAATTTTTCACTCAGTTTCAAATAATCGTGAAATATCTTCTAAATTTAATGGAGTAAATTTCAATGGATCCAATGGAAAAGAAATATTTTTAGAAGCTTCTGACACTTATAAACTATCTGTTTTTGAGATAAAGCAAGAAACAGAACCATTTGATTTCATTTTGGAGAGTAATTTACTTAGTTTCATTAATTCTTTTAATCCTGAAGAAGATAAATCTATTGTTTTTTATTACAGAAAAGATAATAAAGATAGCTTTAGTACAGAAATGTTGATTTCAATGGATAACTTTATGATTAGTTACACATCGGTTAATGAAAAATTTCCAGAGGTAAACTACTTTTTTGAATTTGAACCTGAAACTAAAATAGTTGTTCAAAAAAATGAATTAAAAGATGCACTTCAAAGAATTCAAACTTTGGCTCAAAATGAAAGAACTTTTTTATGCGATATGCAAATTAACAGTTCTGAATTAAAAATAAGAGCTATTGTTAATAATATCGGAAATTCTCTTGAGGAAATTTCTTGTCTTAAATTTGAAGGTTATAAACTTAATATTTCTTTTAACCCAAGTTCTCTATTAGATCACATAGAGTCTTTTGAATCAAATGAAATAAATTTTGATTTCCAAGGAAATAGTAAGTATTTTTTGATAACCTCTAAAAGTGAACCTGAACTTAAGCAAATATTGGTTCCTTCAAGATAATGAATCTTTACGATCTTTTAGAACTACCAACTACAGCATCAATAAAAGAAATAAAAATTGCTTATAAAAGATTAGCAAAGCGTTATCACCCTGATGTAAATAAATTAGGTTCGCAAACTTTTGTTGAAATTAATAATGCTTATTCAATATTAAGTGATCCTAACCAAAAGGAAAAATATGATTCAATGCTGAAAGTTAATGATTTTCAAAATCGCATCAAAAATTTAGATATTAGTGTTAGATGACATGAAAATTTCATGGAAGAACTCGAACTTCGTAAGAACTGAGAATTTGATTTTTTTTCATCTGATGAAGATTTCTTTTATTCTCCATTTACAAAAA" test_kmer = "TAAGTTATTATTTAGTTAATACT" right_kmer = "AGTTAATACTTTTAACAATATTA" print("Task 1. Get kmer frequency") # raw_input("\nReady?") for i in range(len(sequence)-k+1): kmer = sequence[i:i+k] print("Position %s kmer %s freq = %s" % (i, kmer, index[kmer])) print("Task 2. Iter read by read, print the first 20 reads") # raw_input("\nReady?") for i, read in enumerate(index.iter_reads()): if i == 20: break print(i, read) print("Task 3. Iter reads by kmer, returs (start, next_read_start, read, pos_if_uniq|None, all_poses)") # raw_input("\nReady?") for read in iter_reads_by_kmer(test_kmer, index): print(read) print("Task 4. Get distances in reads for two kmers, returns a list of (rid, left_kmer_pos, right_kmer_pos) tuples.") # raw_input("\nReady?") print(get_left_right_distances(test_kmer, right_kmer, index)) print("Task 5. Get layout for kmer, returns (max_pos, reads, lefts, rights, rids, starts), for details see source code") # raw_input("\nReady?") max_pos, reads, lefts, rights, rids, starts = get_layout_for_kmer(right_kmer, index) print("Central layout:") for read in reads: print(read) print("Left flanks:") print(lefts) print("Right flanks:") print(rights) print("Task 6. Iter reads by sequence, returтs (start, next_read_start, read, pos_if_uniq|None, all_poses)") # raw_input("\nReady?") sequence = "AATATTATTAAGGTATTTAAAAAATACTATTATAGTATTTAACATA" for read in iter_reads_by_sequence(sequence, index): print(read) print("Task 7. Iter reads by kmer with reads as SE, returns (start, next_read_start, subread, kmere_pos, -1|0|1 for spring_pos, was_reversed, poses_in_read)") # raw_input("\nReady?") user_reads = set() sequence = "AATATTATTAAGGTATTTAAAAAATACTATTATAGTATTTAACATA" for rid, nextrid, read, pos, spring_pos, was_reversed, poses in get_reads_se_by_kmer(kmer, index, user_reads, k=23): print(rid, read, pos)
4,386
2,039
"""Implements test for pytest-resource-path Fixtures with pytest.ini.""" from pathlib import Path import pytest def test_resource_path_ini(resource_path, request): """Fixture resource_path should be following absolute path.""" assert resource_path == Path(str(request.fspath)).parents[1] / Path( "data/test_package/test_module_something/test_resource_path_ini" ) def test_resource_path_root_ini(resource_path_root, request): """Fixture resource_path_root should be following absolute path.""" assert resource_path_root == Path(str(request.fspath)).parents[1] / Path("data") @pytest.fixture(scope="package") def resource_path_root_scope_package_ini(resource_path_root): yield resource_path_root # Reason: To define fixture in same module. pylint: disable=redefined-outer-name def test_resource_path_root_scope_package_ini(resource_path_root_scope_package_ini, request): assert resource_path_root_scope_package_ini == Path(str(request.fspath)).parents[1] / Path("data")
1,013
321
# Illustrates combining exception / error handling # with file access print('Start') try: with open('myfile2.txt', 'r') as f: lines = f.readlines() for line in lines: print(line, end='') except FileNotFoundError as err: print('oops') print(err) print('Done')
301
93
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Mon Sep 11 13:30:53 2017 @author: laoj """ import numpy as np import pymc3 as pm import theano.tensor as tt from pymc3.distributions.distribution import Discrete, draw_values, generate_samples, infer_shape from pymc3.distributions.dist_math import bound, logpow, factln, Cholesky from pymc3.math import tround #%% n scaler, p 1D #n = 183 n = np.array([[106], [143], [102], [116], [183], [150]]) p = np.array([[ 0.21245365, 0.41223126, 0.37531509], [ 0.13221011, 0.50537169, 0.3624182 ], [ 0.08813779, 0.54447146, 0.36739075], [ 0.18932804, 0.4630365, 0.34763546], [ 0.11006472, 0.49227755, 0.39765773], [ 0.17886852, 0.41098834, 0.41014314]]) # p = np.array([ 0.21245365, 0.41223126, 0.37531509]) n = tt.as_tensor_variable(n) p = tt.as_tensor_variable(p) n = np.squeeze(n) n = tt.shape_padright(n) if n.ndim == 1 else tt.as_tensor_variable(n) n.ndim n * p #%% n = np.array([[106], [143], [102], [116], [183], [150]]) #n = 183 p = np.array([[ 0.21245365, 0.41223126, 0.37531509], [ 0.13221011, 0.50537169, 0.3624182 ], [ 0.08813779, 0.54447146, 0.36739075], [ 0.18932804, 0.4630365, 0.34763546], [ 0.11006472, 0.49227755, 0.39765773], [ 0.17886852, 0.41098834, 0.41014314]]) #p = np.array([[ 0.21245365, 0.41223126, 0.37531509]]) #n = tt.as_tensor_variable(n) p = tt.as_tensor_variable(p) #%% class Multinomial(Discrete): def __init__(self, n, p, *args, **kwargs): super(Multinomial, self).__init__(*args, **kwargs) p = p / tt.sum(p, axis=-1, keepdims=True) n = np.squeeze(n) # works also if n is a tensor if len(self.shape) > 1: m = self.shape[-2] try: assert n.shape == (m,) except (AttributeError, AssertionError): n = n * tt.ones(m) self.n = tt.shape_padright(n) self.p = p if p.ndim > 1 else tt.shape_padleft(p) elif n.ndim == 1: self.n = tt.shape_padright(n) self.p = p if p.ndim > 1 else tt.shape_padleft(p) else: # n is a scalar, p is a 1d array self.n = tt.as_tensor_variable(n) self.p = tt.as_tensor_variable(p) self.mean = self.n * self.p mode = tt.cast(tt.round(self.mean), 'int32') diff = self.n - tt.sum(mode, axis=-1, keepdims=True) inc_bool_arr = tt.abs_(diff) > 0 mode = tt.inc_subtensor(mode[inc_bool_arr.nonzero()], diff[inc_bool_arr.nonzero()]) self.mode = mode def _random(self, n, p, size=None): original_dtype = p.dtype # Set float type to float64 for numpy. This change is related to numpy issue #8317 (https://github.com/numpy/numpy/issues/8317) p = p.astype('float64') # Now, re-normalize all of the values in float64 precision. This is done inside the conditionals if size == p.shape: size = None if (p.ndim == 1) and (n.ndim == 0): p = p / p.sum() randnum = np.random.multinomial(n, p.squeeze(), size=size) else: p = p / p.sum(axis=1, keepdims=True) if n.shape[0] > p.shape[0]: randnum = np.asarray([ np.random.multinomial(nn, p.squeeze(), size=size) for nn in n ]) elif n.shape[0] < p.shape[0]: randnum = np.asarray([ np.random.multinomial(n.squeeze(), pp, size=size) for pp in p ]) else: randnum = np.asarray([ np.random.multinomial(nn, pp, size=size) for (nn, pp) in zip(n, p) ]) return randnum.astype(original_dtype) def random(self, point=None, size=None): n, p = draw_values([self.n, self.p], point=point) samples = generate_samples(self._random, n, p, dist_shape=self.shape, size=size) return samples def logp(self, x): n = self.n p = self.p return bound( tt.sum(factln(n)) - tt.sum(factln(x)) + tt.sum(x * tt.log(p)), tt.all(x >= 0), tt.all(tt.eq(tt.sum(x, axis=-1, keepdims=True), n)), tt.all(p <= 1), tt.all(tt.eq(tt.sum(p, axis=-1), 1)), tt.all(tt.ge(n, 0)), broadcast_conditions=False ) Multinomial.dist(1,np.ones(3)/3,shape=(6, 3)).mode.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).p.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).n.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).mean.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).random() #%% counts =np.asarray([[19, 50, 37], [21, 67, 55], [11, 53, 38], [17, 54, 45], [24, 93, 66], [27, 53, 70]]) Multinomial.dist(n,p,shape=(6, 3)).logp(x=counts).eval() #%% with pm.Model() as model: like = Multinomial('obs_ABC', n, p, observed=counts, shape=counts.shape) #%% paramall = ( [[.25, .25, .25, .25], 4, 2], [[.25, .25, .25, .25], (1, 4), 3], # 3: expect to fail # [[.25, .25, .25, .25], (10, 4)], [[.25, .25, .25, .25], (10, 1, 4), 5], # 5: expect to fail # [[[.25, .25, .25, .25]], (2, 4), [7, 11]], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (2, 4), 13], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (2, 4), [17, 19]], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (1, 2, 4), [23, 29]], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (10, 2, 4), [31, 37]], ) for p, shape, n in paramall: with pm.Model() as model: m = Multinomial('m', n=n, p=np.asarray(p), shape=shape) print(m.random().shape) #%% counts =np.asarray([[19, 50, 37], [21, 67, 55], [11, 53, 38], [17, 54, 45], [24, 93, 66], [27, 53, 70]]) n = np.array([[106], [143], [102], [116], [183], [150]]) sparsity=1 #not zero beta=np.ones(counts.shape) #input for dirichlet with pm.Model() as model: theta=pm.Dirichlet('theta',beta/sparsity, shape = counts.shape) transition=pm.Multinomial('transition',n,theta,observed=counts) trace=pm.sample(1000) #%% import numpy as np import pymc3 as pm import theano.tensor as tt def norm_simplex(p): """Sum-to-zero transformation.""" return (p.T / p.sum(axis=-1)).T def ccmodel(beta, x): """Community composition model.""" return norm_simplex(tt.exp(tt.dot(x, tt.log(beta)))) class DirichletMultinomial(pm.Discrete): """Dirichlet Multinomial Model """ def __init__(self, alpha, *args, **kwargs): super(DirichletMultinomial, self).__init__(*args, **kwargs) self.alpha = alpha def logp(self, x): alpha = self.alpha n = tt.sum(x, axis=-1) sum_alpha = tt.sum(alpha, axis=-1) const = (tt.gammaln(n + 1) + tt.gammaln(sum_alpha)) - tt.gammaln(n + sum_alpha) series = tt.gammaln(x + alpha) - (tt.gammaln(x + 1) + tt.gammaln(alpha)) result = const + tt.sum(series, axis=-1) return result def as_col(x): if isinstance(x, tt.TensorVariable): return x.dimshuffle(0, 'x') else: return np.asarray(x).reshape(-1, 1) def as_row(x): if isinstance(x, tt.TensorVariable): return x.dimshuffle('x', 0) else: return np.asarray(x).reshape(1, -1) n, k, r = 25, 10, 2 x = np.random.randint(0, 1000, size=(n, k)) y = np.random.randint(0, 1000, size=n) design = np.vstack((np.ones(25), np.random.randint(2, size=n))).T with pm.Model() as model: # Community composition pi = pm.Dirichlet('pi', np.ones(k), shape=(r, k)) comp = pm.Deterministic('comp', ccmodel(pi, design)) # Inferred population density of observed taxa (hierarchical model) rho = pm.Normal('rho', shape=r) tau = pm.Lognormal('tau') dens = pm.Lognormal('dens', tt.dot(design, rho), tau=tau, shape=n) # Community composition *with* the spike expected_recovery = as_col(1 / dens) _comp = norm_simplex(tt.concatenate((comp, expected_recovery), axis=1)) # Variability mu = pm.Lognormal('mu') # Data obs = DirichletMultinomial('obs', _comp * mu, observed=tt.concatenate((x, as_col(y)), axis=1)) pm.sample(1000)
8,455
3,796
from useful import * from os import system def remove_implication(formula): while ">" in formula: operator = formula.find(">") print(formula, operator) subform_left = get_subform_left(formula, operator) subform_right = get_subform_right(formula, operator) formula = get_remove_implication(formula, subform_left, subform_right, operator) return formula def get_remove_implication(formula, subform_left, subform_right, operator): # ...(A>B)... |-> ...(-A#B)... no_modification_right = formula[operator + len(subform_right) + 1:] no_modification_left = formula[:operator - len(subform_left)] return f"{no_modification_left}-{subform_left}#{subform_right}{no_modification_right}" def morgan_law(formula): while "-(" in formula: index = formula.find("-(") print(formula, index) operator = get_operator(formula, index + 1) subform_left = get_subform_left(formula, operator) subform_right = get_subform_right(formula, operator) formula = get_morgan_law(formula, subform_left, subform_right, operator) return formula def get_morgan_law(formula, subform_left, subform_right, operator): # ...-(A&B)... |-> ...(-A#-B)... # ...-(A#B)... |-> ...(-A&-B)... match formula[operator]: case "#": new_operator = "&" case "&": new_operator = "#" no_modification_right = formula[operator + len(subform_right) + 1:] no_modification_left = formula[:operator - len(subform_left) - 1 - 1] return f"{no_modification_left}(-{subform_left}{new_operator}-{subform_right}{no_modification_right}" def remove_double_negation(formula): # --A |-> A formula = formula.replace("--", "") return formula def distributivity(formula): index = 0 while index < len(formula): # Existir "#(" ou ")#" é apenas a primeira condição para se aplicar a distributividade # A segunda condição é existir "#(A&B)" ou "(A&B)#" if "#(" in formula[index:index + 2]: # "#(" operator_and = get_operator(formula, index + 1) if formula[operator_and] == "&": # "#(A&B)" print(formula, index, operator_and) formula, index = get_distributivity_lr(formula, index, operator_and) if ")#" in formula[index:index + 2]: # "(#" len_subform_left = len(get_subform_left(formula, index + 1)) operator_and = get_operator(formula, index + 1 - len_subform_left) if formula[operator_and] == "&": # "(A&B)#" print(formula, index + 1, operator_and) formula, index = get_distributivity_rl(formula, index + 1, operator_and) index += 1 return formula def get_distributivity_lr(formula, operator_or, operator_and): # ...(A#(B&C))... |-> ...((A#B)&(A#C))... # Parenteses externo da fórmula subform_left = get_subform_left(formula, operator_or) no_modification_left = formula[:operator_or - len(subform_left)] subform_right = get_subform_right(formula, operator_or) no_modification_right = formula[operator_or + len(subform_right) + 1:] # Parenteses interno da fórmula subform_middle = get_subform_left(formula, operator_and) subform_right = get_subform_right(formula, operator_and) return f"{no_modification_left}({subform_left}#{subform_middle})&({subform_left}#{subform_right}){no_modification_right}", 0 def get_distributivity_rl(formula, operator_or, operator_and): # ...((A&B)#C)... |-> ...((A#C)&(B#C))... # Parenteses externo da fórmula subform_left = get_subform_left(formula, operator_or) no_modification_left = formula[:operator_or - len(subform_left)] subform_right = get_subform_right(formula, operator_or) no_modification_right = formula[operator_or + len(subform_right) + 1:] # Parenteses interno da fórmula subform_left = get_subform_left(formula, operator_and) subform_middle = get_subform_right(formula, operator_and) return f"{no_modification_left}({subform_left}#{subform_right})&({subform_middle}#{subform_right}){no_modification_right}", 0 def distributivity_new_aton(formula): index = 0 while index < len(formula): # Existir "#(" ou ")#" é apenas a primeira condição para se aplicar a distributividade # A segunda condição é existir "#(A&B)" ou "(A&B)#" if "#(" in formula[index:index + 2]: # "#(" operator_and = get_operator(formula, index + 1) if formula[operator_and] == "&": # "#(A&B)" print(formula, index, operator_and) formula, index = get_distributivity_new_atom_lr(formula, index, operator_and) if ")#" in formula[index:index + 2]: # "(#" len_subform_left = len(get_subform_left(formula, index + 1)) operator_and = get_operator(formula, index + 1 - len_subform_left) if formula[operator_and] == "&": # "(A&B)#" print(formula, index + 1, operator_and) formula, index = get_distributivity_new_atom_rl(formula, index + 1, operator_and) index += 1 return formula def get_distributivity_new_atom_lr(formula, operator_or, operator_and): # ...(A#(B&C))... |-> ...(((A#p)&((¬p#B)&(¬p#C)))&((¬B#¬C)#p))... # Parenteses externo da fórmula subform_left = get_subform_left(formula, operator_or) no_modification_left = formula[:operator_or - len(subform_left)] subform_right = get_subform_right(formula, operator_or) no_modification_right = formula[operator_or + len(subform_right) + 1:] # Parenteses interno da fórmula subform_middle = get_subform_left(formula, operator_and) subform_right = get_subform_right(formula, operator_and) new_operator = get_unprecedented(formula) return f"{no_modification_left}(({subform_left}#{new_operator})&((¬{new_operator}#{subform_middle})&(¬{new_operator}#{subform_right})))&((¬{subform_middle}#¬{subform_right})#{new_operator}){no_modification_right}", 0 #return f"{no_modification_left}({subform_left}#{new_operator})&(¬{new_operator}#{subform_middle})&(¬{new_operator}#{subform_right})&(¬{subform_middle}#¬{subform_right}#{new_operator}){no_modification_right}", 0 def get_distributivity_new_atom_rl(formula, operator_or, operator_and): # ...((A&B)#C)... |-> ...(((C#p)&((¬p#A)&(¬p#B)))&((¬A#¬B)#p))... # Parenteses externo da fórmula subform_left = get_subform_left(formula, operator_or) no_modification_left = formula[:operator_or - len(subform_left)] subform_right = get_subform_right(formula, operator_or) no_modification_right = formula[operator_or + len(subform_right) + 1:] # Parenteses interno da fórmula subform_left = get_subform_left(formula, operator_and) subform_middle = get_subform_right(formula, operator_and) new_operator = get_unprecedented(formula) return f"{no_modification_left}(({subform_right}#{new_operator})&((¬{new_operator}#{subform_left})&(¬{new_operator}#{subform_middle})))&((¬{subform_left}#¬{subform_middle})#{new_operator}){no_modification_right}", 0 #return f"{no_modification_left}({subform_right}#{new_operator})&(¬{new_operator}#{subform_left})&(¬{new_operator}#{subform_middle})&(¬{subform_left}#¬{subform_middle}#{new_operator}){no_modification_right}", 0 if __name__ == "__main__": system("cls") #system("clear") while(True): formula = input("Fórmula: ") if formula == 'q': break print(formula) print("Removendo implicações: ") A1 = remove_implication(formula) print(A1) print("Aplicando Lei de Morgan: ") A2 = morgan_law(A1) print(A2) print("Removendo dupla negação: ") A3 = remove_double_negation(A2) print(A3) print("Aplicando distributividade: ") A4 = distributivity(A3) print(A4) print("Aplicando distributividade com novo átomo: ") A5 = distributivity_new_aton(A3) print(A5) system("pause")
8,054
2,831
# Copyright 2018 Oliver Berger # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Core features """ import asyncio import concurrent import logging import os import time from implant import core log = logging.getLogger(__name__) class Echo(core.Command): """Demonstrate the basic command API.""" data = core.Parameter(default='ping', description='Meaningful data.') async def local(self, context): # custom protocol # first: send await context.channel.send_iteration("send to remote") # second: receive from_remote = [] async for x in context.channel: from_remote.append(x) log.debug("************ receiving from remote: %s", from_remote) # third: wait for remote to finish and return result remote_result = await context.remote_future result = { 'local_data': self.data, 'from_remote': ''.join(from_remote), } result.update(remote_result) return result remote = core.CommandRemote('implant.commands.remotes.Echo') class SystemLoad(core.Command): async def local(self, context): t, load = await context.remote_future return t, load async def remote(self, context): t, load = time.time(), os.getloadavg() return t, load class Copy(core.Command): src = core.Parameter(description='Source file at local side.') dest = core.Parameter(description='Desatination file at remote side.') def __init__(self, *args, **kwargs): super(Copy, self).__init__(*args, **kwargs) self.executor = concurrent.futures.ThreadPoolExecutor() self.loop = asyncio.get_event_loop() def __del__(self): self.executor.shutdown(wait=True) async def local(self, context): with open(self.src, "rb") as f: while True: data = await self.loop.run_in_executor(self.executor, f.read, 0x8000) if not data: context.channel.send(StopAsyncIteration()) break await context.channel.send(data) result = await context.remote_future return result async def remote(self, context): with open(self.dest, "wb") as f: async for data in context.channel: await self.loop.run_in_executor(self.executor, f.write, data)
2,896
837
from trainer.utils.losses import * from trainer.utils import custom_ssim
73
23
############################################################################## # # Copyright (c) 2005 Zope Foundation and Contributors. # # This software is subject to the provisions of the Zope Public License, # Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution. # THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED # WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS # FOR A PARTICULAR PURPOSE. # ############################################################################## """Names exported by the ZopeTestCase package """ import ZopeLite as Zope2 import utils import layer from ZopeLite import hasProduct from ZopeLite import installProduct from ZopeLite import hasPackage from ZopeLite import installPackage from ZopeLite import _print from ZopeTestCase import folder_name from ZopeTestCase import user_name from ZopeTestCase import user_password from ZopeTestCase import user_role from ZopeTestCase import standard_permissions from ZopeTestCase import ZopeTestCase from ZopeTestCase import FunctionalTestCase from PortalTestCase import portal_name from PortalTestCase import PortalTestCase from sandbox import Sandboxed from functional import Functional from base import TestCase from base import app from base import close from warnhook import WarningsHook from unittest import main from zopedoctest import ZopeDocTestSuite from zopedoctest import ZopeDocFileSuite from zopedoctest import FunctionalDocTestSuite from zopedoctest import FunctionalDocFileSuite import zopedoctest as doctest import transaction import placeless Zope = Zope2
1,693
496
# Copyright (C) 2010-2011 Richard Lincoln # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from CIM14.CPSM.Equipment.LoadModel.EnergyArea import EnergyArea class SubLoadArea(EnergyArea): """The class is the second level in a hierarchical structure for grouping of loads for the purpose of load flow load scaling. """ def __init__(self, LoadGroups=None, LoadArea=None, *args, **kw_args): """Initialises a new 'SubLoadArea' instance. @param LoadGroups: The Loadgroups in the SubLoadArea. @param LoadArea: The LoadArea where the SubLoadArea belongs. """ self._LoadGroups = [] self.LoadGroups = [] if LoadGroups is None else LoadGroups self._LoadArea = None self.LoadArea = LoadArea super(SubLoadArea, self).__init__(*args, **kw_args) _attrs = [] _attr_types = {} _defaults = {} _enums = {} _refs = ["LoadGroups", "LoadArea"] _many_refs = ["LoadGroups"] def getLoadGroups(self): """The Loadgroups in the SubLoadArea. """ return self._LoadGroups def setLoadGroups(self, value): for x in self._LoadGroups: x.SubLoadArea = None for y in value: y._SubLoadArea = self self._LoadGroups = value LoadGroups = property(getLoadGroups, setLoadGroups) def addLoadGroups(self, *LoadGroups): for obj in LoadGroups: obj.SubLoadArea = self def removeLoadGroups(self, *LoadGroups): for obj in LoadGroups: obj.SubLoadArea = None def getLoadArea(self): """The LoadArea where the SubLoadArea belongs. """ return self._LoadArea def setLoadArea(self, value): if self._LoadArea is not None: filtered = [x for x in self.LoadArea.SubLoadAreas if x != self] self._LoadArea._SubLoadAreas = filtered self._LoadArea = value if self._LoadArea is not None: if self not in self._LoadArea._SubLoadAreas: self._LoadArea._SubLoadAreas.append(self) LoadArea = property(getLoadArea, setLoadArea)
3,132
950
#!/usr/bin/env python import sys HATOMS = ["HG", "HD", "HE", "HH"] lines = open(sys.argv[1]).readlines() for line in lines: if line[:6] == "ATOM " or line[:6] == "HETATM": if line[17:20] == "WAT": continue if line[13] == "H": continue if line[12:14] in HATOMS: continue print(line.strip("\n"))
369
142
import Pmw import os import re from FastIndex import FastIndex, timer from PrologFrame import PrologFrame from TerminalFrame import TerminalFrame class OutputBook(Pmw.NoteBook): def __init__(self, master=None): self.app = master Pmw.NoteBook.__init__(self, self.app.bottom,createcommand=self.create_page) self.spec_page = self.add('Specialised File') self.memo_page = self.add('Memo Table') self.gx_page = self.add('Generating Extension') self.output_page = self.add('Output') self.console_page = self.add('Console') #self.new_terminal_page = self.add('Terminal 1') #spec file self.output_spec = PrologFrame(self.spec_page,"",app=self.app) self.output_spec.pack(side="bottom", fill="both", expand="yes") #memo file self.output_memo = PrologFrame(self.memo_page,"",app=self.app) self.output_memo.pack(side="bottom", fill="both", expand="yes") #gx file self.output_gx = PrologFrame(self.gx_page,"",app=self.app) self.output_gx.pack(side="bottom", fill="both", expand="yes") #output self.output_out = PrologFrame(self.output_page,"",app=self.app) self.output_out.pack(side="bottom", fill="both", expand="yes") #console self.output_console = PrologFrame(self.console_page,"", app=self.app) self.output_console.pack(side="bottom", fill="both", expand="yes") self.pack(side="bottom", fill="both", expand="yes") self.output_spec.text.tag_bind("nametag", "<Motion>", self.mouse_over) self.output_spec.text.tag_bind("nametag", "<Leave>", self.mouse_leave) self.terminal_pages = [] self.terminals = [] self.term_count = 0 self.create_new_terminal() def create_page(self, pagename): if pagename.startswith("Terminal"): i = self.terminal_pages.index(pagename) page = self.page(self.index(pagename)) self.terminals[i] = TerminalFrame(page, app=self.app, id=pagename) self.terminals[i].pack(side="bottom", fill="both", expand="yes") self.app.update_completions() #self.selectpage(pagename) def get_console_stream(self): pass def write_to_console(self,string): self.output_console.text.config(state="normal") self.output_console.text.insert("end", string) self.output_console.text.config(state="disabled") def set_font(self, font): self.output_spec.text["font"] = font self.output_memo.text["font"] = font self.output_gx.text["font"] = font self.output_out.text["font"] = font def view_spec_output(self, filename): (root, ext) = os.path.splitext(filename) fast_idx = self.output_spec.load_source(root + ".spec") self.output_memo.load_source(root + ".memo") self.output_gx.load_source(root + ".gx") self.output_out.clear() funcs = self.get_spec_funcs() spec_lines = self.output_spec.text.get(1.0, 'end') pos = 0 while 1: id = next_id(spec_lines, pos) if id == (): break (start, end) = id pos = end (start_idx, end_idx) = fast_idx.get_two_tk_indices_same_line(start, end) self.output_spec.text.tag_add("nametag", start_idx, end_idx) def reset_output(self, gx=None): if gx is None: pass #self.output_gx.clear() elif os.path.exists(gx): self.output_gx.load_source(gx) else: self.output_gx.clear() self.output_memo.clear() self.output_spec.clear() self.output_out.clear() self.selectpage('Specialised File') def set_output_from_file(self, filename): self.output_out.load_source(filename) def set_output(self, text): self.output_out.set_text(text) def view_output(self): self.selectpage('Output') def get_spec_funcs(self): memo_text = self.output_memo.text.get(1.0, 'end').split('\n') funcs = {} for line in memo_text: if line.startswith('table'): (orig, i) = up_to_comma(line[6:]) (pattern, _) = up_to_comma(line[i + 7:]) i = pattern.find('(') if i > 0: name = pattern[:i] else: name = pattern funcs[name] = orig + " --> " + pattern self.funcs = funcs def get_tag_position(self, x, y): index = self.output_spec.text.index("@"+str(x)+","+str(y)+" + 1 char") return self.output_spec.text.tag_prevrange("nametag", index) start = None def mouse_over(self, event): (start, end) = self.get_tag_position(event.x, event.y) predicate = self.output_spec.text.get(start, end) #print "over " + start + ", " + end + " : " + predicate #print self.funcs[predicate] if self.start != start: self.app.balloon.configure(relmouse="both",yoffset=15) self.app.balloon._showBalloon(self.output_spec.text, self.funcs[predicate], False) self.start = start def mouse_leave(self, event): self.app.balloon.configure(relmouse="none",yoffset=1) self.app.balloon.withdraw() self.start = None def create_new_terminal(self): self.term_count += 1 self.terminal_pages.append('Terminal ' + str(self.term_count)) page = self.add(self.terminal_pages[-1]) self.terminals.append(None) def quit(self): for t in self.terminals: if t is not None: t.quit() def kill_terminal(self, term_str): i = self.terminal_pages.index(term_str) self.terminals[i].quit() self.delete(term_str) self.terminals.pop(i) self.terminal_pages.pop(i) def reset_completions(self): for t in self.terminals: if t is not None: t.reset_completions() def add_completions(self, completions): for t in self.terminals: if t is not None: t.add_completions(completions) def up_to_comma(str): bracket_stack = [] i = 0 current_char = str[i] in_string = False in_double_string = False ret_string = '' while len(bracket_stack) > 0 or current_char != ',' or in_string or in_double_string: if current_char == '(' or current_char == '[' or current_char == '{': bracket_stack.append(current_char) elif current_char == ')' or current_char == ']' or current_char == '}': bracket_stack = bracket_stack[:-1] elif current_char == '"': if in_double_string: in_double_string = False elif not in_string: in_double_string = True elif current_char == "'": if in_string: in_string = False elif not in_double_string: in_string = True ret_string = ret_string + current_char i = i + 1 current_char = str[i] return (ret_string.strip(), i) regexp = re.compile('[a-zA-z0-9_]+__[0-9]+') def next_id(string, pos): match = regexp.search(string[pos:]) if match is None: return () else: return (match.start() + pos, match.end() + pos)
7,693
2,453
import sqlite3 with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute( """ CREATE TABLE IF NOT EXISTS Produtos ( id INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT, name TEXT NOT NULL, price REAL, compra_id INTEGER, FOREIGN KEY (compra_id) REFERENCES Compras(id) ); """ ) cursor.execute( """ CREATE TABLE IF NOT EXISTS Compras ( id INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT, date TEXT NOT NULL ); """ ) class Produto(object): def getAll(self): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute("SELECT * FROM Produtos;") return [ { "id": items[0], "name": items[1], "price": items[2], "compra_id": items[3] } for items in cursor.fetchall() ] def getByCompra(self, compraId): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() print(f"SELECT * FROM Produtos WHERE compra_id = {compraId}") cursor.execute(f"SELECT * FROM Produtos WHERE compra_id = {compraId}") return [ { "id": items[0], "name": items[1], "price": items[2], } for items in cursor.fetchall() ] def insert(self, *args): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() print(f"INSERT INTO Produtos (name, price, compra_id) VALUES ('{args[0]}', {args[1]}, {args[2]})") cursor.execute(f"INSERT INTO Produtos (name, price, compra_id) VALUES ('{args[0]}', {args[1]}, {args[2]})") def getById(self, id): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute(f"SELECT * FROM Produtos WHERE id = {id} ;") return [ { "id": items[0], "name": items[1], "price": items[2] } for items in cursor.fetchall() ][0] def update(self, id, *args): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute(f"UPDATE Produtos SET name = {args[0]}, price = {args[1]}, compra_id = {args[2]} WHERE id = {id};") def delete(self, id): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute(f"DELETE FROM Produtos WHERE id = {id}") def deleteByCompra(self, compraId): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() print(f"DELETE FROM Produtos WHERE compra_id = {compraId}") cursor.execute(f"DELETE FROM Produtos WHERE compra_id = {compraId}") class Compra(object): def __init__(self): self.produto = Produto() def getAll(self): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() print("SELECT * FROM Compras;") cursor.execute("SELECT * FROM Compras;") return [ { "id": items[0], "date": items[1], "produtos": self.produto.getByCompra(items[0]) } for items in cursor.fetchall() ] def insert(self, *args): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() print(f"INSERT INTO Compras (date) VALUES ('{args[0]}')") cursor.execute(f"INSERT INTO Compras (date) VALUES ('{args[0]}')") c = self.getAll()[-1] ps = list(args[1]) for p in ps: self.produto.insert(str(p["name"]), p["price"], c["id"]) # return self.getById(c.id) def getById(self, id): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute(f"SELECT * FROM Compras WHERE id = {id} ;") return [ { "id": items[0], "date": items[1], "produtos": self.produto.getByCompra(id) } for items in cursor.fetchall() ][0] def getByDate(self, date): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute(f"SELECT * FROM Compras WHERE date = '{date}' ;") return [ { "id": items[0], "date": items[1], "produtos": self.produto.getByCompra(items[0]) } for items in cursor.fetchall() ] def update(self, id, *args): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute( """ UPDATE Compras SET date = ?, produto_id = ? WHERE id = ?; """, (*args, id) ) def delete(self, id): with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() self.produto.deleteByCompra(self.getById(id)["id"]) print(f"DELETE FROM Compras WHERE id = {id}") cursor.execute(f"DELETE FROM Compras WHERE id = {id}")
5,468
1,583
disable_gtk_binaries = True def gtk_dependent_cc_library(**kwargs): if not disable_gtk_binaries: native.cc_library(**kwargs) def gtk_dependent_cc_binary(**kwargs): if not disable_gtk_binaries: native.cc_binary(**kwargs)
246
87
import sys from gensim import models from gensim.models.doc2vec import LabeledSentence import pickle def corpus_to_sentences(corpus): sentences = [] for idx, (name, doc) in enumerate(corpus.items()): sys.stdout.write('\r前処理中 {}/{}'.format(idx, len(corpus))) sentence = LabeledSentence(words=doc, tags=[name]) sentences.append(sentence) return sentences with open('corpus_text', 'rb') as f: corpus = pickle.load(f) sentences = corpus_to_sentences(corpus) model = models.Doc2Vec(vector_size=400, window=15, alpha=.025, min_alpha=.025, min_count=1, sample=1e-6) model.build_vocab(sentences) print(len(corpus)) model.train(sentences, total_examples=len(corpus), epochs=20) model.save('doc2vec.model')
743
277
# -*- encoding: utf-8 -*- from collections import UserDict from itertools import count import shutil import winreg import uuid PATH = "Software\\Microsoft\\Windows\\CurrentVersion\\Lxss" KEY = winreg.HKEY_CURRENT_USER class RegistryDescriptor(object): def __init__(self, name): self.name = name def __get__(self, instance, clazz): key = instance._key() if key is not None: return self._get_value_by_registry(key) return self._get_value_by_vartable(instance, key) def __set__(self, instance, value): key = instance._key("", winreg.KEY_WRITE) if key is not None: return self._set_value_by_registry(key, value) def _get_value_by_registry(self, key): with key as k: try: value, _ = winreg.QueryValueEx(k, self.name) except FileNotFoundError: return None return value def _set_value_by_registry(self, key, value): if isinstance(value, int): type = winreg.REG_DWORD elif isinstance(value, (list, tuple)): type = winreg.REG_MULTI_SZ else: type = winreg.REG_SZ with key as k: winreg.SetValueEx(k, self.name, 0, type, value) def _get_value_by_vartable(self, instance, key): return vars(instance)[key] class EnvironmentVariables(UserDict): def __init__(self, distribution): super(EnvironmentVariables, self).__init__() self.distribution = distribution self.reload() def _save_values(self): return (f"{v[0]}={v[1]}" for v in self.data.items()) def save(self): self.distribution.default_environment = list(self._save_values()) def reload(self): self.clear() self.update(dict( v.split("=", 1) for v in self.distribution.default_environment )) class Distribution(object): @classmethod def create(cls, name, source_path): guid = "{%s}"%uuid.uuid4() with winreg.CreateKey(KEY, f"{PATH}\\{guid}") as k: winreg.SetValueEx(k, 'State', 0, winreg.REG_DWORD, 1) winreg.SetValueEx(k, 'DistributionName', 0, winreg.REG_SZ, name) winreg.SetValueEx(k, 'BasePath', 0, winreg.REG_SZ, source_path) winreg.SetValueEx(k, 'DefaultUid', 0, winreg.REG_DWORD, 0) winreg.SetValueEx(k, 'Version', 0, winreg.REG_DWORD, 1) winreg.SetValueEx(k, 'KernelCommandLine', 0, winreg.REG_SZ, 'BOOT_IMAGE=/kernel init=/init ro') winreg.SetValueEx(k, 'DefaultEnvironment', 0, winreg.REG_MULTI_SZ, [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", ]) return cls(guid) def __init__(self, guid=""): self.guid = guid def _key(self, sub="", privileges=winreg.KEY_READ): if not self.guid: return None if sub: sub = "\\" + sub return winreg.OpenKey(KEY, PATH+f"\\{self.guid}"+sub, 0, privileges) name = RegistryDescriptor("DistributionName") base_path = RegistryDescriptor("BasePath") default_user = RegistryDescriptor("DefaultUid") default_environment = RegistryDescriptor("DefaultEnvironment") cmdline = RegistryDescriptor("KernelCommandLine") flags = RegistryDescriptor("Flags") package_family_name = RegistryDescriptor("PackageFamilyName") _state = RegistryDescriptor("State") version = RegistryDescriptor("Version") @property def environment(self): return EnvironmentVariables(self) def launch_params(self, params=("/bin/bash",)): return [shutil.which("wsl.exe"), f"{self.guid}"] + list(params) def __repr__(self): return f"<Distribution '{self.name}' guid:{self.guid}>" def delete(self): with Lxss._key('', winreg.KEY_WRITE) as k: winreg.DeleteKey(k, self.guid) @property def state(self): st = self._state if st == 1: return "Ready" elif st == 3: return "Installing" return "Unknown:" + str(st) @state.setter def state(self, value): if isinstance(value, int): self._state = value return value = value.lower() if value == "ready": self._state = 1 elif value == "installing": self._state = 3 else: self._state = value def __enter__(self): self._state = 3 return self def __exit__(self, *exc): self._state = 1 return False class _Lxss(object): def _key(self, sub="", privileges=winreg.KEY_READ): if sub: sub = "\\" + sub return winreg.OpenKey(KEY, PATH+sub, 0, privileges) default_distribution = RegistryDescriptor("DefaultDistribution") @property def default(self): return Distribution(self.default_distribution) @default.setter def default(self, value): self.default_distribution = value.guid def __iter__(self): for i in count(): with self._key() as k: try: name = winreg.EnumKey(k, i) except OSError as e: if e.winerror != 259: raise break yield Distribution(name) def get(self, value, default=None): for distribution in self: if value.startswith("{") and value.endswith("}"): if distribution.guid.lower() == value.lower(): return distribution else: if distribution.name == value: return distribution return default def __getitem__(self, value): value = self.get(value, None) if value is None: raise KeyError("Unknown distribution") return value Lxss = _Lxss()
6,206
1,982
# changelog.py -- Python module for Debian changelogs # Copyright (C) 2006-7 James Westby <jw+debian@jameswestby.net> # Copyright (C) 2008 Canonical Ltd. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA # The parsing code is based on that from dpkg which is: # Copyright 1996 Ian Jackson # Copyright 2005 Frank Lichtenheld <frank@lichtenheld.de> # and licensed under the same license as above. """This module implements facilities to deal with Debian changelogs.""" from __future__ import absolute_import import os import pwd import re import socket import warnings import sys import six from debian import debian_support # Python 3 doesn't have StandardError, but let's avoid changing our # exception inheritance hierarchy for Python 2. try: _base_exception_class = StandardError except NameError: _base_exception_class = Exception class ChangelogParseError(_base_exception_class): """Indicates that the changelog could not be parsed""" is_user_error = True def __init__(self, line): self._line=line def __str__(self): return "Could not parse changelog: "+self._line class ChangelogCreateError(_base_exception_class): """Indicates that changelog could not be created, as all the information required was not given""" class VersionError(_base_exception_class): """Indicates that the version does not conform to the required format""" is_user_error = True def __init__(self, version): self._version=version def __str__(self): return "Could not parse version: "+self._version # TODO(jsw): Remove this in favor of using debian_support.Version directly. I # don't think we gain anything by using this empty subclass. class Version(debian_support.Version): """Represents a version of a Debian package.""" # debian_support.Version now has all the functionality we need class ChangeBlock(object): """Holds all the information about one block from the changelog.""" def __init__(self, package=None, version=None, distributions=None, urgency=None, urgency_comment=None, changes=None, author=None, date=None, other_pairs=None, encoding='utf-8'): self._raw_version = None self._set_version(version) self.package = package self.distributions = distributions self.urgency = urgency or "unknown" self.urgency_comment = urgency_comment or '' self._changes = changes self.author = author self.date = date self._trailing = [] self.other_pairs = other_pairs or {} self._encoding = encoding self._no_trailer = False self._trailer_separator = " " def _set_version(self, version): if version is not None: self._raw_version = str(version) def _get_version(self): return Version(self._raw_version) version = property(_get_version, _set_version) def other_keys_normalised(self): norm_dict = {} for (key, value) in other_pairs.items(): key = key[0].upper() + key[1:].lower() m = xbcs_re.match(key) if m is None: key = "XS-%s" % key norm_dict[key] = value return norm_dict def changes(self): return self._changes def add_trailing_line(self, line): self._trailing.append(line) def add_change(self, change): if self._changes is None: self._changes = [change] else: #Bit of trickery to keep the formatting nicer with a blank #line at the end if there is one changes = self._changes changes.reverse() added = False for i in range(len(changes)): m = blankline.match(changes[i]) if m is None: changes.insert(i, change) added = True break changes.reverse() if not added: changes.append(change) self._changes = changes def _get_bugs_closed_generic(self, type_re): changes = six.u(' ').join(self._changes) bugs = [] for match in type_re.finditer(changes): closes_list = match.group(0) for match in re.finditer(r"\d+", closes_list): bugs.append(int(match.group(0))) return bugs @property def bugs_closed(self): return self._get_bugs_closed_generic(closes) @property def lp_bugs_closed(self): return self._get_bugs_closed_generic(closeslp) def _format(self): # TODO(jsw): Switch to StringIO or a list to join at the end. block = "" if self.package is None: raise ChangelogCreateError("Package not specified") block += self.package + " " if self._raw_version is None: raise ChangelogCreateError("Version not specified") block += "(" + self._raw_version + ") " if self.distributions is None: raise ChangelogCreateError("Distribution not specified") block += self.distributions + "; " if self.urgency is None: raise ChangelogCreateError("Urgency not specified") block += "urgency=" + self.urgency + self.urgency_comment for (key, value) in self.other_pairs.items(): block += ", %s=%s" % (key, value) block += '\n' if self.changes() is None: raise ChangelogCreateError("Changes not specified") for change in self.changes(): block += change + "\n" if not self._no_trailer: if self.author is None: raise ChangelogCreateError("Author not specified") if self.date is None: raise ChangelogCreateError("Date not specified") block += " -- " + self.author + self._trailer_separator \ + self.date + "\n" for line in self._trailing: block += line + "\n" return block if sys.version >= '3': __str__ = _format def __bytes__(self): return str(self).encode(self._encoding) else: __unicode__ = _format def __str__(self): return unicode(self).encode(self._encoding) topline = re.compile(r'^(\w%(name_chars)s*) \(([^\(\) \t]+)\)' r'((\s+%(name_chars)s+)+)\;' % {'name_chars': '[-+0-9a-z.]'}, re.IGNORECASE) blankline = re.compile(r'^\s*$') change = re.compile(r'^\s\s+.*$') endline = re.compile(r'^ -- (.*) <(.*)>( ?)((\w+\,\s*)?\d{1,2}\s+\w+\s+' r'\d{4}\s+\d{1,2}:\d\d:\d\d\s+[-+]\d{4}\s*)$') endline_nodetails = re.compile(r'^ --(?: (.*) <(.*)>( ?)((\w+\,\s*)?\d{1,2}' r'\s+\w+\s+\d{4}\s+\d{1,2}:\d\d:\d\d\s+[-+]\d{4}' r'))?\s*$') keyvalue= re.compile(r'^([-0-9a-z]+)=\s*(.*\S)$', re.IGNORECASE) value_re = re.compile(r'^([-0-9a-z]+)((\s+.*)?)$', re.IGNORECASE) xbcs_re = re.compile('^X[BCS]+-', re.IGNORECASE) emacs_variables = re.compile(r'^(;;\s*)?Local variables:', re.IGNORECASE) vim_variables = re.compile('^vim:', re.IGNORECASE) cvs_keyword = re.compile(r'^\$\w+:.*\$') comments = re.compile(r'^\# ') more_comments = re.compile(r'^/\*.*\*/') closes = re.compile(r'closes:\s*(?:bug)?\#?\s?\d+(?:,\s*(?:bug)?\#?\s?\d+)*', re.IGNORECASE) closeslp = re.compile(r'lp:\s+\#\d+(?:,\s*\#\d+)*', re.IGNORECASE) old_format_re1 = re.compile(r'^(\w+\s+\w+\s+\d{1,2} \d{1,2}:\d{1,2}:\d{1,2}' r'\s+[\w\s]*\d{4})\s+(.*)\s+(<|\()(.*)(\)|>)') old_format_re2 = re.compile(r'^(\w+\s+\w+\s+\d{1,2},?\s*\d{4})\s+(.*)' r'\s+(<|\()(.*)(\)|>)') old_format_re3 = re.compile(r'^(\w[-+0-9a-z.]*) \(([^\(\) \t]+)\)\;?', re.IGNORECASE) old_format_re4 = re.compile(r'^([\w.+-]+)(-| )(\S+) Debian (\S+)', re.IGNORECASE) old_format_re5 = re.compile('^Changes from version (.*) to (.*):', re.IGNORECASE) old_format_re6 = re.compile(r'^Changes for [\w.+-]+-[\w.+-]+:?\s*$', re.IGNORECASE) old_format_re7 = re.compile(r'^Old Changelog:\s*$', re.IGNORECASE) old_format_re8 = re.compile(r'^(?:\d+:)?\w[\w.+~-]*:?\s*$') class Changelog(object): """Represents a debian/changelog file.""" # TODO(jsw): Avoid masking the 'file' built-in. def __init__(self, file=None, max_blocks=None, allow_empty_author=False, strict=False, encoding='utf-8'): """Initializer. Args: file: The contents of the changelog, either as a str, unicode object, or an iterator of lines (each of which is either a str or unicode) max_blocks: The maximum number of blocks to parse from the input. (Default: no limit) allow_empty_author: Whether to allow an empty author in the trailer line of a change block. (Default: False) strict: Whether to raise an exception if there are errors. (Default: use a warning) encoding: If the input is a str or iterator of str, the encoding to use when interpreting the input. """ self._encoding = encoding self._blocks = [] self.initial_blank_lines = [] if file is not None: self.parse_changelog(file, max_blocks=max_blocks, allow_empty_author=allow_empty_author, strict=strict) def _parse_error(self, message, strict): if strict: raise ChangelogParseError(message) else: warnings.warn(message) def parse_changelog(self, file, max_blocks=None, allow_empty_author=False, strict=True, encoding=None): first_heading = "first heading" next_heading_or_eof = "next heading of EOF" start_of_change_data = "start of change data" more_changes_or_trailer = "more change data or trailer" slurp_to_end = "slurp to end" encoding = encoding or self._encoding if file is None: self._parse_error('Empty changelog file.', strict) return self._blocks = [] self.initial_blank_lines = [] current_block = ChangeBlock(encoding=encoding) changes = [] state = first_heading old_state = None if isinstance(file, bytes): file = file.decode(encoding) if isinstance(file, six.string_types): # Make sure the changelog file is not empty. if len(file.strip()) == 0: self._parse_error('Empty changelog file.', strict) return file = file.splitlines() for line in file: if not isinstance(line, six.text_type): line = line.decode(encoding) # Support both lists of lines without the trailing newline and # those with trailing newlines (e.g. when given a file object # directly) line = line.rstrip('\n') if state == first_heading or state == next_heading_or_eof: top_match = topline.match(line) blank_match = blankline.match(line) if top_match is not None: if (max_blocks is not None and len(self._blocks) >= max_blocks): return current_block.package = top_match.group(1) current_block._raw_version = top_match.group(2) current_block.distributions = top_match.group(3).lstrip() pairs = line.split(";", 1)[1] all_keys = {} other_pairs = {} for pair in pairs.split(','): pair = pair.strip() kv_match = keyvalue.match(pair) if kv_match is None: self._parse_error("Invalid key-value " "pair after ';': %s" % pair, strict) continue key = kv_match.group(1) value = kv_match.group(2) if key.lower() in all_keys: self._parse_error("Repeated key-value: " "%s" % key.lower(), strict) all_keys[key.lower()] = value if key.lower() == "urgency": val_match = value_re.match(value) if val_match is None: self._parse_error("Badly formatted " "urgency value: %s" % value, strict) else: current_block.urgency = val_match.group(1) comment = val_match.group(2) if comment is not None: current_block.urgency_comment = comment else: other_pairs[key] = value current_block.other_pairs = other_pairs state = start_of_change_data elif blank_match is not None: if state == first_heading: self.initial_blank_lines.append(line) else: self._blocks[-1].add_trailing_line(line) else: emacs_match = emacs_variables.match(line) vim_match = vim_variables.match(line) cvs_match = cvs_keyword.match(line) comments_match = comments.match(line) more_comments_match = more_comments.match(line) if ((emacs_match is not None or vim_match is not None) and state != first_heading): self._blocks[-1].add_trailing_line(line) old_state = state state = slurp_to_end continue if (cvs_match is not None or comments_match is not None or more_comments_match is not None): if state == first_heading: self.initial_blank_lines.append(line) else: self._blocks[-1].add_trailing_line(line) continue if ((old_format_re1.match(line) is not None or old_format_re2.match(line) is not None or old_format_re3.match(line) is not None or old_format_re4.match(line) is not None or old_format_re5.match(line) is not None or old_format_re6.match(line) is not None or old_format_re7.match(line) is not None or old_format_re8.match(line) is not None) and state != first_heading): self._blocks[-1].add_trailing_line(line) old_state = state state = slurp_to_end continue self._parse_error("Unexpected line while looking " "for %s: %s" % (state, line), strict) if state == first_heading: self.initial_blank_lines.append(line) else: self._blocks[-1].add_trailing_line(line) elif (state == start_of_change_data or state == more_changes_or_trailer): change_match = change.match(line) end_match = endline.match(line) end_no_details_match = endline_nodetails.match(line) blank_match = blankline.match(line) if change_match is not None: changes.append(line) state = more_changes_or_trailer elif end_match is not None: if end_match.group(3) != ' ': self._parse_error("Badly formatted trailer " "line: %s" % line, strict) current_block._trailer_separator = end_match.group(3) current_block.author = "%s <%s>" \ % (end_match.group(1), end_match.group(2)) current_block.date = end_match.group(4) current_block._changes = changes self._blocks.append(current_block) changes = [] current_block = ChangeBlock(encoding=encoding) state = next_heading_or_eof elif end_no_details_match is not None: if not allow_empty_author: self._parse_error("Badly formatted trailer " "line: %s" % line, strict) continue current_block._changes = changes self._blocks.append(current_block) changes = [] current_block = ChangeBlock(encoding=encoding) state = next_heading_or_eof elif blank_match is not None: changes.append(line) else: cvs_match = cvs_keyword.match(line) comments_match = comments.match(line) more_comments_match = more_comments.match(line) if (cvs_match is not None or comments_match is not None or more_comments_match is not None): changes.append(line) continue self._parse_error("Unexpected line while looking " "for %s: %s" % (state, line), strict) changes.append(line) elif state == slurp_to_end: if old_state == next_heading_or_eof: self._blocks[-1].add_trailing_line(line) else: changes.append(line) else: assert False, "Unknown state: %s" % state if ((state != next_heading_or_eof and state != slurp_to_end) or (state == slurp_to_end and old_state != next_heading_or_eof)): self._parse_error("Found eof where expected %s" % state, strict) current_block._changes = changes current_block._no_trailer = True self._blocks.append(current_block) def get_version(self): """Return a Version object for the last version""" return self._blocks[0].version def set_version(self, version): """Set the version of the last changelog block version can be a full version string, or a Version object """ self._blocks[0].version = Version(version) version = property(get_version, set_version, doc="Version object for last changelog block""") ### For convenience, let's expose some of the version properties full_version = property(lambda self: self.version.full_version) epoch = property(lambda self: self.version.epoch) debian_version = property(lambda self: self.version.debian_revision) debian_revision = property(lambda self: self.version.debian_revision) upstream_version = property(lambda self: self.version.upstream_version) def get_package(self): """Returns the name of the package in the last version.""" return self._blocks[0].package def set_package(self, package): self._blocks[0].package = package package = property(get_package, set_package, doc="Name of the package in the last version") def get_versions(self): """Returns a list of version objects that the package went through.""" return [block.version for block in self._blocks] versions = property(get_versions, doc="List of version objects the package went through") def _raw_versions(self): return [block._raw_version for block in self._blocks] def _format(self): pieces = [] pieces.append(six.u('\n').join(self.initial_blank_lines)) for block in self._blocks: pieces.append(six.text_type(block)) return six.u('').join(pieces) if sys.version >= '3': __str__ = _format def __bytes__(self): return str(self).encode(self._encoding) else: __unicode__ = _format def __str__(self): return unicode(self).encode(self._encoding) def __iter__(self): return iter(self._blocks) def __getitem__(self, n): """ select a changelog entry by number, version string, or Version :param n: integer or str representing a version or Version object """ if type(n) is int: return self._blocks[n] elif type(n) is str: return self[Version(n)] return self._blocks[self.versions.index(n)] def __len__(self): return len(self._blocks) def set_distributions(self, distributions): self._blocks[0].distributions = distributions distributions = property(lambda self: self._blocks[0].distributions, set_distributions) def set_urgency(self, urgency): self._blocks[0].urgency = urgency urgency = property(lambda self: self._blocks[0].urgency, set_urgency) def add_change(self, change): self._blocks[0].add_change(change) def set_author(self, author): self._blocks[0].author = author author = property(lambda self: self._blocks[0].author, set_author) def set_date(self, date): self._blocks[0].date = date date = property(lambda self: self._blocks[0].date, set_date) def new_block(self, **kwargs): kwargs.setdefault('encoding', self._encoding) block = ChangeBlock(**kwargs) block.add_trailing_line('') self._blocks.insert(0, block) def write_to_open_file(self, file): file.write(self.__str__()) def get_maintainer(): """Get the maintainer information in the same manner as dch. This function gets the information about the current user for the maintainer field using environment variables of gecos informations as approriate. It uses the same methods as dch to get the information, namely DEBEMAIL, DEBFULLNAME, EMAIL, NAME, /etc/mailname and gecos. :returns: a tuple of the full name, email pair as strings. Either of the pair may be None if that value couldn't be determined. """ env = os.environ regex = re.compile(r"^(.*)\s+<(.*)>$") # Split email and name if 'DEBEMAIL' in env: match_obj = regex.match(env['DEBEMAIL']) if match_obj: if not 'DEBFULLNAME' in env: env['DEBFULLNAME'] = match_obj.group(1) env['DEBEMAIL'] = match_obj.group(2) if 'DEBEMAIL' not in env or 'DEBFULLNAME' not in env: if 'EMAIL' in env: match_obj = regex.match(env['EMAIL']) if match_obj: if not 'DEBFULLNAME' in env: env['DEBFULLNAME'] = match_obj.group(1) env['EMAIL'] = match_obj.group(2) # Get maintainer's name if 'DEBFULLNAME' in env: maintainer = env['DEBFULLNAME'] elif 'NAME' in env: maintainer = env['NAME'] else: # Use password database if no data in environment variables try: maintainer = re.sub(r',.*', '', pwd.getpwuid(os.getuid()).pw_gecos) except (KeyError, AttributeError): maintainer = None # Get maintainer's mail address if 'DEBEMAIL' in env: email = env['DEBEMAIL'] elif 'EMAIL' in env: email = env['EMAIL'] else: addr = None if os.path.exists('/etc/mailname'): f = open('/etc/mailname') try: addr = f.readline().strip() finally: f.close() if not addr: addr = socket.getfqdn() if addr: user = pwd.getpwuid(os.getuid()).pw_name if not user: addr = None else: addr = "%s@%s" % (user, addr) if addr: email = addr else: email = None return (maintainer, email)
25,235
7,301
# Define a procedure, median, that takes three # numbers as its inputs, and returns the median # of the three numbers. # Make sure your procedure has a return statement. def bigger(a,b): if a > b: return a else: return b def biggest(a,b,c): return bigger(a,bigger(b,c)) def median(a, b ,c): if (b >= a and a >= c) or (c >= a and a >= b): return a if (a >= b and b >= c) or (c >= b and b >= a): return b if (a >= c and c >= b) or (b >= c and c >= a): return c print(median(1,2,3)) #>>> 2 print(median(9,3,6)) #>>> 6 print(median(7,8,7)) #>>> 7
667
275
import requests import os import subprocess import gidgethub from gidgethub import sansio AUTOMERGE_LABEL = ":robot: automerge" def comment_on_pr(issue_number, message): """ Leave a comment on a PR/Issue """ request_headers = sansio.create_headers( "miss-islington", oauth_token=os.getenv("GH_AUTH") ) issue_comment_url = ( f"https://api.github.com/repos/python/cpython/issues/{issue_number}/comments" ) data = {"body": message} response = requests.post(issue_comment_url, headers=request_headers, json=data) if response.status_code == requests.codes.created: print(f"Commented at {response.json()['html_url']}, message: {message}") else: print(response.status_code) print(response.text) return response def assign_pr_to_core_dev(issue_number, coredev_login): """ Assign the PR to a core dev. Should be done when miss-islington failed to backport. """ request_headers = sansio.create_headers( "miss-islington", oauth_token=os.getenv("GH_AUTH") ) edit_issue_url = ( f"https://api.github.com/repos/python/cpython/issues/{issue_number}" ) data = {"assignees": [coredev_login]} response = requests.patch(edit_issue_url, headers=request_headers, json=data) if response.status_code == requests.codes.created: print(f"Assigned PR {issue_number} to {coredev_login}") else: print(response.status_code) print(response.text) return response async def leave_comment(gh, pr_number, message): """ Leave a comment on a PR/Issue """ issue_comment_url = f"/repos/python/cpython/issues/{pr_number}/comments" data = {"body": message} await gh.post(issue_comment_url, data=data) def is_cpython_repo(): cmd = "git log -r 7f777ed95a19224294949e1b4ce56bbffcb1fe9f" try: subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) except subprocess.SubprocessError: return False return True async def get_gh_participants(gh, pr_number): pr_url = f"/repos/python/cpython/pulls/{pr_number}" pr_result = await gh.getitem(pr_url) created_by = pr_result["user"]["login"] merged_by = None if pr_result["merged_by"] and pr_result["merged_by"]["login"] != "miss-islington": merged_by = pr_result["merged_by"]["login"] participants = "" if created_by == merged_by or merged_by is None: participants = f"@{created_by}" else: participants = f"@{created_by} and @{merged_by}" return participants def get_participants(created_by, merged_by): participants = "" if created_by == merged_by or merged_by == "miss-islington": participants = f"@{created_by}" else: participants = f"@{created_by} and @{merged_by}" return participants def normalize_title(title, body): """Normalize the title if it spills over into the PR's body.""" if not (title.endswith("…") and body.startswith("…")): return title else: # Being paranoid in case \r\n is used. return title[:-1] + body[1:].partition("\r\n")[0] def normalize_message(body): """Normalize the message body to make it commit-worthy. Mostly this just means removing HTML comments, but also removes unwanted leading or trailing whitespace. Returns the normalized body. """ while "<!--" in body: body = body[: body.index("<!--")] + body[body.index("-->") + 3 :] return "\n\n" + body.strip() # Copied over from https://github.com/python/bedevere async def is_core_dev(gh, username): """Check if the user is a CPython core developer.""" org_teams = "/orgs/python/teams" team_name = "python core" async for team in gh.getiter(org_teams): if team["name"].lower() == team_name: break else: raise ValueError(f"{team_name!r} not found at {org_teams!r}") # The 'teams' object only provides a URL to a deprecated endpoint, # so manually construct the URL to the non-deprecated team membership # endpoint. membership_url = f"/teams/{team['id']}/memberships/{username}" try: await gh.getitem(membership_url) except gidgethub.BadRequest as exc: if exc.status_code == 404: return False raise else: return True def pr_is_awaiting_merge(pr_labels): label_names = [label["name"] for label in pr_labels] if ( "DO-NOT-MERGE" not in label_names and "awaiting merge" in label_names and "CLA signed" in label_names ): return True return False def pr_is_automerge(pr_labels): for label in pr_labels: if label["name"] == AUTOMERGE_LABEL: return True return False async def get_pr_for_commit(gh, sha): prs_for_commit = await gh.getitem( f"/search/issues?q=type:pr+repo:python/cpython+sha:{sha}" ) if prs_for_commit["total_count"] > 0: # there should only be one pr_for_commit = prs_for_commit["items"][0] return pr_for_commit return None
5,089
1,696
import aiosqlite import sqlite3 import asyncio import nonebot from nonebot.log import logger driver: nonebot.Driver = nonebot.get_driver() config: nonebot.config.Config = driver.config @driver.on_startup async def init_db(): config.db = await aiosqlite.connect("src/static/Kiba.db") logger.info("Kiba Kernel -> Starting to Create \"Kiba Database\"") try: await config.db.executescript( "create table group_poke_table (group_id bigint primary key not null, last_trigger_time int, triggered int, disabled bit, strategy text);" "create table user_poke_table (user_id bigint, group_id bigint, triggered int);" "create table guess_table (group_id bigint, enabled bit);" "create table waiting_table (shop text, location text, wait int, updated text);" "create table plp_table (id bigint, user_id bigint, nickname text, message text, is_picture bit, view bigint, reply bigint);" "create table plp_reply_table (id bigint, plpid bigint, userid bigint, nickname text, message text);" "create table group_plp_table (group_id bigint, disableinsert int, disabletake int, disablereply int, slowmode int, limit int, time bigint);" "create table plp_blacklist_table (id bigint, lastbanner bigint, disableinsert int, disabletake int, disablereply int);" "create table gld_table (qq bigint, uid bigint);" "create table sign_table (no bigint, id bigint, day int);" "create table acard_table (id bigint, times int, six int, five int, four int, three int, two int, one int);" ) logger.info("Kiba Kernel -> Create \"Kiba Database\" successfully") except Exception as e: logger.info(f"Kiba Kernel --Skip-> Database Created....Skipped Creating Databases. \n[SKIP ERR]{e}") pass @driver.on_shutdown async def free_db(): await config.db.close()
1,918
569
from typing import Callable, Optional, Sequence, Tuple, Union import numpy from dexp.processing.utils.nd_slice import nd_split_slices, remove_margin_slice from dexp.processing.utils.normalise import Normalise from dexp.utils import xpArray from dexp.utils.backends import Backend def scatter_gather_i2i( function: Callable, image: xpArray, tiles: Union[int, Tuple[int, ...]], margins: Optional[Union[int, Tuple[int, ...]]] = None, normalise: bool = False, clip: bool = False, to_numpy: bool = True, internal_dtype: Optional[numpy.dtype] = None, ) -> xpArray: """ Image-2-image scatter-gather. 'Scatters' computation of a given unary function by splitting the input array into tiles, computing using a given backend, and reassembling the tiles into a single array of same shape as the inpout that is either backed by the same backend than that of the input image, or that is backed by numpy -- usefull when the compute backend cannot hold the whole input and output images in memory. Parameters ---------- function : unary function image : input image (can be any backend, numpy ) tiles : tile sizes to cut input image into, can be a single integer or a tuple of integers. margins : margins to add to each tile, can be a single integer or a tuple of integers. if None, no margins are added. normalise : normalises the input image. clip : clip after normalisation/denormalisation to_numpy : should the result be a numpy array? Very usefull when the compute backend cannot hold the whole input and output images in memory. internal_dtype : internal dtype for computation Returns ------- Result of applying the unary function to the input image, if to_numpy==True then the image is """ if internal_dtype is None: internal_dtype = image.dtype if type(tiles) == int: tiles = (tiles,) * image.ndim # If None is passed for a tile that means that we don't tile along that axis, we als clip the tile size: tiles = tuple((length if tile is None else min(length, tile)) for tile, length in zip(tiles, image.shape)) if margins is None: margins = (0,) * image.ndim if type(margins) == int: margins = (margins,) * image.ndim if to_numpy: result = numpy.empty(shape=image.shape, dtype=internal_dtype) else: result = Backend.get_xp_module(image).empty_like(image, dtype=internal_dtype) # Normalise: norm = Normalise(Backend.to_backend(image), do_normalise=normalise, clip=clip, quantile=0.005) # image shape: shape = image.shape # We compute the slices objects to cut the input and target images into batches: tile_slices = list(nd_split_slices(shape, chunks=tiles, margins=margins)) tile_slices_no_margins = list(nd_split_slices(shape, chunks=tiles)) # Zipping together slices with and without margins: slices = zip(tile_slices, tile_slices_no_margins) # Number of tiles: number_of_tiles = len(tile_slices) if number_of_tiles == 1: # If there is only one tile, let's not be complicated about it: result = norm.backward(function(norm.forward(image))) if to_numpy: result = Backend.to_numpy(result, dtype=internal_dtype) else: result = Backend.to_backend(result, dtype=internal_dtype) else: _scatter_gather_loop( norm.backward, function, image, internal_dtype, norm.forward, result, shape, slices, to_numpy ) return result def _scatter_gather_loop( denorm_fun: Callable, function: Callable, image: xpArray, internal_dtype: numpy.dtype, norm_fun: Callable, result: Callable, shape: Tuple[int, ...], slices: Sequence[Tuple[slice, ...]], to_numpy: bool, ) -> None: for tile_slice, tile_slice_no_margins in slices: image_tile = image[tile_slice] image_tile = Backend.to_backend(image_tile, dtype=internal_dtype) image_tile = denorm_fun(function(norm_fun(image_tile))) if to_numpy: image_tile = Backend.to_numpy(image_tile, dtype=internal_dtype) else: image_tile = Backend.to_backend(image_tile, dtype=internal_dtype) remove_margin_slice_tuple = remove_margin_slice(shape, tile_slice, tile_slice_no_margins) image_tile = image_tile[remove_margin_slice_tuple] result[tile_slice_no_margins] = image_tile # Dask turned out not too work great here, HUGE overhead compared to the light approach above. # def scatter_gather_dask(backend: Backend, # function, # image, # chunks, # margins=None): # boundary=None # trim=True # align_arrays=True # # image_d = from_array(image, chunks=chunks, asarray=False) # # def function_numpy(_image): # print(_image.shape) # return backend.to_numpy(function(_image)) # # #func, *args, depth=None, boundary=None, trim=True, align_arrays=True, **kwargs # computation= map_overlap(function_numpy, # image_d, # depth=margins, # boundary=boundary, # trim=trim, # align_arrays=align_arrays, # dtype=image.dtype # ) # # #computation.visualize(filename='transpose.png') # result = computation.compute() # # return result
5,492
1,708
from .backend import Backend from .thread import HttpPool
58
15
#!/usr/local/bin/python3 def main(): # Test suite return if __name__ == '__main__': main()
107
42
import fnmatch import glob import os import sys import setuptools.command.build_ext APPLE = sys.platform == 'darwin' BASE_DIR = os.path.dirname(__file__) ABOUT = {} def _read_about(): with open(os.path.join(BASE_DIR, 'omnibus', '__about__.py'), 'rb') as f: src = f.read() if sys.version_info[0] > 2: src = src.decode('UTF-8') exec(src, ABOUT) _read_about() EXCLUDED_STATIC_FILE_PATHS = [ '*.py', '*/__pycache__/*', '*/tests/*', '*/_ext/cc/*', '*/_ext/cy/*', '*/_ext/rs/*', ] def _get_static_files(path): return [filepath for (dirpath, dirnames, filenames) in os.walk(path, followlinks=True) for filename in filenames for filepath in [os.path.join(dirpath, filename)] if not any(fnmatch.fnmatch(filepath, pat) for pat in EXCLUDED_STATIC_FILE_PATHS)] PACKAGE_DATA = [ '.revision', ] + _get_static_files('omnibus') INSTALL_REQUIRES = [ 'toolz>=0.9.0', ] EXTRAS_REQUIRE = { 'bintrees': ['bintrees>=0.2.7'], 'cytoolz': ['cytoolz>=0.9.0'], 'docker': ['docker>=3.7.0'], 'sortedcontainers': ['sortedcontainers>=2.1.0'], } DEBUG = 'DEBUG' in os.environ EXT_MODULES = [] try: import Cython except ImportError: pass else: import Cython.Build import Cython.Compiler.Options EXT_MODULES.extend([ *[ setuptools.Extension( 'omnibus._ext.cc.' + os.path.basename(fpath).rpartition('.')[0], sources=[fpath] ) for fpath in glob.glob('omnibus/_ext/cc/*.cc') ], *Cython.Build.cythonize( [ setuptools.Extension( 'omnibus._ext.cy.' + os.path.basename(fpath).rpartition('.')[0], sources=[fpath], language='c++', ) for fpath in glob.glob('omnibus/_ext/cy/**/*.pyx', recursive=True) ], language_level=3, gdb_debug=DEBUG, compiler_directives={ **Cython.Compiler.Options.get_directive_defaults(), 'embedsignature': True, 'binding': True, }, ), ]) if APPLE: EXT_MODULES.extend([ setuptools.Extension( 'omnibus._ext.m.' + os.path.basename(fpath).rpartition('.')[0], sources=[fpath], extra_link_args=[ '-framework', 'AppKit', '-framework', 'CoreFoundation', ] ) for fpath in glob.glob('omnibus/_ext/m/*.m') ]) if __name__ == '__main__': setuptools.setup( name=ABOUT['__title__'], version=ABOUT['__version__'], description=ABOUT['__description__'], author=ABOUT['__author__'], url=ABOUT['__url__'], python_requires='>=3.7', classifiers=[ 'Intended Audience :: Developers', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: ' + '.'.join(map(str, sys.version_info[:2])), 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python', ], # zip_safe=True, setup_requires=['setuptools'], packages=setuptools.find_packages( include=['omnibus', 'omnibus.*'], exclude=['tests', '*.tests', '*.tests.*'], ), py_modules=['omnibus'], package_data={'omnibus': PACKAGE_DATA}, include_package_data=True, entry_points={}, install_requires=INSTALL_REQUIRES, extras_require=EXTRAS_REQUIRE, ext_modules=EXT_MODULES, )
3,759
1,250
from day1 import * import unittest def test_get_fuel_requirements_for_mass(): assert get_fuel_requirements_for_mass(12) == 2 assert get_fuel_requirements_for_mass(14) == 2 assert get_fuel_requirements_for_mass(1969) == 654 assert get_fuel_requirements_for_mass(100756) == 33583 def test_sum_fuel_requirements_for_all_modules_on_spacecraft(): assert sum_fuel_requirements_for_all_modules_on_spacecraft([12, 14, 1969, 100756]) == 34241 def test_get_fuel_requirements_for_mass_recursive(): assert get_fuel_requirements_for_mass_recursive(14) == 2 assert get_fuel_requirements_for_mass_recursive(1969) == 966 assert get_fuel_requirements_for_mass_recursive(100756) == 50346
689
302
import AppKit from PyObjCTools.TestSupport import TestCase, min_os_level class TestNSPDFInfo(TestCase): @min_os_level("10.9") def testMethods(self): self.assertResultIsBOOL(AppKit.NSPDFInfo.isFileExtensionHidden) self.assertArgIsBOOL(AppKit.NSPDFInfo.setFileExtensionHidden_, 0)
305
102
from collections import OrderedDict, defaultdict import numpy as np ''' generate a id to length dic ''' def gen_sid_len(sentences): sid2len = OrderedDict() for i, sent in enumerate(sentences): sid2len[i] = len(sent) return sid2len def batch_slice(data, batch_size): # data is a list of sentences of the same length batch_num = int(np.ceil(len(data) / float(batch_size))) for i in xrange(batch_num): cur_batch_size = batch_size if i < batch_num - 1 else len(data) - batch_size * i # cur_batch_size is the end-point of the batch sents = data[i * batch_size: i * batch_size + cur_batch_size] yield sents def data_iter(sents_id2length_dic, batch_size, shuffle=True): """ randomly permute data, then sort by source length, and partition into batches ensure that the length of source sentences in each batch is decreasing """ buckets = defaultdict(list) for (sent_id, sent_len) in sents_id2length_dic.iteritems(): buckets[sent_len].append(sent_id) batched_data = [] for (sent_len, sent_ids_smlen) in buckets.iteritems(): # sent_ids_smlen is a list of sentences of the same length if shuffle: np.random.shuffle(sent_ids_smlen) # pdb.set_trace() ''' 'extend' expecting a iterable finishes the iteration ''' batched_data.extend(list(batch_slice(sent_ids_smlen, batch_size))) if shuffle: np.random.shuffle(batched_data) for batch in batched_data: """ sent_ids in the same batch are of the same length """ yield batch
1,636
522
from ._schema import DefaultSchema from ._utils import get_minutes, get_yields, normalize_string class TheHappyFoodie(DefaultSchema): @classmethod def host(cls): return "thehappyfoodie.co.uk" def title(self): return self.soup.find("h1", {"class": "main-title"}).get_text() def total_time(self): return get_minutes( self.soup.find("div", {"class": "recipe__data__total-time"}) ) or sum( [ get_minutes( self.soup.find("div", {"class": "recipe__data__prep-time"}) ), get_minutes( self.soup.find("div", {"class": "recipe__data__cook-time"}) ), ] ) def yields(self): return get_yields( self.soup.find("div", {"class": "recipe__data__yield"}).get_text() ) def ingredients(self): ingredients = self.soup.find( "table", {"class": "recipe__ingredients-table"} ).findAll("tr") ingredients = [ ( ingredient.find( "td", {"class": "recipe__ingredients__amount"} ).get_text(), ingredient.find( "td", {"class": "recipe__ingredients__name"} ).get_text(), ) for ingredient in ingredients ] return [ normalize_string("{} {}".format(amount, name)) for amount, name in ingredients ] def instructions(self): instructions = self.soup.find("div", {"class": "recipe__instructions"}).findAll( "p" ) return "\n".join( normalize_string(instruction.get_text()) for instruction in instructions )
1,793
535
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright [2017] Tatarnikov Viktor [viktor@tatarnikov.org] # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. try: from urllib.parse import unquote except ImportError: from urllib import unquote from flask import url_for, current_app from werkzeug.utils import import_string class DebugToolbar(object): _cached_panel_classes = {} def __init__(self, request, jinja_env): self.jinja_env = jinja_env self.request = request self.panels = [] self.template_context = { 'static_path': '/static/debug_panel/' } self.create_panels() def create_panels(self): """ Populate debug panels """ activated = self.request.cookies.get('fldt_active', '') activated = unquote(activated).split(';') for panel_class in self._iter_panels(current_app): panel_instance = panel_class(jinja_env=self.jinja_env, context=self.template_context) if panel_instance.dom_id() in activated: panel_instance.is_active = True self.panels.append(panel_instance) def render_toolbar(self): context = self.template_context.copy() context.update({'panels': self.panels}) template = self.jinja_env.get_template('base.html') return template.render(**context) @classmethod def load_panels(cls, app): for panel_class in cls._iter_panels(app): # just loop to make sure they've been loaded pass @classmethod def _iter_panels(cls, app): for panel_path in app._config.get("web", "debug-toolbar", "panels", "").split(";"): panel_class = cls._import_panel(app, "WEB.DebugToolbar.Panels.{}".format(panel_path)) if panel_class is not None: yield panel_class @classmethod def _import_panel(cls, app, path): cache = cls._cached_panel_classes try: return cache[path] except KeyError: pass try: panel_class = import_string(path) except ImportError as e: app.logger.warning('Disabled %s due to ImportError: %s', path, e) panel_class = None cache[path] = panel_class return panel_class
2,866
861
# -*- coding: utf-8 -*- # !/usr/bin/python __author__ = 'ma_keling' # Version : 1.0.0 # Start Time : 2018-11-29 # Update Time : # Change Log : ## 1. ## 2. ## 3. import time import arcpy import math def express_arcpy_error(): severity = arcpy.GetMaxSeverity() if severity == 2: # If the tool returned an error arcpy.AddError("Error occurred \n{0}".format(arcpy.GetMessages(2))) elif severity == 1: # If the tool returned no errors, but returned a warning arcpy.AddWarning("Warning raised \n{0}".format(arcpy.GetMessages(1))) else: # If the tool did not return an error or a warning arcpy.AddMessage(arcpy.GetMessages()) # Description: Loop layers and calculate lod for every feature in the layer. def calculate_lods_for_feature(in_layers, fieldName): try: startTime = time.time() timeStampName = time.strftime('%Y_%m_%d %H:%M:%S', time.localtime(time.time())) arcpy.AddMessage("Start compute lods at: {0}".format(timeStampName)) for layer in in_layers: arcpy.AddMessage("Calculating lod for layer : {0}.".format(layer)) add_field(layer, fieldName, 9) cursor = arcpy.da.UpdateCursor(layer, ['SHAPE@', 'OID@', fieldName]) lyr_path = layer.dataSource desc = arcpy.Describe(lyr_path) extent = desc.extent arcpy.AddMessage("Original dataset extent:" + str(desc.extent)) ext_wm = extent.projectAs(arcpy.SpatialReference(102100)) arcpy.AddMessage("New WebMercator extent:" + str(ext_wm)) start_level, start_compute_resolution = confirm_level(ext_wm) if desc.shapeType == "Polygon": baselength, basearea = get_length_area_from_pixel(96, 295828764) lod_area = basearea / math.pow(4, start_level - 1) arcpy.AddMessage("start lod area: " + str(lod_area)) for row in cursor: lod = calculate_lod_for_polygon(row[0], baselength, lod_area,start_level) row[2] = lod cursor.updateRow(row) elif desc.shapeType == "Point": count = get_count(layer) arcpy.AddMessage("Total Points:" + str(count)) if count < 200000: arcpy.AddMessage("Input point sets too small for optimized, skip!") continue else: n = math.ceil(math.log(count / (512 * 512 / 16), 4)) arcpy.AddMessage("n:" + str(n)) for row in cursor: oid = row[1] lod = calculate_lod_for_point(oid,start_level,n) row[2] = lod cursor.updateRow(row) elif desc.shapeType == 'Polyline': baselength = get_length_from_pixel(96, 295828764) lod_length = baselength / math.pow(2, start_level - 1) for row in cursor: lod = calculate_lod_for_polyline(row[0],lod_length,start_level) row[2] = lod cursor.updateRow(row) endTime = time.time() print("Compute finished, elapsed: {0} Seconds.eter..".format(str(endTime - startTime))) arcpy.AddMessage("Compute finished, elapsed: {0} Seconds.eter..".format(str(endTime - startTime))) print("\n") arcpy.AddMessage("\n") except arcpy.ExecuteError: express_arcpy_error() # Description: Compute the total records for a featureclass def get_count(layer): fields = ['SHAPE@'] count = 0 with arcpy.da.SearchCursor(layer, fields) as cursor: for row in cursor: count += 1 return count # Description: get the start level based on layer extent def confirm_level(extent): width = extent.width height = extent.height arcpy.AddMessage("width:"+str(width) +" height:"+ str(height)) length = max(width, height) base_resolution = 78271.516964011724 base_tile_resolution = base_resolution * 512 for level in range(21): start_compute_resolution = base_tile_resolution if length >= base_tile_resolution: arcpy.AddMessage("level:" + str(level)) arcpy.AddMessage("base tile resolution:" + str(base_tile_resolution)) return level, start_compute_resolution else: base_tile_resolution /= 2 # Description: Add a new field with name lod to a table def add_field(inFeatures,fieldName,fieldPrecision): try: startTime = time.time() timeStampName = time.strftime('%Y_%m_%d %H:%M:%S', time.localtime(time.time())) print("start add new field for: ", inFeatures, " at: ", timeStampName) arcpy.AddMessage("start add new field for: {0} at: {1}".format(str(inFeatures), str(timeStampName))) # Execute AddField for new field arcpy.AddField_management(inFeatures, fieldName, "Text", fieldPrecision, field_alias=fieldName, field_is_nullable="NULLABLE") endTime = time.time() print(inFeatures, "Add field:", fieldName, "finished, elapsed: ", str(endTime - startTime) + ' Seconds.eter..') arcpy.AddMessage( "Add field: {0} finished, elapsed: {1} Seconds.eter..".format(fieldName, str(endTime - startTime))) print("\n") arcpy.AddMessage("\n") except arcpy.ExecuteError: express_arcpy_error() # Description: Compute get area and length per pixel based on dpi and scale def get_length_area_from_pixel(dpi,scale): pixel = 1 / dpi * 0.025 length = scale * pixel area = length * length return length,area # Description: Compute get length per pixel based on dpi and scale def get_length_from_pixel(dpi,scale): pixel = 1 / dpi * 0.025 length = scale * pixel return length # Description: Calculate lod for every polygon def calculate_lod_for_polygon(feature,baselength, basearea, start_level): try: if feature: area = feature.getArea("GEODESIC", "SQUAREMETERS") # length = feature.getLength("GEODESIC", "METERS") lod = start_level for i in range(20): # arcpy.AddMessage(str(i) + ":" + str(basearea) + "___"+str(area)) # arcpy.AddMessage(str(i) + ":" + str(baselength) + "___" + str(length)) if area >= basearea : return str(lod) else: lod += 1 basearea /= 4 baselength /= 2 return str(lod) else: print(type(feature)) return "19" except arcpy.ExecuteError: express_arcpy_error() # Description: Calculate lod for every point def calculate_lod_for_point(id, start_level, n): try: for i in range(n): if id % math.pow(4, n - i) == 0: return start_level else: start_level += 1 return start_level except arcpy.ExecuteError: express_arcpy_error() # Description: Calculate lod for every polyline def calculate_lod_for_polyline(feature,baselength, start_level): try: if feature: length = feature.getLength("GEODESIC", "METERS") lod = start_level for i in range(20): # arcpy.AddMessage(str(i) + ":" + str(basearea) + "___"+str(area)) # arcpy.AddMessage(str(i) + ":" + str(baselength) + "___" + str(length)) if length >= baselength: return lod else: lod += 1 baselength /= 2 return lod else: print(type(feature)) except arcpy.ExecuteError: express_arcpy_error() def execute(): in_map = arcpy.GetParameter(0) arcpy.AddMessage("Input map : {0}.".format(in_map)) in_layers = arcpy.GetParameter(1) field_name = "lod" calculate_lods_for_feature(in_layers, field_name) # execute()
8,372
2,758
# -*- coding: utf-8 -*- """ Created on Sat Feb 8 07:38:05 2020 @author: Ham Self Challenge: Count Cluster of Non-0s Given a 1-dimension array of integers, determine how many 'clusters' of non-0 in the array. A 'cluster' is a group of consecutive non-0 values. Scoring: a solution needs to be a 1-liner; i.e. NO point if implementing with a traditional 'for' loop! Sample Input (see STDIN_SIO) A : [ 9, 0, 0, 22, 0, 0, 39, 11, 3, 0, \ 0, 24, 1, 0, 50, 23, 3, 44, 0, 23, \ 25, 6, 36, 19, 10, 23, 0, 37, 4, 1, \ 7, 12, 0, 0, 49 ] Expected Output: 8 """ import itertools STDIN_SIO = """ 9, 0, 0, 22, 0, 0, 39, 11, 3, 0, \ 0, 24, 1, 0, 50, 23, 3, 44, 0, 23, \ 2, 8, 20, 35, 0, 40, 34, 26, 36, 0, \ 35, 19, 20, 18, 11, 43, 19, 21, 40, 0, \ 14, 0, 14, 0, 0, 25, 35, 24, 49, 15, \ 13, 3, 0, 10, 31, 25, 27, 37, 27, 43, \ 44, 27, 8, 43, 0, 0, 33, 25, 19, 47, \ 0, 29, 5, 2, 12, 8, 7, 0, 16, 36, \ 0, 6, 17, 35, 36, 21, 0, 9, 1, 0, \ 43, 29, 39, 15, 18, 0, 34, 26, 48, 0, \ 34, 35, 7, 10, 0, 0, 15, 5, 12, 26, \ 0, 37, 30, 33, 27, 34, 9, 37, 22, 0, \ 0, 24, 30, 0, 0, 38, 23, 25, 0, 30, \ 39, 24, 31, 0, 6, 19, 25, 0, 28, 15, \ 8, 0, 48, 0, 35, 41, 0, 24, 1, 41, \ 31, 0, 35, 21, 15, 26, 15, 27, 4, 0, \ 8, 4, 0, 0, 2, 42, 18, 0, 28, 18, \ 49, 34, 5, 10, 41, 48, 26, 14, 45, 44, \ 9, 0, 49, 50, 24, 0, 0, 0, 23, 0, \ 17, 0, 47, 31, 0, 42, 0, 0, 0, 40, \ 46, 22, 50, 32, 20, 3, 44, 22, 0, 37, \ 25, 0, 19, 26, 14, 23, 27, 41, 0, 1, \ 13, 0, 48, 20, 37, 8, 0, 18, 0, 26, \ 12, 19, 32, 19, 22, 0, 0, 0, 0, 0, \ 16, 0, 0, 43, 0, 10, 5, 0, 6, 26, \ 0, 24, 40, 29, 0, 43, 18, 27, 0, 0, \ 37, 0, 46, 35, 17, 0, 20, 44, 29, 29, \ 40, 33, 22, 27, 0, 0, 38, 21, 4, 0, \ 0, 15, 31, 48, 36, 10, 0, 41, 0, 45, \ 39, 0, 11, 9, 3, 38, 16, 0, 11, 22, \ 37, 0, 3, 44, 10, 12, 47, 22, 32, 7, \ 24, 1, 0, 22, 25, 0, 14, 0, 0, 0, \ 23, 0, 36, 1, 42, 46, 0, 48, 0, 33, \ 5, 27, 45, 0, 15, 29, 0, 50, 2, 31, \ 25, 6, 36, 19, 10, 23, 0, 37, 4, 1, \ 7, 12, 0, 0, 49 """.strip() def count_non_0_clusters_1(arr): """Translate each non-0 to an 'A' char, and 0 to a space. Then join together to become a string. Then split(), then return number of tokens. """ return len("".join(["A" if e else " " for e in arr]).split()) def count_non_0_clusters_2(arr): """groupby() partitions into groups as: [[True , [list of non-0]], [False, [list of 0s]], [True , [list of non-0]], [False, [list of 0s]], ... [True , [list of non-0]]] (Old) Next, the list comprenhension iterates thru each tuple, then collects the 1st element in each tuple if True. Finally, return the len/count of Trues: return len([t[0] for t in itertools.groupby(...) if t[0]]) Next, the list comprenhension iterates thru each tuple, then collects the 1st element in each tuple. Then return the count() of True elements. """ return [t[0] for t in itertools.groupby(arr, lambda e: bool(e))].count(True) if __name__ == '__main__': a = list(map(int, STDIN_SIO.split(","))) # Nicely print it, 10 entries per line, with continuation # so can copy-n-paste back into STDIN_SIO #print(len(a)) #for i in range(0, (len(a) // 10) * 10, 10): # print("%3u," * 10 % tuple(a[i:i+10]), end=" \\\n") #j = a[(len(a) // 10) * 10:] #print("%3u," * (len(j) - 1) % tuple(j[:-1]), end="") #print("%3u" % j[-1]) print("count_*_1() returns", count_non_0_clusters_1(a), "clusters of non-0") print("count_*_2() returns", count_non_0_clusters_2(a), "clusters of non-0")
3,894
2,364
"""Tests for node configuration.""" import json import logging import time from pathlib import Path import allure import pytest from _pytest.tmpdir import TempdirFactory from cardano_clusterlib import clusterlib from cardano_node_tests.utils import cluster_management from cardano_node_tests.utils import cluster_nodes from cardano_node_tests.utils import configuration from cardano_node_tests.utils import helpers LOGGER = logging.getLogger(__name__) @pytest.fixture(scope="module") def create_temp_dir(tmp_path_factory: TempdirFactory): """Create a temporary dir.""" p = Path(tmp_path_factory.getbasetemp()).joinpath(helpers.get_id_for_mktemp(__file__)).resolve() p.mkdir(exist_ok=True, parents=True) return p @pytest.fixture def temp_dir(create_temp_dir: Path): """Change to a temporary dir.""" with helpers.change_cwd(create_temp_dir): yield create_temp_dir # use the "temp_dir" fixture for all tests automatically pytestmark = pytest.mark.usefixtures("temp_dir") @pytest.fixture(scope="module") def epoch_length_start_cluster(tmp_path_factory: TempdirFactory) -> Path: """Update *epochLength* to 1200.""" pytest_globaltemp = helpers.get_pytest_globaltemp(tmp_path_factory) # need to lock because this same fixture can run on several workers in parallel with helpers.FileLockIfXdist(f"{pytest_globaltemp}/startup_files_epoch_1200.lock"): destdir = pytest_globaltemp / "startup_files_epoch_1200" destdir.mkdir(exist_ok=True) # return existing script if it is already generated by other worker destdir_ls = list(destdir.glob("start-cluster*")) if destdir_ls: return destdir_ls[0] startup_files = cluster_nodes.get_cluster_type().cluster_scripts.copy_scripts_files( destdir=destdir ) with open(startup_files.genesis_spec) as fp_in: genesis_spec = json.load(fp_in) genesis_spec["epochLength"] = 1500 with open(startup_files.genesis_spec, "w") as fp_out: json.dump(genesis_spec, fp_out) return startup_files.start_script @pytest.fixture(scope="module") def slot_length_start_cluster(tmp_path_factory: TempdirFactory) -> Path: """Update *slotLength* to 0.3.""" pytest_globaltemp = helpers.get_pytest_globaltemp(tmp_path_factory) # need to lock because this same fixture can run on several workers in parallel with helpers.FileLockIfXdist(f"{pytest_globaltemp}/startup_files_slot_03.lock"): destdir = pytest_globaltemp / "startup_files_slot_03" destdir.mkdir(exist_ok=True) # return existing script if it is already generated by other worker destdir_ls = list(destdir.glob("start-cluster*")) if destdir_ls: return destdir_ls[0] startup_files = cluster_nodes.get_cluster_type().cluster_scripts.copy_scripts_files( destdir=destdir ) with open(startup_files.genesis_spec) as fp_in: genesis_spec = json.load(fp_in) genesis_spec["slotLength"] = 0.3 with open(startup_files.genesis_spec, "w") as fp_out: json.dump(genesis_spec, fp_out) return startup_files.start_script @pytest.fixture def cluster_epoch_length( cluster_manager: cluster_management.ClusterManager, epoch_length_start_cluster: Path ) -> clusterlib.ClusterLib: return cluster_manager.get( singleton=True, cleanup=True, start_cmd=str(epoch_length_start_cluster) ) @pytest.fixture def cluster_slot_length( cluster_manager: cluster_management.ClusterManager, slot_length_start_cluster: Path ) -> clusterlib.ClusterLib: return cluster_manager.get( singleton=True, cleanup=True, start_cmd=str(slot_length_start_cluster) ) def check_epoch_length(cluster_obj: clusterlib.ClusterLib) -> None: end_sec = 15 end_sec_padded = end_sec + 15 # padded to make sure tip got updated cluster_obj.wait_for_new_epoch() epoch = cluster_obj.get_epoch() sleep_time = cluster_obj.epoch_length_sec - end_sec time.sleep(sleep_time) assert epoch == cluster_obj.get_epoch() time.sleep(end_sec_padded) assert epoch + 1 == cluster_obj.get_epoch() @pytest.mark.run(order=3) @pytest.mark.skipif( bool(configuration.TX_ERA), reason="different TX eras doesn't affect this test, pointless to run", ) class TestBasic: """Basic tests for node configuration.""" @allure.link(helpers.get_vcs_link()) def test_epoch_length(self, cluster_epoch_length: clusterlib.ClusterLib): """Test the *epochLength* configuration.""" cluster = cluster_epoch_length assert cluster.slot_length == 0.2 assert cluster.epoch_length == 1500 check_epoch_length(cluster) @allure.link(helpers.get_vcs_link()) @pytest.mark.run(order=2) def test_slot_length(self, cluster_slot_length: clusterlib.ClusterLib): """Test the *slotLength* configuration.""" cluster = cluster_slot_length assert cluster.slot_length == 0.3 assert cluster.epoch_length == 1000 check_epoch_length(cluster)
5,133
1,702
#!/usr/bin/env python from __future__ import print_function import matplotlib matplotlib.use('Agg') import os import heapq import numpy as np import matplotlib.pyplot as plt from mpl_toolkits.mplot3d import Axes3D import matplotlib.colors as colors import matplotlib.cm as cm from numpy import * from random import sample, seed, randint from os.path import getsize as getFileSize import math import random import csv from cycler import cycler from io import StringIO #np.set_printoptions(threshold=np.nan) from collections import Counter from matplotlib.colors import LogNorm from mpl_toolkits.axes_grid1 import AxesGrid from astropy import units as u from astropy import cosmology import matplotlib.ticker as mtick import PlotScripts import ReadScripts import AllVars import GalaxyPhotoion as photo import ObservationalData as Obs import gnedin_analytic as ga from mpi4py import MPI import sys comm = MPI.COMM_WORLD rank = comm.Get_rank() size = comm.Get_size() AllVars.Set_Params_Kali() AllVars.Set_Constants() PlotScripts.Set_Params_Plot() output_format = ".png" # For the Tiamat extended results there is a weird hump when calculating the escape fraction. # This hump occurs at a halo mass of approximately 10.3. # The calculation of fesc skips this hump range (defined from kink_low to kink_high) kink_low = 10.3 kink_high = 10.30000001 m_low = 7.0 # We only sum the photons coming from halos within the mass range m_low < Halo Mass < m_high m_high = 15.0 m_gal_low = 3.0 m_gal_high = 12.0 m_low_SAGE = pow(10, m_low)/1.0e10 * AllVars.Hubble_h m_high_SAGE = pow(10, m_high)/1.0e10 * AllVars.Hubble_h bin_width = 0.2 NB = int((m_high - m_low) / bin_width) NB_gal = int((m_gal_high - m_gal_low) / bin_width) fej_low = 0.0 fej_high = 1.0 fej_bin_width = 0.05 NB_fej = int((fej_high - fej_low) / fej_bin_width) def raise_list_power(my_list, n): return [pow(x, n) for x in my_list] def raise_power_list(my_list, n): return [pow(n, x) for x in my_list] def calculate_beta(MUV, z): ''' Calculation of the dust attenuation parameter Beta. Fit values are from Bouwens (2015) ApJ 793, 115. For z = 5 and 6, Bouwens uses a piece-wise linear relationship and a linear relationship for higher redshift. ## Parameters ---------- MUV : `float' A value of the absolute magnitude in the UV (generally M1600) in the AB magnitude system. z : `float' Redshift the attenuation is calculated at. Returns ------ beta : `float' Value of the UV continuum paramaeter beta. ''' if (z >= 4.5 and z < 5.5): # z = 5 fits. if (MUV > -18.8): dB = -0.08 else: dB = -0.17 B = -2.05 offset = 18.8 elif (z >= 5.5 and z < 6.5): # z = 6 fits. if (MUV > -18.8): dB = -0.08 else: dB = -0.24 B = -2.22 offset = 18.8 elif (z >= 6.5 and z < 7.5): # z = 7 fits. dB = -0.20 B = -2.05 offset = 19.5 elif (z >= 7.5 and z < 8.5): # z = 8 fits. dB = -0.15 B = -2.13 offset = 19.5 elif (z >= 8.5 and z < 9.5): # z = 9 fits. dB = -0.16 B = -2.19 offset = 19.5 elif (z >= 9.5 and z < 10.5): # z = 10 fits. dB = -0.16 B = -2.16 offset = 19.5 beta = dB * (MUV + offset) + B return beta def multiply(array): ''' Performs element wise multiplication. Parameters ---------- array : `~numpy.darray' The array to be multiplied. Returns ------- total : `float' Total of the elements multiplied together. ''' total = 1 for i in range(0, len(array)): total *= array[i] return total ## def Sum_Log(array): ''' Performs an element wise sum of an array who's elements are in log-space. Parameters ---------- array : array Array with elements in log-space. Returns ------ sum_total : float Value of the elements taken to the power of 10 and summed. Units ----- All units are kept the same as the inputs. ''' sum_total = 0.0 for i in range(0, len(array)): sum_total += 10**array[i] return sum_total ## def Std_Log(array, mean): ''' Calculates the standard deviation of an array with elements in log-space. Parameters ---------- array : array Array with elements in log-space. mean : float Mean of the array (not in log). Returns ------ std : float Standard deviation of the input array taken to the power of 10. Units ----- All units are kept the same as the inputs. ''' sum_total = 0.0 for i in range(0, len(array)): sum_total += (10**array[i] - mean)**2 sum_total *= 1.0/len(array) std = np.sqrt(sum_total) return std ### def collect_across_tasks(mean_per_task, std_per_task, N_per_task, SnapList, BinSnapList=[], binned=False, m_bin_low=0.0, m_bin_high=0.0, my_bin_width=bin_width): """ Reduces arrays that are unique to each task onto the master task. The dimensions of the input arrays will change slightly if we are collecting a statistics that is binned across e.g., halo mass or galaxy stellar mass. Parameters ---------- mean_per_task, std_per_task, N_per_task: Nested 2D (or 3D if binned == True) arrays of floats. Outer length is equal to the number of models. Inner length is equal to the number of snapshots the data has been calculated for. Most inner length is equal to the number of bins. Contains the mean/standard deviation/number of objects unique for each task. SnapList: Nested 2D arrays of integers. Outer length is equal to the number of models. Contains the snapshot numbers the data has been calculated for each model. BinSnapList: Nested 2D arrays of integers. Outer length is equal to the number of models. Often statistics are calculated for ALL snapshots but we only wish to plot for a subset of snapshots. This variable allows the binned data to be collected for only a subset of the snapshots. binned: Boolean. Dictates whether the collected data is a 2D or 3D array with the inner-most array being binned across e.g., halo mass. Returns ---------- master_mean, master_std, master_N: Nested 2D (or 3D if binned == True) arrays of floats. Shape is identical to the input mean_per_task etc. If rank == 0 these contain the collected statistics. Otherwise these will be none. master_bin_middle: Array of floats. Contains the location of the middle of the bins for the data. """ master_mean = [] master_std = [] master_N = [] master_bin_middle = [] for model_number in range(0, len(SnapList)): master_mean.append([]) master_std.append([]) master_N.append([]) master_bin_middle.append([]) # If we're collecting a binned statistic (e.g., binned across halo mass), then we need to perform the collecting per snapshot. if binned: count = 0 for snapshot_idx in range(len(SnapList[model_number])): if SnapList[model_number][snapshot_idx] == BinSnapList[model_number][count]: master_mean[model_number], master_std[model_number], master_N[model_number] = calculate_pooled_stats(master_mean[model_number], master_std[model_number], master_N[model_number], mean_per_task[model_number][snapshot_idx], std_per_task[model_number][snapshot_idx], N_per_task[model_number][snapshot_idx]) master_bin_middle[model_number].append(np.arange(m_bin_low, m_bin_high+my_bin_width, my_bin_width)[:-1] + my_bin_width* 0.5) count += 1 if count == len(BinSnapList[model_number]): break else: master_mean[model_number], master_std[model_number], master_N[model_number] = calculate_pooled_stats(master_mean[model_number], master_std[model_number], master_N[model_number], mean_per_task[model_number], std_per_task[model_number], N_per_task[model_number]) if rank == 0: master_mean[model_number] = master_mean[model_number][0] master_std[model_number] = master_std[model_number][0] master_N[model_number] = master_N[model_number][0] return master_mean, master_std, master_N, master_bin_middle ### def calculate_pooled_stats(mean_pool, std_pool, N_pool, mean_local, std_local, N_local): ''' Calculates the pooled mean and standard deviation from multiple processors and appends it to an input array. Formulae taken from https://en.wikipedia.org/wiki/Pooled_variance As we only care about these stats on the rank 0 process, we make use of junk inputs/outputs for other ranks. NOTE: Since the input data may be an array (e.g. pooling the mean/std for a stellar mass function). Parameters ---------- mean_pool, std_pool, N_pool : array of floats. Arrays that contain the current pooled means/standard deviation/number of data points (for rank 0) or just a junk input (for other ranks). mean_local, mean_std : float or array of floats. The non-pooled mean and standard deviation unique for each process. N_local : floating point number or array of floating point numbers. Number of data points used to calculate the mean/standard deviation that is going to be added to the pool. NOTE: Use floating point here so we can use MPI.DOUBLE for all MPI functions. Returns ------- mean_pool, std_pool : array of floats. Original array with the new pooled mean/standard deviation appended (for rank 0) or the new pooled mean/standard deviation only (for other ranks). Units ----- All units are the same as the input. All inputs MUST BE real-space (not log-space). ''' if isinstance(mean_local, list) == True: if len(mean_local) != len(std_local): print("len(mean_local) = {0} \t len(std_local) = {1}".format(len(mean_local), len(std_local))) raise ValueError("Lengths of mean_local and std_local should be equal") if ((type(mean_local).__module__ == np.__name__) == True or (isinstance(mean_local, list) == True)): # Checks to see if we are dealing with arrays. N_times_mean_local = np.multiply(N_local, mean_local) N_times_var_local = np.multiply(N_local, np.multiply(std_local, std_local)) N_local = np.array(N_local).astype(float) N_times_mean_local = np.array(N_times_mean_local).astype(np.float32) if rank == 0: # Only rank 0 holds the final arrays so only it requires proper definitions. N_times_mean_pool = np.zeros_like(N_times_mean_local) N_pool_function = np.zeros_like(N_local) N_times_var_pool = np.zeros_like(N_times_var_local) N_times_mean_pool = N_times_mean_pool.astype(np.float64) # Recast everything to double precision then use MPI.DOUBLE. N_pool_function = N_pool_function.astype(np.float64) N_times_var_pool = N_times_var_pool.astype(np.float64) else: N_times_mean_pool = None N_pool_function = None N_times_var_pool = None comm.Barrier() N_times_mean_local = N_times_mean_local.astype(np.float64) N_local = N_local.astype(np.float64) N_times_var_local = N_times_var_local.astype(np.float64) comm.Reduce([N_times_mean_local, MPI.DOUBLE], [N_times_mean_pool, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the arrays across processors. comm.Reduce([N_local, MPI.DOUBLE],[N_pool_function, MPI.DOUBLE], op = MPI.SUM, root = 0) comm.Reduce([N_times_var_local, MPI.DOUBLE], [N_times_var_pool, MPI.DOUBLE], op = MPI.SUM, root = 0) else: N_times_mean_local = N_local * mean_local N_times_var_local = N_local * std_local * std_local N_times_mean_pool = comm.reduce(N_times_mean_local, op = MPI.SUM, root = 0) N_pool_function = comm.reduce(N_local, op = MPI.SUM, root = 0) N_times_var_pool = comm.reduce(N_times_var_local, op = MPI.SUM, root = 0) if rank == 0: mean_pool_function = np.zeros((len(N_pool_function))) std_pool_function = np.zeros((len(N_pool_function))) for i in range(0, len(N_pool_function)): if N_pool_function[i] == 0: mean_pool_function[i] = 0.0 else: mean_pool_function[i] = np.divide(N_times_mean_pool[i], N_pool_function[i]) if N_pool_function[i] < 3: std_pool_function[i] = 0.0 else: std_pool_function[i] = np.sqrt(np.divide(N_times_var_pool[i], N_pool_function[i])) mean_pool.append(mean_pool_function) std_pool.append(std_pool_function) N_pool.append(N_pool_function) return mean_pool, std_pool, N_pool else: return mean_pool, std_pool, N_pool_function # Junk return because non-rank 0 doesn't care. ## def StellarMassFunction(SnapList, SMF, simulation_norm, FirstFile, LastFile, NumFile, ResolutionLimit_mean, model_tags, observations, paper_plot, output_tag): ''' Calculates the stellar mass function for given galaxies with the option to overplot observations by Song et al. (2013) at z = 6, 7, 8 and/or Baldry et al. (2008) at z = 0.1. Parallel compatible. NOTE: The plotting assumes the redshifts we are plotting at are (roughly) the same for each model. Parameters --------- SnapList : Nested 'array-like`, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots that we plot the stellar mass function at for each model. SMF : Nested 2-dimensional array, SMF[model_number0][snapshot0] = [bin0galaxies, ..., binNgalaxies], with length equal to the number of bins (NB_gal). The count of galaxies within each stellar mass bin. Bounds are given by 'm_gal_low' and 'm_gal_high' in bins given by 'bin_width'. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali FirstFile, LastFile, NumFile : array of integers with length equal to the number of models. The file numbers for each model that were read in (defined by the range between [FirstFile, LastFile] inclusive) and the TOTAL number of files for this model (we may only be plotting a subset of the volume). ResolutionLimit_mean : array of floats with the same shape as SMF. This is the mean stellar mass for a halo with len (number of N-body simulation particles) between 'stellar_mass_halolen_lower' and 'stellar_mass_halolen_upper'. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. observations : int Denotes whether we want to overplot observational results. 0 : Don't plot anything. 1 : Plot Song et al. (2016) at z = 6, 7, 8. 2 : Plot Baldry et al. (2008) at z = 0.1. 3 : Plot both of these. paper_plot : int Denotes whether we want to split the plotting over three panels (z = 6, 7, 8) for the paper or keep it all to one figure. output_tag : string Name of the file that will be generated. File will be saved in the current directory with the output format defined by the 'output_format' variable at the beggining of the file. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Stellar Mass is in units of log10(Msun). ''' ## Empty array initialization ## title = [] normalization_array = [] redshift_labels = [] counts_array = [] bin_middle_array = [] for model_number in range(0, len(SnapList)): counts_array.append([]) bin_middle_array.append([]) redshift_labels.append([]) #### for model_number in range(0, len(SnapList)): # Does this for each of the models. ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() box_factor = (LastFile[model_number] - FirstFile[model_number] + 1.0)/(NumFile[model_number]) # This factor allows us to take a sub-volume of the box and scale the results to represent the entire box. print("We are creating the stellar mass function using {0:.4f} of the box's volume.".format(box_factor)) norm = pow(AllVars.BoxSize,3) / pow(AllVars.Hubble_h, 3) * bin_width * box_factor normalization_array.append(norm) #### for snapshot_idx in range(0, len(SnapList[model_number])): # Loops for each snapshot in each model. tmp = 'z = %.2f' %(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) # Assigns a redshift label. redshift_labels[model_number].append(tmp) ## We perform the plotting on Rank 0 so only this rank requires the final counts array. ## if rank == 0: counts_total = np.zeros_like(SMF[model_number][snapshot_idx]) else: counts_total = None comm.Reduce([SMF[model_number][snapshot_idx], MPI.FLOAT], [counts_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. if rank == 0: counts_array[model_number].append(counts_total) bin_middle_array[model_number].append(np.arange(m_gal_low, m_gal_high+bin_width, bin_width)[:-1] + bin_width * 0.5) #### ## Plotting ## if rank == 0: # Plot only on rank 0. if paper_plot == 0: f = plt.figure() ax = plt.subplot(111) for model_number in range(0, len(SnapList)): for snapshot_idx in range(0, len(SnapList[model_number])): if model_number == 0: # We assume the redshifts for each model are the same, we only want to put a legend label for each redshift once. title = redshift_labels[model_number][snapshot_idx] else: title = '' plt.plot(bin_middle_array[model_number][snapshot_idx], counts_array[model_number][snapshot_idx] / normalization_array[model_number], color = PlotScripts.colors[snapshot_idx], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = title, linewidth = PlotScripts.global_linewidth) #print(np.min(np.log10(ResolutionLimit_mean))) #ax.axvline(np.max(np.log10(ResolutionLimit_mean)), color = 'k', linewidth = PlotScripts.global_linewidth, linestyle = '--') #ax.text(np.max(np.log10(ResolutionLimit_mean)) + 0.1, 1e-3, "Resolution Limit", color = 'k') for model_number in range(0, len(SnapList)): # Place legend labels for each of the models. NOTE: Placed after previous loop for proper formatting of labels. plt.plot(1e100, 1e100, color = 'k', linestyle = PlotScripts.linestyles[model_number], label = model_tags[model_number], rasterized=True, linewidth = PlotScripts.global_linewidth) ## Adjusting axis labels/limits. ## plt.yscale('log', nonposy='clip') plt.axis([6, 11.5, 1e-6, 1e-0]) ax.set_xlabel(r'$\log_{10}\ m_{\mathrm{*}} \:[M_{\odot}]$', fontsize = PlotScripts.global_fontsize) ax.set_ylabel(r'$\Phi\ [\mathrm{Mpc}^{-3}\: \mathrm{dex}^{-1}]$', fontsize = PlotScripts.global_fontsize) ax.xaxis.set_minor_locator(plt.MultipleLocator(0.25)) ax.set_xticks(np.arange(6.0, 12.0)) if (observations == 1 or observations == 3): # If we wanted to plot Song. Obs.Get_Data_SMF() delta = 0.05 caps = 5 ## Song (2016) Plotting ## plt.errorbar(Obs.Song_SMF_z6[:,0], 10**Obs.Song_SMF_z6[:,1], yerr= (10**Obs.Song_SMF_z6[:,1] - 10**Obs.Song_SMF_z6[:,3], 10**Obs.Song_SMF_z6[:,2] - 10**Obs.Song_SMF_z6[:,1]), xerr = 0.25, capsize = caps, elinewidth = PlotScripts.global_errorwidth, alpha = 1.0, lw=2.0, marker='o', ls='none', label = 'Song 2015, z = 6', color = PlotScripts.colors[0], rasterized=True) plt.errorbar(Obs.Song_SMF_z7[:,0], 10**Obs.Song_SMF_z7[:,1], yerr= (10**Obs.Song_SMF_z7[:,1] - 10**Obs.Song_SMF_z7[:,3], 10**Obs.Song_SMF_z7[:,2] - 10**Obs.Song_SMF_z7[:,1]), xerr = 0.25, capsize = caps, alpha=0.75, elinewidth = PlotScripts.global_errorwidth, lw=1.0, marker='o', ls='none', label = 'Song 2015, z = 7', color = PlotScripts.colors[1], rasterized=True) plt.errorbar(Obs.Song_SMF_z8[:,0], 10**Obs.Song_SMF_z8[:,1], yerr= (10**Obs.Song_SMF_z8[:,1] - 10**Obs.Song_SMF_z8[:,3], 10**Obs.Song_SMF_z8[:,2] - 10**Obs.Song_SMF_z8[:,1]), xerr = 0.25, capsize = caps, alpha=0.75, elinewidth = PlotScripts.global_errorwidth, lw=1.0, marker='o', ls='none', label = 'Song 2015, z = 8', color = PlotScripts.colors[2], rasterized=True) #### if ((observations == 2 or observations == 3) and rank == 0): # If we wanted to plot Baldry. Baldry_xval = np.log10(10 ** Obs.Baldry_SMF_z0[:, 0] /AllVars.Hubble_h/AllVars.Hubble_h) Baldry_xval = Baldry_xval - 0.26 # convert back to Chabrier IMF Baldry_yvalU = (Obs.Baldry_SMF_z0[:, 1]+Obs.Baldry_SMF_z0[:, 2]) * AllVars.Hubble_h*AllVars.Hubble_h*AllVars.Hubble_h Baldry_yvalL = (Obs.Baldry_SMF_z0[:, 1]-Obs.Baldry_SMF_z0[:, 2]) * AllVars.Hubble_h*AllVars.Hubble_h*AllVars.Hubble_h plt.fill_between(Baldry_xval, Baldry_yvalU, Baldry_yvalL, facecolor='purple', alpha=0.25, label='Baldry et al. 2008 (z=0.1)') #### leg = plt.legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile = './%s%s' %(output_tag, output_format) plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() if (paper_plot == 1): fig, ax = plt.subplots(nrows=1, ncols=3, sharex=False, sharey=True, figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 for model_number in range(0, len(SnapList)): for count in range(len(SnapList[model_number])): w = np.where((counts_array[model_number][count] > 0))[0] ax[count].plot(bin_middle_array[model_number][count][w], counts_array[model_number][count][w] / normalization_array[model_number], color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = r"$\mathbf{SAGE}$", linewidth = PlotScripts.global_linewidth) tick_locs = np.arange(6.0, 12.0) ax[count].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[count].set_xlim([6.8, 10.3]) ax[count].tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax[count].tick_params(which = 'major', length = PlotScripts.global_ticklength) ax[count].tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) ax[count].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize - delta_fontsize) ax[count].xaxis.set_minor_locator(plt.MultipleLocator(0.25)) #ax[count].set_xticks(np.arange(6.0, 12.0)) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax[count].spines[axis].set_linewidth(PlotScripts.global_axiswidth) # Since y-axis is shared, only need to do this once. ax[0].set_yscale('log', nonposy='clip') ax[0].set_yticklabels([r"$\mathbf{10^{-5}}$",r"$\mathbf{10^{-5}}$",r"$\mathbf{10^{-4}}$", r"$\mathbf{10^{-3}}$", r"$\mathbf{10^{-2}}$",r"$\mathbf{10^{-1}}$"]) ax[0].set_ylim([1e-5, 1e-1]) #ax[0].set_ylabel(r'\mathbf{$\log_{10} \Phi\ [\mathrm{Mpc}^{-3}\: \mathrm{dex}^{-1}]}$', ax[0].set_ylabel(r'$\mathbf{log_{10} \: \Phi\ [Mpc^{-3}\: dex^{-1}]}$', fontsize = PlotScripts.global_labelsize - delta_fontsize) Obs.Get_Data_SMF() PlotScripts.Plot_SMF_z6(ax[0], errorwidth=ewidth, capsize=caps) PlotScripts.Plot_SMF_z7(ax[1], errorwidth=ewidth, capsize=caps) PlotScripts.Plot_SMF_z8(ax[2], errorwidth=ewidth, capsize=caps) #### ax[0].text(0.7, 0.9, r"$\mathbf{z = 6}$", transform = ax[0].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) ax[1].text(0.7, 0.9, r"$\mathbf{z = 7}$", transform = ax[1].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) ax[2].text(0.7, 0.9, r"$\mathbf{z = 8}$", transform = ax[2].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) #leg = ax[0,0].legend(loc=2, bbox_to_anchor = (0.2, -0.5), numpoints=1, labelspacing=0.1) leg = ax[0].legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize - 2) plt.tight_layout() outputFile = "{0}_paper{1}".format(output_tag, output_format) plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() ## def plot_fesc_galaxy(SnapList, PlotSnapList, simulation_norm, mean_galaxy_fesc, std_galaxy_fesc, N_galaxy_fesc, mean_halo_fesc, std_halo_fesc, N_halo_fesc, ResolutionLimit_mean, model_tags, paper_plots, mass_global, fesc_global, Ngamma_global, output_tag): """ Plots the escape fraction as a function of stellar/halo mass. Parallel compatible. Accepts 3D arrays of the escape fraction binned into Stellar Mass bins to plot the escape fraction for multiple models. Mass units are log(Msun) Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_galaxy_fesc, std_galaxy_fesc, N_galaxy_fesc : Nested 3-dimensional array, mean_galaxy_fesc[model_number0][snapshot0] = [bin0_meanfesc, ..., binN_meanfesc], with length equal to the number of models. Mean/Standard deviation for fesc in each stellar mass bin, for each [model_number] and [snapshot_number]. N_galaxy_fesc is the number of galaxies placed into each mass bin. mean_halo_fesc, std_halo_fesc, N_halo_fesc Nested 3-dimensional array, mean_halo_fesc[model_number0][snapshot0] = [bin0_meanfesc, ..., binN_meanfesc], with length equal to the number of models. Identical to previous except using the halo virial mass for the binning rather than stellar mass. ResolutionLimit_mean : array of floats with the same shape as mean_galaxy_fesc. This is the mean stellar mass for a halo with len (number of N-body simulation particles) between 'stellar_mass_halolen_lower' and 'stellar_mass_halolen_upper'. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. paper_plots: Integer. Flag to denote whether we should plot a full, 4 panel plot for the RSAGE paper. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are log(Msun). """ def adjust_stellarmass_plot(ax): #ax.axhline(0.20, 0, 100, color ='k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') #ax.text(7.8, 0.22, r"$f_\mathrm{esc, base}$", color = 'k', # size = PlotScripts.global_fontsize) ax.set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax.set_ylabel(r'$\mathbf{\langle f_{esc}\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax.set_xlim([6.8, 10]) ax.set_ylim([0.05, 0.45]) #ax.axhline(0.35, 0, 100, color ='k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') #ax.text(9.1, 0.37, r"$f_\mathrm{esc} = 0.35$", color = 'k', # size = PlotScripts.global_fontsize) ax.xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax.yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax.tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax.tick_params(which = 'major', length = PlotScripts.global_ticklength) ax.tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax.spines[axis].set_linewidth(PlotScripts.global_axiswidth) tick_locs = np.arange(6.0, 11.0) ax.set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) tick_locs = np.arange(0.0, 0.80, 0.10) ax.set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ''' labels = ax.yaxis.get_ticklabels() locs = ax.yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) ''' leg = ax.legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') def adjust_paper_plots(ax, model_tags): ax[1,0].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[1,1].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[0,0].set_ylabel(r'$\mathbf{\langle f_{esc}\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax[1,0].set_ylabel(r'$\mathbf{\langle f_{esc}\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax_x = [0, 0, 1, 1] ax_y = [0, 1, 0, 1] for count, (x, y) in enumerate(zip(ax_x, ax_y)): ax[x,y].set_xlim([4.8, 10.4]) ax[x,y].set_ylim([0.00, 0.68]) ax[x,y].yaxis.set_major_locator(mtick.MultipleLocator(0.1)) ax[x,y].xaxis.set_major_locator(mtick.MultipleLocator(1.0)) ax[x,y].yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax[x,y].xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax[x,y].tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax[x,y].tick_params(which = 'major', length = PlotScripts.global_ticklength) ax[x,y].tick_params(which = 'minor', length = PlotScripts.global_ticklength - 2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax[x,y].spines[axis].set_linewidth(PlotScripts.global_axiswidth) print(model_tags[count]) label = model_tags[count] ax[x,y].text(0.05, 0.65, label, transform = ax[x,y].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) tick_locs = np.arange(4.0, 11.0) ax[1,0].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[1,1].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) tick_locs = np.arange(-0.1, 0.80, 0.10) ax[0,0].set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[1,0].set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) print("x") labels = ax[1,0].xaxis.get_ticklabels() locs = ax[1,0].xaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) print("y") labels = ax[1,0].yaxis.get_ticklabels() locs = ax[1,0].yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) print("Plotting fesc as a function of stellar mass.") ## Array initialization ## master_mean_fesc_stellar, master_std_fesc_stellar, master_N_fesc_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_fesc, std_galaxy_fesc, N_galaxy_fesc, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: fig, ax = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) fig2, ax2 = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 count_x = 0 for count, model_number in enumerate(range(0, len(SnapList))): if count == 2: count_x += 1 print("There were a total of {0} galaxies over the entire redshift range.".format(sum(N_halo_fesc[model_number]))) ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## w = np.where((master_N_fesc_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_fesc_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: print(master_mean_fesc_stellar[model_number][snapshot_idx]) ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_fesc_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_fesc_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[0], rasterized = True, label = label, lw = PlotScripts.global_linewidth) #w = np.random.randint(0, # len(mass_global[model_number][snapshot_idx][0]), # size=500) #sc = ax2[count_x, count%2].scatter(mass_global[model_number][snapshot_idx][0][w], # fesc_global[model_number][snapshot_idx][0][w], # c=np.log10(Ngamma_global[model_number][snapshot_idx][0][w]*1.0e50), # alpha = 0.5,cmap='plasma') #plt.colorbar(sc) #ax2[count_x, count%2].hexbin(mass_global[model_number][snapshot_idx], # fesc_global[model_number][snapshot_idx], # C=Ngamma_global[model_number][snapshot_idx]) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: adjust_paper_plots(ax, model_tags) leg = ax[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) #leg = ax2[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) #leg.draw_frame(False) # Don't want a box frame #for t in leg.get_texts(): # Reduce the size of the text # t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) if paper_plots == 1: outputFile = './%s_scatter%s' %(output_tag, output_format) fig2.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig2) ## def plot_reionmod_galaxy(SnapList, PlotSnapList, simulation_norm, mean_galaxy_reionmod, std_galaxy_reionmod, N_galaxy_reionmod, mean_galaxy_reionmod_gnedin, std_galaxy_reionmod_gnedin, model_tags, paper_plots, output_tag): """ """ def adjust_paper_plots(ax, model_tags): ax[1,0].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[1,1].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[0,0].set_ylabel(r'$\mathbf{\langle ReionMod\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax[1,0].set_ylabel(r'$\mathbf{\langle ReionMod\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax_x = [0, 0, 1, 1] ax_y = [0, 1, 0, 1] for count, (x, y) in enumerate(zip(ax_x, ax_y)): ax[x,y].set_xlim([4.8, 10.4]) ax[x,y].set_ylim([0.00, 1.05]) #ax[x,y].yaxis.set_major_locator(mtick.MultipleLocator(0.1)) ax[x,y].xaxis.set_major_locator(mtick.MultipleLocator(1.0)) #ax[x,y].yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax[x,y].xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax[x,y].tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax[x,y].tick_params(which = 'major', length = PlotScripts.global_ticklength) ax[x,y].tick_params(which = 'minor', length = PlotScripts.global_ticklength - 2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax[x,y].spines[axis].set_linewidth(PlotScripts.global_axiswidth) print(model_tags[count]) label = model_tags[count] ax[x,y].text(0.05, 0.65, label, transform = ax[x,y].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) tick_locs = np.arange(4.0, 11.0) ax[1,0].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[1,1].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(-0.1, 0.80, 0.10) #ax[0,0].set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], #fontsize = PlotScripts.global_fontsize) #ax[1,0].set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) def adjust_redshift_panels(ax, redshift_tags): ax[1,0].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[1,1].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[0,0].set_ylabel(r'$\mathbf{\langle ReionMod\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax[1,0].set_ylabel(r'$\mathbf{\langle ReionMod\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax_x = [0, 0, 1, 1] ax_y = [0, 1, 0, 1] for count, (x, y) in enumerate(zip(ax_x, ax_y)): ax[x,y].set_xlim([4.8, 10.4]) ax[x,y].set_ylim([0.00, 1.05]) #ax[x,y].yaxis.set_major_locator(mtick.MultipleLocator(0.1)) ax[x,y].xaxis.set_major_locator(mtick.MultipleLocator(1.0)) #ax[x,y].yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax[x,y].xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax[x,y].tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax[x,y].tick_params(which = 'major', length = PlotScripts.global_ticklength) ax[x,y].tick_params(which = 'minor', length = PlotScripts.global_ticklength - 2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax[x,y].spines[axis].set_linewidth(PlotScripts.global_axiswidth) label = redshift_tags[count] ax[x,y].text(0.05, 0.65, label, transform = ax[x,y].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) tick_locs = np.arange(4.0, 11.0) ax[1,0].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[1,1].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) print("Reionization Modifier as a function of stellar mass.") ## Array initialization ## master_mean_reionmod_stellar, master_std_reionmod_stellar, master_N_reionmod_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_reionmod, std_galaxy_reionmod, N_galaxy_reionmod, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) master_mean_reionmod_gnedin_stellar, master_std_reionmod_gnedin_stellar, master_N_reionmod_gnedin_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_reionmod_gnedin, std_galaxy_reionmod_gnedin, N_galaxy_reionmod, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: fig, ax = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) fig2, ax2 = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 count_x = 0 for count, model_number in enumerate(range(0, len(SnapList))): if count == 2: count_x += 1 plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## w = np.where((master_N_reionmod_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_reionmod_stellar[model_number][snapshot_idx][w] = np.nan master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[0], rasterized = True, label = label, lw = PlotScripts.global_linewidth) ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[1], rasterized = True, label = label, lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break z_labels = [] for model_number in range(0, len(SnapList)): count_x = 0 plot_count = 0 for count, snapshot_idx in enumerate(range(len(SnapList[model_number]))): if count == 2: count_x += 1 if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): label = model_tags[model_number] if (model_number == 0): z_label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" z_labels.append(z_label) ## Plots as a function of stellar mass ## w = np.where((master_N_reionmod_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_reionmod_stellar[model_number][snapshot_idx][w] = np.nan master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx][w] = np.nan if (model_number == 0): print(master_mean_reionmod_stellar[model_number][snapshot_idx]) ax2[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_stellar[model_number][snapshot_idx], color = PlotScripts.colors[model_number], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) if (model_number == 0): ax2[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx], color = 'k', ls = '--', rasterized = True, label = "Gnedin", lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: adjust_paper_plots(ax, model_tags) print(z_labels) adjust_redshift_panels(ax2, z_labels) leg = ax[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') leg = ax2[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) #leg = ax2[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) #leg.draw_frame(False) # Don't want a box frame #for t in leg.get_texts(): # Reduce the size of the text # t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) ## Output ## outputFile = "{0}{1}".format(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) outputFile2 = "{0}_redshiftpanels{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ## def plot_nion_galaxy(SnapList, PlotSnapList, simulation_norm, mean_Ngamma_galaxy, std_Ngamma_galaxy, N_Ngamma_galaxy, model_tags, paper_plots, output_tag): """ Plots the number of ionizing photons emitted (not necessarily escaped) as a function of galaxy stellar mass. Parallel compatible. Accepts 3D arrays of the escape fraction binned into Stellar Mass bins to plot the escape fraction for multiple models. Mass units are log(Msun) Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_galaxy_Ngamma, std_galaxy_Ngamma, N_galaxy_Ngamma : Nested 3-dimensional array, mean_galaxy_Ngamma[model_number0][snapshot0] = [bin0_meanNgamma, ..., binN_meanNgamma], with length equal to the number of models. Mean/Standard deviation for Ngamma in each stellar mass bin, for each [model_number] and [snapshot_number]. N_galaxy_Ngamma is the number of galaxies placed into each mass bin. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. paper_plots: Integer. Flag to denote whether we should plot a full, 4 panel plot for the RSAGE paper. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are log(Msun). Ngamma units are 1.0e50 photons/s. """ def adjust_stellarmass_plot(ax): #ax.axhline(0.20, 0, 100, color ='k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') #ax.text(7.8, 0.22, r"$f_\mathrm{esc, base}$", color = 'k', # size = PlotScripts.global_fontsize) ax.set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax.set_ylabel(r'$\mathbf{\log_{10}\langle f_{esc} N_\gamma\rangle_{M_*}}$', size = PlotScripts.global_labelsize) ax.set_xlim([6.8, 10]) #ax.set_ylim([0.05, 0.45]) #ax.axhline(0.35, 0, 100, color ='k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') #ax.text(9.1, 0.37, r"$f_\mathrm{esc} = 0.35$", color = 'k', # size = PlotScripts.global_fontsize) ax.xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) #ax.yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax.tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax.tick_params(which = 'major', length = PlotScripts.global_ticklength) ax.tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax.spines[axis].set_linewidth(PlotScripts.global_axiswidth) tick_locs = np.arange(6.0, 11.0) ax.set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(0.0, 0.80, 0.10) #ax.set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) ''' labels = ax.yaxis.get_ticklabels() locs = ax.yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) ''' leg = ax.legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') def adjust_paper_plots(ax, z_tags): ax[1,0].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[1,1].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax[0,0].set_ylabel(r'$\mathbf{\Sigma log_{10}\langle f_{esc} N_\gamma\rangle_{M_*}}$', size = PlotScripts.global_labelsize - 10) ax[1,0].set_ylabel(r'$\mathbf{\Sigma log_{10}\langle f_{esc} N_\gamma\rangle_{M_*}}$', size = PlotScripts.global_labelsize - 10) ax_x = [0, 0, 1, 1] ax_y = [0, 1, 0, 1] for count, (x, y) in enumerate(zip(ax_x, ax_y)): ax[x,y].set_xlim([4.8, 10.4]) ax[x,y].set_ylim([47, 55]) #ax[x,y].yaxis.set_major_locator(mtick.MultipleLocator(0.1)) ax[x,y].xaxis.set_major_locator(mtick.MultipleLocator(1.0)) #ax[x,y].yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax[x,y].xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax[x,y].tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax[x,y].tick_params(which = 'major', length = PlotScripts.global_ticklength) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax[x,y].spines[axis].set_linewidth(PlotScripts.global_axiswidth) print(z_tags[count]) label = r"$\mathbf{z = " + \ str(int(round(float(z_tags[count])))) +\ "}$" ax[x,y].text(0.7, 0.8, label, transform = ax[x,y].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) tick_locs = np.arange(4.0, 11.0) ax[1,0].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[1,1].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(0.0, 0.80, 0.10) #ax[0,0].set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) #ax[1,0].set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) print("x") labels = ax[1,0].xaxis.get_ticklabels() locs = ax[1,0].xaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) print("y") labels = ax[1,0].yaxis.get_ticklabels() locs = ax[1,0].yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) print("Plotting Ngamma*fesc as a function of stellar mass.") ## Array initialization ## master_mean_Ngamma_stellar, master_std_Ngamma_stellar, master_N_Ngamma_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_Ngamma_galaxy, std_Ngamma_galaxy, N_Ngamma_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: fig, ax = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 z_tags = np.zeros_like(model_tags, dtype=np.float32) for model_number in range(0, len(SnapList)): count_x = 0 ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for count, snapshot_idx in enumerate(range(0, len(SnapList[model_number]))): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if count == 2: count_x += 1 label = model_tags[model_number] z_tags[count] = float(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) ## Plots as a function of stellar mass ## w = np.where((master_N_Ngamma_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_Ngamma_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], np.log10(master_mean_Ngamma_stellar[model_number][snapshot_idx]*1.0e50), color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], np.log10(master_mean_Ngamma_stellar[model_number][snapshot_idx]*1.0e50), color = PlotScripts.colors[model_number], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: adjust_paper_plots(ax, z_tags) leg = ax[0,0].legend(loc="upper left", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) ## def plot_photo_galaxy(SnapList, PlotSnapList, simulation_norm, mean_photo_galaxy, std_photo_galaxy, N_photo_galaxy, model_tags, paper_plots, output_tag): """ Plots the photoionization rate as a function of galaxy stellar mass. Parallel compatible. Accepts 3D arrays of the escape fraction binned into Stellar Mass bins to plot the escape fraction for multiple models. Mass units are log(Msun) Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_photo_galaxy, std_photo_galaxy, N_photo_galaxy : Nested 3-dimensional array, mean_photo_galaxy[model_number0][snapshot0] = [bin0_meanphoto, ..., binN_meanphoto], with length equal to the number of models. Mean/Standard deviation for Photionization Rate in each stellar mass bin, for each [model_number] and [snapshot_number]. N_photo_galaxy is the number of galaxies placed into each mass bin. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. paper_plots: Integer. Flag to denote whether we should plot a full, 4 panel plot for the RSAGE paper. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are log(Msun). Ngamma units are 1.0e50 photons/s. """ def adjust_stellarmass_plot(ax): ax.set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', size = PlotScripts.global_fontsize) ax.set_ylabel(r'$\mathbf{log_{10} \: \Gamma \: [s^{-1}}$', size = PlotScripts.global_labelsize) ax.set_xlim([4.8, 10]) #ax.set_ylim([0.05, 0.45]) ax.xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) #ax.yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax.tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax.tick_params(which = 'major', length = PlotScripts.global_ticklength) ax.tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax.spines[axis].set_linewidth(PlotScripts.global_axiswidth) #tick_locs = np.arange(4.0, 11.0) #ax.set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(0.0, 0.80, 0.10) #ax.set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) ''' labels = ax.yaxis.get_ticklabels() locs = ax.yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) ''' leg = ax.legend(loc="lower right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') print("Plotting photoionization rate as a function of stellar mass.") ## Array initialization ## master_mean_photo_stellar, master_std_photo_stellar, master_N_photo_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_photo_galaxy, std_photo_galaxy, N_photo_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: pass for model_number in range(0, len(SnapList)): count_x = 0 ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for count, snapshot_idx in enumerate(range(0, len(SnapList[model_number]))): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## w = np.where((master_N_photo_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_photo_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], np.log10(master_mean_photo_stellar[model_number][snapshot_idx]), color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: pass plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break for model_number in range(0, len(SnapList)): ax1.plot(np.nan, np.nan, color = 'k', label = model_tags[model_number], lw = PlotScripts.global_linewidth, ls = PlotScripts.linestyles[model_number]) ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: pass ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) ## ## def plot_sfr_galaxy(SnapList, PlotSnapList, simulation_norm, mean_galaxy_sfr, std_galaxy_sfr, mean_galaxy_ssfr, std_galaxy_ssfr, N_galaxy, model_tags, output_tag): """ Plots the specific star formation rate (sSFR) as a function of stellar mass. Parallel compatible. Accepts 3D arrays of the sSFR binned into Stellar Mass bins. Mass units log(Msun). Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_galaxy_ssfr, std_galaxy_ssfr, N_galaxy_ssfr : Nested 3-dimensional array, mean_galaxy_sfr[model_number0][snapshot0] = [bin0_meanssfr, ..., binN_meanssfr], with length equal to the number of models. Mean/Standard deviation for sSFR in each stellar mass bin, for each [model_number] and [snapshot_number]. N_galaxy_fesc is the number of galaxies placed into each mass bin. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are 1e10 Msun (no h). """ def adjust_sfr_plot(ax): ax.set_xlabel(r'$\log_{10}\ M_*\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax.set_ylabel(r'$\mathbf{\langle \mathrm{SFR}\rangle_{M_*}\:[M_\odot\mathrm{yr}^{-1}]}$', size = PlotScripts.global_labelsize) ax.set_xlim([4.8, 10]) ax.set_ylim([-3, 2]) ax.xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax.yaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax.tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax.tick_params(which = 'major', length = PlotScripts.global_ticklength) ax.tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax.spines[axis].set_linewidth(PlotScripts.global_axiswidth) tick_locs = np.arange(6.0, 11.0) ax.set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(0.0, 0.80, 0.10) #ax.set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) labels = ax.yaxis.get_ticklabels() locs = ax.yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) leg = ax.legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') def adjust_ssfr_plot(ax): ax.set_xlabel(r'$\log_{10}\ M_*\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax.set_ylabel(r'$\mathbf{\langle\mathrm{sSFR}\rangle_{M_*}\:[\mathrm{yr^{-1}}}$', size = PlotScripts.global_labelsize) ax.set_xlim([4.8, 10]) ax.set_ylim([-9, -4]) ax.xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax.yaxis.set_minor_locator(mtick.MultipleLocator(0.1)) ax.tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax.tick_params(which = 'major', length = PlotScripts.global_ticklength) ax.tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax.spines[axis].set_linewidth(PlotScripts.global_axiswidth) tick_locs = np.arange(6.0, 11.0) ax.set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(0.0, 0.80, 0.10) #ax.set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) labels = ax.yaxis.get_ticklabels() locs = ax.yaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) leg = ax.legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') print("Plotting sSFR as a function of stellar mass.") ## Array initialization ## master_mean_sfr_stellar, master_std_sfr_stellar, master_N_sfr_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_sfr, std_galaxy_sfr, N_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) master_mean_ssfr_stellar, master_std_ssfr_stellar, master_N_ssfr_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_ssfr, std_galaxy_ssfr, N_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: fig = plt.figure() ax1 = fig.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(0, len(SnapList)): ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_sfr_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) ax2.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_ssfr_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break #for model_number in range(0, len(SnapList)): # Just plot some garbage to get the legend labels correct. #ax1.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) #ax3.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ## Stellar Mass plots ## adjust_sfr_plot(ax1) adjust_ssfr_plot(ax2) ## Output ## outputFile = "./{0}SFR{1}".format(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) outputFile = "./{0}sSFR{1}".format(output_tag, output_format) fig2.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) ## ## def plot_fej_Ngamma(SnapList, PlotSnapList, simulation_norm, mean_Ngamma_fej, std_Ngamma_fej, N_fej, model_tags, output_tag): def adjust_plot(ax): ax.set_xlabel(r'$\mathbf{f_\mathrm{ej}}$', size = PlotScripts.global_fontsize) ax.set_ylabel(r'$\mathbf{\log_{10}\langle N_\gamma\rangle_{f_{ej}}}$', size = PlotScripts.global_labelsize) ax.set_xlim([0.0, 1.0]) #ax.set_ylim([0.05, 0.45]) #ax.axhline(0.35, 0, 100, color ='k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') #ax.text(9.1, 0.37, r"$f_\mathrm{esc} = 0.35$", color = 'k', # size = PlotScripts.global_fontsize) ax.xaxis.set_minor_locator(mtick.MultipleLocator(0.10)) #ax.yaxis.set_minor_locator(mtick.MultipleLocator(0.05)) ax.tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax.tick_params(which = 'major', length = PlotScripts.global_ticklength) ax.tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax.spines[axis].set_linewidth(PlotScripts.global_axiswidth) #tick_locs = np.arange(6.0, 11.0) #ax.set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) #tick_locs = np.arange(0.0, 0.80, 0.10) #ax.set_yticklabels([r"$\mathbf{%.2f}$" % x for x in tick_locs], # fontsize = PlotScripts.global_fontsize) labels = ax.xaxis.get_ticklabels() locs = ax.xaxis.get_ticklocs() for label, loc in zip(labels, locs): print("{0} {1}".format(label, loc)) leg = ax.legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') ## Array initialization ## master_mean_Ngamma_fej, master_std_Ngamma_fej, master_N_Ngamma_fej, master_bin_middle_fej = \ collect_across_tasks(mean_Ngamma_fej, std_Ngamma_fej, N_fej, SnapList, PlotSnapList, True, fej_low, fej_high, fej_bin_width) if rank == 0: fig = plt.figure() ax1 = fig.add_subplot(111) ax2 = ax1.twinx() for model_number in range(0, len(SnapList)): ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): label = model_tags[model_number] w = np.where((master_N_Ngamma_fej[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_Ngamma_fej[model_number][snapshot_idx][w] = np.nan ax1.plot(master_bin_middle_fej[model_number][snapshot_idx], np.log10(master_mean_Ngamma_fej[model_number][snapshot_idx]*1.0e50), color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) #ax1.plot(master_bin_middle_fej[model_number][snapshot_idx], # np.log10(master_mean_Ngamma_fej[model_number][snapshot_idx]*1.0e50 # * master_N_Ngamma_fej[model_number][snapshot_idx]), # color = PlotScripts.colors[plot_count], # ls = PlotScripts.linestyles[model_number], # rasterized = True, label = label, #lw = PlotScripts.global_linewidth) ''' ax2.plot(master_bin_middle_fej[model_number][snapshot_idx], np.log10(master_N_Ngamma_fej[model_number][snapshot_idx]), color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) ''' plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break adjust_plot(ax1) leg = ax1.legend(loc="upper center", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) def plot_ejectedfraction(SnapList, PlotSnapList, simulation_norm, mean_mvir_ejected, std_mvir_ejected, N_ejected, mean_ejected_z, std_ejected_z, N_z, model_tags, output_tag): ''' Plots the ejected fraction as a function of the halo mass. Parallel compatible. Accepts a 3D array of the ejected fraction so we can plot for multiple models and redshifts. Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. mean_mvir_ejected, std_mvir_ejected, N_ejected : Nested 3-dimensional array, mean_mvir_ejected[model_number0][snapshot0] = [bin0_meanejected, ..., binN_meanejected], with length equal to the number of models. Mean/Standard deviation for the escape fraction binned into Halo Mass bins. N_ejected is the number of data points in each bin. Bounds are given by 'm_low' and 'm_high' in bins given by 'bin_width'. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Halo Mass is in units of log10(Msun). ''' print("Plotting the Ejected Fraction as a function of halo mass.") master_mean_ejected_halo, master_std_ejected_halo, master_N_ejected_halo, master_bin_middle_halo = \ collect_across_tasks(mean_mvir_ejected, std_mvir_ejected, N_ejected, SnapList, PlotSnapList, True, m_low, m_high) master_mean_ejected_z, master_std_ejected_z, master_N_ejected_z, _ = \ collect_across_tasks(mean_ejected_z, std_ejected_z, N_z, SnapList) if rank == 0: fig1 = plt.figure() ax1 = fig1.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(0, len(SnapList)): if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() for snapshot_idx in range(0, len(PlotSnapList[model_number])): label = AllVars.SnapZ[PlotSnapList[model_number][snapshot_idx]] ax1.plot(master_bin_middle_halo[model_number][snapshot_idx], master_mean_ejected_halo[model_number][snapshot_idx], color = PlotScripts.colors[snapshot_idx], linestyle = PlotScripts.linestyles[model_number], label = label, lw = PlotScripts.global_linewidth) ax2.plot((AllVars.t_BigBang - AllVars.Lookback_Time[SnapList[model_number]]) * 1.0e3, master_mean_ejected_z[model_number], color = PlotScripts.colors[model_number], label = model_tags[model_number], ls = PlotScripts.linestyles[model_number], lw = PlotScripts.global_linewidth) for model_number in range(0, len(SnapList)): # Just plot some garbage to get the legend labels correct. ax1.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ax1.set_xlabel(r'$\log_{10}\ M_{\mathrm{vir}}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$\mathrm{Ejected \: Fraction}$', size = PlotScripts.global_fontsize) ax1.set_xlim([8.0, 12]) ax1.set_ylim([-0.05, 1.0]) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(0.1)) ax1.yaxis.set_minor_locator(mtick.MultipleLocator(0.025)) leg = ax1.legend(loc=1, numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') outputFile = "./{0}{1}".format(output_tag, output_format) fig1.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig1) ax2.set_xlabel(r"$\mathbf{Time \: since \: Big \: Bang \: [Myr]}$", fontsize = PlotScripts.global_labelsize) tick_locs = np.arange(200.0, 1000.0, 100.0) tick_labels = [r"$\mathbf{%d}$" % x for x in tick_locs] ax2.xaxis.set_major_locator(mtick.MultipleLocator(100)) ax2.set_xticklabels(tick_labels, fontsize = PlotScripts.global_fontsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_ylabel(r'$\mathbf{Mean f_{ej}}$', fontsize = PlotScripts.global_labelsize) ax3 = ax2.twiny() t_plot = (AllVars.t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$\mathbf{%d}$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax3.set_xlabel(r"$\mathbf{z}$", fontsize = PlotScripts.global_labelsize) ax3.set_xlim(PlotScripts.time_xlim) ax3.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax3.set_xticklabels(z_labels, fontsize = PlotScripts.global_fontsize) # But label them as redshifts. leg = ax2.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile2 = "./{0}_z{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ## def plot_mvir_fesc(SnapList, mass_central, fesc, model_tags, output_tag): title = [] redshift_labels = [] mean_fesc_array = [] std_fesc_array = [] mean_halomass_array = [] std_halomass_array = [] bin_middle_array = [] for model_number in range(0, len(SnapList)): redshift_labels.append([]) mean_fesc_array.append([]) std_fesc_array.append([]) mean_halomass_array.append([]) std_halomass_array.append([]) bin_middle_array.append([]) print("Plotting fesc against Mvir") binwidth = 0.1 Frequency = 1 for model_number in range(0, len(SnapList)): for snapshot_idx in range(0, len(SnapList[model_number])): print("Doing Snapshot {0}".format(SnapList[model_number][snapshot_idx])) tmp = 'z = %.2f' %(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) redshift_labels[model_number].append(tmp) minimum_mass = np.floor(min(mass_central[model_number][snapshot_idx])) - 10*binwidth maximum_mass = np.floor(max(mass_central[model_number][snapshot_idx])) + 10*binwidth minimum_mass = 6.0 maximum_mass = 12.0 binning_minimum = comm.allreduce(minimum_mass, op = MPI.MIN) binning_maximum = comm.allreduce(maximum_mass, op = MPI.MAX) halomass_nonlog = [10**x for x in mass_central[model_number][snapshot_idx]] (mean_fesc, std_fesc, N, bin_middle) = AllVars.Calculate_2D_Mean(mass_central[model_number][snapshot_idx], fesc[model_number][snapshot_idx], binwidth, binning_minimum, binning_maximum) mean_fesc_array[model_number], std_fesc_array[model_number] = calculate_pooled_stats(mean_fesc_array[model_number], std_fesc_array[model_number], mean_fesc, std_fesc, N) mean_halomass_array[model_number], std_halomass_array[model_number] = calculate_pooled_stats(mean_halomass_array[model_number], std_halomass_array[model_number], np.mean(halomass_nonlog), np.std(halomass_nonlog), len(mass_central[model_number][snapshot_idx])) ## If want to do mean/etc of halo mass need to update script. ## bin_middle_array[model_number].append(bin_middle) mean_halomass_array[model_number] = np.log10(mean_halomass_array[model_number]) if rank == 0: f = plt.figure() ax1 = plt.subplot(111) for model_number in range(0, len(SnapList)): for snapshot_idx in range(0, len(SnapList[model_number])): if model_number == 0: title = redshift_labels[model_number][snapshot_idx] else: title = '' mean = mean_fesc_array[model_number][snapshot_idx] std = std_fesc_array[model_number][snapshot_idx] bin_middle = bin_middle_array[model_number][snapshot_idx] ax1.plot(bin_middle, mean, color = colors[snapshot_idx], linestyle = linestyles[model_number], rasterized = True, label = title) #ax1.scatter(mean_halomass_array[model_number][snapshot_idx], np.mean(~np.isnan(mean)), color = colors[snapshot_idx], marker = 'o', rasterized = True, s = 40, lw = 3) if (len(SnapList) == 1): ax1.fill_between(bin_middle, np.subtract(mean,std), np.add(mean,std), color = colors[snapshot_idx], alpha = 0.25) ax1.set_xlabel(r'$\log_{10}\ M_{\mathrm{vir}}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$f_\mathrm{esc}$', size = PlotScripts.global_fontsize) #ax1.set_xlim([8.5, 12]) #ax1.set_ylim([0.0, 1.0]) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(0.1)) # ax1.yaxis.set_minor_locator(mtick.MultipleLocator(0.1)) # ax1.set_yscale('log', nonposy='clip') # for model_number in range(0, len(SnapList)): # ax1.plot(1e100, 1e100, color = 'k', ls = linestyles[model_number], label = model_tags[model_number], rasterized=True) leg = ax1.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') outputFile = './' + output_tag + output_format plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to'.format(outputFile)) plt.close() ## def plot_mvir_Ngamma(SnapList, mean_mvir_Ngamma, std_mvir_Ngamma, N_Ngamma, model_tags, output_tag,fesc_prescription=None, fesc_normalization=None, fitpath=None): ''' Plots the number of ionizing photons (pure ngamma times fesc) as a function of halo mass. Parallel compatible. The input data has been binned as a function of halo virial mass (Mvir), with the bins defined at the top of the file (m_low, m_high, bin_width). Accepts 3D arrays to plot ngamma for multiple models. Parameters ---------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. mean_mvir_Ngamma, std_mvir_Ngamma, N_Ngamma : Nested 2-dimensional array, mean_mvir_Ngamma[model_number0][snapshot0] = [bin0_meanNgamma, ..., binN_meanNgamma], with length equal to the number of bins. Mean/Standard deviation/number of data points in each halo mass (Mvir) bin. The number of photons is in units of 1.0e50 s^-1. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. fesc_prescription : int (optional) If this parameter is defined, we will save the Mvir-Ngamma results in a text file (not needed if not saving). Number that controls what escape fraction prescription was used to generate the escape fractions. 0 : Constant, fesc = Constant. 1 : Scaling with Halo Mass, fesc = A*Mh^B. 2 : Scaling with ejected fraction, fesc = fej*A + B. fesc_normalization : float (if fesc_prescription == 0) or `numpy.darray' with length 2 (if fesc_prescription == 1 or == 2) (optional). If this parameter is defined, we will save the Mvir-Ngamma results in a text file (not needed if not saving). Parameter not needed if you're not saving the Mvir-Ngamma results. If fesc_prescription == 0, gives the constant value for the escape fraction. If fesc_prescription == 1 or == 2, gives A and B with the form [A, B]. fitpath : string (optional) If this parameter is defined, we will save the Mvir-Ngamma results in a text file (not needed if not saving). Defines the base path for where we are saving the results. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Ngamma is in units of 1.0e50 s^-1. ''' print("Plotting ngamma*fesc against the halo mass") ## Array initialization. ## title = [] redshift_labels = [] mean_ngammafesc_array = [] std_ngammafesc_array = [] mean_halomass_array = [] std_halomass_array = [] bin_middle_array = [] for model_number in range(0, len(SnapList)): redshift_labels.append([]) mean_ngammafesc_array.append([]) std_ngammafesc_array.append([]) mean_halomass_array.append([]) std_halomass_array.append([]) bin_middle_array.append([]) for model_number in range(0, len(SnapList)): for snapshot_idx in range(0, len(SnapList[model_number])): print("Doing Snapshot {0}".format(SnapList[model_number][snapshot_idx])) tmp = 'z = %.2f' %(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) redshift_labels[model_number].append(tmp) N = N_Ngamma[model_number][snapshot_idx] mean_ngammafesc_array[model_number], std_ngammafesc_array[model_number] = calculate_pooled_stats(mean_ngammafesc_array[model_number], std_ngammafesc_array[model_number], mean_mvir_Ngamma[model_number][snapshot_idx], std_mvir_Ngamma[model_number][snapshot_idx], N) # Collate the values from all processors. bin_middle_array[model_number].append(np.arange(m_low, m_high+bin_width, bin_width)[:-1] + bin_width * 0.5) if rank == 0: f = plt.figure() ax1 = plt.subplot(111) for model_number in range(0, len(SnapList)): count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if model_number == 0: title = redshift_labels[model_number][snapshot_idx] else: title = '' mean = np.zeros((len(mean_ngammafesc_array[model_number][snapshot_idx])), dtype = np.float32) std = np.zeros((len(mean_ngammafesc_array[model_number][snapshot_idx])), dtype=np.float32) for i in range(0, len(mean)): if(mean_ngammafesc_array[model_number][snapshot_idx][i] < 1e-10): mean[i] = np.nan std[i] = np.nan else: mean[i] = np.log10(mean_ngammafesc_array[model_number][snapshot_idx][i] * 1.0e50) # Remember that the input data is in units of 1.0e50 s^-1. std[i] = 0.434 * std_ngammafesc_array[model_number][snapshot_idx][i] / mean_ngammafesc_array[model_number][snapshot_idx][i] # We're plotting in log space so the standard deviation is 0.434*log10(std)/log10(mean). bin_middle = bin_middle_array[model_number][snapshot_idx] if (count < 4): # Only plot at most 5 lines. ax1.plot(bin_middle, mean, color = PlotScripts.colors[snapshot_idx], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = title, linewidth = PlotScripts.global_linewidth) count += 1 ## In this block we save the Mvir-Ngamma results to a file. ## if (fesc_prescription == None or fesc_normalization == None or fitpath == None): raise ValueError("You've specified you want to save the Mvir-Ngamma results but haven't provided an escape fraction prescription, normalization and base path name") # Note: All the checks that escape fraction normalization was written correctly were performed in 'calculate_fesc()', hence it will be correct by this point and we don't need to double check. if (fesc_prescription[model_number] == 0): # Slightly different naming scheme for the constant case (it only has a float for fesc_normalization). fname = "%s/fesc%d_%.3f_z%.3f.txt" %(fitpath, fesc_prescription[model_number], fesc_normalization[model_number], AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) elif (fesc_prescription[model_number] == 1 or fesc_prescription[model_number] == 2): fname = "%s/fesc%d_A%.3eB%.3f_z%.3f.txt" %(fitpath, fesc_prescription[model_number], fesc_normalization[model_number][0], fesc_normalization[model_number][1], AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) f = open(fname, "w+") if not os.access(fname, os.W_OK): print("The filename is {0}".format(fname)) raise ValueError("Can't write to this file.") for i in range(0, len(bin_middle)): f.write("%.4f %.4f %.4f %d\n" %(bin_middle[i], mean[i], std[i], N_Ngamma[model_number][snapshot_idx][i])) f.close() print("Wrote successfully to file {0}".format(fname)) ## for model_number in range(0, len(SnapList)): # Just plot some garbage to get the legend labels correct. ax1.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ax1.set_xlabel(r'$\log_{10}\ M_{\mathrm{vir}}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$\log_{10}\ \dot{N}_\gamma \: f_\mathrm{esc} \: [\mathrm{s}^{-1}]$', size = PlotScripts.global_fontsize) ax1.set_xlim([8.5, 12]) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(0.1)) leg = ax1.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') outputFile = './' + output_tag + output_format plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to'.format(outputFile)) plt.close() def bin_Simfast_halos(RedshiftList, SnapList, halopath, fitpath, fesc_prescription, fesc_normalization, GridSize, output_tag): for model_number in range(0, len(fesc_prescription)): for halo_z_idx in range(0, len(RedshiftList)): snapshot_idx = min(range(len(SnapList)), key=lambda i: abs(SnapList[i]-RedshiftList[halo_z_idx])) # This finds the index of the simulation redshift that most closely matches the Halo redshift. print("Binning Halo redshift {0}".format(RedshiftList[halo_z_idx])) print("For the Halo redshift {0:.3f} the nearest simulation redshift is {1:.3f}".format(RedshiftList[halo_z_idx], SnapList[snapshot_idx])) if (fesc_prescription[model_number] == 0): fname = "%s/fesc%d_%.3f_z%.3f.txt" %(fitpath, fesc_prescription[model_number], fesc_normalization[model_number], AllVars.SnapZ[snapshot_idx]) elif (fesc_prescription[model_number] == 1 or fesc_prescription[model_number] == 2): fname = "%s/fesc%d_A%.3eB%.3f_z%.3f.txt" %(fitpath, fesc_prescription[model_number], fesc_normalization[model_number][0], fesc_normalization[model_number][1], AllVars.SnapZ[snapshot_idx]) print("Reading in file {0}".format(fname)) ## Here we read in the results from the Mvir-Ngamma binning. ## f = open(fname, 'r') fit_mvir, fit_mean, fit_std, fit_N = np.loadtxt(f, unpack = True) f.close() ## Here we read in the halos created by Simfast21 ## # The data file has the structure: # long int N_halos # Then an entry for each halo: # float Mass # float x, y, z positions. # NOTE: The x,y,z positions are the grid indices but are still floats (because Simfast21 is weird like that). Halodesc_full = [ ('Halo_Mass', np.float32), ('Halo_x', np.float32), ('Halo_y', np.float32), ('Halo_z', np.float32) ] names = [Halodesc_full[i][0] for i in range(len(Halodesc_full))] formats = [Halodesc_full[i][1] for i in range(len(Halodesc_full))] Halo_Desc = np.dtype({'names':names, 'formats':formats}, align=True) fname = "%s/halonl_z%.3f_N%d_L100.0.dat.catalog" %(halopath, RedshiftList[halo_z_idx], GridSize) f = open(fname, 'rb') N_Halos = np.fromfile(f, count = 1, dtype = np.long) Halos = np.fromfile(f, count = N_Halos, dtype = Halo_Desc) binned_nion = np.zeros((GridSize*GridSize*GridSize), dtype = float32) # This grid will contain the ionizing photons that results from the binning. binned_Halo_Mass = np.digitize(np.log10(Halos['Halo_Mass']), fit_mvir) # Places the Simfast21 halos into the correct halo mass bins defined by the Mvir-Ngamma results. binned_Halo_Mass[binned_Halo_Mass == len(fit_mvir)] = len(fit_mvir) - 1 # Fixes up the edge case. ## Fore each Halo we now assign it an ionizing flux. ## # This flux is determined by drawing a random number from a normal distribution with mean and standard deviation given by the Mvir-Ngamma results. # NOTE: Remember the Mvir-Ngamma results are in units of log10(s^-1). fit_nan = 0 for i in range(0, N_Halos): if(np.isnan(fit_mean[binned_Halo_Mass[i]]) == True or np.isnan(fit_std[binned_Halo_Mass[i]]) == True): # This halo had mass that was not covered by the Mvir-Ngamma fits. fit_nan += 1 continue nion_halo = np.random.normal(fit_mean[binned_Halo_Mass[i]], fit_std[binned_Halo_Mass[i]]) ## Because of how Simfast21 does their binning, we have some cases where the Halos are technically outside the box. Just fix them up. ## x_grid = int(Halos['Halo_x'][i]) if x_grid >= GridSize: x_grid = GridSize - 1 if x_grid < 0: x_grid = 0 y_grid = int(Halos['Halo_y'][i]) if y_grid >= GridSize: y_grid = GridSize - 1 if y_grid < 0: y_grid = 0 z_grid = int(Halos['Halo_z'][i]) if z_grid >= GridSize: z_grid = GridSize - 1 if z_grid < 0: z_grid = 0 idx = x_grid * GridSize*GridSize + y_grid * GridSize + z_grid binned_nion[idx] += pow(10, nion_halo)/1.0e50 # print"We had %d halos (out of %d, so %.4f fraction) that had halo mass that was not covered by the Mvir-Ngamma results." %(fit_nan, N_Halos, float(fit_nan)/float(N_Halos)) # print "There were %d cells with a non-zero ionizing flux." %(len(binned_nion[binned_nion != 0])) binned_nion = binned_nion.reshape((GridSize,GridSize,GridSize)) cut_slice = 0 cut_width = 512 nion_slice = binned_nion[:,:, cut_slice:cut_slice+cut_width].mean(axis=-1)*1.0e50 ax1 = plt.subplot(211) im = ax1.imshow(np.log10(nion_slice), interpolation='bilinear', origin='low', extent =[0,AllVars.BoxSize,0,AllVars.BoxSize], cmap = 'Purples', vmin = 48, vmax = 53) cbar = plt.colorbar(im, ax = ax1) cbar.set_label(r'$\mathrm{log}_{10}N_{\gamma} [\mathrm{s}^{-1}]$') ax1.set_xlabel(r'$\mathrm{x} (h^{-1}Mpc)$') ax1.set_ylabel(r'$\mathrm{y} (h^{-1}Mpc)$') ax1.set_xlim([0.0, AllVars.BoxSize]) ax1.set_ylim([0.0, AllVars.BoxSize]) title = r"$z = %.3f$" %(RedshiftList[halo_z_idx]) ax1.set_title(title) ax2 = plt.subplot(212) w = np.where((Halos['Halo_z'][:] > cut_slice) & (Halos['Halo_z'][:] <= cut_slice + cut_width))[0] x_plot = Halos['Halo_x'] * float(AllVars.BoxSize)/float(GridSize) y_plot = Halos['Halo_y'] * float(AllVars.BoxSize)/float(GridSize) z_plot = Halos['Halo_z'][w] * float(AllVars.BoxSize)/float(GridSize) ax2.scatter(x_plot[w], y_plot[w], s = 2, alpha = 0.5) ax2.set_xlabel(r'$\mathrm{x} (h^{-1}Mpc)$') ax2.set_ylabel(r'$\mathrm{y} (h^{-1}Mpc)$') ax2.set_xlim([0.0, AllVars.BoxSize]) ax2.set_ylim([0.0, AllVars.BoxSize]) tmp = "z%.3f" %(RedshiftList[halo_z_idx]) plt.tight_layout() outputFile = './' + output_tag + tmp + output_format plt.savefig(outputFile) # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() def plot_photoncount(SnapList, sum_nion, simulation_norm, FirstFile, LastFile, NumFiles, model_tags, output_tag): ''' Plots the ionizing emissivity as a function of redshift. We normalize the emissivity to Mpc^-3 and this function allows the read-in of only a subset of the volume. Parallel compatible. Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model, defines the x-axis we plot against. sum_nion : Nested 1-dimensional array, sum_nion[z0, z1, ..., zn], with length equal to the number of redshifts. Number of escape ionizing photons (i.e., photon rate times the local escape fraction) at each redshift. In units of 1.0e50 s^-1. simulation_norm : array of ints with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation FirstFile, LastFile, NumFile : array of integers with length equal to the number of models. The file numbers for each model that were read in (defined by the range between [FirstFile, LastFile] inclusive) and the TOTAL number of files for this model (we may only be plotting a subset of the volume). model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- sum_nion is in units of 1.0e50 s^-1. ''' print("Plotting the ionizing emissivity.") sum_array = [] for model_number in range(0, len(SnapList)): if(simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") sum_array.append([]) for snapshot_idx in range(0, len(SnapList[model_number])): nion_sum_snapshot = comm.reduce(sum_nion[model_number][snapshot_idx], op = MPI.SUM, root = 0) if rank == 0: sum_array[model_number].append(nion_sum_snapshot * 1.0e50 / (pow(AllVars.BoxSize / AllVars.Hubble_h,3) * (float(LastFile[model_number] - FirstFile[model_number] + 1) / float(NumFiles[model_number])))) if (rank == 0): ax1 = plt.subplot(111) for model_number in range(0, len(SnapList)): if(simulation_norm[model_number] == 0): cosmo = AllVars.Set_Params_Mysim() if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") t = np.empty(len(SnapList[model_number])) for snapshot_idx in range(0, len(SnapList[model_number])): t[snapshot_idx] = (AllVars.t_BigBang - cosmo.lookback_time(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]).value) * 1.0e3 t = [t for t, N in zip(t, sum_array[model_number]) if N > 1.0] sum_array[model_number] = [x for x in sum_array[model_number] if x > 1.0] print("The total number of ionizing photons for model {0} is {1} s^1 Mpc^-3".format(model_number, sum(sum_array[model_number]))) print(np.log10(sum_array[model_number])) ax1.plot(t, np.log10(sum_array[model_number]), color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[model_number], label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) #ax1.fill_between(t, np.subtract(mean,std), np.add(mean,std), color = colors[model_number], alpha = 0.25) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(PlotScripts.time_tickinterval)) #ax1.yaxis.set_minor_locator(mtick.MultipleLocator(0.025)) ax1.set_xlim(PlotScripts.time_xlim) ax1.set_ylim([48.5, 51.5]) ax2 = ax1.twiny() t_plot = (AllVars.t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$%d$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax2.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax2.set_xticklabels(z_labels) # But label them as redshifts. ax1.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$\sum f_\mathrm{esc}\dot{N}_\gamma \: [\mathrm{s}^{-1}\mathrm{Mpc}^{-3}]$', fontsize = PlotScripts.global_fontsize) plot_time = 1 bouwens_z = np.arange(6,16) # Redshift range for the observations. bouwens_t = (AllVars.t_BigBang - cosmo.lookback_time(bouwens_z).value) * 1.0e3 # Corresponding values for what we will plot on the x-axis. bouwens_1sigma_lower = [50.81, 50.73, 50.60, 50.41, 50.21, 50.00, 49.80, 49.60, 49.39, 49.18] # 68% Confidence Intervals for the ionizing emissitivity from Bouwens 2015. bouwens_1sigma_upper = [51.04, 50.85, 50.71, 50.62, 50.56, 50.49, 50.43, 50.36, 50.29, 50.23] bouwens_2sigma_lower = [50.72, 50.69, 50.52, 50.27, 50.01, 49.75, 49.51, 49.24, 48.99, 48.74] # 95% CI. bouwens_2sigma_upper = [51.11, 50.90, 50.74, 50.69, 50.66, 50.64, 50.61, 50.59, 50.57, 50.55] if plot_time == 1: ax1.fill_between(bouwens_t, bouwens_1sigma_lower, bouwens_1sigma_upper, color = 'k', alpha = 0.2) ax1.fill_between(bouwens_t, bouwens_2sigma_lower, bouwens_2sigma_upper, color = 'k', alpha = 0.4, label = r"$\mathrm{Bouwens \: et \: al. \: (2015)}$") else: ax1.fill_between(bouwens_z, bouwens_1sigma_lower, bouwens_1sigma_upper, color = 'k', alpha = 0.2) ax1.fill_between(bouwens_z, bouwens_2sigma_lower, bouwens_2sigma_upper, color = 'k', alpha = 0.4, label = r"$\mathrm{Bouwens \: et \: al. \: (2015)}$") # ax1.text(0.075, 0.965, '(a)', horizontalalignment='center', verticalalignment='center', transform = ax.transAxes) ax1.text(350, 50.0, r"$68\%$", horizontalalignment='center', verticalalignment = 'center', fontsize = PlotScripts.global_labelsize) ax1.text(350, 50.8, r"$95\%$", horizontalalignment='center', verticalalignment = 'center', fontsize = PlotScripts.global_labelsize) leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) plt.tight_layout() outputFile = './{0}{1}'.format(output_tag, output_format) plt.savefig(outputFile) # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() ## def plot_singleSFR(galaxies_filepath_array, merged_galaxies_filepath_array, number_snapshots, simulation_norm, model_tags, output_tag): SFR_gal = [] SFR_ensemble = [] ejected_gal = [] ejected_ensemble = [] infall_gal = [] infall_ensemble = [] ejectedmass_gal = [] ejectedmass_ensemble = [] N_random = 1 ax1 = plt.subplot(111) # ax3 = plt.subplot(122) #ax5 = plt.subplot(133) look_for_alive = 1 #idx_array = [20004, 20005, 20016] #halonr_array = [7381] halonr_array = [389106] #halonr_array = [36885] for model_number in range(0, len(model_tags)): if(simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") SFR_gal.append([]) SFR_ensemble.append([]) ejected_gal.append([]) ejected_ensemble.append([]) infall_gal.append([]) infall_ensemble.append([]) ejectedmass_gal.append([]) ejectedmass_ensemble.append([]) GG, Gal_Desc = ReadScripts.ReadGals_SAGE_DelayedSN(galaxies_filepath_array[model_number], 0, number_snapshots[model_number], comm) # Read in the correct galaxy file. G_Merged, Merged_Desc = ReadScripts.ReadGals_SAGE_DelayedSN(merged_galaxies_filepath_array[model_number], 0, number_snapshots[model_number], comm) # Also need the merged galaxies. G = ReadScripts.Join_Arrays(GG, G_Merged, Gal_Desc) # Then join them together for all galaxies that existed at this Redshift. if look_for_alive == 1: G.GridHistory[G.GridHistory >= 0] = 1 G.GridHistory[G.GridHistory < 0] = 0 alive = np.sum(G.GridHistory, axis = 1) # print "The galaxy that was present in the most snapshots is %d which was in %d snaps" %(np.argmax(alive), np.amax(alive)) most_alive = alive.argsort()[-10:][::-1] # Finds the 3 galaxies alive for the most snapshots. Taken from https://stackoverflow.com/questions/6910641/how-to-get-indices-of-n-maximum-values-in-a-numpy-array # print G.HaloNr[most_alive] t = np.empty((number_snapshots[model_number])) for snapshot_idx in range(0, number_snapshots[model_number]): w = np.where((G.GridHistory[:, snapshot_idx] != -1) & (G.GridStellarMass[:, snapshot_idx] > 0.0) & (G.GridStellarMass[:, snapshot_idx] < 1e5) & (G.GridFoFMass[:, snapshot_idx] >= m_low_SAGE) & (G.GridFoFMass[:, snapshot_idx] <= m_high_SAGE))[0] # Only include those galaxies that existed at the current snapshot, had positive (but not infinite) stellar/Halo mass and Star formation rate. SFR_ensemble[model_number].append(np.mean(G.GridSFR[w,snapshot_idx])) ejected_ensemble[model_number].append(np.mean(G.GridOutflowRate[w, snapshot_idx])) infall_ensemble[model_number].append(np.mean(G.GridInfallRate[w, snapshot_idx])) t[snapshot_idx] = (t_BigBang - cosmo.lookback_time(AllVars.SnapZ[snapshot_idx]).value) * 1.0e3 for p in range(0, N_random): random_idx = (np.where((G.HaloNr == halonr_array[p]))[0])[0] SFR_gal[model_number].append(G.GridSFR[random_idx]) # Remember the star formation rate history of the galaxy. ejected_gal[model_number].append(G.GridOutflowRate[random_idx]) infall_gal[model_number].append(G.GridInfallRate[random_idx]) ejectedmass_gal[model_number].append(G.GridEjectedMass[random_idx]) #SFR_gal[model_number][p][SFR_gal[model_number][p] < 1.0e-15] = 1 for snapshot_idx in range(0, number_snapshots[model_number]): if snapshot_idx == 0: pass elif(G.GridHistory[random_idx, snapshot_idx] == -1): SFR_gal[model_number][p][snapshot_idx] = SFR_gal[model_number][p][snapshot_idx - 1] # SFR_ensemble[model_number] = np.nan_to_num(SFR_ensemble[model_number]) # SFR_ensemble[model_number][SFR_ensemble[model_number] < 1.0e-15] = 1 # ejected_ensemble[model_number][ejected_ensemble[model_number] < 1.0e-15] = 1 ax1.plot(t, SFR_ensemble[model_number], color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[model_number], label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ax1.plot(t, ejected_ensemble[model_number], color = PlotScripts.colors[1], linestyle = PlotScripts.linestyles[model_number], linewidth = PlotScripts.global_linewidth, alpha = 1.0) #ax5.plot(t, infall_ensemble[model_number], color = PlotScripts.colors[2], linestyle = PlotScripts.linestyles[model_number], linewidth = PlotScripts.global_linewidth, alpha = 1.0) #ax5.plot(t, ejectedmass_ensemble[model_number], color = PlotScripts.colors[2], linestyle = PlotScripts.linestyles[model_number], linewidth = PlotScripts.global_linewidth, alpha = 1.0) for p in range(0, N_random): ax1.plot(t, SFR_gal[model_number][p], color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[model_number], alpha = 0.5, linewidth = 1) ax1.plot(t, ejected_gal[model_number][p], color = PlotScripts.colors[1], linestyle = PlotScripts.linestyles[model_number], alpha = 0.5, linewidth = 1) #ax5.plot(t, infall_gal[model_number][p], color = PlotScripts.colors[2], linestyle = PlotScripts.linestyles[model_number], alpha = 0.5, linewidth = 1) #ax5.plot(t, ejectedmass_gal[model_number][p], color = PlotScripts.colors[2], linestyle = PlotScripts.linestyles[model_number], alpha = 0.5, linewidth = 1) #ax1.plot(t, SFR_gal[model_number][p], color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[model_number], alpha = 1.0, linewidth = 1, label = model_tags[model_number]) #ax1.plot(t, ejected_gal[model_number][p], color = PlotScripts.colors[1], linestyle = PlotScripts.linestyles[model_number], alpha = 1.0, linewidth = 1, label = model_tags[model_number]) ax1.plot(np.nan, np.nan, color = 'r', linestyle = '-', label = "SFR") ax1.plot(np.nan, np.nan, color = 'b', linestyle = '-', label = "Outflow") # exit() #ax1.plot(np.nan, np.nan, color = PlotScripts.colors[0], label = 'SFR') #ax1.plot(np.nan, np.nan, color = PlotScripts.colors[1], label = 'Outflow') ax1.set_yscale('log', nonposy='clip') ax1.set_ylabel(r"$\mathrm{Mass \: Flow} \: [\mathrm{M}_\odot \mathrm{yr}^{-1}]$") ax1.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_fontsize) ax1.set_xlim(PlotScripts.time_xlim) ax1.set_ylim([1e-6, 1e3]) ''' ax3.set_yscale('log', nonposy='clip') ax3.set_ylabel(r"$\mathrm{Outflow \: Rate} \: [\mathrm{M}_\odot \mathrm{yr}^{-1}]$") ax3.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_fontsize) ax3.set_xlim(PlotScripts.time_xlim) ax3.set_ylim([1e-8, 1e3]) ax5.set_yscale('log', nonposy='clip') #ax5.set_ylabel(r"$\mathrm{Infall \: Rate} \: [\mathrm{M}_\odot \mathrm{yr}^{-1}]$") ax5.set_ylabel(r"$\mathrm{Ejected Mass} [\mathrm{M}_\odot]$") ax5.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_fontsize) ax5.set_xlim(PlotScripts.time_xlim) #ax5.set_ylim([1e-8, 1e3]) ax5.set_ylim([1e6, 1e10]) ''' ax2 = ax1.twiny() #ax4 = ax3.twiny() #ax6 = ax5.twiny() t_plot = (t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$%d$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax2.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax2.set_xticklabels(z_labels) # But label them as redshifts. ''' ax4.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax4.set_xlim(PlotScripts.time_xlim) ax4.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax4.set_xticklabels(z_labels) # But label them as redshifts. ax6.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax6.set_xlim(PlotScripts.time_xlim) ax6.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax6.set_xticklabels(z_labels) # But label them as redshifts. ''' plt.tight_layout() leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile = './Halo%d_mlow%.2f_%s%s' %(halonr_array[0], m_low_SAGE, output_tag, output_format) plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() ## def plot_quasars_count(SnapList, PlotList, N_quasars_z, N_quasars_boost_z, N_gal_z, mean_quasar_activity, std_quasar_activity, N_halo, N_merger_halo, N_gal, N_merger_galaxy, fesc_prescription, simulation_norm, FirstFile, LastFile, NumFile, model_tags, output_tag): ''' Parameters --------- SnapList : Nested 'array-like` of ints, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots that we plot the quasar density at for each model. PlotList : Nested array of ints, PlotList[model_number0]= [plotsnapshot0_model0, ..., plotsnapshotN_model0], with length equal to the number of models. Snapshots that will be plotted for the quasar activity as a function of halo mass. N_quasars_z : Nested array of floats, N_quasars_z[model_number0] = [N_quasars_z0, N_quasars_z1, ..., N_quasars_zN]. Outer array has length equal to the number of models, inner array has length equal to length of the model's SnapList. Number of quasars, THAT WENT OFF, during the given redshift. N_quasars_boost_z : Nested array of floats, N_quasars_boost_z[model_number0] = [N_quasars_boost_z0, N_quasars_boost_z1, ..., N_quasars_boost_zN]. Outer array has length equal to the number of models, inner array has length equal to length of the model's SnapList. Number of galaxies that had their escape fraction boosted by quasar activity. N_gal_z : Nested array of floats, N_gal_z[model_number0] = [N_gal_z0, N_gal_z1, ..., N_gal_zN]. Outer array has length equal to the number of models, inner array has length equal to length of the model's SnapList. Number of galaxies at each redshift. mean_quasar_activity, std_quasar_activity : Nested 2-dimensional array of floats, mean_quasar_activity[model_number0][snapshot0] = [bin0quasar_activity, ..., binNquasar_activity]. Outer array has length equal to the number of models, inner array has length equal to the length of the model's snaplist and most inner array has length equal to the number of halo bins (NB). Mean/std fraction of galaxies that had quasar go off during each snapshot as a function of halo mass. NOTE : This is for quasars going off, not for galaxies that have their escape fraction being boosted. fesc_prescription : Array with length equal to the number of models. Denotes what escape fraction prescription each model used. Quasars are only tracked when fesc_prescription == 3. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali FirstFile, LastFile, NumFile : array of integers with length equal to the number of models. The file numbers for each model that were read in (defined by the range between [FirstFile, LastFile] inclusive) and the TOTAL number of files for this model (we may only be plotting a subset of the volume). model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. File will be saved in the current directory with the output format defined by the 'output_format' variable at the beggining of the file. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- No relevant units. ''' print("Plotting quasar count/density") if rank == 0: fig = plt.figure() ax1 = fig.add_subplot(111) ax6 = ax1.twinx() fig2 = plt.figure() ax3 = fig2.add_subplot(111) ax5 = ax3.twinx() fig3 = plt.figure() ax7 = fig3.add_subplot(111) fig4 = plt.figure() ax50 = fig4.add_subplot(111) fig5 = plt.figure() ax55 = fig5.add_subplot(111) fig6 = plt.figure() ax56 = fig6.add_subplot(111) mean_quasar_activity_array = [] std_quasar_activity_array = [] N_quasar_activity_array = [] N_gal_halo_array = [] N_gal_array = [] merger_counts_halo_array = [] merger_counts_galaxy_array = [] bin_middle_halo_array = [] bin_middle_galaxy_array = [] for model_number in range(0, len(SnapList)): # Does this for each of the models. if (fesc_prescription[model_number] != 3): # Want to skip the models that didn't count quasars. continue ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif (simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() mean_quasar_activity_array.append([]) std_quasar_activity_array.append([]) N_quasar_activity_array.append([]) N_gal_halo_array.append([]) N_gal_array.append([]) merger_counts_halo_array.append([]) merger_counts_galaxy_array.append([]) bin_middle_halo_array.append([]) bin_middle_galaxy_array.append([]) box_factor = (LastFile[model_number] - FirstFile[model_number] + 1.0)/(NumFile[model_number]) # This factor allows us to take a sub-volume of the box and scale the results to represent the entire box. print("We are plotting the quasar density using {0:.4f} of the box's volume.".format(box_factor)) norm = pow(AllVars.BoxSize,3) / pow(AllVars.Hubble_h, 3) * box_factor #### ## We perform the plotting on Rank 0 so only this rank requires the final counts array. ## if rank == 0: quasars_total = np.zeros_like((N_quasars_z[model_number])) boost_total = np.zeros_like(N_quasars_boost_z[model_number]) gal_count_total = np.zeros_like(N_gal_z[model_number]) else: quasars_total = None boost_total = None gal_count_total = None N_quasars_tmp = np.array((N_quasars_z[model_number])) # So we can use MPI.Reduce() comm.Reduce([N_quasars_tmp, MPI.DOUBLE], [quasars_total, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the number of quasars and passes back to rank 0. N_quasars_boost_tmp = np.array(N_quasars_boost_z[model_number]) # So we can use MPI.Reduce() comm.Reduce([N_quasars_boost_tmp, MPI.DOUBLE], [boost_total, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the number of galaxies that had their fesc boosted. N_gal_tmp = np.array(N_gal_z[model_number]) # So we can use MPI.Reduce() comm.Reduce([N_gal_tmp, MPI.DOUBLE], [gal_count_total, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the number of total galaxies. for snapshot_idx in range(len(SnapList[model_number])): mean_quasar_activity_array[model_number], std_quasar_activity_array[model_number], N_quasar_activity_array[model_number] = calculate_pooled_stats(mean_quasar_activity_array[model_number], std_quasar_activity_array[model_number], N_quasar_activity_array[model_number], mean_quasar_activity[model_number][snapshot_idx], std_quasar_activity[model_number][snapshot_idx], N_halo[model_number][snapshot_idx]) if rank == 0: merger_count_halo_total = np.zeros_like((N_merger_halo[model_number][snapshot_idx])) N_gal_halo_total = np.zeros_like((N_halo[model_number][snapshot_idx])) merger_count_galaxy_total = np.zeros_like((N_merger_galaxy[model_number][snapshot_idx])) N_gal_total = np.zeros_like((N_gal[model_number][snapshot_idx])) else: merger_count_halo_total = None N_gal_halo_total = None merger_count_galaxy_total = None N_gal_total = None comm.Reduce([N_merger_halo[model_number][snapshot_idx], MPI.FLOAT], [merger_count_halo_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. comm.Reduce([N_halo[model_number][snapshot_idx], MPI.FLOAT], [N_gal_halo_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. comm.Reduce([N_merger_galaxy[model_number][snapshot_idx], MPI.FLOAT], [merger_count_galaxy_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. comm.Reduce([N_gal[model_number][snapshot_idx], MPI.FLOAT], [N_gal_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. if rank == 0: merger_counts_halo_array[model_number].append(merger_count_halo_total) N_gal_halo_array[model_number].append(N_gal_halo_total) merger_counts_galaxy_array[model_number].append(merger_count_galaxy_total) N_gal_array[model_number].append(N_gal_total) bin_middle_halo_array[model_number].append(np.arange(m_low, m_high+bin_width, bin_width)[:-1] + bin_width * 0.5) bin_middle_galaxy_array[model_number].append(np.arange(m_gal_low, m_gal_high+bin_width, bin_width)[:-1] + bin_width * 0.5) if rank == 0: plot_count = 0 stop_plot = 0 title = model_tags[model_number] t = np.empty(len(SnapList[model_number])) ZZ = np.empty(len(SnapList[model_number])) for snapshot_idx in range(0, len(SnapList[model_number])): t[snapshot_idx] = (AllVars.t_BigBang - AllVars.Lookback_Time[SnapList[model_number][snapshot_idx]]) * 1.0e3 ZZ[snapshot_idx] = AllVars.SnapZ[SnapList[model_number][snapshot_idx]] if (stop_plot == 0): # print("Snapshot {0} PlotSnapshot " #"{1}".format(SnapList[model_number][snapshot_idx], PlotList[model_number][plot_count])) if (SnapList[model_number][snapshot_idx] == PlotList[model_number][plot_count]): label = "z = {0:.2f}".format(AllVars.SnapZ[PlotList[model_number][plot_count]]) ax7.plot(bin_middle_halo_array[model_number][snapshot_idx], mean_quasar_activity_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) #ax50.plot(bin_middle_halo_array[model_number][snapshot_idx], merger_counts_array[model_number][snapshot_idx] / gal_count_total[snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) ax50.plot(bin_middle_halo_array[model_number][snapshot_idx], merger_counts_halo_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) #ax50.plot(bin_middle_halo_array[model_number][snapshot_idx], merger_counts_array[model_number][snapshot_idx] / N_gal_halo_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) #ax55.plot(bin_middle_galaxy_array[model_number][snapshot_idx], merger_counts_galaxy_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) ax55.plot(bin_middle_galaxy_array[model_number][snapshot_idx], merger_counts_galaxy_array[model_number][snapshot_idx] / N_gal_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) print("plot_count = {0} len(PlotList) = {1}".format(plot_count, len(PlotList[model_number]))) plot_count += 1 print("plot_count = {0} len(PlotList) = {1}".format(plot_count, len(PlotList[model_number]))) if (plot_count == len(PlotList[model_number])): stop_plot = 1 print("For Snapshot {0} at t {3} there were {1} total mergers compared to {2} total galaxies.".format(snapshot_idx, np.sum(merger_counts_galaxy_array[model_number][snapshot_idx]), np.sum(gal_count_total[snapshot_idx]), t[snapshot_idx])) if (np.sum(gal_count_total[snapshot_idx]) > 0.0 and np.sum(merger_counts_galaxy_array[model_number][snapshot_idx]) > 0.0): ax56.scatter(t[snapshot_idx], np.sum(merger_counts_galaxy_array[model_number][snapshot_idx]) / np.sum(gal_count_total[snapshot_idx]), color = 'r', rasterized = True) #ax56.scatter(t[snapshot_idx], quasars_total[snapshot_idx] / np.sum(gal_count_total[snapshot_idx]), color = 'r', rasterized = True) ax1.plot(t, quasars_total / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[0], rasterized = True, linewidth = PlotScripts.global_linewidth) p = np.where((ZZ < 15))[0] #ax1.plot(ZZ[p], quasars_total[p] / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[0], rasterized = True, linewidth = PlotScripts.global_linewidth) ax3.plot(t, boost_total, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[0], rasterized = True, label = title, linewidth = PlotScripts.global_linewidth) w = np.where((gal_count_total > 0.0))[0] # Since we're doing a division, need to only plot those redshifts that actually have galaxies. ax5.plot(t[w], np.divide(boost_total[w], gal_count_total[w]), color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[1], rasterized = True, linewidth = PlotScripts.global_linewidth) ax6.plot(t[w], gal_count_total[w] / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[1], rasterized = True, linewidth = PlotScripts.global_linewidth) #ax6.plot(ZZ[p], gal_count_total[p] / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[1], rasterized = True, linewidth = PlotScripts.global_linewidth) ax1.plot(np.nan, np.nan, color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[0], label = "Quasar Ejection Density") ax1.plot(np.nan, np.nan, color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[1], label = "Galaxy Density") ax3.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[0], label = "Count") ax3.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[1], label = "Fraction of Galaxies") ax7.set_xlabel(r'$\log_{10}\ M_\mathrm{vir}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax7.set_ylabel(r'$\mathrm{Mean \: Quasar \: Activity}$', size = PlotScripts.global_fontsize) ax50.set_xlabel(r'$\log_{10}\ M_\mathrm{vir}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) #ax50.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) ax50.set_ylabel(r'$\mathrm{Number \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) ax55.set_xlabel(r'$\log_{10}\ M_\mathrm{*}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax55.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) #ax55.set_ylabel(r'$\mathrm{Number \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) ax56.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_labelsize) ax56.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) #ax56.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Quasar \: Activity}$', size = PlotScripts.global_fontsize) ax56.set_yscale('log', nonposy='clip') ax50.axvline(np.log10(32.0*AllVars.PartMass / AllVars.Hubble_h), color = 'k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') ax1.xaxis.set_minor_locator(mtick.MultipleLocator(PlotScripts.time_tickinterval)) ax1.set_xlim(PlotScripts.time_xlim) ax1.set_yscale('log', nonposy='clip') ax3.xaxis.set_minor_locator(mtick.MultipleLocator(PlotScripts.time_tickinterval)) ax3.set_xlim(PlotScripts.time_xlim) ax3.set_yscale('log', nonposy='clip') ## Create a second axis at the top that contains the corresponding redshifts. ## ## The redshift defined in the variable 'z_plot' will be displayed. ## ax2 = ax1.twiny() ax4 = ax3.twiny() ax57 = ax56.twiny() t_plot = (AllVars.t_BigBang - AllVars.cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding time values on the bottom. z_labels = ["$%d$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax2.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax2.set_xticklabels(z_labels) # But label them as redshifts. ax4.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax4.set_xlim(PlotScripts.time_xlim) ax4.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax4.set_xticklabels(z_labels) # But label them as redshifts. ax57.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax57.set_xlim(PlotScripts.time_xlim) ax57.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax57.set_xticklabels(z_labels) # But label them as redshifts. ax1.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_labelsize) #ax1.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax1.set_ylabel(r'$N_\mathrm{Quasars} \: [\mathrm{Mpc}^{-3}]$', fontsize = PlotScripts.global_fontsize) ax6.set_ylabel(r'$N_\mathrm{Gal} \: [\mathrm{Mpc}^{-3}]$', fontsize = PlotScripts.global_fontsize) ax3.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_labelsize) ax3.set_ylabel(r'$N_\mathrm{Boosted}$', fontsize = PlotScripts.global_fontsize) ax5.set_ylabel(r'$\mathrm{Fraction \: Boosted}$', fontsize = PlotScripts.global_fontsize) leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax3.legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax7.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax50.legend(loc='upper right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax55.legend(loc='upper right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) fig.tight_layout() fig2.tight_layout() fig3.tight_layout() fig5.tight_layout() fig6.tight_layout() outputFile1 = './{0}_quasardensity{1}'.format(output_tag, output_format) outputFile2 = './{0}_boostedcount{1}'.format(output_tag, output_format) outputFile3 = './{0}_quasar_activity_halo{1}'.format(output_tag, output_format) outputFile4 = './{0}_mergercount_global{1}'.format(output_tag, output_format) outputFile5 = './{0}_mergercount_global_stellarmass{1}'.format(output_tag, output_format) outputFile6 = './{0}_mergercount_total{1}'.format(output_tag, output_format) fig.savefig(outputFile1) # Save the figure fig2.savefig(outputFile2) # Save the figure fig3.savefig(outputFile3) # Save the figure fig4.savefig(outputFile4) # Save the figure fig5.savefig(outputFile5) # Save the figure fig6.savefig(outputFile6) # Save the figure print("Saved to {0}".format(outputFile1)) print("Saved to {0}".format(outputFile2)) print("Saved to {0}".format(outputFile3)) print("Saved to {0}".format(outputFile4)) print("Saved to {0}".format(outputFile5)) print("Saved to {0}".format(outputFile6)) plt.close(fig) plt.close(fig2) plt.close(fig3) ## def plot_photon_quasar_fraction(snapshot, filenr, output_tag, QuasarFractionalPhoton, QuasarActivityToggle, NumSubsteps): ax1 = plt.subplot(111) counts, bin_edges, bin_middle = AllVars.Calculate_Histogram(QuasarFractionalPhoton, 0.05, 0, 0, 1) ax1.plot(bin_middle, counts, lw = PlotScripts.global_linewidth, color = 'r') ax1.axvline(np.mean(QuasarFractionalPhoton[QuasarFractionalPhoton != 0]), lw = 0.5, ls = '-') ax1.set_yscale('log', nonposy='clip') ax1.set_xlabel(r"$\mathrm{Fractional \: Photon \: Boost}$") ax1.set_ylabel(r"$\mathrm{Count}$") ax1.set_ylim([1e1, 1e5]) outputFile1 = './photonfraction/file{0}_snap{1}_{2}{3}'.format(filenr, snapshot, output_tag, output_format) plt.tight_layout() plt.savefig(outputFile1) print("Saved to {0}".format(outputFile1)) plt.close() ### def plot_quasar_substep(snapshot, filenr, output_tag, substep): ax1 = plt.subplot(111) counts, bin_edges, bin_middle = AllVars.Calculate_Histogram(substep, 0.1, 0, 0, 10) ax1.plot(bin_middle, counts, lw = PlotScripts.global_linewidth, color = 'r') ax1.axvline(np.mean(substep[substep != -1]), lw = 0.5, ls = '-') ax1.set_yscale('log', nonposy='clip') ax1.set_xlabel(r"$\mathrm{Substep \: Quasar \: Activity}$") ax1.set_ylabel(r"$\mathrm{Count}$") # ax1.set_ylim([1e1, 1e5]) outputFile1 = './substep_activity/file{0}_snap{1}_{2}{3}'.format(filenr, snapshot, output_tag, output_format) plt.tight_layout() plt.savefig(outputFile1) print("Saved to {0}".format(outputFile1)) plt.close() ### def plot_post_quasar_SFR(PlotSnapList, model_number, Gal, output_tag): ax1 = plt.subplot(111) ax2 = ax1.twinx() count = 0 snapshot_thickness = 20 # How many snapshots before/after the quasar event do we want to track? for snapshot_idx in PlotSnapList[model_number]: w = np.where((G.QuasarActivity[:, snapshot_idx] == 1) & (G.LenHistory[:, snapshot_idx] > 200.0) & (G.GridStellarMass[:, snapshot_idx] > 0.001))[0] w_slice_gridhistory = G.GridHistory[w,snapshot_idx-snapshot_thickness:snapshot_idx+snapshot_thickness] potential_gal = [] for i in range(len(w_slice_gridhistory)): ww = np.where((w_slice_gridhistory[i] >= 0))[0] if (len(ww) == snapshot_thickness * 2): potential_gal.append(w[i]) if (len(potential_gal) == 0): return count += 1 print("There were {0} galaxies that had an energetic quasar wind event at snapshot {1} (z = {2:.3f})".format(len(potential_gal), snapshot_idx, AllVars.SnapZ[snapshot_idx])) chosen_gal = potential_gal[1] lenhistory_array = np.empty((int(snapshot_thickness*2 + 1))) SFR_array = np.empty((int(snapshot_thickness*2 + 1))) gridhistory_array = np.empty((int(snapshot_thickness*2 + 1))) coldgas_array = np.empty((int(snapshot_thickness*2 + 1))) t = np.empty((int(snapshot_thickness*2 + 1))) for i in range(-snapshot_thickness, snapshot_thickness+1): #print("SFR {0} {1}".format(snapshot_idx + i, G.GridSFR[chosen_gal, snapshot_idx+i])) #print("ColdGas {0} {1}".format(snapshot_idx + i, G.GridColdGas[chosen_gal, snapshot_idx+i])) lenhistory_array[i+snapshot_thickness] = (G.LenHistory[chosen_gal, snapshot_idx+i]) SFR_array[i+snapshot_thickness] = (G.GridSFR[chosen_gal, snapshot_idx+i]) #- (G.GridSFR[chosen_gal, snapshot_idx]) gridhistory_array[i+snapshot_thickness] = (G.GridHistory[chosen_gal, snapshot_idx+i]) coldgas_array[i+snapshot_thickness] = (G.GridColdGas[chosen_gal, snapshot_idx+i] * 1.0e10 / AllVars.Hubble_h) #- (G.GridColdGas[chosen_gal, snapshot_idx]) t[i+snapshot_thickness] = (-AllVars.Lookback_Time[snapshot_idx+i] + AllVars.Lookback_Time[snapshot_idx]) * 1.0e3 print("Len History {0}".format(lenhistory_array)) print("Grid History {0}".format(gridhistory_array)) print("Cold Gas {0}".format(coldgas_array)) print("SFR {0}".format(SFR_array)) stellarmass_text = r"$log M_* = {0:.2f} \: M_\odot$".format(np.log10(G.GridStellarMass[chosen_gal, snapshot_idx] * 1.0e10 / AllVars.Hubble_h)) Ndym_text = "Dynamical Time = {0:.2f} Myr".format(G.DynamicalTime[chosen_gal, snapshot_idx]) z_text = "z = {0:.2f}".format(AllVars.SnapZ[snapshot_idx]) ax1.text(0.05, 0.95, z_text, transform = ax1.transAxes, fontsize = PlotScripts.global_fontsize - 4) ax1.text(0.05, 0.9, stellarmass_text, transform = ax1.transAxes, fontsize = PlotScripts.global_fontsize - 4) ax1.text(0.05, 0.85, Ndym_text, transform = ax1.transAxes, fontsize = PlotScripts.global_fontsize - 4) ax1.plot(t, SFR_array, color = 'r', lw = PlotScripts.global_linewidth) ax2.plot(t, coldgas_array, color = 'b', lw = PlotScripts.global_linewidth) ax1.set_xlabel(r"$\mathrm{Time \: Since \: Quasar \: Event \: [Myr]}$", size = PlotScripts.global_labelsize - 10) # ax1.set_ylabel(r"$\mathrm{Fractional \: SFR \: Relative \: To \: SFR_{Quasar}}$", size = PlotScripts.global_labelsize - 10) # ax2.set_ylabel(r"$\mathrm{Difference \: Cold \: Gas \: Mass \: Relative \: To \: Cold_{Quasar}}$", size = PlotScripts.global_labelsize - 10) ax1.set_ylabel(r"$\mathrm{SFR} \: [\mathrm{M}_\odot \mathrm{yr}^{-1}]$", size = PlotScripts.global_labelsize - 10) ax2.set_ylabel(r"$\mathrm{Cold \: Gas \: Mass \: [\mathrm{M}_\odot]}$",size = PlotScripts.global_labelsize - 10) ax1.set_yscale('log', nonposy='clip') ax2.set_yscale('log', nonposy='clip') ax1.plot(np.nan, np.nan, color = 'r', label = r"$\mathrm{SFR}$") ax1.plot(np.nan, np.nan, color = 'b', label = r"$\mathrm{Cold \: Gas}$") leg = ax1.legend(loc='upper right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile = "{0}_galaxy{2}{1}".format(output_tag, output_format, chosen_gal) plt.tight_layout() plt.savefig(outputFile) print("Saved to {0}".format(outputFile)) plt.close() exit() ### def plot_stellarmass_blackhole(SnapList, simulation_norm, mean_galaxy_BHmass, std_galaxy_BHmass, N_galaxy_BHmass, FirstFile, LastFile, NumFile, model_tags, output_tag): master_mean_SMBH, master_std_SMBH, master_N, master_bin_middle = \ collect_across_tasks(mean_galaxy_BHmass, std_galaxy_BHmass, N_galaxy_BHmass, SnapList, SnapList, True, m_gal_low, m_gal_high) if rank == 0: fig = plt.figure() ax1 = fig.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(0, len(SnapList)): ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() box_factor = (LastFile[model_number] - FirstFile[model_number] + 1.0)/(NumFile[model_number]) # This factor allows us to take a sub-volume of the box and scale the results to represent the entire box. norm = pow(AllVars.BoxSize,3) / pow(AllVars.Hubble_h, 3) * bin_width * box_factor for snapshot_idx in range(0, len(SnapList[model_number])): w = np.where((master_N[model_number][snapshot_idx] > 0.0))[0] mean = np.log10(master_mean_SMBH[model_number][snapshot_idx][w]) upper = np.log10(np.add(master_mean_SMBH[model_number][snapshot_idx][w], master_std_SMBH[model_number][snapshot_idx][w])) lower = np.log10(np.subtract(master_mean_SMBH[model_number][snapshot_idx][w], master_std_SMBH[model_number][snapshot_idx][w])) label = "z = {0:.2f}" \ .format(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) ax1.plot(master_bin_middle[model_number][snapshot_idx][w], mean, label = label, color = PlotScripts.colors[snapshot_idx], ls = PlotScripts.linestyles[model_number], lw = PlotScripts.global_linewidth, rasterized = True) #ax1.fill_between(bin_middle_stellar_array[model_number][snapshot_idx][w], lower, upper, color = PlotScripts.colors[model_number], alpha = 0.25) ax2.plot(master_bin_middle[model_number][snapshot_idx][w], master_N[model_number][snapshot_idx][w] / norm, label = label, ls = PlotScripts.linestyles[model_number], lw = PlotScripts.global_linewidth, rasterized = True) Obs.Get_Data_SMBH() PlotScripts.plot_SMBH_z8(ax1) ax1.set_xlabel(r"$\log_{10}\mathrm{M}_* [\mathrm{M}_\odot]$", size = PlotScripts.global_fontsize) ax1.set_ylabel(r"$\log_{10}\mathrm{M}_\mathrm{BH} [\mathrm{M}_\odot]$", size = PlotScripts.global_fontsize) ax2.set_xlabel(r"$\log_{10}\mathrm{M}_\mathrm{BH} [\mathrm{M}_\odot]$", size = PlotScripts.global_fontsize) ax2.set_ylabel(r'$\Phi\ [\mathrm{Mpc}^{-3}\: \mathrm{dex}^{-1}]$', fontsize = PlotScripts.global_fontsize) ax2.set_yscale('log', nonposy='clip') ax1.set_xticks(np.arange(7.0, 12.0)) ax1.set_yticks(np.arange(3.0, 12.0)) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax1.yaxis.set_minor_locator(mtick.MultipleLocator(0.25)) ax1.set_xlim([7.0, 10.25]) ax1.set_ylim([3.0, 8.0]) leg = ax1.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax2.legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile = "{0}{1}".format(output_tag, output_format) plt.tight_layout() fig.savefig(outputFile) print("Saved to {0}".format(outputFile)) plt.close(fig) outputFile2 = "{0}_MF{1}".format(output_tag, output_format) plt.tight_layout() fig2.savefig(outputFile2) print("Saved to {0}".format(outputFile2)) plt.close(fig2) ### def plot_reionmod(PlotSnapList, SnapList, simulation_norm, mean_reionmod_halo, std_reionmod_halo, N_halo, mean_reionmod_z, std_reionmod_z, N_reionmod, plot_z, model_tags, output_tag): """ Plot the reionization modifier as a function of halo mass and redshift. Parameters ---------- PlotSnapList, SnapList: 2D Nested arrays of integers. Outer length is equal to the number of models and inner length is number of snapshots we're plotting/calculated for. PlotSnapList contains the snapshots for each model we will plot for the halo mass figure. SnapList contains the snapshots for each model that we have performed calculations for. These aren't equal because we don't want to plot halo curves for ALL redshifts. simulation_norm: Array of integers. Length is equal to the number of models. Contains the simulation identifier for each model. Used to set the parameters of each model. mean_reionmod_halo, std_reionmod_halo: 3D Nested arrays of floats. Most outer length is equal to the number of models, next length is number of snapshots for each model, then inner-most length is the number of halo mass- bins (given by NB). Contains the mean/standard deviation values for the reionization modifier as a function of halo mass. NOTE: These are unique for each task. N_halo: 3D Nested arrays of floats. Lengths are identical to mean_reionmod_halo. Contains the number of halos in each halo mass bin. NOTE: These are unique for each task. mean_reionmod_z, std_reionmod_z: 2D Nested arrays of floats. Outer length is equal to the number of models, inner length is the number of snapshots for each model. NOTE: This inner length can be different to the length of PlotSnapList as we don't necessarily need to plot for every snapshot we calculate. Contains the mean/standard deviation values for the rieonization modifier as a function of redshift. NOTE: These are unique for each task. N_reionmod: 2D Nested arrays of floats. Lengths are identical to mean_reionmod_z. Contains the number of galaxies at each redshift that have non-negative reionization modifier. A negative reionization modifier is a galaxy who didn't have infall/stripping during the snapshot. NOTE: These are unique for each task. plot_z: Boolean. Denotes whether we want to plot the reionization modifier as a function of redshift. Useful because we often only calculate statistics for a subset of the snapshots to decrease computation time. For these runs, we don't want to plot for something that requires ALL snapshots. model_tags: Array of strings. Length is equal to the number of models. Contains the legend labels for each model. output_tag: String. The prefix for the output file. Returns ---------- None. Plot is saved in current directory as "./<output_tag>.<output_format>" """ master_mean_reionmod_halo, master_std_reionmod_halo, master_N_reionmod_halo, master_bin_middle = collect_across_tasks(mean_reionmod_halo, std_reionmod_halo, N_halo, SnapList, PlotSnapList, True, m_low, m_high) if plot_z: master_mean_reionmod_z, master_std_reionmod_z, master_N_reionmod_z, _ = collect_across_tasks(mean_reionmod_z, std_reionmod_z, N_reionmod) if rank == 0: fig1 = plt.figure() ax1 = fig1.add_subplot(111) if plot_z: fig2 = plt.figure() ax10 = fig2.add_subplot(111) for model_number in range(len(PlotSnapList)): if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() for snapshot_idx in range(len((PlotSnapList[model_number]))): if snapshot_idx == 0: label = model_tags[model_number] else: label = "" nonzero_bins = np.where(master_N_reionmod_halo[model_number][snapshot_idx] > 0.0)[0] ax1.plot(master_bin_middle[model_number][snapshot_idx][nonzero_bins], master_mean_reionmod_halo[model_number][snapshot_idx][nonzero_bins], label = label, ls = PlotScripts.linestyles[model_number], color = PlotScripts.colors[snapshot_idx]) if plot_z: ax10.plot((AllVars.t_BigBang - AllVars.Lookback_Time[SnapList[model_number]])*1.0e3, master_mean_reionmod_z[model_number], color = PlotScripts.colors[model_number], label = model_tags[model_number], ls = PlotScripts.linestyles[model_number], lw = 3) for count, snapshot_idx in enumerate(PlotSnapList[model_number]): #label = r"$\mathbf{z = " + str(int(round(AllVars.SnapZ[snapshot_idx]))) + "}$" label = r"$\mathbf{z = " + str(AllVars.SnapZ[snapshot_idx]) + "}$" ax1.plot(np.nan, np.nan, ls = PlotScripts.linestyles[0], color = PlotScripts.colors[count], label = label) ax1.set_xlim([8.5, 11.5]) ax1.set_ylim([0.0, 1.05]) ax1.set_xlabel(r'$\mathbf{log_{10} \: M_{vir} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax1.set_ylabel(r'$\mathbf{Mean ReionMod}$', fontsize = PlotScripts.global_labelsize) leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile1 = "./{0}_halo{1}".format(output_tag, output_format) fig1.savefig(outputFile1, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile1)) plt.close(fig1) if plot_z: ax10.set_xlabel(r"$\mathbf{Time \: since \: Big \: Bang \: [Myr]}$", fontsize = PlotScripts.global_labelsize) tick_locs = np.arange(200.0, 1000.0, 100.0) tick_labels = [r"$\mathbf{%d}$" % x for x in tick_locs] ax10.xaxis.set_major_locator(mtick.MultipleLocator(100)) ax10.set_xticklabels(tick_labels, fontsize = PlotScripts.global_fontsize) ax10.set_xlim(PlotScripts.time_xlim) ax10.set_ylabel(r'$\mathbf{Mean ReionMod}$', fontsize = PlotScripts.global_labelsize) ax11 = ax10.twiny() t_plot = (AllVars.t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$\mathbf{%d}$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax11.set_xlabel(r"$\mathbf{z}$", fontsize = PlotScripts.global_labelsize) ax11.set_xlim(PlotScripts.time_xlim) ax11.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax11.set_xticklabels(z_labels, fontsize = PlotScripts.global_fontsize) # But label them as redshifts. leg = ax10.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile2 = "./{0}_z{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ## def plot_dust(PlotSnapList, SnapList, simulation_norm, mean_dust_galaxy, std_dust_galaxy, N_galaxy, mean_dust_halo, std_dust_halo, N_halo, plot_z, model_tags, output_tag): """ """ master_mean_dust_galaxy, master_std_dust_galaxy, master_N_dust_galaxy, master_bin_middle_galaxy = \ collect_across_tasks(mean_dust_galaxy, std_dust_galaxy, N_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) master_mean_dust_halo, master_std_dust_halo, master_N_dust_halo, master_bin_middle_halo = \ collect_across_tasks(mean_dust_halo, std_dust_halo, N_halo, SnapList, PlotSnapList, True, m_low, m_high) if rank == 0: fig1 = plt.figure() ax1 = fig1.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(len(PlotSnapList)): if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() for snapshot_idx in range(len((PlotSnapList[model_number]))): if snapshot_idx == 0: label = model_tags[model_number] else: label = "" nonzero_bins = np.where(master_N_dust_galaxy[model_number][snapshot_idx] > 0.0)[0] ax1.plot(master_bin_middle_galaxy[model_number][snapshot_idx][nonzero_bins], master_mean_dust_galaxy[model_number][snapshot_idx][nonzero_bins], label = label, ls = PlotScripts.linestyles[model_number], color = PlotScripts.colors[snapshot_idx]) nonzero_bins = np.where(master_N_dust_halo[model_number][snapshot_idx] > 0.0)[0] ax2.plot(master_bin_middle_halo[model_number][snapshot_idx][nonzero_bins], master_mean_dust_halo[model_number][snapshot_idx][nonzero_bins], label = label, ls = PlotScripts.linestyles[model_number], color = PlotScripts.colors[snapshot_idx]) print(master_mean_dust_halo[model_number][snapshot_idx]) for count, snapshot_idx in enumerate(PlotSnapList[model_number]): #label = r"$\mathbf{z = " + str(int(round(AllVars.SnapZ[snapshot_idx]))) + "}$" label = r"$\mathbf{z = " + str(AllVars.SnapZ[snapshot_idx]) + "}$" ax1.plot(np.nan, np.nan, ls = PlotScripts.linestyles[0], color = PlotScripts.colors[count], label = label) ax2.plot(np.nan, np.nan, ls = PlotScripts.linestyles[0], color = PlotScripts.colors[count], label = label) ax1.set_xlim([2.0, 10.5]) #ax1.set_ylim([1.0, 6.0]) ax1.set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax1.set_ylabel(r'$\mathbf{log_{10} \: \langle M_{Dust}\rangle_{M*}}$', fontsize = PlotScripts.global_labelsize) leg = ax1.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile1 = "./{0}_galaxy{1}".format(output_tag, output_format) fig1.savefig(outputFile1, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile1)) plt.close(fig1) ax2.set_xlim([6.8, 11.5]) #ax2.set_ylim([1.0, 6.0]) ax2.set_xlabel(r'$\mathbf{log_{10} \: M_{vir} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax2.set_ylabel(r'$\mathbf{log_{10} \: \langle M_{Dust}\rangle_{Mvir}}$', fontsize = PlotScripts.global_labelsize) leg = ax2.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile2 = "./{0}_halo{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) def plot_dust_scatter(SnapList, mass_gal, mass_halo, mass_dust, output_tag): fig1 = plt.figure() ax1 = fig1.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) fig3 = plt.figure() ax3 = fig3.add_subplot(111, projection='3d') fig4 = plt.figure() ax4 = fig4.add_subplot(111) ax1.scatter(mass_gal, mass_dust) ax2.scatter(mass_halo, mass_dust) #ax3.scatter(mass_gal, mass_halo, mass_dust) hb = ax4.hexbin(mass_halo, mass_dust, bins='log', cmap='inferno') ax1.set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax1.set_ylabel(r'$\mathbf{log_{10} \: M_{Dust}}$', fontsize = PlotScripts.global_labelsize) ax2.set_xlabel(r'$\mathbf{log_{10} \: M_{vir} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax2.set_ylabel(r'$\mathbf{log_{10} \: M_{Dust}}$', fontsize = PlotScripts.global_labelsize) ax4.set_xlabel(r'$\mathbf{log_{10} \: M_{vir} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax4.set_ylabel(r'$\mathbf{log_{10} \: M_{Dust}}$', fontsize = PlotScripts.global_labelsize) cb = fig4.colorbar(hb, ax=ax4) cb.set_label('log10(N)') outputFile1 = "./{0}_galaxy{1}".format(output_tag, output_format) fig1.savefig(outputFile1, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile1)) plt.close(fig1) outputFile2 = "./{0}_halo{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) #outputFile3 = "./{0}_3D{1}".format(output_tag, output_format) #fig3.savefig(outputFile3, bbox_inches='tight') # Save the figure #print('Saved file to {0}'.format(outputFile3)) #plt.close(fig3) outputFile4 = "./{0}_hexbin{1}".format(output_tag, output_format) fig4.savefig(outputFile4, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile4)) plt.close(fig4) ### Here ends the plotting functions. ### ### Here begins the functions that calculate various properties for the galaxies (fesc, Magnitude etc). ### def Calculate_HaloPartStellarMass(halo_part, stellar_mass, bound_low, bound_high): ''' Calculates the stellar mass for galaxies whose host halos contain a specified number of particles. Parameters ---------- halo_part : array Array containing the number of particles inside each halo. stellar_mass : array Array containing the Stellar Mass for each galaxy (entries align with HaloPart). Units of log10(Msun). bound_low, bound_high : int We calculate the Stellar Mass of galaxies whose host halo has, bound_low <= halo_part <= bound_high. Return ----- mass, mass_std : float Mean and standard deviation stellar mass of galaxies whose host halo has number of particles between the specified bounds. Units of log10(Msun) Units ----- Input Stellar Mass is in units of log10(Msun). Output mean/std Stellar Mass is in units of log10(Msun). ''' w = np.where((halo_part >= bound_low) & (halo_part <= bound_high))[0] # Find the halos with particle number between the bounds. mass = np.mean(10**(stellar_mass[w])) mass_std = np.std(10**(stellar_mass[w])) return np.log10(mass), np.log10(mass_std) ## def calculate_UV_extinction(z, L, M): ''' Calculates the observed UV magnitude after dust extinction is accounted for. Parameters ---------- z : float Redshift we are calculating the extinction at. L, M : array, length equal to the number of galaxies at this snapshot. Array containing the UV luminosities and magnitudes. Returns ------- M_UV_obs : array, length equal to the number of galaxies at this snapshot. Array containing the observed UV magnitudes. Units ----- Luminosities are in units of log10(erg s^-1 A^-1). Magnitudes are in the AB system. ''' M_UV_bins = np.arange(-24, -16, 0.1) A_mean = np.zeros((len(MUV_bins))) # A_mean is the average UV extinction for a given UV bin. for j in range(0, len(M_UV_bins)): beta = calculate_beta(M_UV_bins[j], AllVars.SnapZ[current_snap]) # Fits the beta parameter for the current redshift/UV bin. dist = np.random.normal(beta, 0.34, 10000) # Generates a normal distribution with mean beta and standard deviation of 0.34. A = 4.43 + 1.99*dist A[A < 0] = 0 # Negative extinctions don't make sense. A_Mean[j] = np.mean(A) indices = np.digitize(M, M_UV_bins) # Bins the simulation magnitude into the MUV bins. Note that digitize defines an index i if bin[i-1] <= x < bin[i] whereas I prefer bin[i] <= x < bin[i+1] dust = A_Mean[indices] flux = AllVars.Luminosity_to_Flux(L, 10.0) # Calculate the flux from a distance of 10 parsec, units of log10(erg s^-1 A^-1 cm^-2). flux_observed = flux - 0.4*dust f_nu = ALlVars.spectralflux_wavelength_to_frequency(10**flux_observed, 1600) # Spectral flux desnity in Janksy. M_UV_obs(-2.5 * np.log10(f_nu) + 8.90) # AB Magnitude from http://www.astro.ljmu.ac.uk/~ikb/convert-units/node2.html return M_UV_obs ## def update_cumulative_stats(mean_pool, std_pool, N_pool, mean_local, std_local, N_local): ''' Update the cumulative statistics (such as Stellar Mass Function, Mvir-Ngamma, fesc-z) that are saved across files. Pooled mean formulae taken : from https://www.ncbi.nlm.nih.gov/books/NBK56512/ Pooled variance formulae taken from : https://en.wikipedia.org/wiki/Pooled_variance Parameters ---------- mean_pool, std_pool, N_pool : array of floats with length equal to the number of bins (e.g. the mass bins for the Stellar Mass Function). The current mean, standard deviation and number of data points within in each bin. This is the array that will be updated in this function. mean_local, std_local, N_local : array of floats with length equal to the number of bins. The mean, standard deviation and number of data points within in each bin that will be added to the pool. Returns ------- mean_pool, std_pool, N_pool : (See above) The updated arrays with the local values added and accounted for within the pools. Units ----- All units are kept the same as the input units. Values are in real-space (not log-space). ''' N_times_mean_local = np.multiply(N_local, mean_local) N_times_var_local = np.multiply(N_local - 1, np.multiply(std_local, std_local)) # Actually N - 1 because of Bessel's Correction # https://en.wikipedia.org/wiki/Bessel%27s_correction). # N_times_mean_pool = np.add(N_times_mean_local, np.multiply(N_pool, mean_pool)) N_times_var_pool = np.add(N_times_var_local, np.multiply(N_pool - 1, np.multiply(std_pool, std_pool))) N_pool = np.add(N_local, N_pool) ''' print(mean_local) print(type(mean_local)) print((type(mean_local).__module__ == np.__name__)) print(isinstance(mean_local, list)) print(isinstance(mean_local,float64)) print(isinstance(mean_local,float32)) ''' if (((type(mean_local).__module__ == np.__name__) == True or (isinstance(mean_local, list) == True)) and isinstance(mean_local, float) == False and isinstance(mean_local, int) == False and isinstance(mean_local,float32) == False and isinstance(mean_local, float64) == False): # Checks to see if we are dealing with arrays. for i in range(0, len(N_pool)): if(N_pool[i] == 0): # This case is when we have no data points in the bin. mean_pool[i] = 0.0 else: mean_pool[i] = N_times_mean_pool[i]/N_pool[i] if(N_pool[i] < 3): # In this instance we don't have enough data points to properly calculate the standard deviation. std_pool[i] = 0.0 else: std_pool[i] = np.sqrt(N_times_var_pool[i]/ (N_pool[i] - 2)) # We have -2 because there is two instances of N_pool contains two 'N - 1' terms. else: mean_pool = N_times_mean_pool / N_pool if(N_pool < 3): std_pool = 0.0 else: std_pool = np.sqrt(N_times_var_pool / (N_pool - 2)) return mean_pool, std_pool ### Here ends the functions that deal with galaxy data manipulation. ### ################################# if __name__ == '__main__': np.seterr(divide='ignore') number_models = 4 galaxies_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/galaxies/const_0.3_z5.782" merged_galaxies_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/galaxies/const_0.3_MergedGalaxies" photo_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/grids/cifog/const_0.3_photHI" zreion_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/grids/cifog/const_0.3_reionization_redshift" galaxies_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/galaxies/fej_alpha0.40_beta0.05_z5.782" merged_galaxies_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/galaxies/fej_alpha0.40_beta0.05_MergedGalaxies" photo_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/grids/cifog/fej_alpha0.40_beta0.05_photHI" zreion_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/grids/cifog/fej_alpha0.40_beta0.05_reionization_redshift" galaxies_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/galaxies/MHneg_1e8_1e12_0.99_0.05_z5.782" merged_galaxies_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/galaxies/MHneg_1e8_1e12_0.99_0.05_MergedGalaxies" photo_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/grids/cifog/MHneg_1e8_1e12_0.99_0.05_photHI" zreion_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/grids/cifog/MHneg_1e8_1e12_0.99_0.05_reionization_redshift" galaxies_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/galaxies/MHpos_1e8_1e12_0.01_0.50_z5.782" merged_galaxies_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/galaxies/MHpos_1e8_1e12_0.01_0.50_MergedGalaxies" photo_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/grids/cifog/MHpos_1e8_1e12_0.01_0.50_photHI" zreion_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/grids/cifog/MHpos_1e8_1e12_0.01_0.50_reionization_redshift" galaxies_filepath_array = [galaxies_model1, galaxies_model2, galaxies_model3, galaxies_model4] photo_array = [photo_model1, photo_model2, photo_model3, photo_model4] zreion_array = [zreion_model1, zreion_model2, zreion_model3, zreion_model4] GridSize_array = [256, 256, 256, 256] precision_array = [2, 2, 2, 2] merged_galaxies_filepath_array = [merged_galaxies_model1, merged_galaxies_model2, merged_galaxies_model3, merged_galaxies_model4] number_substeps = [10, 10, 10, 10] # How many substeps does each model have (specified by STEPS variable within SAGE). number_snapshots = [99, 99, 99, 99] # Number of snapshots in the simulation (we don't have to do calculations for ALL snapshots). # Tiamat extended has 164 snapshots. FirstFile = [0, 0, 0, 0] # The first file number THAT WE ARE PLOTTING. #LastFile = [63, 63, 63, 63] # The last file number THAT WE ARE PLOTTING. LastFile = [0, 0, 0, 0] # The last file number THAT WE ARE PLOTTING. NumFile = [64, 64, 64, 64] # The number of files for this simulation (plotting a subset of these files is allowed). same_files = [0, 0, 0, 0] # In the case that model 1 and model 2 (index 0 and 1) have the same files, we don't want to read them in a second time. # This array will tell us if we should keep the files for the next model or otherwise throw them away. # The files will be kept until same_files[current_model_number] = 0. # For example if we had 5 models we were plotting and model 1, 2, 3 shared the same files and models 4, 5 shared different files, # Then same_files = [1, 1, 0, 1, 0] would be the correct values. done_model = np.zeros((number_models)) # We use this to keep track of if we have done a model already. model_tags = [r"$\mathbf{f_\mathrm{esc} \: Constant}$", r"$\mathbf{f_\mathrm{esc} \: \propto \: f_\mathrm{ej}}$", r"$\mathbf{f_\mathrm{esc} \: \propto \: M_\mathrm{H}^{-1}}$", r"$\mathbf{f_\mathrm{esc} \: \propto \: M_\mathrm{H}}$"] ## Constants used for each model. ## # Need to add an entry for EACH model. # halo_cut = [32, 32, 32, 32] # Only calculate properties for galaxies whose host halos have at least this many particles. # For Tiamat, z = [6, 7, 8] are snapshots [78, 64, 51] # For Kali, z = [6, 7, 8] are snapshots [93, 76, 64] #SnapList = [np.arange(0,99), np.arange(0,99)] # These are the snapshots over which the properties are calculated. NOTE: If the escape fraction is selected (fesc_prescription == 3) then this should be ALL the snapshots in the simulation as this prescriptions is temporally important. #SnapList = [np.arange(20,99), np.arange(20, 99), np.arange(20, 99)] SnapList = [[33, 50, 76, 93], [33, 50, 76, 93], [33, 50, 76, 93], [33, 50, 76, 93]] #SnapList = [[64], # [64], # [64], # [64]] #SnapList = [[33, 50, 64, 76, 93]] #SnapList = [[64], [64]] #SnapList = [np.arange(20,99)] #PlotSnapList = [[30, 50, 64, 76, 93]] #PlotSnapList = [[93, 76, 64], [93, 76, 64]] #SnapList = [[93, 76, 64], [93, 76, 64]] PlotSnapList = SnapList simulation_norm = [5, 5, 5, 5] # Changes the constants (cosmology, snapshot -> redshift mapping etc) for each simulation. # 0 for MySim (Manodeep's old one). # 1 for Mini-Millennium. # 2 for Tiamat (up to z =5). # 3 for extended Tiamat (down to z = 1.6ish). # 4 for Britton's Sim Pip # 5 for Manodeep's new simulation Kali. stellar_mass_halolen_lower = [32, 95, 95, 95] # These limits are for the number of particles in a halo. stellar_mass_halolen_upper = [50, 105, 105, 105] # We calculate the average stellar mass for galaxies whose host halos have particle count between these limits. calculate_observed_LF = [0, 0, 0, 0] # Determines whether we want to account for dust extinction when calculating the luminosity function of each model. paper_plots = 1 ############################################################################################################## ## Do a few checks to ensure all the arrays were specified properly. ## for model_number in range(0,number_models): assert(LastFile[model_number] - FirstFile[model_number] + 1 >= size) if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") if (number_snapshots[model_number] != len(AllVars.SnapZ)): # Here we do a check to ensure that the simulation we've defined correctly matches the number of snapshots we have also defined. print("The number_snapshots array is {0}".format(number_snapshots)) print("The simulation_norm array is {0}".format(simulation_norm)) print("The number of snapshots for model_number {0} has {1} but you've said there is only {2}".format(model_number, len(AllVars.SnapZ), number_snapshots[model_number])) raise ValueError("Check either that the number of snapshots has been defined properly and that the normalization option is correct.") ###################################################################### ##################### SETTING UP ARRAYS ############################## ###################################################################### ### The arrays are set up in a 3 part process. ### ### This is because our arrays are 3D nested to account for the model number and snapshots. ### # First set up the outer most array. # ## Arrays for functions of stellar mass. ## SMF = [] # Stellar Mass Function. mean_fesc_galaxy_array = [] # Mean escape fraction as a function of stellar mass. std_fesc_galaxy_array = [] # Same as above but standard devation. N_galaxy_array = [] # Number of galaxies as a function of stellar mass. mean_BHmass_galaxy_array = [] # Black hole mass as a function of stellar mass. std_BHmass_galaxy_array = [] # Same as above but standard deviation. mergers_galaxy_array = [] # Number of mergers as a function of halo mass. mean_dust_galaxy_array = [] # Mean dust mass as a function of stellar mass. std_dust_galaxy_array = [] # Same as above but standard deviation. mean_sfr_galaxy_array = [] # Mean star formation rate as a # function of stellar mass std_sfr_galaxy_array = [] # Same as above but standard deviation. mean_ssfr_galaxy_array = [] # Mean specific star formation rate as a # function of stellar mass std_ssfr_galaxy_array = [] # Same as above but standard deviation. mean_Ngamma_galaxy_array = [] # Mean number of ionizing photons emitted as # a function of stellar mass. std_Ngamma_galaxy_array = [] # Same as above but standard deviation. mean_photo_galaxy_array = [] # Mean photoionization rate. std_photo_galaxy_array = [] # Std photoionization rate. mean_reionmod_galaxy_array = [] # Mean reionization modifier using RSAGE. std_reionmod_galaxy_array = [] # Std. mean_gnedin_reionmod_galaxy_array = [] # Mean reionization modifier using Gnedin analytic prescription. std_gnedin_reionmod_galaxy_array = [] # Std. ## Arrays for functions of halo mass. ## mean_ejected_halo_array = [] # Mean ejected fractions as a function of halo mass. std_ejected_halo_array = [] # Same as above but standard deviation. mean_fesc_halo_array = [] # Mean escape fraction as a function of halo mass. std_fesc_halo_array = [] # Same as above but standard deviation. mean_Ngamma_halo_array = [] # Mean number of ionizing photons THAT ESCAPE as a function of halo mass. std_Ngamma_halo_array = [] # Same as above but standard deviation. N_halo_array = [] # Number of galaxies as a function of halo mass. mergers_halo_array = [] # Number of mergers as a function of halo mass. mean_quasar_activity_array = [] # Mean fraction of galaxies that have quasar actvitity as a function of halo mas. std_quasar_activity_array = [] # Same as above but standard deviation. mean_reionmod_halo_array = [] # Mean reionization modifier as a function of halo mass. std_reionmod_halo_array = [] # Same as above but for standard deviation. mean_dust_halo_array = [] # Mean dust mass as a function of halo mass. std_dust_halo_array = [] # Same as above but standard deviation. ## Arrays for functions of redshift. ## sum_Ngamma_z_array = [] # Total number of ionizing photons THAT ESCAPE as a functio of redshift. mean_fesc_z_array = [] # Mean number of ionizing photons THAT ESCAPE as a function of redshift. std_fesc_z_array = [] # Same as above but standard deviation. N_z = [] # Number of galaxies as a function of redshift. galaxy_halo_mass_mean = [] # Mean galaxy mass as a function of redshift. N_quasars_z = [] # This tracks how many quasars went off during a specified snapshot. N_quasars_boost_z = [] # This tracks how many galaxies are having their escape fraction boosted by quasar activity. dynamicaltime_quasars_mean_z = [] # Mean dynamical time of galaxies that have a quasar event as a function of redshift. dynamicaltime_quasars_std_z = [] # Same as above but standard deviation. dynamicaltime_all_mean_z = [] # Mean dynamical time of all galaxies. dynamicaltime_all_std_z = [] # Same as above but standard deviation. mean_reionmod_z = [] # Mean reionization modifier as a function of redshift. std_reionmod_z = [] # Same as above but for standard deviation. N_reionmod_z = [] # Number of galaxies with a non-negative reionization modifier. mean_ejected_z = [] # Mean ejected fraction as a function of redshift. std_ejected_z = [] # Same as above but for standard deviation. ## Arrays that aren't functions of other variables. ## Ngamma_global = [] mass_global = [] fesc_global = [] ## Arrays as a function of fej ## mean_Ngamma_fej = [] std_Ngamma_fej = [] N_fej = [] ## Now the outer arrays have been defined, set up the next nest level for the number of models. ## for model_number in range(0,number_models): ## Galaxy Arrays ## SMF.append([]) mean_fesc_galaxy_array.append([]) std_fesc_galaxy_array.append([]) N_galaxy_array.append([]) mean_BHmass_galaxy_array.append([]) std_BHmass_galaxy_array.append([]) mergers_galaxy_array.append([]) mean_dust_galaxy_array.append([]) std_dust_galaxy_array.append([]) mean_sfr_galaxy_array.append([]) std_sfr_galaxy_array.append([]) mean_ssfr_galaxy_array.append([]) std_ssfr_galaxy_array.append([]) mean_Ngamma_galaxy_array.append([]) std_Ngamma_galaxy_array.append([]) mean_photo_galaxy_array.append([]) std_photo_galaxy_array.append([]) mean_reionmod_galaxy_array.append([]) std_reionmod_galaxy_array.append([]) mean_gnedin_reionmod_galaxy_array.append([]) std_gnedin_reionmod_galaxy_array.append([]) ## Halo arrays. ## mean_ejected_halo_array.append([]) std_ejected_halo_array.append([]) mean_fesc_halo_array.append([]) std_fesc_halo_array.append([]) mean_Ngamma_halo_array.append([]) std_Ngamma_halo_array.append([]) N_halo_array.append([]) mergers_halo_array.append([]) mean_quasar_activity_array.append([]) std_quasar_activity_array.append([]) mean_reionmod_halo_array.append([]) std_reionmod_halo_array.append([]) mean_dust_halo_array.append([]) std_dust_halo_array.append([]) ## Redshift arrays. ## sum_Ngamma_z_array.append([]) mean_fesc_z_array.append([]) std_fesc_z_array.append([]) N_z.append([]) galaxy_halo_mass_mean.append([]) N_quasars_z.append([]) N_quasars_boost_z.append([]) dynamicaltime_quasars_mean_z.append([]) dynamicaltime_quasars_std_z.append([]) dynamicaltime_all_mean_z.append([]) dynamicaltime_all_std_z.append([]) mean_reionmod_z.append([]) std_reionmod_z.append([]) N_reionmod_z.append([]) mean_ejected_z.append([]) std_ejected_z.append([]) ## Arrays that aren't functions ## Ngamma_global.append([]) mass_global.append([]) fesc_global.append([]) ## Arrays as a function of fej ## mean_Ngamma_fej.append([]) std_Ngamma_fej.append([]) N_fej.append([]) ## And then finally set up the inner most arrays ## ## NOTE: We do the counts as float so we can keep consistency when we're calling MPI operations (just use MPI.FLOAT rather than deciding if we need to use MPI.INT) for snapshot_idx in range(len(SnapList[model_number])): ## For the arrays that are functions of stellar/halo mass, the inner most level will be an array with the statistic binned across mass ## ## E.g. SMF[model_number][snapshot_idx] will return an array whereas N_z[model_number][snapshot_idx] will return a float. ## ## Functions of stellar mass arrays. ## SMF[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_fesc_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_fesc_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) N_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_BHmass_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_BHmass_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mergers_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_dust_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_dust_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_sfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_sfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_ssfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_ssfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_Ngamma_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_Ngamma_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_photo_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_photo_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_gnedin_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_gnedin_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) ## Function of halo mass arrays. ## mean_ejected_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_ejected_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_fesc_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_fesc_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_Ngamma_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_Ngamma_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) N_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mergers_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_quasar_activity_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_quasar_activity_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_reionmod_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_reionmod_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_dust_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_dust_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) ## Function of Redshift arrays. ## sum_Ngamma_z_array[model_number].append(0.0) mean_fesc_z_array[model_number].append(0.0) std_fesc_z_array[model_number].append(0.0) N_z[model_number].append(0.0) galaxy_halo_mass_mean[model_number].append(0.0) N_quasars_z[model_number].append(0.0) N_quasars_boost_z[model_number].append(0.0) dynamicaltime_quasars_mean_z[model_number].append(0.0) dynamicaltime_quasars_std_z[model_number].append(0.0) dynamicaltime_all_mean_z[model_number].append(0.0) dynamicaltime_all_std_z[model_number].append(0.0) mean_reionmod_z[model_number].append(0.0) std_reionmod_z[model_number].append(0.0) N_reionmod_z[model_number].append(0.0) mean_ejected_z[model_number].append(0.0) std_ejected_z[model_number].append(0.0) Ngamma_global[model_number].append([]) mass_global[model_number].append([]) fesc_global[model_number].append([]) ## Arrays as a function of fej. ## mean_Ngamma_fej[model_number].append(np.zeros((NB_fej), dtype = np.float32)) std_Ngamma_fej[model_number].append(np.zeros((NB_fej), dtype = np.float32)) N_fej[model_number].append(np.zeros((NB_fej), dtype = np.float32)) ###################################################################### #################### ALL ARRAYS SETUP ################################ ###################################################################### ## Now it's (finally) time to read in all the data and do the actual work. ## for model_number in range(number_models): if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") if (done_model[model_number] == 1): # If we have already done this model (i.e., we kept the files and skipped this loop), move along. assert(FirstFile[model_number] == FirstFile[model_number - 1]) assert(LastFile[model_number] == LastFile[model_number - 1]) continue for fnr in range(FirstFile[model_number] + rank, LastFile[model_number]+1, size): # Divide up the input files across the processors. GG, Gal_Desc = ReadScripts.ReadGals_SAGE(galaxies_filepath_array[model_number], fnr, number_snapshots[model_number], comm) # Read galaxies G_Merged, _ = ReadScripts.ReadGals_SAGE(merged_galaxies_filepath_array[model_number], fnr, number_snapshots[model_number], comm) # Also need the merged galaxies. G = ReadScripts.Join_Arrays(GG, G_Merged, Gal_Desc) # Then join them together for all galaxies. keep_files = 1 # Flips to 0 when we are done with this file. current_model_number = model_number # Used to differentiate between outer model_number and the inner model_number because we can keep files across model_numbers. while(keep_files == 1): ## Just a few definitions to cut down the clutter a smidge. ## current_halo_cut = halo_cut[current_model_number] NumSubsteps = number_substeps[current_model_number] do_observed_LF = calculate_observed_LF[current_model_number] for snapshot_idx in range(0, len(SnapList[current_model_number])): # Now let's calculate stats for each required redshift. current_snap = SnapList[current_model_number][snapshot_idx] # Get rid of some clutter. w_gal = np.where((G.GridHistory[:, current_snap] != -1) & (G.GridStellarMass[:, current_snap] > 0.0) & (G.LenHistory[:, current_snap] > current_halo_cut) & (G.GridSFR[:, current_snap] >= 0.0) & (G.GridFoFMass[:, current_snap] >= 0.0))[0] # Only include those galaxies that existed at the current snapshot, had positive (but not infinite) stellar/Halo mass and Star formation rate. Ensure the galaxies also resides in a halo that is sufficiently resolved. w_merged_gal = np.where((G_Merged.GridHistory[:, current_snap] != -1) & (G_Merged.GridStellarMass[:, current_snap] > 0.0) & (G_Merged.LenHistory[:, current_snap] > current_halo_cut) & (G_Merged.GridSFR[:, current_snap] >= 0.0) & (G_Merged.GridFoFMass[:, current_snap] >= 0.0) & (G_Merged.LenMergerGal[:,current_snap] > current_halo_cut))[0] print("There were {0} galaxies for snapshot {1} (Redshift {2:.3f}) model {3}.".format(len(w_gal), current_snap, AllVars.SnapZ[current_snap], current_model_number)) if (len(w_gal) == 0): continue mass_gal = np.log10(G.GridStellarMass[w_gal, current_snap] * 1.0e10 / AllVars.Hubble_h) # Msun. Log Units. w_SFR = w_gal[np.where((G.GridSFR[w_gal, current_snap] > 0.0))[0]] mass_SFR_gal = np.log10(G.GridStellarMass[w_SFR, current_snap] * \ 1.0e10 / AllVars.Hubble_h) SFR_gal = np.log10(G.GridSFR[w_SFR,current_snap]) sSFR_gal = SFR_gal - mass_SFR_gal halo_part_count = G.LenHistory[w_gal, current_snap] metallicity_gal = G.GridZ[w_gal, current_snap] metallicity_tremonti_gal = np.log10(G.GridZ[w_gal, current_snap] / 0.02) + 9.0 # Using the Tremonti relationship for metallicity. mass_central = np.log10(G.GridFoFMass[w_gal, current_snap] * 1.0e10 / AllVars.Hubble_h) # Msun. Log Units. ejected_fraction = G.EjectedFraction[w_gal, current_snap] w_dust = np.where(((G.GridDustColdGas[w_gal, current_snap] +G.GridDustHotGas[w_gal, current_snap] +G.GridDustEjectedMass[w_gal, current_snap]) > 0.0) & (G.GridType[w_gal, current_snap] == 0))[0] total_dust_gal = np.log10((G.GridDustColdGas[w_gal[w_dust], current_snap] +G.GridDustHotGas[w_gal[w_dust], current_snap] +G.GridDustEjectedMass[w_gal[w_dust], current_snap]) * 1.0e10 / AllVars.Hubble_h) mass_gal_dust = np.log10(G.GridStellarMass[w_gal[w_dust], current_snap] * 1.0e10 / AllVars.Hubble_h) mass_centralgal_dust = np.log10(G.GridFoFMass[w_gal[w_dust], current_snap] * 1.0e10 / AllVars.Hubble_h) fesc = G.Gridfesc[w_gal, current_snap] fesc[fesc < 0.0] = 0.0 Ngamma_gal = G.GridNgamma_HI[w_gal, current_snap] # 1.0e50 # photons/s. if model_number < 3: Ngamma_gal += 50.0 # Old versions of SAGE incorrectly # subtracted 50. Ngamma_gal *= fesc reionmod = G.GridReionMod[w_gal, current_snap] mass_reionmod_central = mass_central[reionmod > -1] mass_reionmod_gal = mass_gal[reionmod > -1] reionmod = reionmod[reionmod > -1] # Some satellite galaxies that don't have HotGas and hence won't be stripped. As a result reionmod = -1 for these. Ignore them. mass_BH = G.GridBHMass[w_gal, current_snap] * 1.0e10 / AllVars.Hubble_h # Msun. Not log units. L_UV = SFR_gal + 39.927 # Using relationship from STARBURST99, units of erg s^-1 A^-1. Log Units. M_UV = AllVars.Luminosity_to_ABMag(L_UV, 1600) if (do_observed_LF == 1): # Calculate the UV extinction if requested. M_UV_obs = calculate_UV_extinction(AllVars.SnapZ[current_snap], L_UV, M_UV[snap_idx]) galaxy_halo_mass_mean_local, galaxy_halo_mass_std_local = Calculate_HaloPartStellarMass(halo_part_count, mass_gal, stellar_mass_halolen_lower[current_model_number], stellar_mass_halolen_upper[current_model_number]) # This is the average stellar mass for galaxies whose halos have the specified number of particles. galaxy_halo_mass_mean[current_model_number][snapshot_idx] += pow(10, galaxy_halo_mass_mean_local) / (LastFile[current_model_number] + 1) # Adds to the average of the mean. photofield_path = "{0}_{1:03d}".format(photo_array[current_model_number], current_snap) #photo_gal = photo.calc_gal_photoion(G.GridHistory[w_gal, current_snap], # photofield_path, # GridSize_array[current_model_number], # precision_array[current_model_number]) #zreion_path = "{0}".format(zreion_array[current_model_number]) #zreion_gal = photo.calc_gal_zreion(G.GridHistory[w_gal, current_snap], # zreion_path, # GridSize_array[current_model_number], # precision_array[current_model_number]) z_0 = 8.0 z_r = 7.0 gnedin_mfilt = ga.get_filter_mass(np.array(AllVars.SnapZ[current_snap]), z_0, z_r) gnedin_reionmod_gal = 1.0 / pow(1.0 + 0.26*pow(10, gnedin_mfilt - mass_central), 3.0) ########################################### ######## BASE PROPERTIES CALCULATED ####### ########################################### # Time to calculate relevant statistics. ### Functions of Galaxies/Stellar Mass ### ## Stellar Mass Function ## (counts_local, bin_edges, bin_middle) = AllVars.Calculate_Histogram(mass_gal, bin_width, 0, m_gal_low, m_gal_high) # Bin the Stellar Mass SMF[current_model_number][snapshot_idx] += counts_local ## Escape Fraction ## (mean_fesc_galaxy_local, std_fesc_galaxy_local, N_local, sum_fesc_galaxy, bin_middle) = AllVars.Calculate_2D_Mean(mass_gal, fesc, bin_width, m_gal_low, m_gal_high) (mean_fesc_galaxy_array[current_model_number][snapshot_idx], std_fesc_galaxy_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_fesc_galaxy_array[current_model_number][snapshot_idx], std_fesc_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_fesc_galaxy_local, std_fesc_galaxy_local, N_local) ## Black Hole Mass ## (mean_BHmass_galaxy_local, std_BHmass_galaxy_local, N_local, sum_BHmass_galaxy, bin_middle) = AllVars.Calculate_2D_Mean(mass_gal, mass_BH, bin_width, m_gal_low, m_gal_high) (mean_BHmass_galaxy_array[current_model_number][snapshot_idx], std_BHmass_galaxy_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_BHmass_galaxy_array[current_model_number][snapshot_idx], std_BHmass_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_BHmass_galaxy_local, std_BHmass_galaxy_local, N_local) ## Total Dust Mass ## (mean_dust_galaxy_local, std_dust_galaxy_local, N_local, sum_dust_galaxy, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal_dust, total_dust_gal, bin_width, m_gal_low, m_gal_high) (mean_dust_galaxy_array[current_model_number][snapshot_idx], std_dust_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_dust_galaxy_array[current_model_number][snapshot_idx], std_dust_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_dust_galaxy_local, std_dust_galaxy_local, N_local) ## Star Formation Rate ## (mean_sfr_galaxy_local, std_sfr_galaxy_local, N_local, sum_sfr_galaxy, bin_middle) = AllVars.Calculate_2D_Mean( mass_SFR_gal, SFR_gal, bin_width, m_gal_low, m_gal_high) (mean_sfr_galaxy_array[current_model_number][snapshot_idx], std_sfr_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_sfr_galaxy_array[current_model_number][snapshot_idx], std_sfr_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_sfr_galaxy_local, std_sfr_galaxy_local, N_local) ## Specific Star Formation Rate ## (mean_ssfr_galaxy_local, std_ssfr_galaxy_local, N_local, sum_ssfr_galaxy, bin_middle) = AllVars.Calculate_2D_Mean( mass_SFR_gal, sSFR_gal, bin_width, m_gal_low, m_gal_high) (mean_ssfr_galaxy_array[current_model_number][snapshot_idx], std_ssfr_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_ssfr_galaxy_array[current_model_number][snapshot_idx], std_ssfr_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_ssfr_galaxy_local, std_ssfr_galaxy_local, N_local) ## Number of Ionizing Photons ## (mean_Ngamma_galaxy_local, std_Ngamma_galaxy_local, N_local, sum_Ngamma_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal, Ngamma_gal, bin_width, m_gal_low, m_gal_high) (mean_Ngamma_galaxy_array[current_model_number][snapshot_idx], std_Ngamma_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_Ngamma_galaxy_array[current_model_number][snapshot_idx], std_Ngamma_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_Ngamma_galaxy_local, std_Ngamma_galaxy_local, N_local) ## Photoionization rate ## ''' (mean_photo_galaxy_local, std_photo_galaxy_local, N_local, sum_photo_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal, photo_gal, bin_width, m_gal_low, m_gal_high) (mean_photo_galaxy_array[current_model_number][snapshot_idx], std_photo_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_photo_galaxy_array[current_model_number][snapshot_idx], std_photo_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_photo_galaxy_local, std_photo_galaxy_local, N_local) ''' ## RSAGE Reionization Modifier ## (mean_reionmod_galaxy_local, std_reionmod_galaxy_local, N_local, sum_reionmod_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_reionmod_gal, reionmod, bin_width, m_gal_low, m_gal_high) (mean_reionmod_galaxy_array[current_model_number][snapshot_idx], std_reionmod_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_reionmod_galaxy_array[current_model_number][snapshot_idx], std_reionmod_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_reionmod_galaxy_local, std_reionmod_galaxy_local, N_local) ## Gnedin Reionization Modifier ## (mean_gnedin_reionmod_galaxy_local, std_gnedin_reionmod_galaxy_local, N_local, sum_gnedin_reionmod_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal, gnedin_reionmod_gal, bin_width, m_gal_low, m_gal_high) (mean_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx], std_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx], std_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_gnedin_reionmod_galaxy_local, std_gnedin_reionmod_galaxy_local, N_local) N_galaxy_array[current_model_number][snapshot_idx] += N_local ### Functions of Halos/Halo Mass ### ## Ejected Fraction ## (mean_ejected_halo_local, std_ejected_halo_local, N_local, sum_ejected_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_central, ejected_fraction, bin_width, m_low, m_high) (mean_ejected_halo_array[current_model_number][snapshot_idx], std_ejected_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_ejected_halo_array[current_model_number][snapshot_idx], std_ejected_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_ejected_halo_local, std_ejected_halo_local, N_local) # Then update the running total. ## Quasar Fraction ## (mean_quasar_activity_local, std_quasar_activity_local,N_local, sum_quasar_activity_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_central, G.QuasarActivity[w_gal, current_snap], bin_width, m_low, m_high) (mean_quasar_activity_array[current_model_number][snapshot_idx], std_quasar_activity_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_quasar_activity_array[current_model_number][snapshot_idx], std_quasar_activity_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_quasar_activity_local, std_quasar_activity_local, N_local) # Then update the running total. ## fesc Value ## (mean_fesc_halo_local, std_fesc_halo_local, N_local, sum_fesc_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_central, fesc, bin_width, m_low, m_high) (mean_fesc_halo_array[current_model_number][snapshot_idx], std_fesc_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_fesc_halo_array[current_model_number][snapshot_idx], std_fesc_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_fesc_halo_local, std_fesc_halo_local, N_local) # Then update the running total. ## Ngamma ## #(mean_Ngamma_halo_local, std_Ngamma_halo_local, N_local, sum_Ngamma_halo, bin_middle) \ #= AllVars.Calculate_2D_Mean(mass_central, ionizing_photons, bin_width, m_low, m_high) #mean_Ngamma_halo_local = np.divide(mean_Ngamma_halo_local, 1.0e50) ## Divide out a constant to keep the numbers manageable. #std_Ngamma_halo_local = np.divide(std_Ngamma_halo_local, 1.0e50) #(mean_Ngamma_halo_array[current_model_number][snapshot_idx], std_Ngamma_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_Ngamma_halo_array[current_model_number][snapshot_idx], std_Ngamma_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_Ngamma_halo_local, std_Ngamma_halo_local, N_local) # Then update the running total. ## Reionization Modifier ## (mean_reionmod_halo_local, std_reionmod_halo_local, N_local, sum_reionmod_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_reionmod_central, reionmod, bin_width, m_low, m_high) (mean_reionmod_halo_array[current_model_number][snapshot_idx], std_reionmod_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_reionmod_halo_array[current_model_number][snapshot_idx], std_reionmod_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_reionmod_halo_local, std_reionmod_halo_local, N_local) # Then update the running total. ## Total Dust Mass ## (mean_dust_halo_local, std_dust_halo_local, N_local, sum_dust_halo, bin_middle) = AllVars.Calculate_2D_Mean( mass_centralgal_dust, total_dust_gal, bin_width, m_low, m_high) (mean_dust_halo_array[current_model_number][snapshot_idx], std_dust_halo_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_dust_halo_array[current_model_number][snapshot_idx], std_dust_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_dust_halo_local, std_dust_halo_local, N_local) N_halo_array[current_model_number][snapshot_idx] += N_local ### Functions of redshift ### ## Ngamma ## #sum_Ngamma_z_array[current_model_number][snapshot_idx] += np.sum(np.divide(ionizing_photons, 1.0e50)) # Remember that we're dividing out a constant! ## fesc Value ## (mean_fesc_z_array[current_model_number][snapshot_idx], std_fesc_z_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_fesc_z_array[current_model_number][snapshot_idx], std_fesc_z_array[current_model_number][snapshot_idx], N_z[current_model_number][snapshot_idx], np.mean(fesc), np.std(fesc), len(w_gal)) # Updates the mean escape fraction for this redshift. ## Reionization Modifier ## (mean_reionmod_z[current_model_number][snapshot_idx], std_reionmod_z[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_reionmod_z[current_model_number][snapshot_idx], std_reionmod_z[current_model_number][snapshot_idx], N_reionmod_z[current_model_number][snapshot_idx], np.mean(reionmod), np.std(reionmod), len(reionmod)) N_reionmod_z[current_model_number][snapshot_idx] += len(reionmod) ## Ejected Fraction ## (mean_ejected_z[current_model_number][snapshot_idx],std_ejected_z[current_model_number][snapshot_idx]) \ = update_cumulative_stats(mean_ejected_z[current_model_number][snapshot_idx], std_ejected_z[current_model_number][snapshot_idx], N_z[current_model_number][snapshot_idx], np.mean(ejected_fraction), np.std(ejected_fraction), len(w_gal)) N_z[current_model_number][snapshot_idx] += len(w_gal) #### Arrays that are just kept across snapshots ## Ngamma_global[current_model_number][snapshot_idx].append(Ngamma_gal) mass_global[current_model_number][snapshot_idx].append(mass_gal) fesc_global[current_model_number][snapshot_idx].append(fesc) #### Arrays that are function of fej ## (mean_Ngamma_fej_local, std_Ngamma_fej_local, N_local, sum_Ngamma_fej_local, bin_middle) = AllVars.Calculate_2D_Mean( ejected_fraction, Ngamma_gal, fej_bin_width, fej_low, fej_high) (mean_Ngamma_fej[current_model_number][snapshot_idx], std_Ngamma_fej[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_Ngamma_fej[current_model_number][snapshot_idx], std_Ngamma_fej[current_model_number][snapshot_idx], N_fej[current_model_number][snapshot_idx], mean_Ngamma_fej_local, std_Ngamma_fej_local, N_local) N_fej[current_model_number][snapshot_idx] += N_local done_model[current_model_number] = 1 if (current_model_number < number_models): keep_files = same_files[current_model_number] # Decide if we want to keep the files loaded or throw them out. current_model_number += 1 # Update the inner loop model number. #StellarMassFunction(PlotSnapList, SMF, simulation_norm, FirstFile, # LastFile, NumFile, galaxy_halo_mass_mean, model_tags, # 1, paper_plots, "wtf") #plot_reionmod(PlotSnapList, SnapList, simulation_norm, mean_reionmod_halo_array, #std_reionmod_halo_array, N_halo_array, mean_reionmod_z, #std_reionmod_z, N_reionmod_z, False, model_tags, #"reionmod_selfcon") #plot_dust_scatter(SnapList, mass_gal_dust, mass_centralgal_dust, total_dust_gal, # "dust_scatter") #plot_dust(PlotSnapList, SnapList, simulation_norm, mean_dust_galaxy_array, # std_dust_galaxy_array, N_galaxy_array, mean_dust_halo_array, # std_dust_halo_array, N_halo_array, False, model_tags, # "dustmass_total") #plot_stellarmass_blackhole(PlotSnapList, simulation_norm, mean_BHmass_galaxy_array, # std_BHmass_galaxy_array, N_galaxy_array, # FirstFile, LastFile, NumFile, # model_tags, "StellarMass_BHMass") #plot_ejectedfraction(SnapList, PlotSnapList, simulation_norm, # mean_ejected_halo_array, std_ejected_halo_array, # N_halo_array, mean_ejected_z, std_ejected_z, N_z, # model_tags, "ejectedfraction") #plot_quasars_count(SnapList, PlotSnapList, N_quasars_z, N_quasars_boost_z, N_z, mean_quasar_activity_array, std_quasar_activity_array, N_halo_array, mergers_halo_array, SMF, mergers_galaxy_array, fesc_prescription, simulation_norm, FirstFile, LastFile, NumFile, model_tags, "SN_Prescription") plot_fesc_galaxy(SnapList, PlotSnapList, simulation_norm, mean_fesc_galaxy_array, std_fesc_galaxy_array, N_galaxy_array, mean_fesc_halo_array, std_fesc_halo_array, N_halo_array, galaxy_halo_mass_mean, model_tags, paper_plots, mass_global, fesc_global, Ngamma_global, "fesc_paper") plot_reionmod_galaxy(SnapList, PlotSnapList, simulation_norm, mean_reionmod_galaxy_array, std_reionmod_galaxy_array, N_galaxy_array, mean_gnedin_reionmod_galaxy_array, std_gnedin_reionmod_galaxy_array, model_tags, paper_plots, "reionmod") exit() #plot_nion_galaxy(SnapList, PlotSnapList, simulation_norm, # mean_Ngamma_galaxy_array, std_Ngamma_galaxy_array, # N_galaxy_array, model_tags, # paper_plots, "Ngamma") ''' plot_photo_galaxy(SnapList, PlotSnapList, simulation_norm, mean_photo_galaxy_array, std_photo_galaxy_array, N_galaxy_array, model_tags, paper_plots, "photo") ''' plot_sfr_galaxy(SnapList, PlotSnapList, simulation_norm, mean_sfr_galaxy_array, std_sfr_galaxy_array, mean_ssfr_galaxy_array, std_ssfr_galaxy_array, N_galaxy_array, model_tags, "sSFR") #plot_fej_Ngamma(SnapList, PlotSnapList, simulation_norm, # mean_Ngamma_fej, std_Ngamma_fej, # N_fej, model_tags, "Ngamma_fej") #plot_photoncount(SnapList, sum_Ngamma_z_array, simulation_norm, FirstFile, LastFile, NumFile, model_tags, "Ngamma_test") ## PARALELL COMPATIBLE #plot_mvir_Ngamma(SnapList, mean_Ngamma_halo_array, std_Ngamma_halo_array, N_halo_array, model_tags, "Mvir_Ngamma_test", fesc_prescription, fesc_normalization, "/lustre/projects/p004_swin/jseiler/tiamat/halo_ngamma/") ## PARALELL COMPATIBLE
241,016
81,302
from multiprocessing import Queue, Value from time import sleep from access_face_vision.source.camera import Camera from access_face_vision.utils import create_parser from access_face_vision import access_logger LOG_LEVEL = 'debug' logger, log_que, que_listener = access_logger.set_main_process_logger(LOG_LEVEL) def test_camera(): logger.info('Starting Camera test') cmd_args = create_parser() camera = Camera(cmd_args, Queue(), log_que, LOG_LEVEL, Value('i',0), draw_frames=True) camera.start() sleep(60) camera.stop() logger.info('Camera test completed') que_listener.stop() if __name__ == '__main__': test_camera()
661
215
from utils.deserializer.protobuf_deserializer import ProtoLoader from pathlib import Path import pandas as pd import pytest PROTOFILES_DIR_PATH = Path(__file__).parent.joinpath("protofilesdir").absolute().__str__() INVALID_PATH = "some/wrong/path" @pytest.mark.parametrize('filepath', ["test_file.pb", "test_file_1.txt", "test_file_2.xml"]) def test_should_return_single_df_sequence_regardless_file_extension(filepath): loader = ProtoLoader(PROTOFILES_DIR_PATH) sequence = loader.get_single_sequence(filepath) assert isinstance(sequence, pd.DataFrame) def test_should_return_not_none_when_directory_not_empty(): loader = ProtoLoader(PROTOFILES_DIR_PATH) seq_list = loader.get_list_of_sequences() assert seq_list is not None def test_should_return_correct_length_of_seq_list(): loader = ProtoLoader(PROTOFILES_DIR_PATH) seq_list = loader.get_list_of_sequences() assert len(seq_list) == 3 def test_should_return_empty_list_when_directory_empty(): loader = ProtoLoader(PROTOFILES_DIR_PATH + INVALID_PATH) seq_list = loader.get_list_of_sequences() assert len(seq_list) == 0 def test_should_check_for_list_when_directory_empty(): loader = ProtoLoader(PROTOFILES_DIR_PATH + INVALID_PATH) seq_list = loader.get_list_of_sequences() assert isinstance(seq_list, list) def test_should_return_list_of_sequences(): loader = ProtoLoader(PROTOFILES_DIR_PATH) seq_list = loader.get_list_of_sequences() for seq in seq_list: assert isinstance(seq, pd.DataFrame)
1,537
531
#!/usr/bin/env python # encoding: utf-8 """ turbine.py Created by Andrew Ning and Katherine Dykes on 2014-01-13. Copyright (c) NREL. All rights reserved. """ from openmdao.main.api import Assembly, Component from openmdao.main.datatypes.api import Float, Array, Enum, Bool, Int from openmdao.lib.drivers.api import FixedPointIterator import numpy as np #from rotorse.rotor import RotorSE #from towerse.tower import TowerSE #from commonse.rna import RNAMass, RotorLoads from drivewpact.drive import DriveWPACT from drivewpact.hub import HubWPACT from commonse.csystem import DirectionVector from commonse.utilities import interp_with_deriv, hstack, vstack from drivese.drive import Drive4pt, Drive3pt from drivese.drivese_utils import blade_moment_transform, blade_force_transform from drivese.hub import HubSE, Hub_System_Adder_drive from SEAMLoads.SEAMLoads import SEAMLoads from SEAMTower.SEAMTower import SEAMTower from SEAMAero.SEAM_AEP import SEAM_PowerCurve from SEAMRotor.SEAMRotor import SEAMBladeStructure # from SEAMGeometry.SEAMGeometry import SEAMGeometry def connect_io(top, cls): cls_name = cls.name for name in cls.list_inputs(): try: top.connect(name, cls_name + '.%s' % name) except: # print 'failed connecting', cls_name, name pass for name in cls.list_outputs(): try: top.connect(cls_name + '.%s' % name, name) except: pass def configure_turbine(assembly, with_new_nacelle=True, flexible_blade=False, with_3pt_drive=False): """a stand-alone configure method to allow for flatter assemblies Parameters ---------- assembly : Assembly an openmdao assembly to be configured with_new_nacelle : bool False uses the default implementation, True uses an experimental implementation designed to smooth out discontinities making in amenable for gradient-based optimization flexible_blade : bool if True, internally solves the coupled aero/structural deflection using fixed point iteration. Note that the coupling is currently only in the flapwise deflection, and is primarily only important for highly flexible blades. If False, the aero loads are passed to the structure but there is no further iteration. """ #SEAM variables ---------------------------------- #d2e = Float(0.73, iotype='in', desc='Dollars to Euro ratio' assembly.add('rated_power',Float(3000., iotype='in', units='kW', desc='Turbine rated power', group='Global')) assembly.add('hub_height', Float(100., iotype='in', units='m', desc='Hub height', group='Global')) assembly.add('rotor_diameter', Float(110., iotype='in', units='m', desc='Rotor diameter', group='Global')) # assembly.add('site_type',Enum('onshore', values=('onshore', 'offshore'), iotype='in', desc='Site type', group='Global')) assembly.add('tower_bottom_diameter', Float(4., iotype='in', desc='Tower bottom diameter', group='Global')) assembly.add('tower_top_diameter', Float(2., iotype='in', desc='Tower top diameter', group='Global')) assembly.add('project_lifetime', Float(iotype = 'in', desc='Operating years', group='Global')) assembly.add('rho_steel', Float(7.8e3, iotype='in', desc='density of steel', group='Tower')) assembly.add('lifetime_cycles', Float(1.e7, iotype='in', desc='Equivalent lifetime cycles', group='Rotor')) assembly.add('stress_limit_extreme_tower', Float(iotype='in', units='MPa', desc='Tower ultimate strength', group='Tower')) assembly.add('stress_limit_fatigue_tower', Float(iotype='in', units='MPa', desc='Tower fatigue strength', group='Tower')) assembly.add('safety_factor_tower', Float(iotype='in', desc='Tower loads safety factor', group='Tower')) assembly.add('PMtarget_tower', Float(1., iotype='in', desc='', group='Tower')) assembly.add('wohler_exponent_tower', Float(4., iotype='in', desc='Tower fatigue Wohler exponent', group='Tower')) assembly.add('tower_z', Array(iotype='out', desc='Tower discretization')) assembly.add('tower_wall_thickness', Array(iotype='out', units='m', desc='Tower wall thickness')) assembly.add('tower_mass', Float(iotype='out', units='kg', desc='Tower mass')) assembly.add('tsr', Float(iotype='in', units='m', desc='Design tip speed ratio', group='Aero')) assembly.add('F', Float(iotype='in', desc='Rotor power loss factor', group='Aero')) assembly.add('wohler_exponent_blade_flap', Float(iotype='in', desc='Wohler Exponent blade flap', group='Rotor')) assembly.add('nSigma4fatFlap', Float(iotype='in', desc='', group='Loads')) assembly.add('nSigma4fatTower', Float(iotype='in', desc='', group='Loads')) assembly.add('dLoad_dU_factor_flap', Float(iotype='in', desc='', group='Loads')) assembly.add('dLoad_dU_factor_tower', Float(iotype='in', desc='', group='Loads')) assembly.add('blade_edge_dynload_factor_ext', Float(iotype='in', desc='Extreme dynamic edgewise loads factor', group='Loads')) assembly.add('blade_edge_dynload_factor_fat', Float(iotype='in', desc='Fatigue dynamic edgewise loads factor', group='Loads')) assembly.add('PMtarget_blades', Float(1., iotype='in', desc='', group='Rotor')) assembly.add('max_tipspeed', Float(iotype='in', desc='Maximum tip speed', group='Aero')) assembly.add('n_wsp', Int(iotype='in', desc='Number of wind speed bins', group='Aero')) assembly.add('min_wsp', Float(0.0, iotype = 'in', units = 'm/s', desc = 'min wind speed', group='Aero')) assembly.add('max_wsp', Float(iotype = 'in', units = 'm/s', desc = 'max wind speed', group='Aero')) assembly.add('turbulence_int', Float(iotype='in', desc='Reference turbulence intensity', group='Plant_AEP')) # assembly.add('WeibullInput', Bool(True, iotype='in', desc='Flag for Weibull input', group='AEP')) assembly.add('weibull_C', Float(iotype = 'in', units='m/s', desc = 'Weibull scale factor', group='AEP')) assembly.add('weibull_k', Float(iotype = 'in', desc='Weibull shape or form factor', group='AEP')) assembly.add('blade_sections', Int(iotype='in', desc='number of sections along blade', group='Rotor')) assembly.add('wohler_exponent_blade_flap', Float(iotype='in', desc='Blade flap fatigue Wohler exponent', group='Rotor')) assembly.add('MaxChordrR', Float(iotype='in', units='m', desc='Spanwise position of maximum chord', group='Rotor')) assembly.add('tif_blade_root_flap_ext', Float(1., iotype='in', desc='Technology improvement factor flap extreme', group='Rotor')) assembly.add('tif_blade_root_edge_ext', Float(1., iotype='in', desc='Technology improvement factor edge extreme', group='Rotor')) assembly.add('tif_blade_root_flap_fat', Float(1., iotype='in', desc='Technology improvement factor flap LEQ', group='Rotor')) assembly.add('sc_frac_flap', Float(iotype='in', desc='spar cap fraction of chord', group='Rotor')) assembly.add('sc_frac_edge', Float(iotype='in', desc='spar cap fraction of thickness', group='Rotor')) assembly.add('safety_factor_blade', Float(iotype='in', desc='Blade loads safety factor', group='Rotor')) assembly.add('stress_limit_extreme_blade', Float(iotype='in', units='MPa', desc='Blade ultimate strength', group='Rotor')) assembly.add('stress_limit_fatigue_blade', Float(iotype='in', units='MPa', desc='Blade fatigue strength', group='Rotor')) assembly.add('AddWeightFactorBlade', Float(iotype='in', desc='Additional weight factor for blade shell', group='Rotor')) assembly.add('blade_material_density', Float(iotype='in', units='kg/m**3', desc='Average density of blade materials', group='Rotor')) assembly.add('blade_mass', Float(iotype = 'out', units = 'kg', desc = 'Blade mass')) # assembly.add('mean_wsp', Float(iotype = 'in', units = 'm/s', desc = 'mean wind speed', group='Aero')) # [m/s] assembly.add('air_density', Float(iotype = 'in', units = 'kg/m**3', desc = 'density of air', group='Plant_AEP')) # [kg / m^3] assembly.add('max_Cp', Float(iotype = 'in', desc = 'max CP', group='Aero')) assembly.add('gearloss_const', Float(iotype = 'in', desc = 'Gear loss constant', group='Drivetrain')) assembly.add('gearloss_var', Float(iotype = 'in', desc = 'Gear loss variable', group='Drivetrain')) assembly.add('genloss', Float(iotype = 'in', desc = 'Generator loss', group='Drivetrain')) assembly.add('convloss', Float(iotype = 'in', desc = 'Converter loss', group='Drivetrain')) # Outputs assembly.add('rated_wind_speed', Float(units = 'm / s', iotype='out', desc='wind speed for rated power')) assembly.add('ideal_power_curve', Array(iotype='out', units='kW', desc='total power before losses and turbulence')) assembly.add('power_curve', Array(iotype='out', units='kW', desc='total power including losses and turbulence')) assembly.add('wind_curve', Array(iotype='out', units='m/s', desc='wind curve associated with power curve')) assembly.add('aep', Float(iotype = 'out', units='mW*h', desc='Annual energy production in mWh')) assembly.add('total_aep', Float(iotype = 'out', units='mW*h', desc='AEP for total years of production')) # END SEAM Variables ---------------------- # Add SEAM components and connections assembly.add('loads', SEAMLoads()) assembly.add('tower_design', SEAMTower(21)) assembly.add('blade_design', SEAMBladeStructure()) assembly.add('aep_calc', SEAM_PowerCurve()) assembly.driver.workflow.add(['loads', 'tower_design', 'blade_design', 'aep_calc']) assembly.connect('loads.tower_bottom_moment_max', 'tower_design.tower_bottom_moment_max') assembly.connect('loads.tower_bottom_moment_leq', 'tower_design.tower_bottom_moment_leq') assembly.connect('loads.blade_root_flap_max', 'blade_design.blade_root_flap_max') assembly.connect('loads.blade_root_edge_max', 'blade_design.blade_root_edge_max') assembly.connect('loads.blade_root_flap_leq', 'blade_design.blade_root_flap_leq') assembly.connect('loads.blade_root_edge_leq', 'blade_design.blade_root_edge_leq') connect_io(assembly, assembly.aep_calc) connect_io(assembly, assembly.loads) connect_io(assembly, assembly.tower_design) connect_io(assembly, assembly.blade_design) # End SEAM add components and connections ------------- if with_new_nacelle: assembly.add('hub',HubSE()) assembly.add('hubSystem',Hub_System_Adder_drive()) if with_3pt_drive: assembly.add('nacelle', Drive3pt()) else: assembly.add('nacelle', Drive4pt()) else: assembly.add('nacelle', DriveWPACT()) assembly.add('hub', HubWPACT()) assembly.driver.workflow.add(['hub', 'nacelle']) if with_new_nacelle: assembly.driver.workflow.add(['hubSystem']) # connections to hub and hub system assembly.connect('blade_design.blade_mass', 'hub.blade_mass') assembly.connect('loads.blade_root_flap_max', 'hub.rotor_bending_moment') assembly.connect('rotor_diameter', ['hub.rotor_diameter']) assembly.connect('blade_design.blade_root_diameter', 'hub.blade_root_diameter') assembly.add('blade_number',Int(3,iotype='in',desc='number of blades', group='Aero')) assembly.connect('blade_number', 'hub.blade_number') if with_new_nacelle: assembly.connect('rated_power','hub.machine_rating') assembly.connect('rotor_diameter', ['hubSystem.rotor_diameter']) assembly.connect('nacelle.MB1_location','hubSystem.MB1_location') # TODO: bearing locations assembly.connect('nacelle.L_rb','hubSystem.L_rb') assembly.add('rotor_tilt', Float(5.0, iotype='in', desc='rotor tilt', group='Rotor')) assembly.connect('rotor_tilt','hubSystem.shaft_angle') assembly.connect('hub.hub_diameter','hubSystem.hub_diameter') assembly.connect('hub.hub_thickness','hubSystem.hub_thickness') assembly.connect('hub.hub_mass','hubSystem.hub_mass') assembly.connect('hub.spinner_mass','hubSystem.spinner_mass') assembly.connect('hub.pitch_system_mass','hubSystem.pitch_system_mass') # connections to nacelle #TODO: fatigue option variables assembly.connect('rotor_diameter', 'nacelle.rotor_diameter') assembly.connect('1.5 * aep_calc.rated_torque', 'nacelle.rotor_torque') assembly.connect('loads.max_thrust', 'nacelle.rotor_thrust') assembly.connect('aep_calc.rated_speed', 'nacelle.rotor_speed') assembly.connect('rated_power', 'nacelle.machine_rating') assembly.add('generator_speed',Float(1173.7,iotype='in',units='rpm',desc='speed of generator', group='Drivetrain')) # - should be in nacelle assembly.connect('generator_speed/aep_calc.rated_speed', 'nacelle.gear_ratio') assembly.connect('tower_top_diameter', 'nacelle.tower_top_diameter') assembly.connect('blade_number * blade_design.blade_mass + hub.hub_system_mass', 'nacelle.rotor_mass') # assuming not already in rotor force / moments # variable connections for new nacelle if with_new_nacelle: assembly.connect('blade_number','nacelle.blade_number') assembly.connect('rotor_tilt','nacelle.shaft_angle') assembly.connect('333.3 * rated_power / 1000.0','nacelle.shrink_disc_mass') assembly.connect('blade_design.blade_root_diameter','nacelle.blade_root_diameter') #moments - ignoring for now (nacelle will use internal defaults) #assembly.connect('rotor.Mxyz_0','moments.b1') #assembly.connect('rotor.Mxyz_120','moments.b2') #assembly.connect('rotor.Mxyz_240','moments.b3') #assembly.connect('rotor.Pitch','moments.pitch_angle') #assembly.connect('rotor.TotalCone','moments.cone_angle') assembly.connect('1.5 * aep_calc.rated_torque','nacelle.rotor_bending_moment_x') #accounted for in ratedConditions.Q #assembly.connect('moments.My','nacelle.rotor_bending_moment_y') #assembly.connect('moments.Mz','nacelle.rotor_bending_moment_z') #forces - ignoring for now (nacelle will use internal defaults) #assembly.connect('rotor.Fxyz_0','forces.b1') #assembly.connect('rotor.Fxyz_120','forces.b2') #assembly.connect('rotor.Fxyz_240','forces.b3') #assembly.connect('rotor.Pitch','forces.pitch_angle') #assembly.connect('rotor.TotalCone','forces.cone_angle') assembly.connect('loads.max_thrust','nacelle.rotor_force_x') #assembly.connect('forces.Fy','nacelle.rotor_force_y') #assembly.connect('forces.Fz','nacelle.rotor_force_z') class Turbine_SE_SEAM(Assembly): def configure(self): configure_turbine(self) if __name__ == '__main__': turbine = Turbine_SE_SEAM() #=========== SEAM inputs turbine.AddWeightFactorBlade = 1.2 turbine.blade_material_density = 2100.0 turbine.tower_bottom_diameter = 6. turbine.tower_top_diameter = 3.78 turbine.blade_edge_dynload_factor_ext = 2.5 turbine.blade_edge_dynload_factor_fat = 0.75 turbine.F = 0.777 turbine.MaxChordrR = 0.2 turbine.project_lifetime = 20.0 turbine.lifetime_cycles = 10000000.0 turbine.blade_sections = 21 turbine.PMtarget_tower = 1.0 turbine.PMtarget_blades = 1.0 turbine.safety_factor_blade = 1.1 turbine.safety_factor_tower = 1.5 turbine.stress_limit_extreme_tower = 235.0 turbine.stress_limit_fatigue_tower = 14.885 turbine.stress_limit_extreme_blade = 200.0 turbine.stress_limit_fatigue_blade = 27.0 turbine.tif_blade_root_flap_ext = 1.0 turbine.tif_blade_root_flap_fat = 1.0 turbine.tif_blade_root_edge_ext = 1.0 turbine.weibull_C = 11.0 turbine.weibull_k = 2.0 turbine.wohler_exponent_blade_flap = 10.0 turbine.wohler_exponent_tower = 4.0 turbine.dLoad_dU_factor_flap = 0.9 turbine.dLoad_dU_factor_tower = 0.8 turbine.hub_height = 90.0 turbine.max_tipspeed = 80.0 turbine.n_wsp = 26 turbine.min_wsp = 0.0 turbine.max_wsp = 25.0 turbine.nSigma4fatFlap = 1.2 turbine.nSigma4fatTower = 0.8 turbine.rated_power = 5000.0 turbine.rho_steel = 7800.0 turbine.rotor_diameter = 126.0 turbine.sc_frac_edge = 0.8 turbine.sc_frac_flap = 0.3 turbine.tsr = 8.0 turbine.air_density = 1.225 turbine.turbulence_int = 0.16 turbine.max_Cp = 0.49 turbine.gearloss_const = 0.01 # Fraction turbine.gearloss_var = 0.014 # Fraction turbine.genloss = 0.03 # Fraction turbine.convloss = 0.03 # Fraction #============== # === nacelle ====== turbine.blade_number = 3 # turbine level that must be added for SEAM turbine.rotor_tilt = 5.0 # turbine level that must be added for SEAM turbine.generator_speed = 1173.7 turbine.nacelle.L_ms = 1.0 # (Float, m): main shaft length downwind of main bearing in low-speed shaft turbine.nacelle.L_mb = 2.5 # (Float, m): main shaft length in low-speed shaft turbine.nacelle.h0_front = 1.7 # (Float, m): height of Ibeam in bedplate front turbine.nacelle.h0_rear = 1.35 # (Float, m): height of Ibeam in bedplate rear turbine.nacelle.drivetrain_design = 'geared' turbine.nacelle.crane = True # (Bool): flag for presence of crane turbine.nacelle.bevel = 0 # (Int): Flag for the presence of a bevel stage - 1 if present, 0 if not turbine.nacelle.gear_configuration = 'eep' # (Str): tring that represents the configuration of the gearbox (stage number and types) turbine.nacelle.Np = [3, 3, 1] # (Array): number of planets in each stage turbine.nacelle.ratio_type = 'optimal' # (Str): optimal or empirical stage ratios turbine.nacelle.shaft_type = 'normal' # (Str): normal or short shaft length #turbine.nacelle.shaft_angle = 5.0 # (Float, deg): Angle of the LSS inclindation with respect to the horizontal turbine.nacelle.shaft_ratio = 0.10 # (Float): Ratio of inner diameter to outer diameter. Leave zero for solid LSS turbine.nacelle.carrier_mass = 8000.0 # estimated for 5 MW turbine.nacelle.mb1Type = 'CARB' # (Str): Main bearing type: CARB, TRB or SRB turbine.nacelle.mb2Type = 'SRB' # (Str): Second bearing type: CARB, TRB or SRB turbine.nacelle.yaw_motors_number = 8.0 # (Float): number of yaw motors turbine.nacelle.uptower_transformer = True turbine.nacelle.flange_length = 0.5 #m turbine.nacelle.gearbox_cm = 0.1 turbine.nacelle.hss_length = 1.5 turbine.nacelle.overhang = 5.0 #TODO - should come from turbine configuration level turbine.nacelle.check_fatigue = 0 #0 if no fatigue check, 1 if parameterized fatigue check, 2 if known loads inputs # ================= # === run === turbine.run() print 'mass rotor blades (kg) =', turbine.blade_number * turbine.blade_design.blade_mass print 'mass hub system (kg) =', turbine.hubSystem.hub_system_mass print 'mass nacelle (kg) =', turbine.nacelle.nacelle_mass print 'mass tower (kg) =', turbine.tower_design.tower_mass # =================
18,966
6,798
"""add run_type Revision ID: 5dd2ba8222b1 Revises: 079a74c15e8b Create Date: 2021-07-22 23:53:04.043651 """ from alembic import op import sqlalchemy as sa from sqlalchemy.dialects import postgresql # revision identifiers, used by Alembic. revision = '5dd2ba8222b1' down_revision = '079a74c15e8b' branch_labels = None depends_on = None def upgrade(): op.add_column('experiment_runs', sa.Column('run_type', sa.Text(), nullable=True), schema='triage_metadata') op.execute("UPDATE triage_metadata.experiment_runs SET run_type='experiment' WHERE run_type IS NULL") op.alter_column('experiment_runs', 'experiment_hash', nullable=True, new_column_name='run_hash', schema='triage_metadata') op.drop_constraint('experiment_runs_experiment_hash_fkey', 'experiment_runs', type_='foreignkey', schema='triage_metadata') op.execute("ALTER TABLE triage_metadata.experiment_runs RENAME TO triage_runs") op.create_table('retrain', sa.Column('retrain_hash', sa.Text(), nullable=False), sa.Column('config', postgresql.JSONB(astext_type=sa.Text()), nullable=True), sa.Column('prediction_date', sa.DateTime(), nullable=True), sa.PrimaryKeyConstraint('retrain_hash'), schema='triage_metadata', ) op.alter_column('models', 'built_in_experiment_run', nullable=False, new_column_name='built_in_triage_run', schema='triage_metadata') op.execute("CREATE TABLE triage_metadata.deprecated_models_built_by_experiment AS SELECT model_id, model_hash, built_by_experiment FROM triage_metadata.models") op.drop_column('models', 'built_by_experiment', schema='triage_metadata') op.create_table('retrain_models', sa.Column('retrain_hash', sa.String(), nullable=False), sa.Column('model_hash', sa.String(), nullable=False), sa.ForeignKeyConstraint(['retrain_hash'], ['triage_metadata.retrain.retrain_hash'], ), sa.PrimaryKeyConstraint('retrain_hash', 'model_hash'), schema='triage_metadata' ) def downgrade(): op.execute("ALTER TABLE triage_metadata.triage_runs RENAME TO experiment_runs") op.drop_column('experiment_runs', 'run_type', schema='triage_metadata') op.alter_column('experiment_runs', 'run_hash', nullable=True, new_column_name='experiment_hash', schema='triage_metadata') op.create_foreign_key('experiment_runs_experiment_hash_fkey', 'experiment_runs', 'experiments', ['experiment_hash'], ['experiment_hash'], source_schema='triage_metadata', referent_schema='triage_metadata') op.drop_table('retrain_models', schema='triage_metadata') op.drop_table('retrain', schema='triage_metadata') op.add_column('models', sa.Column('built_by_experiment', sa.Text(), nullable=True), schema='triage_metadata') op.alter_column('models', 'built_in_triage_run', nullable=False, new_column_name='built_in_experiment_run', schema='triage_metadata')
2,880
990
from .config import add_panopticfcn_config from .panoptic_seg import PanopticFCN from .build_solver import build_lr_scheduler
126
43
from http.client import HTTPResponse from django.shortcuts import render from django.http import HttpResponse # Create your views here. # def index(request): # return HttpResponse("Hello World!") def index(request): return render(request, "hello/index.html") def brian(request): return HttpResponse("Hello Brian") def david(request): return HttpResponse("Hello David") # def greet(request, name): # return HttpResponse(f"Hello, {name.capitalize()}!") def greet(request, name): return render(request, "hello/greet.html", { "name": name.capitalize() })
592
177
## # Copyright (c) 2010-2017 Apple Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ## """ Tests for L{txdav.common.datastore.upgrade.migrate}. """ from twext.enterprise.adbapi2 import Pickle from twext.enterprise.dal.syntax import Delete from twext.python.filepath import CachingFilePath from txweb2.http_headers import MimeType from twisted.internet.defer import inlineCallbacks, Deferred, returnValue from twisted.internet.protocol import Protocol from twisted.protocols.amp import AMP, Command, String from twisted.python.modules import getModule from twisted.python.reflect import qual, namedAny from twisted.trial.unittest import TestCase from twistedcaldav import customxml, caldavxml from twistedcaldav.config import config from twistedcaldav.ical import Component from txdav.base.propertystore.base import PropertyName from txdav.caldav.datastore.test.common import CommonTests from txdav.carddav.datastore.test.common import CommonTests as ABCommonTests from txdav.common.datastore.file import CommonDataStore from txdav.common.datastore.sql_tables import schema from txdav.common.datastore.test.util import SQLStoreBuilder from txdav.common.datastore.test.util import ( populateCalendarsFrom, StubNotifierFactory, resetCalendarMD5s, populateAddressBooksFrom, resetAddressBookMD5s, deriveValue, withSpecialValue, CommonCommonTests ) from txdav.common.datastore.upgrade.migrate import UpgradeToDatabaseStep, \ StoreSpawnerService, swapAMP from txdav.xml import element import copy class CreateStore(Command): """ Create a store in a subprocess. """ arguments = [('delegateTo', String())] class PickleConfig(Command): """ Unpickle some configuration in a subprocess. """ arguments = [('delegateTo', String()), ('config', Pickle())] class StoreCreator(AMP): """ Helper protocol. """ @CreateStore.responder def createStore(self, delegateTo): """ Create a store and pass it to the named delegate class. """ swapAMP(self, namedAny(delegateTo)(SQLStoreBuilder.childStore())) return {} @PickleConfig.responder def pickleConfig(self, config, delegateTo): # from twistedcaldav.config import config as globalConfig # globalConfig._data = config._data swapAMP(self, namedAny(delegateTo)(config)) return {} class StubSpawner(StoreSpawnerService): """ Stub spawner service which populates the store forcibly. """ def __init__(self, config=None): super(StubSpawner, self).__init__() self.config = config @inlineCallbacks def spawnWithStore(self, here, there): """ 'here' and 'there' are the helper protocols 'there' will expect to be created with an instance of a store. """ master = yield self.spawn(AMP(), StoreCreator) yield master.callRemote(CreateStore, delegateTo=qual(there)) returnValue(swapAMP(master, here)) @inlineCallbacks def spawnWithConfig(self, config, here, there): """ Similar to spawnWithStore except the child process gets a configuration object instead. """ master = yield self.spawn(AMP(), StoreCreator) subcfg = copy.deepcopy(self.config) del subcfg._postUpdateHooks[:] yield master.callRemote(PickleConfig, config=subcfg, delegateTo=qual(there)) returnValue(swapAMP(master, here)) class HomeMigrationTests(CommonCommonTests, TestCase): """ Tests for L{UpgradeToDatabaseStep}. """ av1 = Component.fromString("""BEGIN:VCALENDAR VERSION:2.0 CALSCALE:GREGORIAN PRODID:-//calendarserver.org//Zonal//EN BEGIN:VAVAILABILITY ORGANIZER:mailto:user01@example.com UID:1@example.com DTSTAMP:20061005T133225Z DTEND:20140101T000000Z BEGIN:AVAILABLE UID:1-1@example.com DTSTAMP:20061005T133225Z SUMMARY:Monday to Friday from 9:00 to 17:00 DTSTART:20130101T090000Z DTEND:20130101T170000Z RRULE:FREQ=WEEKLY;BYDAY=MO,TU,WE,TH,FR END:AVAILABLE END:VAVAILABILITY END:VCALENDAR """) @inlineCallbacks def setUp(self): """ Set up two stores to migrate between. """ yield super(HomeMigrationTests, self).setUp() yield self.buildStoreAndDirectory( extraUids=( u"home1", u"home2", u"home3", u"home_defaults", u"home_no_splits", u"home_splits", u"home_splits_shared", ) ) self.sqlStore = self.store # Add some files to the file store. self.filesPath = CachingFilePath(self.mktemp()) self.filesPath.createDirectory() fileStore = self.fileStore = CommonDataStore( self.filesPath, {"push": StubNotifierFactory()}, self.directory, True, True ) self.upgrader = UpgradeToDatabaseStep(self.fileStore, self.sqlStore) requirements = CommonTests.requirements extras = deriveValue(self, "extraRequirements", lambda t: {}) requirements = self.mergeRequirements(requirements, extras) yield populateCalendarsFrom(requirements, fileStore) md5s = CommonTests.md5s yield resetCalendarMD5s(md5s, fileStore) self.filesPath.child("calendars").child( "__uids__").child("ho").child("me").child("home1").child( ".some-extra-data").setContent("some extra data") requirements = ABCommonTests.requirements yield populateAddressBooksFrom(requirements, fileStore) md5s = ABCommonTests.md5s yield resetAddressBookMD5s(md5s, fileStore) self.filesPath.child("addressbooks").child( "__uids__").child("ho").child("me").child("home1").child( ".some-extra-data").setContent("some extra data") # Add some properties we want to check get migrated over txn = self.fileStore.newTransaction() home = yield txn.calendarHomeWithUID("home_defaults") cal = yield home.calendarWithName("calendar_1") props = cal.properties() props[PropertyName.fromElement(caldavxml.SupportedCalendarComponentSet)] = caldavxml.SupportedCalendarComponentSet( caldavxml.CalendarComponent(name="VEVENT"), caldavxml.CalendarComponent(name="VTODO"), ) props[PropertyName.fromElement(element.ResourceType)] = element.ResourceType( element.Collection(), caldavxml.Calendar(), ) props[PropertyName.fromElement(customxml.GETCTag)] = customxml.GETCTag.fromString("foobar") inbox = yield home.calendarWithName("inbox") props = inbox.properties() props[PropertyName.fromElement(customxml.CalendarAvailability)] = customxml.CalendarAvailability.fromString(str(self.av1)) props[PropertyName.fromElement(caldavxml.ScheduleDefaultCalendarURL)] = caldavxml.ScheduleDefaultCalendarURL( element.HRef.fromString("/calendars/__uids__/home_defaults/calendar_1"), ) yield txn.commit() def mergeRequirements(self, a, b): """ Merge two requirements dictionaries together, modifying C{a} and returning it. @param a: Some requirements, in the format of L{CommonTests.requirements}. @type a: C{dict} @param b: Some additional requirements, to be merged into C{a}. @type b: C{dict} @return: C{a} @rtype: C{dict} """ for homeUID in b: homereq = a.setdefault(homeUID, {}) homeExtras = b[homeUID] for calendarUID in homeExtras: calreq = homereq.setdefault(calendarUID, {}) calendarExtras = homeExtras[calendarUID] calreq.update(calendarExtras) return a @withSpecialValue( "extraRequirements", { "home1": { "calendar_1": { "bogus.ics": ( getModule("twistedcaldav").filePath.sibling("zoneinfo") .child("EST.ics").getContent(), CommonTests.metadata1 ) } } } ) @inlineCallbacks def test_unknownTypeNotMigrated(self): """ The only types of calendar objects that should get migrated are VEVENTs and VTODOs. Other component types, such as free-standing VTIMEZONEs, don't have a UID and can't be stored properly in the database, so they should not be migrated. """ yield self.upgrader.stepWithResult(None) txn = self.sqlStore.newTransaction() self.addCleanup(txn.commit) self.assertIdentical( None, (yield (yield (yield ( yield txn.calendarHomeWithUID("home1") ).calendarWithName("calendar_1")) ).calendarObjectWithName("bogus.ics")) ) @inlineCallbacks def test_upgradeCalendarHomes(self): """ L{UpgradeToDatabaseService.startService} will do the upgrade, then start its dependent service by adding it to its service hierarchy. """ # Create a fake directory in the same place as a home, but with a non-existent uid fake_dir = self.filesPath.child("calendars").child("__uids__").child("ho").child("me").child("foobar") fake_dir.makedirs() # Create a fake file in the same place as a home,with a name that matches the hash uid prefix fake_file = self.filesPath.child("calendars").child("__uids__").child("ho").child("me").child("home_file") fake_file.setContent("") yield self.upgrader.stepWithResult(None) txn = self.sqlStore.newTransaction() self.addCleanup(txn.commit) for uid in CommonTests.requirements: if CommonTests.requirements[uid] is not None: self.assertNotIdentical( None, (yield txn.calendarHomeWithUID(uid)) ) # Successfully migrated calendar homes are deleted self.assertFalse(self.filesPath.child("calendars").child( "__uids__").child("ho").child("me").child("home1").exists()) # Want metadata preserved home = (yield txn.calendarHomeWithUID("home1")) calendar = (yield home.calendarWithName("calendar_1")) for name, metadata, md5 in ( ("1.ics", CommonTests.metadata1, CommonTests.md5Values[0]), ("2.ics", CommonTests.metadata2, CommonTests.md5Values[1]), ("3.ics", CommonTests.metadata3, CommonTests.md5Values[2]), ): object = (yield calendar.calendarObjectWithName(name)) self.assertEquals(object.getMetadata(), metadata) self.assertEquals(object.md5(), md5) @withSpecialValue( "extraRequirements", { "nonexistent": { "calendar_1": { } } } ) @inlineCallbacks def test_upgradeCalendarHomesMissingDirectoryRecord(self): """ Test an upgrade where a directory record is missing for a home; the original home directory will remain on disk. """ yield self.upgrader.stepWithResult(None) txn = self.sqlStore.newTransaction() self.addCleanup(txn.commit) for uid in CommonTests.requirements: if CommonTests.requirements[uid] is not None: self.assertNotIdentical( None, (yield txn.calendarHomeWithUID(uid)) ) self.assertIdentical(None, (yield txn.calendarHomeWithUID(u"nonexistent"))) # Skipped calendar homes are not deleted self.assertTrue(self.filesPath.child("calendars").child( "__uids__").child("no").child("ne").child("nonexistent").exists()) @inlineCallbacks def test_upgradeExistingHome(self): """ L{UpgradeToDatabaseService.startService} will skip migrating existing homes. """ startTxn = self.sqlStore.newTransaction("populate empty sample") yield startTxn.calendarHomeWithUID("home1", create=True) yield startTxn.commit() yield self.upgrader.stepWithResult(None) vrfyTxn = self.sqlStore.newTransaction("verify sample still empty") self.addCleanup(vrfyTxn.commit) home = yield vrfyTxn.calendarHomeWithUID("home1") # The default calendar is still there. self.assertNotIdentical(None, (yield home.calendarWithName("calendar"))) # The migrated calendar isn't. self.assertIdentical(None, (yield home.calendarWithName("calendar_1"))) @inlineCallbacks def test_upgradeAttachments(self): """ L{UpgradeToDatabaseService.startService} upgrades calendar attachments as well. """ # Need to tweak config and settings to setup dropbox to work self.patch(config, "EnableDropBox", True) self.patch(config, "EnableManagedAttachments", False) self.sqlStore.enableManagedAttachments = False txn = self.sqlStore.newTransaction() cs = schema.CALENDARSERVER yield Delete( From=cs, Where=cs.NAME == "MANAGED-ATTACHMENTS" ).on(txn) yield txn.commit() txn = self.fileStore.newTransaction() committed = [] def maybeCommit(): if not committed: committed.append(True) return txn.commit() self.addCleanup(maybeCommit) @inlineCallbacks def getSampleObj(): home = (yield txn.calendarHomeWithUID("home1")) calendar = (yield home.calendarWithName("calendar_1")) object = (yield calendar.calendarObjectWithName("1.ics")) returnValue(object) inObject = yield getSampleObj() someAttachmentName = "some-attachment" someAttachmentType = MimeType.fromString("application/x-custom-type") attachment = yield inObject.createAttachmentWithName( someAttachmentName, ) transport = attachment.store(someAttachmentType) someAttachmentData = "Here is some data for your attachment, enjoy." transport.write(someAttachmentData) yield transport.loseConnection() yield maybeCommit() yield self.upgrader.stepWithResult(None) committed = [] txn = self.sqlStore.newTransaction() outObject = yield getSampleObj() outAttachment = yield outObject.attachmentWithName(someAttachmentName) allDone = Deferred() class SimpleProto(Protocol): data = '' def dataReceived(self, data): self.data += data def connectionLost(self, reason): allDone.callback(self.data) self.assertEquals(outAttachment.contentType(), someAttachmentType) outAttachment.retrieve(SimpleProto()) allData = yield allDone self.assertEquals(allData, someAttachmentData) @inlineCallbacks def test_upgradeAddressBookHomes(self): """ L{UpgradeToDatabaseService.startService} will do the upgrade, then start its dependent service by adding it to its service hierarchy. """ yield self.upgrader.stepWithResult(None) txn = self.sqlStore.newTransaction() self.addCleanup(txn.commit) for uid in ABCommonTests.requirements: if ABCommonTests.requirements[uid] is not None: self.assertNotIdentical( None, (yield txn.addressbookHomeWithUID(uid)) ) # Successfully migrated addressbook homes are deleted self.assertFalse(self.filesPath.child("addressbooks").child( "__uids__").child("ho").child("me").child("home1").exists()) # Want metadata preserved home = (yield txn.addressbookHomeWithUID("home1")) adbk = (yield home.addressbookWithName("addressbook")) for name, md5 in ( ("1.vcf", ABCommonTests.md5Values[0]), ("2.vcf", ABCommonTests.md5Values[1]), ("3.vcf", ABCommonTests.md5Values[2]), ): object = (yield adbk.addressbookObjectWithName(name)) self.assertEquals(object.md5(), md5) @inlineCallbacks def test_upgradeProperties(self): """ L{UpgradeToDatabaseService.startService} will do the upgrade, then start its dependent service by adding it to its service hierarchy. """ yield self.upgrader.stepWithResult(None) txn = self.sqlStore.newTransaction() self.addCleanup(txn.commit) # Want metadata preserved home = (yield txn.calendarHomeWithUID("home_defaults")) cal = (yield home.calendarWithName("calendar_1")) inbox = (yield home.calendarWithName("inbox")) # Supported components self.assertEqual(cal.getSupportedComponents(), "VEVENT") self.assertTrue(cal.properties().get(PropertyName.fromElement(caldavxml.SupportedCalendarComponentSet)) is None) # Resource type removed self.assertTrue(cal.properties().get(PropertyName.fromElement(element.ResourceType)) is None) # Ctag removed self.assertTrue(cal.properties().get(PropertyName.fromElement(customxml.GETCTag)) is None) # Availability self.assertEquals(str(home.getAvailability()), str(self.av1)) self.assertTrue(inbox.properties().get(PropertyName.fromElement(customxml.CalendarAvailability)) is None) # Default calendar self.assertTrue(home.isDefaultCalendar(cal)) self.assertTrue(inbox.properties().get(PropertyName.fromElement(caldavxml.ScheduleDefaultCalendarURL)) is None) def test_fileStoreFromPath(self): """ Verify that fileStoreFromPath() will return a CommonDataStore if the given path contains either "calendars" or "addressbooks" sub-directories. Otherwise it returns None """ # No child directories docRootPath = CachingFilePath(self.mktemp()) docRootPath.createDirectory() step = UpgradeToDatabaseStep.fileStoreFromPath(docRootPath) self.assertEquals(step, None) # "calendars" child directory exists childPath = docRootPath.child("calendars") childPath.createDirectory() step = UpgradeToDatabaseStep.fileStoreFromPath(docRootPath) self.assertTrue(isinstance(step, CommonDataStore)) childPath.remove() # "addressbooks" child directory exists childPath = docRootPath.child("addressbooks") childPath.createDirectory() step = UpgradeToDatabaseStep.fileStoreFromPath(docRootPath) self.assertTrue(isinstance(step, CommonDataStore)) childPath.remove()
19,426
5,590
from netapp.netapp_object import NetAppObject class AliasesInfo(NetAppObject): """ A list of WWPNs and their aliases generated according to the input - alias, WWPN or nothing. """ _vserver = None @property def vserver(self): """ Vserver containing the alias """ return self._vserver @vserver.setter def vserver(self, val): if val != None: self.validate('vserver', val) self._vserver = val _aliases_wwpn = None @property def aliases_wwpn(self): """ The FCP WWPN for which the alias is given """ return self._aliases_wwpn @aliases_wwpn.setter def aliases_wwpn(self, val): if val != None: self.validate('aliases_wwpn', val) self._aliases_wwpn = val _aliases_alias = None @property def aliases_alias(self): """ The 32-character alias for a given FCP WWPN """ return self._aliases_alias @aliases_alias.setter def aliases_alias(self, val): if val != None: self.validate('aliases_alias', val) self._aliases_alias = val @staticmethod def get_api_name(): return "aliases-info" @staticmethod def get_desired_attrs(): return [ 'vserver', 'aliases-wwpn', 'aliases-alias', ] def describe_properties(self): return { 'vserver': { 'class': basestring, 'is_list': False, 'required': 'optional' }, 'aliases_wwpn': { 'class': basestring, 'is_list': False, 'required': 'required' }, 'aliases_alias': { 'class': basestring, 'is_list': False, 'required': 'required' }, }
1,770
547
import numpy as np import code from imagernn.utils import merge_init_structs, initw, accumNpDicts from imagernn.lstm_generator import LSTMGenerator from imagernn.rnn_generator import RNNGenerator def decodeGenerator(generator): if generator == 'lstm': return LSTMGenerator if generator == 'rnn': return RNNGenerator else: raise Exception('generator %s is not yet supported' % (base_generator_str,)) class GenericBatchGenerator: """ Base batch generator class. This class is aware of the fact that we are generating sentences from images. """ @staticmethod def init(params, misc): # inputs image_encoding_size = params.get('image_encoding_size', 128) word_encoding_size = params.get('word_encoding_size', 128) hidden_size = params.get('hidden_size', 128) generator = params.get('generator', 'lstm') vocabulary_size = len(misc['wordtoix']) output_size = len(misc['ixtoword']) # these should match though image_size = 4096 # size of CNN vectors hardcoded here if generator == 'lstm': assert image_encoding_size == word_encoding_size, 'this implementation does not support different sizes for these parameters' # initialize the encoder models model = {} model['We'] = initw(image_size, image_encoding_size) # image encoder model['be'] = np.zeros((1,image_encoding_size)) model['Ws'] = initw(vocabulary_size, word_encoding_size) # word encoder update = ['We', 'be', 'Ws'] regularize = ['We', 'Ws'] init_struct = { 'model' : model, 'update' : update, 'regularize' : regularize} # descend into the specific Generator and initialize it # why generate again?? Belter, 20170510 Generator = decodeGenerator(generator) generator_init_struct = Generator.init(word_encoding_size, hidden_size, output_size) merge_init_structs(init_struct, generator_init_struct) return init_struct @staticmethod def forward(batch, model, params, misc, predict_mode = False): """ iterates over items in the batch and calls generators on them """ # we do the encoding here across all images/words in batch in single matrix # multiplies to gain efficiency. The RNNs are then called individually # in for loop on per-image-sentence pair and all they are concerned about is # taking single matrix of vectors and doing the forward/backward pass without # knowing anything about images, sentences or anything of that sort. # encode all images # concatenate as rows. If N is number of image-sentence pairs, # F will be N x image_size F = np.row_stack(x['image']['feat'] for x in batch) We = model['We'] be = model['be'] Xe = F.dot(We) + be # Xe becomes N x image_encoding_size # decode the generator we wish to use generator_str = params.get('generator', 'lstm') Generator = decodeGenerator(generator_str) # encode all words in all sentences (which exist in our vocab) wordtoix = misc['wordtoix'] Ws = model['Ws'] gen_caches = [] Ys = [] # outputs for i,x in enumerate(batch): # take all words in this sentence and pluck out their word vectors # from Ws. Then arrange them in a single matrix Xs # Note that we are setting the start token as first vector # and then all the words afterwards. And start token is the first row of Ws ix = [0] + [ wordtoix[w] for w in x['sentence']['tokens'] if w in wordtoix ] Xs = np.row_stack( [Ws[j, :] for j in ix] ) Xi = Xe[i,:] # forward prop through the RNN gen_Y, gen_cache = Generator.forward(Xi, Xs, model, params, predict_mode = predict_mode) gen_caches.append((ix, gen_cache)) Ys.append(gen_Y) # back up information we need for efficient backprop cache = {} if not predict_mode: # ok we need cache as well because we'll do backward pass cache['gen_caches'] = gen_caches cache['Xe'] = Xe cache['Ws_shape'] = Ws.shape cache['F'] = F cache['generator_str'] = generator_str return Ys, cache @staticmethod def backward(dY, cache): Xe = cache['Xe'] generator_str = cache['generator_str'] dWs = np.zeros(cache['Ws_shape']) gen_caches = cache['gen_caches'] F = cache['F'] dXe = np.zeros(Xe.shape) Generator = decodeGenerator(generator_str) # backprop each item in the batch grads = {} for i in xrange(len(gen_caches)): ix, gen_cache = gen_caches[i] # unpack local_grads = Generator.backward(dY[i], gen_cache) dXs = local_grads['dXs'] # intercept the gradients wrt Xi and Xs del local_grads['dXs'] dXi = local_grads['dXi'] del local_grads['dXi'] accumNpDicts(grads, local_grads) # add up the gradients wrt model parameters # now backprop from dXs to the image vector and word vectors dXe[i,:] += dXi # image vector for n,j in enumerate(ix): # and now all the other words dWs[j,:] += dXs[n,:] # finally backprop into the image encoder dWe = F.transpose().dot(dXe) dbe = np.sum(dXe, axis=0, keepdims = True) accumNpDicts(grads, { 'We':dWe, 'be':dbe, 'Ws':dWs }) return grads @staticmethod def predict(batch, model, params, **kwparams): """ some code duplication here with forward pass, but I think we want the freedom in future """ F = np.row_stack(x['image']['feat'] for x in batch) We = model['We'] be = model['be'] Xe = F.dot(We) + be # Xe becomes N x image_encoding_size generator_str = params['generator'] Generator = decodeGenerator(generator_str) Ys = [] for i,x in enumerate(batch): gen_Y = Generator.predict(Xe[i, :], model, model['Ws'], params, **kwparams) Ys.append(gen_Y) return Ys
5,745
1,878