content
stringlengths
1
1.05M
input_ids
listlengths
1
883k
ratio_char_token
float64
1
22.9
token_count
int64
1
883k
from setuptools import setup setup(name='neural_networks_tfw1', version='0.1', description='Implementing Neural Networks with Tensorflow', packages=['neural_networks_tfw1'], author='Dilay Fidan Ercelik', author_email='dilay.ercelik@gmail.com', zip_safe=False)
[ 6738, 900, 37623, 10141, 1330, 9058, 198, 198, 40406, 7, 3672, 11639, 710, 1523, 62, 3262, 5225, 62, 27110, 86, 16, 3256, 198, 220, 220, 220, 220, 220, 2196, 11639, 15, 13, 16, 3256, 198, 220, 220, 220, 220, 220, 6764, 11639, 3546, ...
2.395161
124
# coding=utf8 import re import json.decoder from collections import namedtuple from json.decoder import JSONDecoder from json.scanner import py_make_scanner from json.decoder import py_scanstring def errmsg_inv(e): assert isinstance(e, ValueError) message = e.message idx = message.rindex(':') errmsg, left = message[:idx], message[idx + 1:] numbers = re.compile(r'\d+').findall(left) parser = e.__dict__.get("parser", "") result = { "parsers": e.__dict__.get("parsers", []), "error": errors.get_decode_error(parser, errmsg), "lineno": int(numbers[0]), "colno": int(numbers[1]), } if len(numbers) == 3: result["pos"] = int(numbers[2]) if len(numbers) > 3: result["endlineno"] = int(numbers[2]) result["endcolno"] = int(numbers[3]) result["pos"] = int(numbers[4]) result["end"] = int(numbers[5]) return result def record_parser_name(parser): return new_parser def make_decoder(): json.decoder.scanstring = record_parser_name(py_scanstring) decoder = JSONDecoder() decoder.parse_object = record_parser_name(decoder.parse_object) decoder.parse_array = record_parser_name(decoder.parse_array) decoder.parse_string = record_parser_name(py_scanstring) decoder.parse_object = record_parser_name(decoder.parse_object) decoder.scan_once = py_make_scanner(decoder) return decoder decoder = make_decoder() DecodeResult = namedtuple('DecodeResult', ['success', 'exception', 'err_info'])
[ 2, 19617, 28, 40477, 23, 198, 198, 11748, 302, 198, 11748, 33918, 13, 12501, 12342, 198, 6738, 17268, 1330, 3706, 83, 29291, 198, 6738, 33918, 13, 12501, 12342, 1330, 19449, 10707, 12342, 198, 6738, 33918, 13, 35836, 1008, 1330, 12972, ...
2.479167
624
#!/usr/bin/python ##################################################################### # Name : capaSystem.py # Description : Read system data and update db for web display # Environment : Tested under Raspberry Pi Rasbian Jessie Summer 17 # Author : Steve Osteen sosteen@gmail.com ###################################################################### import MySQLdb import sys import time from subprocess import Popen, PIPE import logging import logging.handlers log = logging.getLogger('CapaTimeLapseLog') log.setLevel(logging.DEBUG) # prod: logging.ERROR handler = logging.handlers.SysLogHandler(address='/dev/log') formatter = logging.Formatter('%(name)-12s %(levelname)-8s %(message)s') handler.setFormatter(formatter) log.addHandler(handler) if __name__ == '__main__': main()
[ 2, 48443, 14629, 14, 8800, 14, 29412, 198, 29113, 29113, 4242, 2, 198, 2, 6530, 220, 220, 220, 220, 220, 220, 220, 220, 220, 220, 220, 220, 1058, 1451, 64, 11964, 13, 9078, 198, 2, 12489, 220, 220, 220, 220, 220, 1058, 4149, 1080,...
3.309237
249
from django.template.loader import render_to_string from django.core.mail import send_mail from django.conf import settings from django.contrib.auth import get_user_model from django.db.models import Q from wagtail.wagtailcore.models import PageRevision, GroupPagePermission from wagtail.wagtailusers.models import UserProfile # The following will check to see if we can import task from celery - # if not then we definitely haven't installed it try: from celery.decorators import task NO_CELERY = False except: NO_CELERY = True # However, we could have installed celery for other projects. So we will also # check if we have defined the BROKER_URL setting. If not then definitely we # haven't configured it. if NO_CELERY or not hasattr(settings, 'BROKER_URL'): # So if we enter here we will define a different "task" decorator that # just returns the original function and sets its delay attribute to # point to the original function: This way, the send_notification # function will be actually called instead of the the # send_notification.delay()
[ 6738, 42625, 14208, 13, 28243, 13, 29356, 1330, 8543, 62, 1462, 62, 8841, 198, 6738, 42625, 14208, 13, 7295, 13, 4529, 1330, 3758, 62, 4529, 198, 6738, 42625, 14208, 13, 10414, 1330, 6460, 198, 6738, 42625, 14208, 13, 3642, 822, 13, 1...
3.37963
324
#Author: Shu Shi #emaiL: shushi@research.att.com from toscalib.tosca_workbook import ToscaWorkBook from toscalib.tosca_builder import ToscaBuilder import getopt, sys, json, logging if __name__ == "__main__": main()
[ 2, 13838, 25, 32344, 16380, 198, 2, 368, 1872, 43, 25, 427, 17731, 31, 34033, 13, 1078, 13, 785, 198, 198, 6738, 284, 1416, 282, 571, 13, 83, 418, 6888, 62, 1818, 2070, 1330, 40195, 6888, 12468, 10482, 198, 6738, 284, 1416, 282, 5...
2.597701
87
from django.test import TestCase from rest_framework.test import APIClient
[ 6738, 42625, 14208, 13, 9288, 1330, 6208, 20448, 198, 6738, 1334, 62, 30604, 13, 9288, 1330, 3486, 2149, 75, 1153, 628 ]
3.619048
21
import pandas as pd import dateutil from lusidtools.lpt import lpt from lusidtools.lpt import lse from lusidtools.lpt import stdargs from .either import Either import re import urllib.parse rexp = re.compile(r".*page=([^=']{10,}).*") TOOLNAME = "scopes" TOOLTIP = "List scopes" # Standalone tool
[ 11748, 19798, 292, 355, 279, 67, 198, 11748, 3128, 22602, 198, 6738, 300, 385, 312, 31391, 13, 75, 457, 1330, 300, 457, 198, 6738, 300, 385, 312, 31391, 13, 75, 457, 1330, 300, 325, 198, 6738, 300, 385, 312, 31391, 13, 75, 457, 13...
2.581197
117
#!/usr/bin/python # -*- coding: utf-8 -*- """ QQBot -- A conversation robot base on Tencent's SmartQQ Website -- https://github.com/pandolia/qqbot/ Author -- pandolia@yeah.net """ import sys, os p = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) if p not in sys.path: sys.path.insert(0, p) import sys, subprocess, time from apscheduler.schedulers.background import BackgroundScheduler from collections import defaultdict from qqbot.qconf import QConf from qqbot.utf8logger import INFO, CRITICAL, ERROR, WARN from qqbot.qsession import QLogin, RequestError from qqbot.exitcode import RESTART, POLL_ERROR, FRESH_RESTART from qqbot.common import StartDaemonThread, Import from qqbot.qterm import QTermServer from qqbot.mainloop import MainLoop, Put from qqbot.groupmanager import GroupManager for name, slots in QQBot.slotsTable.items(): setattr(QQBot, name, wrap(slots)) QQBotSlot = QQBot.AddSlot QQBotSched = QQBot.AddSched if __name__ == '__main__': bot = QQBot() bot.Login(user='hcj') gl = bot.List('group') ml = bot.List(gl[0]) m = ml[0]
[ 2, 48443, 14629, 14, 8800, 14, 29412, 198, 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 198, 37811, 198, 48, 48, 20630, 220, 220, 1377, 317, 5273, 9379, 2779, 319, 9368, 1087, 338, 10880, 48, 48, 198, 33420, 1377...
2.628916
415
from redgrease import GearsBuilder gb = GearsBuilder() gb.run()
[ 6738, 2266, 16694, 589, 1330, 46680, 32875, 198, 198, 22296, 796, 46680, 32875, 3419, 198, 22296, 13, 5143, 3419, 198 ]
3.25
20
from copy import copy from itertools import groupby import unittest from datetime import datetime, timedelta from typing import List from activitywatch.base import Watcher, Activity, Logger from activitywatch.settings import Settings from activitywatch.utils import floor_datetime, ceil_datetime from activitywatch.filters.split import split_by_interval, overlaps from activitywatch.filters.chunk import chunk_by_tags HOUR = timedelta(hours=1)
[ 6738, 4866, 1330, 4866, 198, 6738, 340, 861, 10141, 1330, 1448, 1525, 198, 11748, 555, 715, 395, 198, 6738, 4818, 8079, 1330, 4818, 8079, 11, 28805, 12514, 198, 6738, 19720, 1330, 7343, 198, 198, 6738, 3842, 8340, 13, 8692, 1330, 12242,...
3.682927
123
voc_it = [ ['a', 'noun', 'c'], ['a', 'preposition', 'a'], ['abbagliante', 'pres_part', 'c'], ['abbagliante', 'adjective', 'c'], ['abbagliante', 'noun', 'c'], ['abbaiare', 'verb', 'c'], ['abbandonare', 'verb', 'a'], ['abbandonato', 'past_part', 'b'], ['abbandonato', 'adjective', 'b'], ['abbandono', 'noun', 'b'], ['abbassare', 'verb', 'a'], ['abbasso', 'adverb', 'c'], ['abbasso', 'exclamation', 'c'], ['abbastanza', 'adverb', 'a'], ['abbattere', 'verb', 'b'], ['abbeverare', 'verb', 'c'], ['abbigliamento', 'noun', 'b'], ['abbinare', 'verb', 'b'], ['abbonamento', 'noun', 'b'], ['abbonare', 'verb', 'c'], ['abbondante', 'pres_part', 'b'], ['abbondante', 'adjective', 'b'], ['abbondare', 'verb', 'c'], ['abbottonare', 'verb', 'c'], ['abbracciare', 'verb', 'a'], ['abbraccio', 'noun', 'b'], ['abbreviare', 'verb', 'c'], ['abbronzare', 'verb', 'c'], ['abete', 'noun', 'c'], ['abile', 'adjective', 'b'], ['abilit', 'noun', 'b'], ['abisso', 'noun', 'b'], ['abitante', 'pres_part', 'b'], ['abitante', 'adjective', 'b'], ['abitante', 'noun', 'b'], ['abitare', 'verb', 'a'], ['abitare', 'noun', 'a'], ['abitazione', 'noun', 'b'], ['abito', 'noun', 'a'], ['abituale', 'adjective', 'b'], ['abituare', 'verb', 'a'], ['abitudine', 'noun', 'a'], ['abolire', 'verb', 'b'], ['abortire', 'verb', 'c'], ['aborto', 'noun', 'c'], ['abruzzese', 'adjective', 'c'], ['abruzzese', 'noun', 'c'], ['abusare', 'verb', 'c'], ['abuso', 'noun', 'b'], ['acca', 'noun', 'c'], ['accademia', 'noun', 'b'], ['accademico', 'adjective', 'b'], ['accademico', 'noun', 'b'], ['accadere', 'verb', 'a'], ['accampamento', 'noun', 'c'], ['accanto', 'adverb', 'a'], ['accappatoio', 'noun', 'c'], ['accarezzare', 'verb', 'b'], ['accattone', 'noun', 'c'], ['accavallare', 'verb', 'c'], ['accecare', 'verb', 'c'], ['accedere', 'verb', 'b'], ['accelerare', 'verb', 'b'], ['acceleratore', 'adjective', 'c'], ['acceleratore', 'noun', 'c'], ['accelerazione', 'noun', 'b'], ['accendere', 'verb', 'a'], ['accendino', 'noun', 'c'], ['accennare', 'verb', 'b'], ['accenno', 'noun', 'c'], ['accentare', 'verb', 'c'], ['accertamento', 'noun', 'b'], ['accertare', 'verb', 'b'], ['acceso', 'past_part', 'b'], ['acceso', 'adjective', 'b'], ['accesso', 'noun', 'a'], ['accessorio', 'adjective', 'b'], ['accessorio', 'noun', 'b'], ['accetta', 'noun', 'c'], ['accettabile', 'adjective', 'b'], ['accettare', 'verb', 'a'], ['acchiappare', 'verb', 'c'], ['acciacco', 'noun', 'c'], ['acciaio', 'noun', 'b'], ['accidente', 'noun', 'b'], ['acciuga', 'noun', 'c'], ['accogliente', 'pres_part', 'c'], ['accogliente', 'adjective', 'c'], ['accoglienza', 'noun', 'b'], ['accogliere', 'verb', 'a'], ['accoltellare', 'verb', 'c'], ['accomodare', 'verb', 'b'], ['accompagnare', 'verb', 'a'], ['acconsentire', 'verb', 'c'], ['accontentare', 'verb', 'b'], ['accorciare', 'verb', 'c'], ['accordare', 'verb', 'b'], ['accordo', 'noun', 'a'], ['accorgersi', 'verb', 'a'], ['accorrere', 'verb', 'c'], ['accostare', 'verb', 'b'], ['accudire', 'verb', 'c'], ['accumulare', 'verb', 'b'], ['accumulatore', 'adjective', 'c'], ['accumulatore', 'noun', 'c'], ['accurato', 'past_part', 'b'], ['accurato', 'adjective', 'b'], ['accusa', 'noun', 'a'], ['accusare', 'verb', 'a'], ['accento', 'noun', 'b'], ['acerbo', 'adjective', 'c'], ['aceto', 'noun', 'c'], ['acido', 'adjective', 'b'], ['acido', 'noun', 'b'], ['acqua', 'noun', 'a'], ['acquarello', 'noun', 'c'], ['acquario', 'noun', 'c'], ['acquasanta', 'noun', 'c'], ['acquisire', 'verb', 'b'], ['acquisizione', 'noun', 'b'], ['acquistare', 'verb', 'a'], ['acquisto', 'noun', 'a'], ['acquolina', 'noun', 'c'], ['acrobata', 'noun', 'c'], ['acuto', 'adjective', 'b'], ['acuto', 'noun', 'b'], ['adattare', 'verb', 'b'], ['adattatore', 'noun', 'c'], ['adatto', 'adjective', 'a'], ['addetto', 'past_part', 'b'], ['addetto', 'adjective', 'b'], ['addetto', 'noun', 'b'], ['addio', 'exclamation', 'b'], ['addio', 'noun', 'b'], ['addirittura', 'adverb', 'a'], ['addizione', 'noun', 'c'], ['addobbare', 'verb', 'c'], ['addolcire', 'verb', 'c'], ['addomesticare', 'verb', 'c'], ['addormentarsi', 'verb', 'b'], ['addormentato', 'past_part', 'c'], ['addormentato', 'adjective', 'c'], ['addossare', 'verb', 'a'], ['addosso', 'adverb', 'c'], ['addosso', 'exclamation', 'c'], ['addrizzare', 'verb', 'c'], ['adeguare', 'verb', 'b'], ['adeguato', 'past_part', 'b'], ['adeguato', 'adjective', 'b'], ['adeguato', 'noun', 'b'], ['aderente', 'pres_part', 'c'], ['aderente', 'adjective', 'c'], ['aderente', 'noun', 'c'], ['aderire', 'verb', 'b'], ['adesione', 'noun', 'b'], ['adesso', 'adverb', 'a'], ['adolescente', 'adjective', 'a'], ['adolescente', 'noun', 'a'], ['adolescenza', 'noun', 'b'], ['adoperare', 'verb', 'b'], ['adorare', 'verb', 'a'], ['adottare', 'verb', 'a'], ['adozione', 'noun', 'b'], ['adriatico', 'adjective', 'c'], ['adulto', 'adjective', 'a'], ['adulto', 'noun', 'a'], ['aereo', 'adjective', 'a'], ['aereo', 'noun', 'a'], ['aereo', 'noun', 'b'], ['aeroplano', 'noun', 'c'], ['aeroporto', 'noun', 'b'], ['afa', 'noun', 'c'], ['affacciare', 'verb', 'b'], ['affamare', 'verb', 'c'], ['affamato', 'past_part', 'c'], ['affamato', 'adjective', 'c'], ['affamato', 'noun', 'c'], ['affannarsi', 'verb', 'c'], ['affannato', 'past_part', 'c'], ['affannato', 'adjective', 'c'], ['affanno', 'noun', 'c'], ['affare', 'noun', 'a'], ['affascinante', 'pres_part', 'b'], ['affascinante', 'adjective', 'b'], ['affascinare', 'verb', 'b'], ['affaticare', 'verb', 'c'], ['affatto', 'adverb', 'a'], ['affermare', 'verb', 'a'], ['affermazione', 'noun', 'b'], ['afferrare', 'verb', 'b'], ['affettare', 'verb', 'c'], ['affettato', 'past_part', 'c'], ['affettato', 'adjective', 'c'], ['affettato', 'noun', 'c'], ['affetto', 'noun', 'b'], ['affetto', 'adjective', 'b'], ['affettuoso', 'adjective', 'b'], ['affezionato', 'past_part', 'c'], ['affezionato', 'adjective', 'c'], ['affiancare', 'verb', 'b'], ['affidamento', 'noun', 'b'], ['affidare', 'verb', 'a'], ['affilato', 'past_part', 'c'], ['affilato', 'adjective', 'c'], ['affinch', 'conjunction', 'b'], ['affittare', 'verb', 'b'], ['affitto', 'noun', 'b'], ['affogare', 'verb', 'c'], ['affollare', 'verb', 'c'], ['affondare', 'verb', 'b'], ['affresco', 'noun', 'b'], ['affrontare', 'verb', 'a'], ['affumicare', 'verb', 'c'], ['africano', 'adjective', 'b'], ['africano', 'noun', 'b'], ['agenda', 'noun', 'b'], ['agente', 'pres_part', 'a'], ['agente', 'adjective', 'a'], ['agente', 'noun', 'a'], ['agenzia', 'noun', 'a'], ['agganciare', 'verb', 'b'], ['aggettivo', 'noun', 'b'], ['aggiornamento', 'noun', 'b'], ['aggiornare', 'verb', 'b'], ['aggirare', 'verb', 'b'], ['aggiungere', 'verb', 'a'], ['aggiustare', 'verb', 'b'], ['aggrapparsi', 'verb', 'b'], ['aggravare', 'verb', 'c'], ['aggredire', 'verb', 'b'], ['aggressione', 'noun', 'b'], ['aggressivo', 'adjective', 'b'], ['agiato', 'past_part', 'c'], ['agiato', 'adjective', 'c'], ['agile', 'adjective', 'c'], ['agio', 'noun', 'b'], ['agire', 'verb', 'a'], ['agitare', 'verb', 'b'], ['agitazione', 'noun', 'b'], ['aglio', 'noun', 'c'], ['agnello', 'noun', 'b'], ['ago', 'noun', 'b'], ['agonia', 'noun', 'c'], ['agosto', 'noun', 'a'], ['agricolo', 'adjective', 'b'], ['agricoltore', 'noun', 'c'], ['agricoltura', 'noun', 'b'], ['agrume', 'noun', 'c'], ['aguzzare', 'verb', 'c'], ['aguzzo', 'adjective', 'c'], ['aiuola', 'noun', 'c'], ['aiutare', 'verb', 'a'], ['aiuto', 'noun', 'a'], ['aiuto', 'exclamation', 'a'], ['ala', 'noun', 'a'], ['alba', 'noun', 'a'], ['albanese', 'adjective', 'b'], ['albanese', 'noun', 'b'], ['albergo', 'noun', 'a'], ['albero', 'noun', 'a'], ['albicocca', 'noun', 'c'], ['albicocca', 'adjective', 'c'], ['album', 'noun', 'a'], ['alcol', 'noun', 'b'], ['alcuno', 'adjective', 'a'], ['alcuno', 'pronoun', 'a'], ['alfabeto', 'noun', 'c'], ['alga', 'noun', 'c'], ['algerino', 'adjective', 'c'], ['algerino', 'noun', 'c'], ['alieno', 'adjective', 'b'], ['alieno', 'noun', 'b'], ['alimentare', 'adjective', 'b'], ['alimentare', 'noun', 'b'], ['alimentare', 'verb', 'b'], ['alimentari', 'noun', 'c'], ['alimentazione', 'noun', 'b'], ['alimento', 'noun', 'b'], ['alito', 'noun', 'c'], ['allacciare', 'verb', 'c'], ['allagare', 'verb', 'c'], ['allargare', 'verb', 'b'], ['allarmare', 'verb', 'c'], ['allarme', 'noun', 'b'], ['allattare', 'verb', 'c'], ['alleanza', 'noun', 'b'], ['allearsi', 'verb', 'c'], ['alleato', 'past_part', 'b'], ['alleato', 'adjective', 'b'], ['alleato', 'noun', 'b'], ['allegato', 'past_part', 'b'], ['allegato', 'adjective', 'b'], ['allegato', 'noun', 'b'], ['alleggerire', 'verb', 'c'], ['allegria', 'noun', 'b'], ['allegro', 'adjective', 'b'], ['allegro', 'adverb', 'b'], ['allegro', 'noun', 'b'], ['allenamento', 'noun', 'b'], ['allenare', 'verb', 'b'], ['allenatore', 'adjective', 'b'], ['allenatore', 'noun', 'b'], ['allentare', 'verb', 'c'], ['allergia', 'noun', 'c'], ['allevare', 'verb', 'b'], ['allievo', 'noun', 'b'], ['allineare', 'verb', 'c'], ['alloggio', 'noun', 'b'], ['allontanare', 'verb', 'a'], ['allora', 'adverb', 'a'], ['allora', 'conjunction', 'a'], ['alluce', 'noun', 'c'], ['alludere', 'verb', 'b'], ['alluminio', 'noun', 'c'], ['allungare', 'verb', 'a'], ['alluvione', 'noun', 'c'], ['almeno', 'adverb', 'a'], ['alquanto', 'adjective', 'b'], ['alquanto', 'pronoun', 'b'], ['alquanto', 'adverb', 'b'], ['altalena', 'noun', 'c'], ['altamente', 'adverb', 'b'], ['altare', 'noun', 'b'], ['alterare', 'verb', 'b'], ['alternare', 'verb', 'b'], ['alternativa', 'noun', 'b'], ['alternativo', 'adjective', 'b'], ['alterno', 'adjective', 'c'], ['altezza', 'noun', 'a'], ['alto', 'adjective', 'a'], ['alto', 'noun', 'a'], ['alto', 'adverb', 'a'], ['altoatesino', 'adjective', 'c'], ['altoatesino', 'noun', 'c'], ['altopiano', 'noun', 'c'], ['altrettanto', 'adjective', 'a'], ['altrettanto', 'pronoun', 'a'], ['altrettanto', 'adverb', 'a'], ['altrimenti', 'adverb', 'a'], ['altro', 'adjective', 'a'], ['altro', 'pronoun', 'a'], ['altro', 'adverb', 'a'], ['altrove', 'adverb', 'b'], ['altrui', 'adjective', 'b'], ['altrui', 'pronoun', 'b'], ['alunno', 'noun', 'b'], ['alveare', 'noun', 'c'], ['alzare', 'verb', 'a'], ['amante', 'pres_part', 'a'], ['amante', 'adjective', 'a'], ['amante', 'noun', 'a'], ['amare', 'verb', 'a'], ['amaro', 'adjective', 'b'], ['amaro', 'noun', 'b'], ['amato', 'past_part', 'b'], ['amato', 'adjective', 'b'], ['amato', 'noun', 'b'], ['ambasciata', 'noun', 'c'], ['ambientale', 'adjective', 'a'], ['ambientare', 'verb', 'b'], ['ambiente', 'noun', 'a'], ['ambiente', 'adjective', 'a'], ['ambito', 'noun', 'a'], ['ambizione', 'noun', 'b'], ['ambulanza', 'noun', 'b'], ['americano', 'adjective', 'a'], ['americano', 'noun', 'a'], ['amicizia', 'noun', 'a'], ['amico', 'adjective', 'a'], ['amico', 'noun', 'a'], ['ammaccare', 'verb', 'c'], ['ammalarsi', 'verb', 'b'], ['ammalato', 'past_part', 'c'], ['ammalato', 'adjective', 'c'], ['ammalato', 'noun', 'c'], ['ammanettare', 'verb', 'c'], ['ammassare', 'verb', 'c'], ['ammasso', 'noun', 'c'], ['ammazzare', 'verb', 'a'], ['ammettere', 'verb', 'a'], ['amministrativo', 'adjective', 'b'], ['amministrativo', 'noun', 'b'], ['amministratore', 'noun', 'b'], ['amministrazione', 'noun', 'a'], ['ammirare', 'verb', 'b'], ['ammissione', 'noun', 'b'], ['ammobiliare', 'verb', 'c'], ['ammoniaca', 'noun', 'c'], ['ammorbidente', 'pres_part', 'c'], ['ammorbidente', 'adjective', 'c'], ['ammorbidente', 'noun', 'c'], ['ammucchiare', 'verb', 'c'], ['ammuffire', 'verb', 'c'], ['amore', 'noun', 'a'], ['amoroso', 'adjective', 'b'], ['amoroso', 'noun', 'b'], ['ampiamente', 'adverb', 'b'], ['ampio', 'adjective', 'a'], ['ampio', 'noun', 'a'], ['amplificatore', 'adjective', 'c'], ['amplificatore', 'noun', 'c'], ['analcolico', 'adjective', 'c'], ['analcolico', 'noun', 'c'], ['analfabeta', 'adjective', 'c'], ['analfabeta', 'noun', 'c'], ['analisi', 'noun', 'a'], ['analitico', 'adjective', 'b'], ['analizzare', 'verb', 'a'], ['analogo', 'adjective', 'b'], ['ananas', 'noun', 'c'], ['anarchico', 'adjective', 'c'], ['anarchico', 'noun', 'c'], ['anatra', 'noun', 'c'], ['anche', 'conjunction', 'a'], ['anche', 'adverb', 'a'], ['anconetano', 'adjective', 'c'], ['anconetano', 'noun', 'c'], ['ancora', 'adverb', 'a'], ['ancora', 'conjunction', 'a'], ['ancorare', 'verb', 'b'], ['andamento', 'noun', 'b'], ['andare', 'verb', 'a'], ['andata', 'noun', 'c'], ['anello', 'noun', 'a'], ['angelo', 'noun', 'a'], ['angolare', 'adjective', 'b'], ['angolare', 'noun', 'b'], ['angolo', 'noun', 'a'], ['angoscia', 'noun', 'b'], ['anima', 'noun', 'a'], ['animale', 'noun', 'a'], ['animale', 'adjective', 'b'], ['animare', 'verb', 'a'], ['animato', 'past_part', 'b'], ['animato', 'adjective', 'b'], ['animato', 'adverb', 'b'], ['animo', 'noun', 'b'], ['animo', 'exclamation', 'b'], ['annacquare', 'verb', 'c'], ['annaffiare', 'verb', 'c'], ['annebbiare', 'verb', 'c'], ['anniversario', 'noun', 'b'], ['anniversario', 'adjective', 'b'], ['anno', 'noun', 'a'], ['annodare', 'verb', 'c'], ['annoiare', 'verb', 'b'], ['annotare', 'verb', 'b'], ['annuale', 'adjective', 'b'], ['annuale', 'noun', 'b'], ['annuire', 'verb', 'b'], ['annullare', 'verb', 'b'], ['annunciare', 'verb', 'a'], ['annuncio', 'noun', 'b'], ['annusare', 'verb', 'c'], ['anonimo', 'adjective', 'b'], ['anonimo', 'noun', 'b'], ['ansia', 'noun', 'a'], ['ansioso', 'adjective', 'b'], ['ansioso', 'noun', 'b'], ['antartico', 'adjective', 'c'], ['antartico', 'noun', 'c'], ['antenna', 'noun', 'b'], ['anteprima', 'noun', 'b'], ['anteriore', 'adjective', 'b'], ['anticalcare', 'adjective', 'c'], ['antichit', 'noun', 'c'], ['anticipare', 'verb', 'b'], ['anticipo', 'noun', 'b'], ['antico', 'adjective', 'a'], ['antico', 'noun', 'a'], ['antipasto', 'noun', 'c'], ['antirughe', 'adjective', 'c'], ['antirughe', 'noun', 'c'], ['antropologia', 'noun', 'b'], ['anulare', 'adjective', 'c'], ['anulare', 'noun', 'c'], ['anzi', 'adverb', 'a'], ['anzi', 'preposition', 'a'], ['anziano', 'adjective', 'a'], ['anziano', 'noun', 'a'], ['anzich', 'conjunction', 'b'], ['aostano', 'adjective', 'c'], ['aostano', 'noun', 'c'], ['ape', 'noun', 'b'], ['aperitivo', 'noun', 'c'], ['aperitivo', 'adjective', 'c'], ['aperto', 'past_part', 'a'], ['aperto', 'adjective', 'a'], ['aperto', 'noun', 'a'], ['aperto', 'adverb', 'a'], ['apertura', 'noun', 'a'], ['aspettativa', 'noun', 'b'], ['apostolo', 'noun', 'c'], ['appalto', 'noun', 'b'], ['appannare', 'verb', 'c'], ['apparato', 'noun', 'b'], ['apparecchiare', 'verb', 'c'], ['apparecchiatura', 'noun', 'c'], ['apparecchio', 'noun', 'b'], ['apparente', 'pres_part', 'b'], ['apparente', 'adjective', 'b'], ['apparentemente', 'adverb', 'b'], ['apparenza', 'noun', 'b'], ['apparire', 'verb', 'a'], ['apparizione', 'noun', 'b'], ['appartamento', 'noun', 'a'], ['appartenenza', 'noun', 'b'], ['appartenere', 'verb', 'a'], ['appassionare', 'verb', 'b'], ['appassionarsi', 'verb', 'c'], ['appassionato', 'past_part', 'b'], ['appassionato', 'adjective', 'b'], ['appassionato', 'noun', 'b'], ['appello', 'noun', 'b'], ['appena', 'adverb', 'a'], ['appena', 'conjunction', 'a'], ['appendere', 'verb', 'b'], ['appendicite', 'noun', 'c'], ['appenninico', 'adjective', 'c'], ['appeso', 'past_part', 'c'], ['appeso', 'adjective', 'c'], ['appeso', 'noun', 'c'], ['appiccicare', 'verb', 'c'], ['appiglio', 'noun', 'c'], ['applauso', 'noun', 'b'], ['applicare', 'verb', 'a'], ['applicazione', 'noun', 'b'], ['appoggiare', 'verb', 'a'], ['appoggio', 'noun', 'b'], ['apposito', 'adjective', 'b'], ['apposta', 'adverb', 'b'], ['apposta', 'adjective', 'b'], ['apprendere', 'verb', 'b'], ['apprendimento', 'noun', 'b'], ['apprendista', 'noun', 'c'], ['apprezzare', 'verb', 'a'], ['approccio', 'noun', 'b'], ['approfittare', 'verb', 'b'], ['approfondimento', 'noun', 'b'], ['approfondire', 'verb', 'b'], ['approvare', 'verb', 'b'], ['approvazione', 'noun', 'b'], ['appuntamento', 'noun', 'a'], ['appuntire', 'verb', 'c'], ['appunto', 'noun', 'b'], ['appunto', 'adverb', 'a'], ['aprile', 'noun', 'a'], ['aprire', 'verb', 'a'], ['apriscatole', 'noun', 'c'], ['aquila', 'noun', 'c'], ['aquilano', 'adjective', 'c'], ['aquilano', 'noun', 'c'], ['aquilone', 'noun', 'c'], ['arabo', 'adjective', 'a'], ['arabo', 'noun', 'a'], ['arachide', 'noun', 'c'], ['aragosta', 'noun', 'c'], ['aranciata', 'noun', 'c'], ['arancio', 'noun', 'c'], ['arare', 'verb', 'c'], ['aratro', 'noun', 'c'], ['arbitro', 'noun', 'b'], ['archeologo', 'noun', 'c'], ['architettare', 'verb', 'b'], ['architetto', 'noun', 'b'], ['architettonico', 'adjective', 'b'], ['architettura', 'noun', 'b'], ['archiviare', 'verb', 'b'], ['archivio', 'noun', 'b'], ['arco', 'noun', 'a'], ['arcobaleno', 'noun', 'c'], ['area', 'noun', 'a'], ['argentino', 'adjective', 'b'], ['argentino', 'noun', 'b'], ['argento', 'noun', 'b'], ['argomentare', 'verb', 'b'], ['argomentazione', 'noun', 'b'], ['argomento', 'noun', 'a'], ['aria', 'noun', 'a'], ['aristocratico', 'adjective', 'c'], ['aristocratico', 'noun', 'c'], ['aritmetica', 'noun', 'c'], ['aritmetico', 'adjective', 'c'], ['aritmetico', 'noun', 'c'], ['arma', 'noun', 'a'], ['armadio', 'noun', 'b'], ['armamento', 'noun', 'c'], ['armare', 'verb', 'b'], ['armato', 'past_part', 'b'], ['armato', 'adjective', 'b'], ['armato', 'noun', 'b'], ['armonia', 'noun', 'b'], ['aroma', 'noun', 'c'], ['arrabbiarsi', 'verb', 'a'], ['arrampicarsi', 'verb', 'b'], ['arredamento', 'noun', 'b'], ['arredare', 'verb', 'c'], ['arrendersi', 'verb', 'b'], ['arrendersi', 'verb', 'c'], ['arrestare', 'verb', 'a'], ['arresto', 'noun', 'b'], ['arricchire', 'verb', 'b'], ['arrivare', 'verb', 'a'], ['arrivederci', 'exclamation', 'b'], ['arrivederci', 'noun', 'b'], ['arrivo', 'noun', 'a'], ['arrosto', 'noun', 'c'], ['arrosto', 'adjective', 'c'], ['arrosto', 'adverb', 'c'], ['arrugginire', 'verb', 'c'], ['arte', 'noun', 'a'], ['arteria', 'noun', 'b'], ['artico', 'adjective', 'c'], ['artico', 'noun', 'c'], ['articolare', 'verb', 'b'], ['articolare', 'noun', 'b'], ['articolazione', 'noun', 'b'], ['articolo', 'noun', 'a'], ['artificiale', 'adjective', 'b'], ['artigianale', 'adjective', 'c'], ['artigiano', 'noun', 'b'], ['artigiano', 'adjective', 'b'], ['artiglieria', 'noun', 'c'], ['artiglio', 'noun', 'c'], ['artista', 'noun', 'a'], ['artistico', 'adjective', 'a'], ['artistico', 'noun', 'a'], ['ascella', 'noun', 'c'], ['ascensore', 'noun', 'b'], ['ascesa', 'noun', 'b'], ['ascesso', 'noun', 'c'], ['ascia', 'noun', 'c'], ['asciugamano', 'noun', 'b'], ['asciugare', 'verb', 'b'], ['asciutto', 'adjective', 'b'], ['asciutto', 'noun', 'b'], ['ascoltare', 'verb', 'a'], ['ascolto', 'noun', 'b'], ['asfaltare', 'verb', 'c'], ['asfalto', 'noun', 'c'], ['asiatico', 'adjective', 'b'], ['asiatico', 'noun', 'b'], ['asilo', 'noun', 'b'], ['asino', 'noun', 'b'], ['asma', 'noun', 'c'], ['asparago', 'noun', 'c'], ['aspettare', 'verb', 'a'], ['aspetto', 'noun', 'a'], ['aspirapolvere', 'noun', 'c'], ['aspirare', 'verb', 'b'], ['aspirazione', 'noun', 'b'], ['aspro', 'adjective', 'b'], ['aspro', 'noun', 'b'], ['assaggiare', 'verb', 'b'], ['assaggio', 'noun', 'c'], ['assai', 'adverb', 'a'], ['assai', 'adjective', 'a'], ['assai', 'noun', 'a'], ['assalire', 'verb', 'c'], ['assaltare', 'verb', 'c'], ['assalto', 'noun', 'b'], ['assaporare', 'verb', 'c'], ['assassinare', 'verb', 'b'], ['assassinio', 'noun', 'c'], ['assassino', 'noun', 'b'], ['assassino', 'adjective', 'b'], ['asse', 'noun', 'b'], ['assediare', 'verb', 'c'], ['assegnare', 'verb', 'b'], ['assegno', 'noun', 'b'], ['assemblea', 'noun', 'b'], ['assente', 'adjective', 'b'], ['assente', 'noun', 'b'], ['assenza', 'noun', 'a'], ['assicurare', 'verb', 'a'], ['assicurazione', 'noun', 'b'], ['assieme', 'adverb', 'a'], ['assieme', 'noun', 'a'], ['assistente', 'pres_part', 'b'], ['assistente', 'adjective', 'b'], ['assistente', 'noun', 'b'], ['assistenza', 'noun', 'b'], ['assistere', 'verb', 'a'], ['associare', 'verb', 'b'], ['associazione', 'noun', 'a'], ['assolutamente', 'adverb', 'a'], ['assoluto', 'adjective', 'a'], ['assoluto', 'noun', 'a'], ['assoluzione', 'noun', 'c'], ['assolvere', 'verb', 'b'], ['assomigliare', 'verb', 'b'], ['assorbente', 'pres_part', 'c'], ['assorbente', 'adjective', 'c'], ['assorbente', 'noun', 'c'], ['assorbire', 'verb', 'b'], ['assordare', 'verb', 'c'], ['assumere', 'verb', 'a'], ['assunzione', 'noun', 'b'], ['assurdo', 'adjective', 'a'], ['assurdo', 'noun', 'a'], ['asta', 'noun', 'b'], ['astemio', 'adjective', 'c'], ['astemio', 'noun', 'c'], ['astratto', 'past_part', 'b'], ['astratto', 'adjective', 'b'], ['astratto', 'noun', 'b'], ['astronave', 'noun', 'c'], ['astuccio', 'noun', 'c'], ['astuto', 'adjective', 'c'], ['astuto', 'noun', 'c'], ['astuzia', 'noun', 'c'], ['ateniese', 'adjective', 'c'], ['ateniese', 'noun', 'c'], ['ateo', 'adjective', 'b'], ['ateo', 'noun', 'b'], ['atlantico', 'adjective', 'c'], ['atleta', 'noun', 'b'], ['atmosfera', 'noun', 'a'], ['atomica', 'noun', 'c'], ['atomico', 'adjective', 'b'], ['atomo', 'noun', 'b'], ['atrio', 'noun', 'c'], ['atroce', 'adjective', 'b'], ['attaccante', 'pres_part', 'c'], ['attaccante', 'adjective', 'c'], ['attaccante', 'noun', 'c'], ['attaccapanni', 'noun', 'c'], ['attaccare', 'verb', 'a'], ['attacco', 'noun', 'a'], ['atteggiamento', 'noun', 'a'], ['atteggiare', 'verb', 'c'], ['attendere', 'verb', 'a'], ['attenere', 'verb', 'b'], ['attentamente', 'adverb', 'b'], ['attentare', 'verb', 'c'], ['attentato', 'noun', 'b'], ['attento', 'adjective', 'a'], ['attenzione', 'noun', 'a'], ['atterraggio', 'noun', 'c'], ['atterrare', 'verb', 'b'], ['attesa', 'noun', 'a'], ['attestare', 'verb', 'b'], ['attimo', 'noun', 'a'], ['attingere', 'verb', 'b'], ['attirare', 'verb', 'b'], ['attivare', 'verb', 'b'], ['attivit', 'noun', 'a'], ['attivo', 'adjective', 'a'], ['attivo', 'noun', 'a'], ['atto', 'noun', 'a'], ['attore', 'noun', 'a'], ['attorno', 'adverb', 'a'], ['attrarre', 'verb', 'b'], ['attraversare', 'verb', 'a'], ['attraverso', 'preposition', 'a'], ['attraverso', 'adverb', 'a'], ['attrazione', 'noun', 'b'], ['attrezzare', 'verb', 'b'], ['attrezzatura', 'noun', 'b'], ['attrezzo', 'noun', 'b'], ['attribuire', 'verb', 'b'], ['attrice', 'noun', 'b'], ['attuale', 'adjective', 'a'], ['attualit', 'noun', 'b'], ['attualmente', 'adverb', 'b'], ['attuare', 'verb', 'b'], ['augurare', 'verb', 'b'], ['augurio', 'noun', 'b'], ['aula', 'noun', 'b'], ['aumentare', 'verb', 'a'], ['aumento', 'noun', 'a'], ['australiano', 'adjective', 'c'], ['australiano', 'noun', 'c'], ['austriaco', 'adjective', 'b'], ['austriaco', 'noun', 'b'], ['autentico', 'adjective', 'b'], ['autentico', 'noun', 'b'], ['autista', 'noun', 'b'], ['auto', 'noun', 'a'], ['autoambulanza', 'noun', 'c'], ['autobotte', 'noun', 'c'], ['autobus', 'noun', 'b'], ['autografo', 'adjective', 'c'], ['autografo', 'noun', 'c'], ['automaticamente', 'adverb', 'b'], ['automatico', 'adjective', 'b'], ['automatico', 'noun', 'b'], ['automobile', 'noun', 'b'], ['automobilista', 'noun', 'c'], ['autonomia', 'noun', 'b'], ['autonomo', 'adjective', 'b'], ['autonomo', 'noun', 'b'], ['autore', 'noun', 'a'], ['autorevole', 'adjective', 'c'], ['autorit', 'noun', 'a'], ['autorizzare', 'verb', 'a'], ['autoscontro', 'noun', 'c'], ['autoscuola', 'noun', 'c'], ['autostop', 'noun', 'c'], ['autostrada', 'noun', 'b'], ['autotreno', 'noun', 'c'], ['autunno', 'noun', 'b'], ['avambraccio', 'noun', 'c'], ['avanguardia', 'noun', 'b'], ['avanti', 'adverb', 'a'], ['avanti', 'adjective', 'a'], ['avanti', 'loc-comando', 'a'], ['avanti', 'preposition', 'a'], ['avanti', 'noun', 'a'], ['avanzare', 'verb', 'a'], ['avanzato', 'past_part', 'b'], ['avanzato', 'adjective', 'b'], ['avanzo', 'noun', 'c'], ['avarizia', 'noun', 'c'], ['avaro', 'adjective', 'c'], ['avaro', 'noun', 'c'], ['avena', 'noun', 'c'], ['avere', 'verb', 'a'], ['aviazione', 'noun', 'c'], ['avvantaggiare', 'verb', 'c'], ['avvelenare', 'verb', 'b'], ['avvelenato', 'past_part', 'c'], ['avvelenato', 'adjective', 'c'], ['avvenimento', 'noun', 'b'], ['avvenire', 'adjective', 'a'], ['avvenire', 'noun', 'a'], ['avventura', 'noun', 'a'], ['avverare', 'verb', 'c'], ['avversario', 'noun', 'b'], ['avvertire', 'verb', 'a'], ['avviamento', 'noun', 'c'], ['avviare', 'verb', 'a'], ['avvicinare', 'verb', 'a'], ['avvio', 'noun', 'b'], ['avvisare', 'verb', 'b'], ['avviso', 'noun', 'b'], ['avvitare', 'verb', 'c'], ['avvocato', 'noun', 'a'], ['avvolgere', 'verb', 'b'], ['azienda', 'noun', 'a'], ['aziendale', 'adjective', 'b'], ['azione', 'noun', 'a'], ['azione', 'noun', 'b'], ['azzardare', 'verb', 'b'], ['azzardo', 'noun', 'c'], ['azzurro', 'noun', 'a'], ['azzurro', 'adjective', 'a'], ['babbo', 'noun', 'b'], ['baby', 'noun', 'b'], ['baby', 'adjective', 'b'], ['babydoll', 'noun', 'c'], ['bacca', 'noun', 'c'], ['baccal', 'noun', 'c'], ['bacheca', 'noun', 'b'], ['baciare', 'verb', 'a'], ['bacinella', 'noun', 'c'], ['bacino', 'noun', 'b'], ['bacio', 'noun', 'a'], ['baco', 'noun', 'c'], ['badare', 'verb', 'b'], ['baffo', 'noun', 'b'], ['bagagliaio', 'noun', 'c'], ['bagaglio', 'noun', 'b'], ['bagnare', 'verb', 'b'], ['bagnato', 'past_part', 'b'], ['bagnato', 'adjective', 'b'], ['bagnato', 'noun', 'b'], ['bagno', 'noun', 'a'], ['bagnoschiuma', 'noun', 'c'], ['balcone', 'noun', 'b'], ['balena', 'noun', 'b'], ['balia', 'noun', 'b'], ['ballare', 'verb', 'a'], ['ballerina', 'noun', 'c'], ['ballerino', 'noun', 'c'], ['ballerino', 'adjective', 'c'], ['balletto', 'noun', 'c'], ['ballo', 'noun', 'b'], ['balsamo', 'noun', 'c'], ['bambina', 'noun', 'a'], ['bambinaia', 'noun', 'c'], ['bambino', 'noun', 'a'], ['bambino', 'adjective', 'a'], ['bambola', 'noun', 'b'], ['banale', 'adjective', 'b'], ['banana', 'noun', 'c'], ['banca', 'noun', 'a'], ['bancarella', 'noun', 'c'], ['bancario', 'adjective', 'b'], ['bancario', 'noun', 'b'], ['banco', 'noun', 'b'], ['bancone', 'noun', 'b'], ['band', 'noun', 'b'], ['banda', 'noun', 'b'], ['bandiera', 'noun', 'b'], ['bando', 'noun', 'b'], ['bar', 'noun', 'a'], ['bara', 'noun', 'b'], ['baracca', 'noun', 'c'], ['barba', 'noun', 'b'], ['barbabietola', 'noun', 'c'], ['barbaro', 'adjective', 'b'], ['barbaro', 'noun', 'b'], ['barca', 'noun', 'a'], ['barella', 'noun', 'c'], ['barese', 'adjective', 'c'], ['barese', 'noun', 'c'], ['barile', 'noun', 'c'], ['barista', 'noun', 'c'], ['barriera', 'noun', 'b'], ['basare', 'verb', 'a'], ['base', 'noun', 'a'], ['basetta', 'noun', 'c'], ['basilica', 'noun', 'b'], ['basilico', 'noun', 'c'], ['basket', 'noun', 'c'], ['basso', 'adjective', 'a'], ['basso', 'noun', 'a'], ['basso', 'adverb', 'a'], ['bastardo', 'adjective', 'b'], ['bastardo', 'noun', 'b'], ['bastare', 'verb', 'a'], ['bastonare', 'verb', 'c'], ['bastone', 'noun', 'b'], ['battaglia', 'noun', 'a'], ['battello', 'noun', 'c'], ['battere', 'verb', 'a'], ['battere', 'noun', 'a'], ['batteria', 'noun', 'b'], ['batterio', 'noun', 'b'], ['batticuore', 'noun', 'c'], ['battipanni', 'noun', 'c'], ['battito', 'noun', 'c'], ['battuta', 'noun', 'a'], ['batuffolo', 'noun', 'c'], ['baule', 'noun', 'c'], ['bava', 'noun', 'c'], ['bavaglio', 'noun', 'c'], ['beato', 'past_part', 'b'], ['beato', 'adjective', 'b'], ['beato', 'noun', 'b'], ['beccare', 'verb', 'b'], ['befana', 'noun', 'c'], ['beffa', 'noun', 'c'], ['beh', 'exclamation', 'a'], ['belare', 'verb', 'c'], ['belga', 'adjective', 'c'], ['belga', 'noun', 'c'], ['bella', 'noun', 'b'], ['bellezza', 'noun', 'a'], ['bello', 'adjective', 'a'], ['bello', 'noun', 'a'], ['bench', 'conjunction', 'b'], ['benda', 'noun', 'c'], ['bene', 'adverb', 'a'], ['bene', 'exclamation', 'a'], ['bene', 'noun', 'a'], ['benedetto', 'past_part', 'b'], ['benedetto', 'adjective', 'b'], ['benedetto', 'noun', 'b'], ['beneficenza', 'noun', 'c'], ['beneficio', 'noun', 'b'], ['benessere', 'noun', 'b'], ['benestante', 'adjective', 'c'], ['benestante', 'noun', 'c'], ['bens', 'conjunction', 'b'], ['bens', 'adverb', 'b'], ['benvenuto', 'adjective', 'b'], ['benvenuto', 'noun', 'b'], ['benzina', 'noun', 'b'], ['benzinaio', 'noun', 'c'], ['bere', 'verb', 'a'], ['bere', 'noun', 'a'], ['berlinese', 'adjective', 'c'], ['berlinese', 'noun', 'c'], ['berretto', 'noun', 'c'], ['bersaglio', 'noun', 'b'], ['besciamella', 'noun', 'c'], ['bestemmia', 'noun', 'c'], ['bestia', 'noun', 'b'], ['bestiale', 'adjective', 'c'], ['bevanda', 'noun', 'b'], ['bevitore', 'noun', 'c'], ['bevuta', 'noun', 'c'], ['bi', 'noun', 'c'], ['bianco', 'adjective', 'a'], ['bianco', 'noun', 'a'], ['bibbia', 'noun', 'b'], ['bibita', 'noun', 'c'], ['biblico', 'adjective', 'b'], ['biblico', 'noun', 'b'], ['bibliografia', 'noun', 'b'], ['biblioteca', 'noun', 'b'], ['bicchiere', 'noun', 'a'], ['bici', 'noun', 'b'], ['bicicletta', 'noun', 'b'], ['bid', 'noun', 'c'], ['bidello', 'noun', 'c'], ['biglia', 'noun', 'c'], ['biglietteria', 'noun', 'c'], ['biglietto', 'noun', 'a'], ['bikini', 'noun', 'c'], ['bilancia', 'noun', 'b'], ['bilancio', 'noun', 'b'], ['biliardo', 'noun', 'c'], ['bimba', 'noun', 'b'], ['bimbo', 'noun', 'b'], ['binario', 'noun', 'c'], ['biografia', 'noun', 'b'], ['biologia', 'noun', 'b'], ['biologico', 'adjective', 'b'], ['biologico', 'noun', 'b'], ['bionda', 'noun', 'b'], ['biondo', 'adjective', 'b'], ['biondo', 'noun', 'b'], ['birichino', 'noun', 'c'], ['birichino', 'adjective', 'c'], ['birillo', 'noun', 'c'], ['birra', 'noun', 'b'], ['bisbigliare', 'verb', 'c'], ['biscia', 'noun', 'c'], ['biscotto', 'adjective', 'b'], ['biscotto', 'noun', 'b'], ['bisnonno', 'noun', 'c'], ['bisognare', 'verb', 'a'], ['bisogno', 'noun', 'a'], ['bistecca', 'noun', 'c'], ['bistecchiera', 'noun', 'c'], ['bisticciare', 'verb', 'c'], ['bit', 'noun', 'b'], ['bizzarro', 'adjective', 'b'], ['bloccare', 'verb', 'a'], ['blocco', 'noun', 'b'], ['blocco', 'noun', 'b'], ['blog', 'noun', 'a'], ['blu', 'adjective', 'a'], ['blu', 'noun', 'a'], ['bocca', 'noun', 'a'], ['bocchino', 'noun', 'c'], ['boccia', 'noun', 'c'], ['bocciare', 'verb', 'b'], ['bocciatura', 'noun', 'c'], ['bocciolo', 'noun', 'c'], ['boccone', 'noun', 'c'], ['boh', 'exclamation', 'b'], ['boia', 'noun', 'c'], ['boia', 'adjective', 'c'], ['bolla', 'noun', 'b'], ['bolletta', 'noun', 'b'], ['bollito', 'past_part', 'c'], ['bollito', 'adjective', 'c'], ['bollito', 'noun', 'c'], ['bollitore', 'noun', 'c'], ['bollo', 'noun', 'c'], ['bolognese', 'adjective', 'c'], ['bolognese', 'noun', 'c'], ['bolzanino', 'adjective', 'c'], ['bolzanino', 'noun', 'c'], ['bomba', 'noun', 'b'], ['bombardare', 'verb', 'b'], ['bombola', 'noun', 'c'], ['bomboniera', 'noun', 'c'], ['bont', 'noun', 'b'], ['bordo', 'noun', 'a'], ['borgata', 'noun', 'c'], ['borghese', 'adjective', 'b'], ['borghese', 'noun', 'b'], ['borghesia', 'noun', 'c'], ['borgo', 'noun', 'b'], ['borotalco', 'noun', 'c'], ['borsa', 'noun', 'a'], ['borsa', 'noun', 'b'], ['borsetta', 'noun', 'c'], ['bosco', 'noun', 'a'], ['bosniaco', 'adjective', 'c'], ['bosniaco', 'noun', 'c'], ['boss', 'noun', 'b'], ['bossolo', 'noun', 'c'], ['botanica', 'noun', 'c'], ['botta', 'noun', 'b'], ['botte', 'noun', 'c'], ['bottega', 'noun', 'b'], ['bottegaio', 'noun', 'c'], ['bottegaio', 'adjective', 'c'], ['bottiglia', 'noun', 'a'], ['botto', 'noun', 'c'], ['bottone', 'noun', 'b'], ['bovino', 'adjective', 'c'], ['bovino', 'noun', 'c'], ['box', 'noun', 'b'], ['boxer', 'noun', 'c'], ['braccialetto', 'noun', 'c'], ['bracciante', 'noun', 'c'], ['braccio', 'noun', 'a'], ['branco', 'noun', 'b'], ['brand', 'noun', 'b'], ['brandello', 'noun', 'c'], ['brano', 'noun', 'a'], ['brasiliano', 'adjective', 'b'], ['brasiliano', 'noun', 'b'], ['bravo', 'adjective', 'a'], ['bravo', 'noun', 'a'], ['bravo', 'exclamation', 'a'], ['bresaola', 'noun', 'c'], ['bretella', 'noun', 'c'], ['breve', 'adjective', 'a'], ['breve', 'adverb', 'a'], ['breve', 'noun', 'a'], ['briciola', 'noun', 'c'], ['brigantaggio', 'noun', 'c'], ['brigante', 'noun', 'c'], ['brillante', 'pres_part', 'b'], ['brillante', 'adjective', 'b'], ['brillante', 'noun', 'b'], ['brillantina', 'noun', 'c'], ['brillare', 'verb', 'b'], ['brina', 'noun', 'c'], ['brioche', 'noun', 'c'], ['britannico', 'adjective', 'b'], ['britannico', 'noun', 'b'], ['brivido', 'noun', 'b'], ['brocca', 'noun', 'c'], ['brogliaccio', 'noun', 'b'], ['bronchite', 'noun', 'c'], ['brontolare', 'verb', 'c'], ['bronzo', 'noun', 'b'], ['bruciare', 'verb', 'a'], ['bruciato', 'past_part', 'b'], ['bruciato', 'adjective', 'b'], ['bruciato', 'noun', 'b'], ['bruciatura', 'noun', 'c'], ['bruco', 'noun', 'c'], ['bruco', 'adjective', 'c'], ['bruschetta', 'noun', 'c'], ['brutale', 'adjective', 'c'], ['brutto', 'adjective', 'a'], ['brutto', 'noun', 'a'], ['brutto', 'adverb', 'a'], ['buca', 'noun', 'b'], ['bucare', 'verb', 'b'], ['bucato', 'noun', 'c'], ['buccia', 'noun', 'c'], ['buco', 'noun', 'a'], ['budino', 'noun', 'c'], ['bufala', 'noun', 'c'], ['bufalo', 'noun', 'c'], ['bufera', 'noun', 'c'], ['buffet', 'noun', 'c'], ['buffo', 'adjective', 'b'], ['buffo', 'noun', 'b'], ['bugia', 'noun', 'b'], ['bugiardo', 'adjective', 'b'], ['bugiardo', 'noun', 'b'], ['buio', 'adjective', 'a'], ['buio', 'noun', 'a'], ['bulgaro', 'adjective', 'c'], ['bulgaro', 'noun', 'c'], ['buonafede', 'noun', 'c'], ['buonasera', 'exclamation', 'b'], ['buongiorno', 'exclamation', 'a'], ['buongusto', 'noun', 'c'], ['buono', 'adjective', 'a'], ['buono', 'noun', 'a'], ['buono', 'adverb', 'a'], ['buonuomo', 'noun', 'c'], ['burattino', 'noun', 'c'], ['burocrazia', 'noun', 'c'], ['burrasca', 'noun', 'c'], ['burro', 'noun', 'b'], ['burrone', 'noun', 'c'], ['business', 'noun', 'b'], ['business', 'adjective', 'b'], ['bussare', 'verb', 'b'], ['bussola', 'noun', 'c'], ['busta', 'noun', 'b'], ['bustina', 'noun', 'c'], ['busto', 'noun', 'c'], ['buttare', 'verb', 'a'], ['cabina', 'noun', 'b'], ['cacao', 'noun', 'c'], ['cacca', 'noun', 'b'], ['caccia', 'noun', 'a'], ['cacciare', 'verb', 'a'], ['cacciatore', 'noun', 'b'], ['cacciavite', 'noun', 'c'], ['cadavere', 'noun', 'a'], ['cadere', 'verb', 'a'], ['cadere', 'noun', 'a'], ['caduta', 'noun', 'b'], ['caff', 'noun', 'a'], ['caff', 'adjective', 'a'], ['caffellatte', 'noun', 'c'], ['caffellatte', 'adjective', 'c'], ['caffettiera', 'noun', 'c'], ['cagare', 'verb', 'b'], ['cagliaritano', 'adjective', 'c'], ['cagliaritano', 'noun', 'c'], ['calabrese', 'adjective', 'c'], ['calabrese', 'noun', 'c'], ['calabrone', 'noun', 'c'], ['calamaro', 'noun', 'c'], ['calamita', 'noun', 'c'], ['calare', 'verb', 'b'], ['calcagno', 'noun', 'c'], ['calciare', 'verb', 'c'], ['calciatore', 'noun', 'b'], ['calcinaccio', 'noun', 'c'], ['calcio', 'noun', 'a'], ['calcolare', 'verb', 'b'], ['calcolatore', 'adjective', 'c'], ['calcolatore', 'noun', 'c'], ['calcolatrice', 'noun', 'c'], ['calcolo', 'noun', 'b'], ['caldo', 'adjective', 'a'], ['caldo', 'noun', 'a'], ['caldo', 'adverb', 'a'], ['calendario', 'noun', 'b'], ['calligrafia', 'noun', 'c'], ['callo', 'noun', 'c'], ['calma', 'noun', 'b'], ['calmare', 'verb', 'b'], ['calmo', 'adjective', 'b'], ['calo', 'noun', 'b'], ['calore', 'noun', 'a'], ['calpestare', 'verb', 'c'], ['calunnia', 'noun', 'c'], ['calvario', 'noun', 'c'], ['calza', 'noun', 'b'], ['calzare', 'verb', 'c'], ['calzatura', 'noun', 'c'], ['calzino', 'noun', 'c'], ['calzolaio', 'noun', 'c'], ['calzoleria', 'noun', 'c'], ['calzone', 'noun', 'c'], ['cambiamento', 'noun', 'a'], ['cambiare', 'verb', 'a'], ['cambio', 'noun', 'a'], ['camera', 'noun', 'a'], ['camerata', 'noun', 'c'], ['cameriere', 'noun', 'b'], ['camicetta', 'noun', 'c'], ['camicia', 'noun', 'b'], ['caminetto', 'noun', 'c'], ['camion', 'noun', 'a'], ['camionista', 'noun', 'c'], ['cammello', 'noun', 'c'], ['cammello', 'adjective', 'c'], ['camminare', 'verb', 'a'], ['camminata', 'noun', 'c'], ['cammino', 'noun', 'b'], ['camomilla', 'noun', 'c'], ['camorra', 'noun', 'b'], ['campagna', 'noun', 'a'], ['campana', 'noun', 'b'], ['campanella', 'noun', 'c'], ['campanello', 'noun', 'b'], ['campanile', 'noun', 'c'], ['campano', 'adjective', 'c'], ['campano', 'noun', 'c'], ['campare', 'verb', 'b'], ['campeggio', 'noun', 'c'], ['campionato', 'noun', 'b'], ['campione', 'noun', 'a'], ['campo', 'noun', 'a'], ['campobassano', 'adjective', 'c'], ['campobassano', 'noun', 'c'], ['camposanto', 'noun', 'c'], ['canadese', 'adjective', 'c'], ['canadese', 'noun', 'c'], ['canaglia', 'noun', 'c'], ['canale', 'noun', 'a'], ['canapa', 'noun', 'c'], ['canarino', 'noun', 'c'], ['canarino', 'adjective', 'c'], ['cancellare', 'verb', 'a'], ['cancellatura', 'noun', 'c'], ['cancello', 'noun', 'b'], ['cancro', 'noun', 'b'], ['candela', 'noun', 'b'], ['candeliere', 'noun', 'c'], ['candidare', 'verb', 'b'], ['candidato', 'past_part', 'a'], ['candidato', 'adjective', 'a'], ['candidato', 'noun', 'a'], ['candido', 'adjective', 'b'], ['cane', 'noun', 'a'], ['canestro', 'noun', 'c'], ['canguro', 'noun', 'c'], ['canna', 'noun', 'b'], ['cannibale', 'adjective', 'c'], ['cannibale', 'noun', 'c'], ['cannuccia', 'noun', 'c'], ['canone', 'noun', 'b'], ['canottiera', 'noun', 'c'], ['canotto', 'noun', 'c'], ['cantante', 'pres_part', 'b'], ['cantante', 'adjective', 'b'], ['cantante', 'noun', 'b'], ['cantare', 'verb', 'a'], ['cantautore', 'noun', 'c'], ['cantiere', 'noun', 'b'], ['cantilena', 'noun', 'c'], ['cantina', 'noun', 'b'], ['canto', 'noun', 'a'], ['canzone', 'noun', 'a'], ['caos', 'noun', 'b'], ['capace', 'adjective', 'a'], ['capacit', 'noun', 'a'], ['capanna', 'noun', 'b'], ['capannone', 'noun', 'b'], ['caparra', 'noun', 'c'], ['capello', 'noun', 'a'], ['capire', 'verb', 'a'], ['capitale', 'adjective', 'a'], ['capitale', 'noun', 'a'], ['capitano', 'noun', 'a'], ['capitare', 'verb', 'a'], ['capitolo', 'noun', 'a'], ['capo', 'noun', 'a'], ['capodanno', 'noun', 'c'], ['capogiro', 'noun', 'c'], ['capolavoro', 'noun', 'b'], ['capoluogo', 'noun', 'c'], ['caporale', 'noun', 'b'], ['caporale', 'adjective', 'b'], ['caposquadra', 'noun', 'c'], ['capotavola', 'noun', 'c'], ['capoufficio', 'noun', 'c'], ['cappa', 'noun', 'c'], ['cappella', 'noun', 'b'], ['cappelliera', 'noun', 'c'], ['cappello', 'noun', 'b'], ['cappero', 'noun', 'c'], ['cappotto', 'noun', 'c'], ['cappuccino', 'adjective', 'c'], ['cappuccino', 'noun', 'c'], ['cappuccino', 'adjective', 'c'], ['cappuccio', 'noun', 'c'], ['capra', 'noun', 'b'], ['capriccio', 'noun', 'b'], ['capriola', 'noun', 'c'], ['carabiniere', 'noun', 'a'], ['caramella', 'noun', 'b'], ['caramella', 'adjective', 'b'], ['carattere', 'noun', 'a'], ['caratteristica', 'noun', 'a'], ['caratteristico', 'adjective', 'b'], ['caratterizzare', 'verb', 'a'], ['carbone', 'noun', 'b'], ['carburante', 'pres_part', 'c'], ['carburante', 'adjective', 'c'], ['carburante', 'noun', 'c'], ['carcassa', 'noun', 'c'], ['carcerato', 'past_part', 'c'], ['carcerato', 'adjective', 'c'], ['carcerato', 'noun', 'c'], ['carcere', 'noun', 'a'], ['carciofino', 'noun', 'c'], ['carciofo', 'noun', 'c'], ['cardellino', 'noun', 'c'], ['cardiaco', 'adjective', 'b'], ['cardiaco', 'noun', 'b'], ['cardigan', 'noun', 'c'], ['cardinale', 'adjective', 'b'], ['cardinale', 'noun', 'b'], ['cardinale', 'adjective', 'b'], ['carenza', 'noun', 'b'], ['carica', 'noun', 'loc-comando'], ['caricare', 'verb', 'a'], ['carico', 'noun', 'a'], ['carico', 'adjective', 'b'], ['carino', 'adjective', 'a'], ['carit', 'noun', 'b'], ['carnagione', 'noun', 'c'], ['carne', 'noun', 'a'], ['carnevale', 'noun', 'c'], ['carnivoro', 'adjective', 'c'], ['carnivoro', 'noun', 'c'], ['carnoso', 'adjective', 'c'], ['carnoso', 'noun', 'c'], ['caro', 'adjective', 'a'], ['caro', 'adverb', 'a'], ['caro', 'noun', 'a'], ['carosello', 'noun', 'c'], ['carovana', 'noun', 'c'], ['carriera', 'noun', 'a'], ['carro', 'noun', 'b'], ['carrozzeria', 'noun', 'c'], ['carta', 'noun', 'a'], ['cartaceo', 'adjective', 'b'], ['cartella', 'noun', 'b'], ['cartello', 'noun', 'b'], ['cartoleria', 'noun', 'c'], ['cartolina', 'noun', 'b'], ['cartone', 'noun', 'b'], ['cartuccia', 'noun', 'c'], ['casa', 'noun', 'a'], ['casalinga', 'noun', 'c'], ['casalingo', 'adjective', 'c'], ['casalingo', 'noun', 'c'], ['cascare', 'verb', 'b'], ['cascata', 'noun', 'c'], ['casco', 'noun', 'c'], ['caserma', 'noun', 'b'], ['casetta', 'noun', 'b'], ['casino', 'noun', 'a'], ['caso', 'noun', 'a'], ['cassa', 'noun', 'a'], ['cassaforte', 'noun', 'c'], ['cassapanca', 'noun', 'c'], ['casseruola', 'noun', 'c'], ['cassetta', 'noun', 'b'], ['cassettiera', 'noun', 'c'], ['cassetto', 'noun', 'b'], ['cassiera', 'noun', 'c'], ['castagna', 'noun', 'c'], ['castagno', 'noun', 'c'], ['castano', 'adjective', 'c'], ['castello', 'noun', 'a'], ['castoro', 'noun', 'c'], ['casuale', 'adjective', 'b'], ['casuale', 'noun', 'b'], ['catalogo', 'noun', 'b'], ['catanzarese', 'adjective', 'c'], ['catanzarese', 'noun', 'c'], ['catarro', 'noun', 'c'], ['catasta', 'noun', 'c'], ['catastrofe', 'noun', 'b'], ['catechismo', 'noun', 'c'], ['categoria', 'noun', 'a'], ['catena', 'noun', 'a'], ['catenaccio', 'noun', 'c'], ['catino', 'noun', 'c'], ['catrame', 'noun', 'c'], ['cattedrale', 'adjective', 'b'], ['cattedrale', 'noun', 'b'], ['cattivo', 'adjective', 'a'], ['cattivo', 'noun', 'a'], ['cattolico', 'adjective', 'a'], ['cattolico', 'noun', 'a'], ['catturare', 'verb', 'b'], ['causa', 'noun', 'a'], ['causare', 'verb', 'a'], ['cavalcare', 'verb', 'b'], ['cavaliere', 'noun', 'a'], ['cavalletta', 'noun', 'c'], ['cavallo', 'noun', 'a'], ['cavare', 'verb', 'b'], ['cavatappi', 'noun', 'c'], ['caverna', 'noun', 'c'], ['caviglia', 'noun', 'b'], ['cavit', 'noun', 'b'], ['cavo', 'adjective', 'b'], ['cavo', 'noun', 'b'], ['cavo', 'noun', 'b'], ['cavolo', 'noun', 'b'], ['cazzata', 'noun', 'b'], ['cazzo', 'noun', 'a'], ['ce', 'pronoun', 'a'], ['ce', 'adverb', 'a'], ['cece', 'noun', 'c'], ['ceco', 'adjective', 'c'], ['ceco', 'noun', 'c'], ['cecoslovacco', 'adjective', 'c'], ['cecoslovacco', 'noun', 'c'], ['cedere', 'verb', 'a'], ['celare', 'verb', 'b'], ['celebrare', 'verb', 'b'], ['celebre', 'adjective', 'b'], ['celeste', 'adjective', 'b'], ['celeste', 'noun', 'b'], ['cella', 'noun', 'b'], ['cellula', 'noun', 'a'], ['cellulare', 'adjective', 'a'], ['cellulare', 'noun', 'a'], ['cemento', 'noun', 'b'], ['cena', 'noun', 'a'], ['cenare', 'verb', 'b'], ['cenere', 'noun', 'b'], ['cenere', 'adjective', 'b'], ['cenno', 'noun', 'b'], ['centesimo', 'adjective', 'b'], ['centesimo', 'noun', 'b'], ['centimetro', 'noun', 'b'], ['centinaio', 'noun', 'a'], ['cento', 'adjective', 'a'], ['cento', 'noun', 'a'], ['centrale', 'adjective', 'a'], ['centrale', 'noun', 'a'], ['centralino', 'noun', 'c'], ['centrare', 'verb', 'b'], ['centro', 'noun', 'a'], ['centroamericano', 'adjective', 'c'], ['centroamericano', 'noun', 'c'], ['ceramica', 'noun', 'b'], ['cercare', 'verb', 'a'], ['cerchio', 'noun', 'b'], ['cereale', 'noun', 'c'], ['cereale', 'adjective', 'c'], ['cerebrale', 'adjective', 'b'], ['cerebrale', 'noun', 'b'], ['cerimonia', 'noun', 'b'], ['cerino', 'noun', 'c'], ['cerniera', 'noun', 'c'], ['cerotto', 'noun', 'c'], ['certamente', 'adverb', 'a'], ['certezza', 'noun', 'a'], ['certificare', 'verb', 'b'], ['certificato', 'past_part', 'b'], ['certificato', 'adjective', 'b'], ['certificato', 'noun', 'b'], ['certo', 'adjective', 'a'], ['certo', 'adjective', 'a'], ['certo', 'pronoun', 'a'], ['certo', 'adverb', 'a'], ['cervello', 'noun', 'a'], ['cervo', 'noun', 'c'], ['cespuglio', 'noun', 'b'], ['cessare', 'verb', 'b'], ['cesso', 'noun', 'b'], ['cestino', 'noun', 'c'], ['cesto', 'noun', 'c'], ['cetriolo', 'noun', 'c'], ['chat', 'noun', 'b'], ['che', 'pronoun', 'a'], ['che', 'adjective', 'a'], ['che', 'noun', 'a'], ['chewingum', 'noun', 'c'], ['chi', 'pronoun', 'a'], ['chiacchiera', 'noun', 'b'], ['chiacchierare', 'verb', 'b'], ['chiamare', 'verb', 'a'], ['chiamata', 'noun', 'b'], ['chiaramente', 'adverb', 'a'], ['chiarezza', 'noun', 'b'], ['chiarire', 'verb', 'a'], ['chiaro', 'adjective', 'a'], ['chiaro', 'noun', 'a'], ['chiaro', 'adverb', 'a'], ['chiasso', 'noun', 'c'], ['chiave', 'noun', 'a'], ['chiazza', 'noun', 'c'], ['chiedere', 'verb', 'a'], ['chiesa', 'noun', 'a'], ['chilo', 'noun', 'b'], ['chilogrammo', 'noun', 'c'], ['chilometro', 'noun', 'a'], ['chimico', 'adjective', 'a'], ['chimico', 'noun', 'a'], ['china', 'noun', 'c'], ['chinare', 'verb', 'b'], ['chinotto', 'noun', 'c'], ['chiodo', 'noun', 'b'], ['chiosco', 'noun', 'b'], ['chirurgia', 'noun', 'b'], ['chirurgico', 'adjective', 'b'], ['chirurgico', 'noun', 'b'], ['chirurgo', 'noun', 'b'], ['chiss', 'adverb', 'a'], ['chitarra', 'noun', 'b'], ['chiudere', 'verb', 'a'], ['chiunque', 'pronoun', 'a'], ['chiuso', 'past_part', 'a'], ['chiuso', 'adjective', 'a'], ['chiuso', 'noun', 'a'], ['chiuso', 'adverb', 'a'], ['chiusura', 'noun', 'b'], ['ci', 'noun', 'c'], ['ci', 'pronoun', 'a'], ['ci', 'adverb', 'a'], ['ciabatta', 'noun', 'c'], ['ciambella', 'noun', 'c'], ['ciao', 'exclamation', 'a'], ['ciascuno', 'adjective', 'a'], ['ciascuno', 'pronoun', 'a'], ['cibare', 'verb', 'c'], ['cibo', 'noun', 'a'], ['cicatrice', 'noun', 'b'], ['ciclismo', 'noun', 'b'], ['ciclista', 'noun', 'c'], ['ciclo', 'noun', 'b'], ['cicogna', 'noun', 'c'], ['cicoria', 'noun', 'c'], ['cieco', 'adjective', 'b'], ['cieco', 'noun', 'b'], ['cielo', 'noun', 'a'], ['cifra', 'noun', 'a'], ['ciglio', 'noun', 'b'], ['cigno', 'noun', 'c'], ['cileno', 'adjective', 'c'], ['cileno', 'noun', 'c'], ['ciliegia', 'noun', 'c'], ['ciliegia', 'adjective', 'c'], ['ciliegio', 'noun', 'c'], ['cilindro', 'noun', 'c'], ['cima', 'noun', 'c'], ['cimice', 'noun', 'c'], ['ciminiera', 'noun', 'c'], ['cimitero', 'noun', 'b'], ['cinema', 'noun', 'a'], ['cinematografico', 'adjective', 'b'], ['cinese', 'adjective', 'a'], ['cinese', 'noun', 'a'], ['cinghia', 'noun', 'c'], ['cinghiale', 'noun', 'c'], ['cinguettare', 'verb', 'c'], ['cinguettio', 'noun', 'c'], ['cinico', 'adjective', 'c'], ['cinico', 'noun', 'c'], ['cinquanta', 'adjective', 'a'], ['cinquanta', 'noun', 'a'], ['cinque', 'adjective', 'a'], ['cinque', 'noun', 'a'], ['cinquecento', 'adjective', 'b'], ['cinquecento', 'noun', 'b'], ['cintura', 'noun', 'b'], ['cinturino', 'noun', 'c'], ['ci', 'pronoun', 'a'], ['ciocca', 'noun', 'c'], ['cioccolatino', 'noun', 'c'], ['cioccolato', 'noun', 'b'], ['cioccolato', 'adjective', 'b'], ['cio', 'conjunction', 'a'], ['ciotola', 'noun', 'c'], ['cipolla', 'noun', 'b'], ['cipresso', 'noun', 'c'], ['cipriota', 'adjective', 'c'], ['cipriota', 'noun', 'c'], ['circa', 'preposition', 'a'], ['circa', 'adverb', 'a'], ['circa', 'noun', 'a'], ['circo', 'noun', 'b'], ['circolare', 'adjective', 'b'], ['circolare', 'noun', 'b'], ['circolare', 'verb', 'b'], ['circolazione', 'noun', 'b'], ['circolo', 'noun', 'b'], ['circondare', 'verb', 'a'], ['circostanza', 'noun', 'a'], ['circuito', 'noun', 'b'], ['citare', 'verb', 'a'], ['citato', 'past_part', 'b'], ['citato', 'adjective', 'b'], ['citato', 'noun', 'b'], ['citazione', 'noun', 'b'], ['citofono', 'noun', 'c'], ['citt', 'noun', 'a'], ['cittadina', 'noun', 'b'], ['cittadinanza', 'noun', 'b'], ['cittadino', 'adjective', 'a'], ['cittadino', 'noun', 'a'], ['ciuffo', 'noun', 'c'], ['civile', 'adjective', 'a'], ['civile', 'noun', 'a'], ['civilt', 'noun', 'b'], ['clacson', 'noun', 'c'], ['clan', 'noun', 'b'], ['clandestino', 'adjective', 'b'], ['clandestino', 'noun', 'b'], ['classe', 'noun', 'a'], ['classico', 'adjective', 'a'], ['classico', 'noun', 'a'], ['classifica', 'noun', 'b'], ['classificare', 'verb', 'b'], ['clero', 'noun', 'c'], ['cliccare', 'verb', 'b'], ['cliente', 'noun', 'a'], ['clima', 'noun', 'b'], ['clinica', 'noun', 'b'], ['clinico', 'adjective', 'b'], ['clinico', 'noun', 'b'], ['clistere', 'noun', 'c'], ['cloro', 'noun', 'c'], ['club', 'noun', 'b'], ['cobra', 'noun', 'c'], ['cocaina', 'noun', 'b'], ['coccinella', 'noun', 'c'], ['coccio', 'noun', 'c'], ['cocciuto', 'adjective', 'c'], ['cocciuto', 'noun', 'c'], ['cocco', 'noun', 'c'], ['coccodrillo', 'noun', 'c'], ['coccola', 'noun', 'c'], ['coccolare', 'verb', 'c'], ['cocomero', 'noun', 'c'], ['coda', 'noun', 'a'], ['codice', 'noun', 'a'], ['coerente', 'adjective', 'b'], ['cofano', 'noun', 'c'], ['cogliere', 'verb', 'a'], ['coglione', 'noun', 'a'], ['cognato', 'noun', 'b'], ['cognato', 'adjective', 'b'], ['cognome', 'noun', 'b'], ['coincidenza', 'noun', 'b'], ['coincidere', 'verb', 'b'], ['coinvolgere', 'verb', 'a'], ['coinvolgimento', 'noun', 'b'], ['colare', 'verb', 'b'], ['colata', 'noun', 'c'], ['colazione', 'noun', 'b'], ['colera', 'noun', 'c'], ['colica', 'noun', 'c'], ['colino', 'noun', 'c'], ['colla', 'noun', 'c'], ['collaborare', 'verb', 'b'], ['collaboratore', 'noun', 'b'], ['collaborazione', 'noun', 'b'], ['collana', 'noun', 'b'], ['collant', 'noun', 'c'], ['collant', 'adjective', 'c'], ['collare', 'noun', 'c'], ['collasso', 'noun', 'c'], ['collaterale', 'adjective', 'b'], ['collaterale', 'noun', 'b'], ['colle', 'noun', 'c'], ['collega', 'noun', 'a'], ['collegamento', 'noun', 'b'], ['collegare', 'verb', 'a'], ['collegio', 'noun', 'b'], ['collera', 'noun', 'c'], ['colletta', 'noun', 'c'], ['collettivo', 'adjective', 'b'], ['collettivo', 'noun', 'b'], ['collezione', 'noun', 'b'], ['collina', 'noun', 'b'], ['collo', 'noun', 'a'], ['collocare', 'verb', 'b'], ['colloquio', 'noun', 'b'], ['colluttorio', 'noun', 'c'], ['colmo', 'noun', 'c'], ['colomba', 'noun', 'b'], ['colombo', 'noun', 'c'], ['colonna', 'noun', 'a'], ['colonnello', 'noun', 'b'], ['colorante', 'pres_part', 'c'], ['colorante', 'adjective', 'c'], ['colorante', 'noun', 'c'], ['colorare', 'verb', 'b'], ['colorato', 'past_part', 'b'], ['colorato', 'adjective', 'b'], ['colore', 'noun', 'a'], ['coloro', 'pronoun', 'a'], ['colosso', 'noun', 'c'], ['colpa', 'noun', 'a'], ['colpevole', 'adjective', 'b'], ['colpevole', 'noun', 'b'], ['colpire', 'verb', 'a'], ['colpo', 'noun', 'a'], ['coltellata', 'noun', 'c'], ['coltello', 'noun', 'a'], ['coltivare', 'verb', 'b'], ['coltivazione', 'noun', 'c'], ['colto', 'adjective', 'b'], ['colto', 'noun', 'b'], ['colui', 'pronoun', 'b'], ['coma', 'noun', 'b'], ['comandamento', 'noun', 'b'], ['comandante', 'pres_part', 'b'], ['comandante', 'adjective', 'b'], ['comandante', 'noun', 'b'], ['comandare', 'verb', 'b'], ['comando', 'noun', 'b'], ['combaciare', 'verb', 'c'], ['combattente', 'pres_part', 'c'], ['combattente', 'adjective', 'c'], ['combattente', 'noun', 'c'], ['combattere', 'verb', 'a'], ['combattimento', 'noun', 'b'], ['combinare', 'verb', 'b'], ['combinazione', 'noun', 'b'], ['come', 'adverb', 'a'], ['come', 'conjunction', 'a'], ['cometa', 'noun', 'c'], ['comfort', 'noun', 'c'], ['comico', 'adjective', 'b'], ['comico', 'noun', 'b'], ['cominciare', 'verb', 'a'], ['cominciare', 'noun', 'a'], ['comitato', 'noun', 'b'], ['comma', 'noun', 'b'], ['commedia', 'noun', 'b'], ['commentare', 'verb', 'a'], ['commento', 'noun', 'a'], ['commerciale', 'adjective', 'a'], ['commerciale', 'noun', 'a'], ['commerciante', 'pres_part', 'b'], ['commerciante', 'adjective', 'b'], ['commerciante', 'noun', 'b'], ['commercio', 'noun', 'b'], ['commettere', 'verb', 'a'], ['commissariato', 'noun', 'b'], ['commissario', 'noun', 'a'], ['commissione', 'noun', 'a'], ['community', 'noun', 'b'], ['commuovere', 'verb', 'b'], ['comodino', 'noun', 'c'], ['comodit', 'noun', 'c'], ['comodo', 'adjective', 'a'], ['comodo', 'noun', 'a'], ['compagnia', 'noun', 'a'], ['compagno', 'noun', 'a'], ['compagno', 'adjective', 'a'], ['comparire', 'verb', 'a'], ['comparsa', 'noun', 'b'], ['compassione', 'noun', 'c'], ['compasso', 'noun', 'c'], ['compatibile', 'adjective', 'b'], ['compatriota', 'noun', 'c'], ['compatto', 'adjective', 'b'], ['compatto', 'noun', 'b'], ['compensare', 'verb', 'b'], ['compenso', 'noun', 'b'], ['competente', 'adjective', 'b'], ['competente', 'noun', 'b'], ['competenza', 'noun', 'b'], ['competere', 'verb', 'b'], ['competizione', 'noun', 'b'], ['compiangere', 'verb', 'c'], ['compiere', 'verb', 'a'], ['compilare', 'verb', 'b'], ['compito', 'noun', 'a'], ['compleanno', 'noun', 'b'], ['complessivo', 'adjective', 'b'], ['complesso', 'noun', 'b'], ['complesso', 'adjective', 'a'], ['completamente', 'adverb', 'a'], ['completare', 'verb', 'b'], ['completo', 'adjective', 'a'], ['completo', 'noun', 'a'], ['complicare', 'verb', 'b'], ['complicato', 'past_part', 'b'], ['complicato', 'adjective', 'b'], ['complice', 'noun', 'b'], ['complice', 'adjective', 'b'], ['complimento', 'noun', 'b'], ['complotto', 'noun', 'c'], ['componente', 'pres_part', 'b'], ['componente', 'adjective', 'b'], ['componente', 'noun', 'b'], ['comporre', 'verb', 'a'], ['comportamento', 'noun', 'a'], ['comportare', 'verb', 'a'], ['composizione', 'noun', 'b'], ['composto', 'past_part', 'b'], ['composto', 'adjective', 'b'], ['composto', 'noun', 'b'], ['comprare', 'verb', 'a'], ['comprendere', 'verb', 'a'], ['comprensibile', 'adjective', 'b'], ['comprensione', 'noun', 'b'], ['comprensivo', 'adjective', 'c'], ['compreso', 'past_part', 'a'], ['compreso', 'adjective', 'a'], ['compromesso', 'noun', 'b'], ['compromettere', 'verb', 'b'], ['computer', 'noun', 'a'], ['comunale', 'adjective', 'b'], ['comunale', 'noun', 'b'], ['comune', 'adjective', 'a'], ['comune', 'noun', 'a'], ['comune', 'noun', 'a'], ['comunicare', 'verb', 'a'], ['comunicazione', 'noun', 'a'], ['comunione', 'noun', 'b'], ['comunismo', 'noun', 'b'], ['comunista', 'adjective', 'a'], ['comunista', 'noun', 'a'], ['comunit', 'noun', 'a'], ['comunque', 'adverb', 'a'], ['comunque', 'conjunction', 'a'], ['con', 'preposition', 'a'], ['conca', 'noun', 'c'], ['concedere', 'verb', 'b'], ['concentrare', 'verb', 'a'], ['concentrazione', 'noun', 'b'], ['concepire', 'noun', 'b'], ['concerto', 'noun', 'a'], ['concessione', 'noun', 'b'], ['concesso', 'past_part', 'b'], ['concesso', 'adjective', 'b'], ['concetto', 'past_part', 'a'], ['concetto', 'adjective', 'a'], ['concetto', 'noun', 'a'], ['concezione', 'noun', 'b'], ['conchiglia', 'noun', 'c'], ['concime', 'noun', 'c'], ['concludere', 'verb', 'a'], ['conclusione', 'noun', 'a'], ['concordare', 'verb', 'b'], ['concorrente', 'pres_part', 'b'], ['concorrente', 'adjective', 'b'], ['concorrente', 'noun', 'b'], ['concorrenza', 'noun', 'b'], ['concorrere', 'verb', 'b'], ['concorso', 'noun', 'b'], ['concreto', 'adjective', 'a'], ['concreto', 'noun', 'a'], ['condanna', 'noun', 'b'], ['condannare', 'verb', 'a'], ['condimento', 'noun', 'c'], ['condividere', 'verb', 'a'], ['condizionare', 'verb', 'b'], ['condizione', 'noun', 'a'], ['condoglianza', 'noun', 'c'], ['condominio', 'noun', 'b'], ['condotta', 'noun', 'b'], ['condurre', 'verb', 'a'], ['conduttore', 'adjective', 'b'], ['conduttore', 'noun', 'b'], ['conduttura', 'noun', 'c'], ['conferenza', 'noun', 'b'], ['conferire', 'verb', 'b'], ['conferma', 'noun', 'b'], ['confermare', 'verb', 'a'], ['confessare', 'verb', 'b'], ['confessione', 'noun', 'b'], ['confessore', 'noun', 'c'], ['confetto', 'noun', 'c'], ['confetto', 'adjective', 'c'], ['confettura', 'noun', 'c'], ['confezione', 'noun', 'b'], ['conficcare', 'verb', 'c'], ['confidare', 'verb', 'b'], ['confidenza', 'noun', 'b'], ['confine', 'noun', 'a'], ['conflitto', 'noun', 'b'], ['confondere', 'verb', 'a'], ['confortare', 'verb', 'c'], ['confrontare', 'verb', 'b'], ['confronto', 'noun', 'a'], ['confusione', 'noun', 'b'], ['confuso', 'past_part', 'b'], ['confuso', 'adjective', 'b'], ['congedo', 'noun', 'c'], ['congelare', 'verb', 'b'], ['congelatore', 'noun', 'c'], ['congestione', 'noun', 'c'], ['congiura', 'noun', 'c'], ['congresso', 'noun', 'b'], ['coniglio', 'noun', 'b'], ['coniugato', 'past_part', 'c'], ['coniugato', 'adjective', 'c'], ['coniugato', 'noun', 'c'], ['coniuge', 'noun', 'b'], ['connessione', 'noun', 'b'], ['connettere', 'verb', 'b'], ['cono', 'noun', 'b'], ['conoscenza', 'noun', 'a'], ['conoscere', 'verb', 'a'], ['conosciuto', 'past_part', 'b'], ['conosciuto', 'adjective', 'b'], ['conosciuto', 'noun', 'b'], ['conquista', 'noun', 'b'], ['conquistare', 'verb', 'a'], ['consapevole', 'adjective', 'b'], ['consapevolezza', 'noun', 'b'], ['consegna', 'noun', 'b'], ['consegnare', 'verb', 'a'], ['conseguente', 'pres_part', 'b'], ['conseguente', 'adjective', 'b'], ['conseguente', 'noun', 'b'], ['conseguenza', 'noun', 'a'], ['conseguire', 'verb', 'b'], ['consenso', 'noun', 'b'], ['consentire', 'verb', 'a'], ['conservare', 'verb', 'a'], ['conservazione', 'noun', 'b'], ['considerare', 'verb', 'a'], ['considerazione', 'noun', 'a'], ['consigliare', 'verb', 'a'], ['consigliere', 'noun', 'b'], ['consiglio', 'noun', 'a'], ['consistente', 'pres_part', 'b'], ['consistente', 'adjective', 'b'], ['consistenza', 'noun', 'b'], ['consistere', 'verb', 'b'], ['consolare', 'verb', 'b'], ['consonante', 'noun', 'c'], ['consorzio', 'noun', 'b'], ['constatare', 'verb', 'b'], ['consueto', 'adjective', 'b'], ['consueto', 'noun', 'b'], ['consulente', 'adjective', 'b'], ['consulente', 'noun', 'b'], ['consulenza', 'noun', 'b'], ['consultare', 'verb', 'b'], ['consumare', 'verb', 'a'], ['consumatore', 'noun', 'b'], ['consumatore', 'adjective', 'b'], ['consumazione', 'noun', 'c'], ['consumo', 'noun', 'b'], ['contachilometri', 'noun', 'c'], ['contadino', 'noun', 'b'], ['contadino', 'adjective', 'b'], ['contagiare', 'verb', 'c'], ['contagio', 'noun', 'c'], ['contagioso', 'adjective', 'c'], ['contagocce', 'noun', 'c'], ['contaminare', 'verb', 'b'], ['contante', 'pres_part', 'b'], ['contante', 'adjective', 'b'], ['contante', 'noun', 'b'], ['contare', 'verb', 'a'], ['contatore', 'noun', 'c'], ['contattare', 'verb', 'b'], ['contatto', 'noun', 'a'], ['conte', 'noun', 'b'], ['contemplare', 'verb', 'b'], ['contemporaneamente', 'adverb', 'b'], ['contemporaneo', 'adjective', 'a'], ['contemporaneo', 'noun', 'a'], ['contenere', 'verb', 'a'], ['contenitore', 'adjective', 'b'], ['contenitore', 'noun', 'b'], ['contentare', 'verb', 'b'], ['contentezza', 'noun', 'c'], ['contento', 'adjective', 'a'], ['contenuto', 'past_part', 'a'], ['contenuto', 'adjective', 'a'], ['contenuto', 'noun', 'a'], ['contestare', 'verb', 'b'], ['contestazione', 'noun', 'b'], ['contesto', 'noun', 'a'], ['continente', 'noun', 'b'], ['continuamente', 'adverb', 'b'], ['continuare', 'verb', 'a'], ['continuazione', 'noun', 'b'], ['continuit', 'noun', 'b'], ['continuo', 'adjective', 'a'], ['continuo', 'noun', 'a'], ['continuo', 'adverb', 'a'], ['conto', 'noun', 'a'], ['contorno', 'noun', 'b'], ['contrabbandiere', 'noun', 'c'], ['contrabbando', 'noun', 'c'], ['contraccambiare', 'verb', 'c'], ['contraddizione', 'noun', 'b'], ['contrario', 'adjective', 'a'], ['contrario', 'noun', 'a'], ['contrarre', 'verb', 'b'], ['contrastare', 'verb', 'b'], ['contrasto', 'noun', 'b'], ['contratto', 'noun', 'a'], ['contribuire', 'verb', 'b'], ['contributo', 'noun', 'b'], ['contro', 'preposition', 'a'], ['contro', 'adverb', 'a'], ['contro', 'noun', 'a'], ['controllare', 'verb', 'a'], ['controllo', 'noun', 'a'], ['controllore', 'noun', 'c'], ['convegno', 'noun', 'b'], ['conveniente', 'pres_part', 'b'], ['conveniente', 'adjective', 'b'], ['convenire', 'verb', 'b'], ['convenzione', 'noun', 'b'], ['conversazione', 'noun', 'a'], ['conversione', 'noun', 'b'], ['convertire', 'verb', 'b'], ['convincente', 'pres_part', 'b'], ['convincente', 'adjective', 'b'], ['convincere', 'verb', 'a'], ['convinto', 'past_part', 'b'], ['convinto', 'adjective', 'b'], ['convinzione', 'noun', 'b'], ['convivenza', 'noun', 'b'], ['convivere', 'verb', 'b'], ['convocare', 'verb', 'b'], ['convulsione', 'noun', 'c'], ['coordinamento', 'noun', 'b'], ['coordinare', 'verb', 'b'], ['coperchio', 'noun', 'c'], ['coperta', 'noun', 'b'], ['copertina', 'noun', 'b'], ['coperto', 'past_part', 'b'], ['coperto', 'adjective', 'b'], ['coperto', 'noun', 'b'], ['copertura', 'noun', 'b'], ['copia', 'noun', 'a'], ['copiare', 'verb', 'b'], ['copione', 'noun', 'b'], ['coppa', 'noun', 'b'], ['coppia', 'noun', 'a'], ['copricostume', 'noun', 'c'], ['copriletto', 'noun', 'c'], ['coprire', 'verb', 'a'], ['copyright', 'noun', 'b'], ['coraggio', 'noun', 'a'], ['coraggio', 'exclamation', 'a'], ['coraggioso', 'adjective', 'b'], ['corallo', 'noun', 'c'], ['corallo', 'adjective', 'c'], ['corazza', 'noun', 'c'], ['corazzata', 'noun', 'c'], ['corazziere', 'noun', 'c'], ['corda', 'noun', 'a'], ['coriandolo', 'noun', 'c'], ['coricare', 'verb', 'c'], ['cornacchia', 'noun', 'c'], ['cornetto', 'noun', 'c'], ['cornice', 'noun', 'b'], ['corno', 'noun', 'b'], ['cornuto', 'adjective', 'c'], ['cornuto', 'noun', 'c'], ['coro', 'noun', 'b'], ['corona', 'noun', 'b'], ['corpo', 'noun', 'a'], ['corporatura', 'noun', 'c'], ['correggere', 'verb', 'a'], ['corrente', 'pres_part', 'a'], ['corrente', 'adjective', 'a'], ['corrente', 'noun', 'a'], ['corrente', 'adverb', 'a'], ['correre', 'verb', 'a'], ['correttamente', 'adverb', 'b'], ['corretto', 'past_part', 'b'], ['corretto', 'adjective', 'b'], ['correzione', 'noun', 'c'], ['corridoio', 'noun', 'b'], ['corridore', 'adjective', 'c'], ['corridore', 'noun', 'c'], ['corriera', 'noun', 'c'], ['corriere', 'noun', 'a'], ['corrispondente', 'pres_part', 'b'], ['corrispondente', 'adjective', 'b'], ['corrispondente', 'noun', 'b'], ['corrispondenza', 'noun', 'b'], ['corrispondere', 'verb', 'a'], ['corruzione', 'noun', 'b'], ['corsa', 'noun', 'a'], ['corsia', 'noun', 'c'], ['corso', 'noun', 'a'], ['corte', 'noun', 'a'], ['corteccia', 'noun', 'c'], ['corteggiare', 'verb', 'c'], ['cortesia', 'noun', 'b'], ['cortile', 'noun', 'b'], ['corto', 'adjective', 'a'], ['corvo', 'noun', 'c'], ['cosa', 'noun', 'a'], ['coscia', 'noun', 'b'], ['cosciente', 'adjective', 'c'], ['coscienza', 'noun', 'a'], ['cos', 'adverb', 'a'], ['cosiddetto', 'adjective', 'a'], ['costa', 'noun', 'a'], ['costante', 'adjective', 'b'], ['costante', 'noun', 'b'], ['costantemente', 'adverb', 'b'], ['costare', 'verb', 'a'], ['costellazione', 'noun', 'b'], ['costituire', 'verb', 'a'], ['costituzionale', 'adjective', 'b'], ['costituzione', 'noun', 'b'], ['costo', 'noun', 'a'], ['costoso', 'adjective', 'b'], ['costringere', 'verb', 'a'], ['costruire', 'verb', 'a'], ['costruttivo', 'adjective', 'b'], ['costruzione', 'noun', 'a'], ['costume', 'noun', 'a'], ['cotoletta', 'noun', 'c'], ['cotone', 'noun', 'b'], ['cottura', 'noun', 'c'], ['covare', 'verb', 'c'], ['covo', 'noun', 'c'], ['cozza', 'noun', 'c'], ['cracker', 'noun', 'c'], ['cranio', 'noun', 'b'], ['cravatta', 'noun', 'b'], ['creare', 'verb', 'a'], ['creativit', 'noun', 'b'], ['creativo', 'adjective', 'b'], ['creativo', 'noun', 'b'], ['creatura', 'noun', 'b'], ['creazione', 'noun', 'b'], ['credente', 'pres_part', 'b'], ['credente', 'adjective', 'b'], ['credente', 'noun', 'b'], ['credenza', 'noun', 'c'], ['credere', 'verb', 'a'], ['credere', 'noun', 'a'], ['credibile', 'adjective', 'b'], ['credito', 'noun', 'a'], ['creditore', 'noun', 'b'], ['credo', 'noun', 'c'], ['crema', 'noun', 'b'], ['crema', 'adjective', 'b'], ['crepaccio', 'noun', 'c'], ['crpe', 'noun', 'c'], ['crescente', 'pres_part', 'b'], ['crescente', 'adjective', 'b'], ['crescente', 'noun', 'b'], ['crescere', 'verb', 'a'], ['crescita', 'noun', 'a'], ['cretino', 'adjective', 'b'], ['cretino', 'noun', 'b'], ['criceto', 'noun', 'c'], ['criminale', 'adjective', 'b'], ['criminale', 'noun', 'b'], ['crimine', 'noun', 'b'], ['criniera', 'noun', 'c'], ['crisantemo', 'noun', 'c'], ['crisi', 'noun', 'a'], ['cristallo', 'noun', 'b'], ['cristianesimo', 'noun', 'b'], ['cristiano', 'adjective', 'a'], ['cristiano', 'noun', 'a'], ['criterio', 'noun', 'b'], ['critica', 'noun', 'a'], ['criticare', 'verb', 'b'], ['critico', 'adjective', 'a'], ['critico', 'noun', 'a'], ['croato', 'adjective', 'c'], ['croato', 'noun', 'c'], ['croce', 'noun', 'b'], ['crocifiggere', 'verb', 'c'], ['crocifisso', 'past_part', 'c'], ['crocifisso', 'adjective', 'c'], ['crocifisso', 'noun', 'c'], ['crollare', 'verb', 'b'], ['cronaca', 'noun', 'b'], ['cronico', 'adjective', 'b'], ['cronico', 'noun', 'b'], ['cronista', 'noun', 'c'], ['crostaceo', 'noun', 'c'], ['crostino', 'noun', 'c'], ['crudele', 'adjective', 'b'], ['crudele', 'noun', 'b'], ['crudo', 'adjective', 'b'], ['crudo', 'noun', 'b'], ['cu', 'noun', 'c'], ['cubo', 'noun', 'b'], ['cubo', 'adjective', 'b'], ['cucchiaio', 'noun', 'b'], ['cuccia', 'noun', 'c'], ['cucciolo', 'noun', 'b'], ['cucina', 'noun', 'a'], ['cucinare', 'verb', 'a'], ['cucire', 'verb', 'b'], ['cucito', 'past_part', 'c'], ['cucito', 'adjective', 'c'], ['cucito', 'noun', 'c'], ['cucitura', 'noun', 'c'], ['cuffia', 'noun', 'b'], ['cugino', 'noun', 'b'], ['cui', 'pronoun', 'a'], ['cullare', 'verb', 'c'], ['culo', 'noun', 'a'], ['culto', 'noun', 'b'], ['cultura', 'noun', 'a'], ['culturale', 'adjective', 'a'], ['cumulo', 'noun', 'c'], ['cuocere', 'verb', 'b'], ['cuoco', 'noun', 'b'], ['cuore', 'noun', 'a'], ['cupo', 'adjective', 'b'], ['cupo', 'noun', 'b'], ['cura', 'noun', 'a'], ['curare', 'verb', 'a'], ['curiosare', 'verb', 'b'], ['curiosit', 'noun', 'b'], ['curioso', 'adjective', 'a'], ['curioso', 'noun', 'a'], ['curriculum', 'noun', 'b'], ['curva', 'noun', 'b'], ['curvo', 'adjective', 'b'], ['curvo', 'noun', 'b'], ['cuscino', 'noun', 'b'], ['custode', 'noun', 'b'], ['custode', 'adjective', 'b'], ['custodia', 'noun', 'b'], ['custodire', 'verb', 'b'], ['da', 'preposition', 'a'], ['dado', 'noun', 'c'], ['danese', 'adjective', 'c'], ['danese', 'noun', 'c'], ['dannato', 'past_part', 'b'], ['dannato', 'adjective', 'b'], ['dannato', 'noun', 'b'], ['danneggiare', 'verb', 'b'], ['danno', 'noun', 'a'], ['dannoso', 'adjective', 'c'], ['danza', 'noun', 'b'], ['dappertutto', 'adverb', 'b'], ['dare', 'verb', 'a'], ['dare', 'noun', 'a'], ['data', 'noun', 'a'], ['dato', 'past_part', 'a'], ['dato', 'adjective', 'a'], ['dato', 'noun', 'a'], ['dattero', 'noun', 'c'], ['davanti', 'adverb', 'a'], ['davanti', 'adjective', 'a'], ['davanti', 'noun', 'a'], ['davanzale', 'noun', 'c'], ['davvero', 'adverb', 'a'], ['dea', 'noun', 'b'], ['debito', 'noun', 'a'], ['debole', 'adjective', 'a'], ['debole', 'noun', 'a'], ['debolezza', 'noun', 'b'], ['decennio', 'noun', 'b'], ['decidere', 'verb', 'a'], ['decina', 'noun', 'a'], ['decisamente', 'adverb', 'b'], ['decisione', 'noun', 'a'], ['decisivo', 'adjective', 'b'], ['deciso', 'past_part', 'b'], ['deciso', 'adjective', 'b'], ['decorare', 'verb', 'b'], ['decorato', 'past_part', 'c'], ['decorato', 'adjective', 'c'], ['decorato', 'noun', 'c'], ['decorazione', 'noun', 'b'], ['decoroso', 'adjective', 'c'], ['decreto', 'noun', 'b'], ['dedica', 'noun', 'c'], ['dedicare', 'verb', 'a'], ['dedurre', 'verb', 'b'], ['deficiente', 'adjective', 'b'], ['deficiente', 'noun', 'b'], ['definire', 'verb', 'a'], ['definitivamente', 'adverb', 'b'], ['definitivo', 'adjective', 'a'], ['definitivo', 'noun', 'a'], ['definizione', 'noun', 'a'], ['deformare', 'verb', 'c'], ['deforme', 'adjective', 'c'], ['deforme', 'noun', 'c'], ['defunto', 'past_part', 'b'], ['defunto', 'adjective', 'b'], ['defunto', 'noun', 'b'], ['degno', 'adjective', 'b'], ['degradare', 'verb', 'b'], ['delegare', 'verb', 'b'], ['delegato', 'past_part', 'b'], ['delegato', 'adjective', 'b'], ['delegato', 'noun', 'b'], ['delegazione', 'noun', 'c'], ['delfino', 'noun', 'c'], ['delicatezza', 'noun', 'c'], ['delicato', 'adjective', 'b'], ['delicato', 'noun', 'b'], ['delinquente', 'pres_part', 'c'], ['delinquente', 'adjective', 'c'], ['delinquente', 'noun', 'c'], ['delirare', 'verb', 'c'], ['delirio', 'noun', 'b'], ['delitto', 'noun', 'b'], ['delizia', 'noun', 'c'], ['delizioso', 'adjective', 'b'], ['deludere', 'verb', 'b'], ['delusione', 'noun', 'b'], ['deluso', 'past_part', 'b'], ['deluso', 'adjective', 'b'], ['deluso', 'noun', 'b'], ['democratico', 'adjective', 'b'], ['democratico', 'noun', 'b'], ['democrazia', 'noun', 'a'], ['democristiano', 'adjective', 'c'], ['democristiano', 'noun', 'c'], ['demoralizzare', 'verb', 'c'], ['denaro', 'noun', 'a'], ['denominare', 'verb', 'b'], ['denso', 'adjective', 'b'], ['dente', 'noun', 'a'], ['dentiera', 'noun', 'c'], ['dentifricio', 'noun', 'c'], ['dentista', 'noun', 'b'], ['dentro', 'adverb', 'a'], ['dentro', 'preposition', 'a'], ['dentro', 'noun', 'a'], ['denuncia', 'noun', 'b'], ['denunciare', 'verb', 'a'], ['deodorante', 'pres_part', 'c'], ['deodorante', 'adjective', 'c'], ['deodorante', 'noun', 'c'], ['depilazione', 'noun', 'c'], ['deporre', 'verb', 'b'], ['depositare', 'verb', 'b'], ['deposito', 'noun', 'b'], ['deposizione', 'noun', 'b'], ['depressione', 'noun', 'b'], ['deprimere', 'verb', 'b'], ['depuratore', 'adjective', 'c'], ['depuratore', 'noun', 'c'], ['deputato', 'past_part', 'b'], ['deputato', 'adjective', 'b'], ['deputato', 'noun', 'b'], ['derivare', 'verb', 'a'], ['derubare', 'verb', 'c'], ['descrivere', 'verb', 'a'], ['descrizione', 'noun', 'a'], ['deserto', 'noun', 'b'], ['deserto', 'adjective', 'b'], ['desiderare', 'verb', 'a'], ['desiderio', 'noun', 'a'], ['design', 'noun', 'b'], ['dessert', 'noun', 'c'], ['destinare', 'verb', 'a'], ['destinazione', 'noun', 'b'], ['destino', 'noun', 'a'], ['destra', 'noun', 'a'], ['destro', 'adjective', 'a'], ['destro', 'noun', 'a'], ['detective', 'noun', 'b'], ['detenere', 'verb', 'b'], ['detenuto', 'past_part', 'c'], ['detenuto', 'adjective', 'c'], ['detenuto', 'noun', 'c'], ['determinare', 'verb', 'a'], ['determinato', 'past_part', 'a'], ['determinato', 'adjective', 'a'], ['determinazione', 'noun', 'b'], ['detersivo', 'adjective', 'c'], ['detersivo', 'noun', 'c'], ['dettagliato', 'past_part', 'b'], ['dettagliato', 'adjective', 'b'], ['dettaglio', 'noun', 'a'], ['dettare', 'verb', 'b'], ['dettato', 'past_part', 'c'], ['dettato', 'adjective', 'c'], ['dettato', 'noun', 'c'], ['devastare', 'verb', 'b'], ['deviare', 'verb', 'c'], ['deviazione', 'noun', 'c'], ['di', 'preposition', 'a'], ['di', 'noun', 'c'], ['diagnosi', 'noun', 'b'], ['dialetto', 'noun', 'a'], ['dialogare', 'verb', 'b'], ['dialogo', 'noun', 'a'], ['diamante', 'noun', 'a'], ['diametro', 'noun', 'b'], ['diario', 'noun', 'b'], ['diario', 'adjective', 'b'], ['diavolo', 'noun', 'a'], ['dibattito', 'noun', 'b'], ['dicembre', 'noun', 'a'], ['dichiarare', 'verb', 'a'], ['dichiarazione', 'noun', 'a'], ['diciotto', 'adjective', 'b'], ['diciotto', 'noun', 'b'], ['dieci', 'adjective', 'a'], ['dieci', 'noun', 'a'], ['diecimila', 'adjective', 'b'], ['diecimila', 'noun', 'b'], ['dieta', 'noun', 'b'], ['dietetico', 'adjective', 'c'], ['dietro', 'preposition', 'a'], ['dietro', 'adverb', 'a'], ['dietro', 'adjective', 'a'], ['dietro', 'noun', 'a'], ['difendere', 'verb', 'a'], ['difensore', 'adjective', 'b'], ['difensore', 'noun', 'b'], ['difesa', 'noun', 'a'], ['difetto', 'noun', 'b'], ['differente', 'pres_part', 'a'], ['differente', 'adjective', 'a'], ['differenza', 'noun', 'a'], ['difficile', 'adjective', 'a'], ['difficile', 'noun', 'a'], ['difficilmente', 'adverb', 'b'], ['difficolt', 'noun', 'a'], ['diffidente', 'adjective', 'c'], ['diffidente', 'noun', 'c'], ['diffidenza', 'noun', 'c'], ['diffondere', 'verb', 'a'], ['diffusione', 'noun', 'b'], ['diffuso', 'past_part', 'b'], ['diffuso', 'adjective', 'b'], ['diga', 'noun', 'c'], ['digestione', 'noun', 'c'], ['digestivo', 'adjective', 'c'], ['digestivo', 'noun', 'c'], ['digitale', 'adjective', 'b'], ['digitale', 'noun', 'b'], ['digiunare', 'verb', 'c'], ['dignit', 'noun', 'b'], ['diluvio', 'noun', 'c'], ['dimagrante', 'pres_part', 'c'], ['dimagrante', 'adjective', 'c'], ['dimensione', 'noun', 'a'], ['dimenticare', 'verb', 'a'], ['dimettere', 'verb', 'b'], ['dimezzare', 'verb', 'c'], ['diminuire', 'verb', 'b'], ['dimostrare', 'verb', 'a'], ['dimostrazione', 'noun', 'b'], ['dinamica', 'noun', 'b'], ['dinamico', 'adjective', 'b'], ['dinosauro', 'noun', 'c'], ['dintorno', 'adverb', 'b'], ['dintorno', 'noun', 'b'], ['dio', 'noun', 'a'], ['dipartimento', 'noun', 'b'], ['dipendente', 'pres_part', 'a'], ['dipendente', 'adjective', 'a'], ['dipendente', 'noun', 'a'], ['dipendenza', 'noun', 'b'], ['dipendere', 'verb', 'a'], ['dipingere', 'verb', 'b'], ['dipinto', 'past_part', 'b'], ['dipinto', 'adjective', 'b'], ['dipinto', 'noun', 'b'], ['diploma', 'noun', 'b'], ['diplomatico', 'adjective', 'b'], ['diplomatico', 'noun', 'b'], ['dire', 'verb', 'a'], ['dire', 'noun', 'a'], ['diretta', 'noun', 'b'], ['direttamente', 'adverb', 'a'], ['diretto', 'past_part', 'a'], ['diretto', 'adjective', 'a'], ['diretto', 'noun', 'a'], ['direttore', 'noun', 'a'], ['direttore', 'adjective', 'a'], ['direttrice', 'noun', 'c'], ['direzione', 'noun', 'a'], ['dirigente', 'adjective', 'b'], ['dirigente', 'noun', 'b'], ['dirigere', 'verb', 'a'], ['diritto', 'noun', 'a'], ['disagio', 'noun', 'b'], ['disastro', 'noun', 'b'], ['disattento', 'adjective', 'c'], ['discarica', 'noun', 'b'], ['discendere', 'verb', 'b'], ['discepolo', 'noun', 'b'], ['discesa', 'noun', 'b'], ['disciplina', 'noun', 'b'], ['disco', 'noun', 'a'], ['discordia', 'noun', 'c'], ['discorso', 'noun', 'a'], ['discoteca', 'noun', 'b'], ['discreto', 'adjective', 'b'], ['discreto', 'noun', 'b'], ['discussione', 'noun', 'a'], ['discusso', 'past_part', 'b'], ['discusso', 'adjective', 'b'], ['discutere', 'verb', 'a'], ['disegnare', 'verb', 'a'], ['disegno', 'noun', 'a'], ['diseredare', 'verb', 'c'], ['disgrazia', 'noun', 'b'], ['disinfettante', 'pres_part', 'c'], ['disinfettante', 'adjective', 'c'], ['disinfettare', 'verb', 'c'], ['disinteresse', 'noun', 'c'], ['disoccupazione', 'noun', 'b'], ['disonesto', 'adjective', 'c'], ['disonesto', 'noun', 'c'], ['disordinato', 'past_part', 'c'], ['disordinato', 'adjective', 'c'], ['disordine', 'noun', 'b'], ['dispari', 'adjective', 'c'], ['dispensa', 'noun', 'c'], ['disperare', 'verb', 'b'], ['disperato', 'past_part', 'b'], ['disperato', 'adjective', 'b'], ['disperazione', 'noun', 'b'], ['disperdere', 'verb', 'b'], ['dispetto', 'noun', 'b'], ['dispettoso', 'adjective', 'c'], ['dispiacere', 'verb', 'a'], ['disponibile', 'adjective', 'a'], ['disponibile', 'noun', 'a'], ['disponibilit', 'noun', 'b'], ['disporre', 'verb', 'a'], ['dispositivo', 'adjective', 'b'], ['dispositivo', 'noun', 'b'], ['disposizione', 'noun', 'a'], ['disprezzo', 'noun', 'b'], ['dissenso', 'noun', 'c'], ['distacco', 'noun', 'b'], ['distante', 'pres_part', 'b'], ['distante', 'adjective', 'b'], ['distante', 'adverb', 'b'], ['distanza', 'noun', 'a'], ['distendere', 'verb', 'b'], ['disteso', 'past_part', 'c'], ['disteso', 'adjective', 'c'], ['disteso', 'noun', 'c'], ['distinguere', 'verb', 'a'], ['distintivo', 'adjective', 'c'], ['distintivo', 'noun', 'c'], ['distinto', 'past_part', 'b'], ['distinto', 'adjective', 'b'], ['distinto', 'noun', 'b'], ['distinzione', 'noun', 'b'], ['distrarre', 'verb', 'b'], ['distratto', 'past_part', 'c'], ['distratto', 'adjective', 'c'], ['distrazione', 'noun', 'c'], ['distretto', 'noun', 'b'], ['distribuire', 'verb', 'a'], ['distributore', 'adjective', 'b'], ['distributore', 'noun', 'b'], ['distribuzione', 'noun', 'b'], ['distruggere', 'verb', 'a'], ['distrutto', 'past_part', 'c'], ['distrutto', 'adjective', 'c'], ['distruzione', 'noun', 'b'], ['disturbare', 'verb', 'b'], ['disturbo', 'noun', 'b'], ['disubbidiente', 'pres_part', 'c'], ['disubbidiente', 'adjective', 'c'], ['disubbidienza', 'noun', 'c'], ['disubbidire', 'verb', 'c'], ['dito', 'noun', 'a'], ['ditta', 'noun', 'b'], ['dittatura', 'noun', 'b'], ['divano', 'noun', 'a'], ['divano-letto', 'noun', 'c'], ['divenire', 'verb', 'a'], ['divenire', 'noun', 'a'], ['diventare', 'verb', 'a'], ['diversamente', 'adverb', 'b'], ['diversit', 'noun', 'b'], ['diverso', 'adjective', 'a'], ['diverso', 'adjective', 'a'], ['diverso', 'pronoun', 'a'], ['divertente', 'pres_part', 'a'], ['divertente', 'adjective', 'a'], ['divertimento', 'noun', 'b'], ['divertire', 'verb', 'a'], ['divertito', 'past_part', 'b'], ['divertito', 'adjective', 'b'], ['dividere', 'verb', 'a'], ['divieto', 'noun', 'b'], ['divinit', 'noun', 'b'], ['divino', 'adjective', 'b'], ['divino', 'noun', 'b'], ['divisa', 'noun', 'b'], ['divisione', 'noun', 'b'], ['divorare', 'verb', 'b'], ['divorziare', 'verb', 'c'], ['divorzio', 'noun', 'b'], ['dizionario', 'noun', 'b'], ['do', 'noun', 'c'], ['doccia', 'noun', 'b'], ['docciaschiuma', 'noun', 'c'], ['docente', 'pres_part', 'b'], ['docente', 'adjective', 'b'], ['docente', 'noun', 'b'], ['docile', 'adjective', 'c'], ['documentare', 'verb', 'b'], ['documentario', 'adjective', 'b'], ['documentario', 'noun', 'b'], ['documentazione', 'noun', 'b'], ['documento', 'noun', 'a'], ['dodici', 'adjective', 'a'], ['dodici', 'noun', 'a'], ['dogana', 'noun', 'c'], ['dolce', 'adjective', 'a'], ['dolce', 'noun', 'a'], ['dolce', 'adverb', 'a'], ['dolcezza', 'noun', 'b'], ['dolcificante', 'pres_part', 'c'], ['dolcificante', 'adjective', 'c'], ['dolcificante', 'noun', 'c'], ['dolciume', 'noun', 'c'], ['dolere', 'verb', 'c'], ['dolersi', 'verb', 'c'], ['dollaro', 'noun', 'a'], ['dolore', 'noun', 'a'], ['doloroso', 'adjective', 'b'], ['domanda', 'noun', 'a'], ['domandare', 'verb', 'a'], ['domani', 'adverb', 'a'], ['domani', 'noun', 'a'], ['domenica', 'noun', 'a'], ['domestica', 'noun', 'c'], ['domestico', 'adjective', 'b'], ['domestico', 'noun', 'b'], ['dominante', 'pres_part', 'b'], ['dominante', 'adjective', 'b'], ['dominante', 'noun', 'b'], ['dominare', 'verb', 'b'], ['dominio', 'noun', 'b'], ['don', 'noun', 'a'], ['donare', 'verb', 'b'], ['dondolare', 'verb', 'c'], ['donna', 'noun', 'a'], ['dono', 'noun', 'b'], ['dopo', 'adverb', 'a'], ['dopo', 'preposition', 'a'], ['dopo', 'conjunction', 'a'], ['dopo', 'adjective', 'a'], ['dopo', 'noun', 'a'], ['dopobarba', 'noun', 'c'], ['doppio', 'adjective', 'a'], ['doppio', 'noun', 'a'], ['doppio', 'adverb', 'a'], ['doppione', 'noun', 'c'], ['dorato', 'past_part', 'b'], ['dorato', 'adjective', 'b'], ['dorato', 'noun', 'b'], ['dormiglione', 'adjective', 'c'], ['dormiglione', 'noun', 'c'], ['dormire', 'verb', 'a'], ['dorso', 'noun', 'b'], ['dose', 'noun', 'b'], ['dotare', 'verb', 'b'], ['dotato', 'past_part', 'b'], ['dotato', 'adjective', 'b'], ['dote', 'noun', 'b'], ['dottore', 'noun', 'a'], ['dottoressa', 'noun', 'b'], ['dottrina', 'noun', 'b'], ['dove', 'adverb', 'a'], ['dove', 'conjunction', 'a'], ['dove', 'noun', 'a'], ['dovere', 'verb', 'a'], ['dovere', 'noun', 'a'], ['dovuto', 'past_part', 'b'], ['dovuto', 'adjective', 'b'], ['dovuto', 'noun', 'b'], ['dozzina', 'noun', 'b'], ['drago', 'noun', 'b'], ['dramma', 'noun', 'b'], ['drammatico', 'adjective', 'b'], ['dritto', 'adjective', 'b'], ['dritto', 'adverb', 'b'], ['dritto', 'noun', 'b'], ['drizzare', 'verb', 'c'], ['droga', 'noun', 'a'], ['drogare', 'verb', 'b'], ['drogato', 'past_part', 'c'], ['drogato', 'adjective', 'c'], ['drogato', 'noun', 'c'], ['dubbio', 'noun', 'a'], ['dubbio', 'adjective', 'b'], ['dubitare', 'verb', 'b'], ['dublinese', 'adjective', 'c'], ['dublinese', 'noun', 'c'], ['due', 'adjective', 'a'], ['due', 'noun', 'a'], ['duecento', 'adjective', 'b'], ['duecento', 'noun', 'b'], ['duello', 'noun', 'b'], ['duemila', 'adjective', 'b'], ['duemila', 'noun', 'b'], ['dunque', 'conjunction', 'a'], ['dunque', 'noun', 'a'], ['duomo', 'noun', 'c'], ['durante', 'pres_part', 'a'], ['durante', 'preposition', 'a'], ['durante', 'noun', 'a'], ['durare', 'verb', 'a'], ['durata', 'noun', 'a'], ['duro', 'adjective', 'a'], ['duro', 'noun', 'a'], ['duro', 'adverb', 'a'], ['e', 'noun', 'c'], ['e', 'conjunction', 'a'], ['ebbene', 'conjunction', 'b'], ['ebraico', 'adjective', 'b'], ['ebraico', 'noun', 'b'], ['ebreo', 'adjective', 'a'], ['ebreo', 'noun', 'a'], ['eccellente', 'pres_part', 'b'], ['eccellente', 'adjective', 'b'], ['eccellenza', 'noun', 'b'], ['eccessivo', 'adjective', 'b'], ['eccesso', 'noun', 'b'], ['eccetera', 'adverb', 'b'], ['eccezionale', 'adjective', 'b'], ['eccezione', 'noun', 'b'], ['eccitare', 'verb', 'b'], ['ecco', 'adverb', 'a'], ['eco', 'noun', 'b'], ['ecologico', 'adjective', 'b'], ['economia', 'noun', 'a'], ['economico', 'adjective', 'a'], ['economico', 'noun', 'a'], ['economista', 'noun', 'b'], ['edicola', 'noun', 'a'], ['edificio', 'noun', 'a'], ['editore', 'noun', 'a'], ['editore', 'adjective', 'a'], ['editoriale', 'adjective', 'b'], ['editoriale', 'noun', 'b'], ['edizione', 'noun', 'a'], ['educare', 'verb', 'b'], ['educativo', 'adjective', 'b'], ['educato', 'past_part', 'c'], ['educato', 'adjective', 'c'], ['educazione', 'noun', 'a'], ['effe', 'noun', 'c'], ['effettivamente', 'adverb', 'a'], ['effettivo', 'adjective', 'b'], ['effettivo', 'noun', 'b'], ['effetto', 'noun', 'a'], ['effettuare', 'verb', 'a'], ['efficace', 'adjective', 'b'], ['efficacia', 'noun', 'b'], ['efficiente', 'adjective', 'b'], ['efficienza', 'noun', 'b'], ['egiziano', 'adjective', 'c'], ['egiziano', 'noun', 'c'], ['egli', 'pronoun', 'a'], ['elaborare', 'verb', 'b'], ['elaborazione', 'noun', 'b'], ['elastico', 'adjective', 'b'], ['elastico', 'noun', 'b'], ['elegante', 'adjective', 'a'], ['eleganza', 'noun', 'b'], ['eleggere', 'verb', 'b'], ['elementare', 'adjective', 'a'], ['elemento', 'noun', 'a'], ['elemosina', 'noun', 'c'], ['elencare', 'verb', 'b'], ['elenco', 'noun', 'a'], ['elettorale', 'adjective', 'b'], ['elettore', 'noun', 'b'], ['elettricista', 'noun', 'c'], ['elettricit', 'noun', 'c'], ['elettrico', 'adjective', 'a'], ['elettrico', 'noun', 'a'], ['elettrodomestico', 'noun', 'c'], ['elettromagnetico', 'adjective', 'b'], ['elettrone', 'noun', 'b'], ['elettronico', 'adjective', 'a'], ['elevare', 'verb', 'b'], ['elevato', 'past_part', 'b'], ['elevato', 'adjective', 'b'], ['elezione', 'noun', 'b'], ['elica', 'noun', 'c'], ['elicottero', 'noun', 'c'], ['eliminare', 'verb', 'a'], ['eliminazione', 'noun', 'b'], ['elle', 'noun', 'c'], ['elmo', 'noun', 'c'], ['e-mail', 'noun', 'a'], ['emanare', 'verb', 'b'], ['emergenza', 'noun', 'b'], ['emergere', 'verb', 'a'], ['emettere', 'verb', 'b'], ['emigrazione', 'noun', 'c'], ['emiliano', 'adjective', 'c'], ['emiliano', 'noun', 'c'], ['emissione', 'noun', 'b'], ['emme', 'noun', 'c'], ['emmenthal', 'noun', 'c'], ['emo', 'noun', 'b'], ['emotivo', 'adjective', 'b'], ['emotivo', 'noun', 'b'], ['emozionante', 'pres_part', 'c'], ['emozionante', 'adjective', 'c'], ['emozionare', 'verb', 'b'], ['emozionato', 'past_part', 'c'], ['emozionato', 'adjective', 'c'], ['emozione', 'noun', 'a'], ['enciclopedia', 'noun', 'c'], ['energetico', 'adjective', 'b'], ['energetico', 'noun', 'b'], ['energia', 'noun', 'a'], ['enne', 'noun', 'c'], ['ennesimo', 'adjective', 'b'], ['enorme', 'adjective', 'a'], ['ente', 'noun', 'a'], ['entit', 'noun', 'b'], ['entrambi', 'pronoun', 'a'], ['entrambi', 'adjective', 'a'], ['entrare', 'verb', 'a'], ['entrare', 'noun', 'a'], ['entrata', 'noun', 'a'], ['entro', 'preposition', 'a'], ['entro', 'adverb', 'a'], ['entusiasmo', 'noun', 'b'], ['entusiasta', 'adjective', 'b'], ['entusiasta', 'noun', 'b'], ['epifania', 'noun', 'c'], ['episodio', 'noun', 'a'], ['epoca', 'noun', 'a'], ['eppure', 'conjunction', 'a'], ['equazione', 'noun', 'b'], ['equilibrio', 'noun', 'a'], ['equino', 'adjective', 'c'], ['equino', 'noun', 'c'], ['equipaggio', 'noun', 'c'], ['equivalere', 'verb', 'b'], ['equivoco', 'adjective', 'b'], ['equivoco', 'noun', 'b'], ['era', 'noun', 'a'], ['erba', 'noun', 'b'], ['erede', 'noun', 'b'], ['eredit', 'noun', 'b'], ['ereditare', 'verb', 'b'], ['ergastolo', 'noun', 'c'], ['ergere', 'verb', 'b'], ['ernia', 'noun', 'c'], ['eroe', 'noun', 'a'], ['eroina', 'noun', 'c'], ['erotico', 'adjective', 'b'], ['erotico', 'noun', 'b'], ['errare', 'verb', 'b'], ['erre', 'noun', 'c'], ['errore', 'noun', 'a'], ['esagerare', 'verb', 'b'], ['esagerato', 'past_part', 'b'], ['esagerato', 'adjective', 'b'], ['esagerato', 'noun', 'b'], ['esagerazione', 'noun', 'c'], ['esagono', 'noun', 'c'], ['esagono', 'adjective', 'c'], ['esaltare', 'verb', 'b'], ['esaltazione', 'noun', 'c'], ['esame', 'noun', 'a'], ['esaminare', 'verb', 'b'], ['esattamente', 'adverb', 'a'], ['esatto', 'adjective', 'a'], ['esatto', 'adverb', 'a'], ['esaurire', 'verb', 'b'], ['esca', 'noun', 'c'], ['eschimese', 'adjective', 'c'], ['eschimese', 'noun', 'c'], ['esclamare', 'verb', 'b'], ['esclamazione', 'noun', 'c'], ['escludere', 'verb', 'a'], ['esclusione', 'noun', 'b'], ['esclusivamente', 'adverb', 'b'], ['esclusivo', 'adjective', 'b'], ['escluso', 'past_part', 'b'], ['escluso', 'adjective', 'b'], ['escluso', 'noun', 'b'], ['esecutivo', 'adjective', 'b'], ['esecutivo', 'noun', 'b'], ['esecuzione', 'noun', 'b'], ['eseguire', 'verb', 'a'], ['esempio', 'noun', 'a'], ['esemplare', 'noun', 'b'], ['esemplare', 'adjective', 'b'], ['esercitare', 'verb', 'b'], ['esercito', 'noun', 'a'], ['esercizio', 'noun', 'a'], ['esibire', 'verb', 'b'], ['esigenza', 'noun', 'a'], ['esigere', 'verb', 'b'], ['esilio', 'noun', 'c'], ['esistente', 'pres_part', 'b'], ['esistente', 'adjective', 'b'], ['esistente', 'noun', 'b'], ['esistenza', 'noun', 'a'], ['esistere', 'verb', 'a'], ['esitare', 'verb', 'b'], ['esito', 'noun', 'b'], ['esordio', 'noun', 'b'], ['espansione', 'noun', 'b'], ['espellere', 'verb', 'b'], ['esperienza', 'noun', 'a'], ['esperimento', 'noun', 'b'], ['esperto', 'past_part', 'a'], ['esperto', 'adjective', 'a'], ['esperto', 'noun', 'a'], ['esplicito', 'adjective', 'b'], ['esplodere', 'verb', 'b'], ['esplorare', 'verb', 'b'], ['esplosione', 'noun', 'b'], ['esplosivo', 'adjective', 'b'], ['esplosivo', 'noun', 'b'], ['esponente', 'pres_part', 'b'], ['esponente', 'noun', 'b'], ['esporre', 'verb', 'a'], ['esposizione', 'noun', 'b'], ['espressione', 'noun', 'a'], ['espresso', 'past_part', 'c'], ['espresso', 'adjective', 'c'], ['espresso', 'noun', 'c'], ['esprimere', 'verb', 'a'], ['essa', 'pronoun', 'a'], ['esse', 'noun', 'c'], ['esse', 'pronoun', 'b'], ['essenza', 'noun', 'b'], ['essenziale', 'adjective', 'b'], ['essenziale', 'noun', 'b'], ['essenzialmente', 'adverb', 'b'], ['essere', 'verb', 'a'], ['essere', 'noun', 'a'], ['essi', 'pronoun', 'a'], ['esso', 'pronoun', 'a'], ['est', 'noun', 'b'], ['est', 'adjective', 'b'], ['estate', 'noun', 'a'], ['estendere', 'verb', 'b'], ['estensione', 'noun', 'b'], ['esterno', 'adjective', 'a'], ['esterno', 'noun', 'a'], ['estero', 'adjective', 'a'], ['estero', 'noun', 'a'], ['estetico', 'adjective', 'b'], ['estivo', 'adjective', 'b'], ['estone', 'adjective', 'c'], ['estone', 'noun', 'c'], ['estraneo', 'adjective', 'b'], ['estraneo', 'noun', 'b'], ['estrarre', 'verb', 'b'], ['estratto', 'past_part', 'b'], ['estratto', 'adjective', 'b'], ['estratto', 'noun', 'b'], ['estrazione', 'noun', 'b'], ['estremamente', 'adverb', 'b'], ['estremit', 'noun', 'b'], ['estremo', 'adjective', 'a'], ['estremo', 'noun', 'a'], ['et', 'noun', 'a'], ['eterno', 'adjective', 'b'], ['eterno', 'noun', 'b'], ['etica', 'noun', 'b'], ['etichetta', 'noun', 'b'], ['etico', 'adjective', 'b'], ['ettaro', 'noun', 'c'], ['etto', 'noun', 'c'], ['euro', 'noun', 'a'], ['europeo', 'adjective', 'a'], ['europeo', 'noun', 'a'], ['evadere', 'verb', 'c'], ['evaporare', 'verb', 'c'], ['evasione', 'noun', 'b'], ['evento', 'noun', 'a'], ['eventuale', 'adjective', 'a'], ['eventualmente', 'adverb', 'b'], ['evidente', 'adjective', 'a'], ['evidentemente', 'adverb', 'a'], ['evidenza', 'noun', 'b'], ['evidenziare', 'verb', 'b'], ['evidenziatore', 'adjective', 'c'], ['evidenziatore', 'noun', 'c'], ['evitare', 'verb', 'a'], ['evocare', 'verb', 'b'], ['evoluzione', 'noun', 'b'], ['ex', 'adjective', 'a'], ['ex', 'noun', 'a'], ['ex', 'preposition', 'a'], ['extra', 'adjective', 'b'], ['extra', 'noun', 'b'], ['fa', 'adverb', 'a'], ['fabbrica', 'noun', 'a'], ['fabbricare', 'verb', 'b'], ['fabbro', 'noun', 'c'], ['faccenda', 'noun', 'b'], ['faccia', 'noun', 'a'], ['facciata', 'noun', 'b'], ['facile', 'adjective', 'a'], ['facile', 'adverb', 'a'], ['facilit', 'noun', 'b'], ['facilitare', 'verb', 'b'], ['facilitazione', 'noun', 'c'], ['facilmente', 'adverb', 'a'], ['facolt', 'noun', 'b'], ['fagiano', 'noun', 'c'], ['falco', 'noun', 'c'], ['falegname', 'noun', 'c'], ['fallimento', 'noun', 'b'], ['fallire', 'verb', 'b'], ['fallito', 'past_part', 'b'], ['fallito', 'adjective', 'b'], ['fallito', 'noun', 'b'], ['falso', 'adjective', 'a'], ['falso', 'adverb', 'a'], ['falso', 'noun', 'a'], ['fama', 'noun', 'b'], ['fame', 'noun', 'a'], ['famiglia', 'noun', 'a'], ['familiare', 'adjective', 'a'], ['familiare', 'noun', 'a'], ['famoso', 'adjective', 'a'], ['fan', 'noun', 'b'], ['fanale', 'noun', 'c'], ['fanciulla', 'noun', 'b'], ['fanciullo', 'adjective', 'c'], ['fanciullo', 'noun', 'c'], ['fango', 'noun', 'b'], ['fangoso', 'adjective', 'c'], ['fantascienza', 'noun', 'b'], ['fantasia', 'noun', 'a'], ['fantasma', 'noun', 'b'], ['fantastico', 'adjective', 'a'], ['fantastico', 'noun', 'a'], ['fanteria', 'noun', 'c'], ['fantino', 'noun', 'c'], ['fantoccio', 'noun', 'c'], ['fare', 'verb', 'a'], ['fare', 'noun', 'a'], ['farfalla', 'noun', 'b'], ['farina', 'noun', 'b'], ['farmacia', 'noun', 'b'], ['farmaco', 'noun', 'b'], ['faro', 'noun', 'c'], ['fascia', 'noun', 'a'], ['fasciatoio', 'noun', 'c'], ['fascicolo', 'noun', 'b'], ['fascino', 'noun', 'b'], ['fascio', 'noun', 'b'], ['fascismo', 'noun', 'b'], ['fascista', 'adjective', 'b'], ['fascista', 'noun', 'b'], ['fase', 'noun', 'a'], ['fastidio', 'noun', 'a'], ['fastidioso', 'adjective', 'b'], ['fata', 'noun', 'b'], ['fatica', 'noun', 'a'], ['faticare', 'verb', 'b'], ['faticoso', 'adjective', 'b'], ['fatto', 'noun', 'a'], ['fattore', 'noun', 'a'], ['fattoria', 'noun', 'b'], ['fattura', 'noun', 'b'], ['fatturato', 'past_part', 'b'], ['fatturato', 'adjective', 'b'], ['fatturato', 'noun', 'b'], ['fauna', 'noun', 'c'], ['fava', 'noun', 'c'], ['favola', 'noun', 'b'], ['favoloso', 'adjective', 'b'], ['favore', 'noun', 'a'], ['favorevole', 'adjective', 'b'], ['favorire', 'verb', 'b'], ['fax', 'noun', 'b'], ['fazzoletto', 'noun', 'b'], ['febbraio', 'noun', 'a'], ['febbre', 'noun', 'b'], ['fecondare', 'verb', 'c'], ['fede', 'noun', 'a'], ['fedele', 'adjective', 'b'], ['fedele', 'noun', 'b'], ['fedelt', 'noun', 'b'], ['federa', 'noun', 'c'], ['federale', 'adjective', 'b'], ['federale', 'noun', 'b'], ['fegato', 'noun', 'b'], ['felice', 'adjective', 'a'], ['felicit', 'noun', 'b'], ['felino', 'noun', 'c'], ['felino', 'adjective', 'c'], ['felpa', 'noun', 'c'], ['femmina', 'noun', 'a'], ['femminile', 'adjective', 'a'], ['femminile', 'noun', 'a'], ['fenomeno', 'noun', 'a'], ['feria', 'noun', 'b'], ['feriale', 'adjective', 'c'], ['ferie', 'noun', 'c'], ['ferire', 'verb', 'b'], ['ferita', 'noun', 'a'], ['ferito', 'past_part', 'b'], ['ferito', 'adjective', 'b'], ['ferito', 'noun', 'b'], ['fermaglio', 'noun', 'c'], ['fermare', 'verb', 'a'], ['fermo', 'adjective', 'a'], ['feroce', 'adjective', 'b'], ['ferragosto', 'noun', 'c'], ['ferramenta', 'noun', 'c'], ['ferro', 'noun', 'a'], ['ferrovia', 'noun', 'b'], ['ferroviario', 'adjective', 'b'], ['ferroviere', 'noun', 'c'], ['fertilizzante', 'pres_part', 'c'], ['fertilizzante', 'adjective', 'c'], ['fertilizzante', 'noun', 'c'], ['fessura', 'noun', 'c'], ['festa', 'noun', 'a'], ['festeggiare', 'verb', 'a'], ['festival', 'noun', 'b'], ['festivo', 'adjective', 'c'], ['fetta', 'noun', 'b'], ['fiaba', 'noun', 'b'], ['fiala', 'noun', 'c'], ['fiamma', 'noun', 'b'], ['fiammifero', 'noun', 'c'], ['fiammifero', 'adjective', 'c'], ['fianco', 'noun', 'a'], ['fiatare', 'verb', 'c'], ['fiato', 'noun', 'b'], ['fibbia', 'noun', 'c'], ['fibra', 'noun', 'b'], ['ficcare', 'verb', 'b'], ['fiction', 'noun', 'b'], ['fidanzamento', 'noun', 'c'], ['fidanzarsi', 'verb', 'b'], ['fidanzata', 'noun', 'b'], ['fidanzato', 'past_part', 'b'], ['fidanzato', 'adjective', 'b'], ['fidanzato', 'noun', 'b'], ['fidarsi', 'verb', 'a'], ['fiducia', 'noun', 'a'], ['fiducioso', 'adjective', 'c'], ['fieno', 'noun', 'c'], ['fiera', 'noun', 'b'], ['fiero', 'adjective', 'b'], ['figlia', 'noun', 'a'], ['figliastro', 'noun', 'c'], ['figlio', 'noun', 'a'], ['figura', 'noun', 'a'], ['figurare', 'verb', 'a'], ['figurina', 'noun', 'c'], ['fila', 'noun', 'a'], ['filante', 'pres_part', 'c'], ['filante', 'adjective', 'c'], ['filante', 'noun', 'c'], ['filare', 'verb', 'b'], ['filastrocca', 'noun', 'c'], ['file', 'noun', 'a'], ['filetto', 'noun', 'c'], ['film', 'noun', 'a'], ['filmato', 'past_part', 'b'], ['filmato', 'adjective', 'b'], ['filmato', 'noun', 'b'], ['filo', 'noun', 'a'], ['filosofia', 'noun', 'a'], ['filosofico', 'adjective', 'b'], ['filosofo', 'noun', 'b'], ['filtrare', 'verb', 'b'], ['filtro', 'noun', 'b'], ['finale', 'adjective', 'a'], ['finale', 'noun', 'a'], ['finalit', 'noun', 'b'], ['finalmente', 'adverb', 'a'], ['finanza', 'noun', 'b'], ['finanziamento', 'noun', 'b'], ['finanziare', 'verb', 'b'], ['finanziario', 'adjective', 'a'], ['finanziatore', 'adjective', 'c'], ['finanziatore', 'noun', 'c'], ['finch', 'conjunction', 'a'], ['fine', 'noun', 'a'], ['fine', 'adjective', 'b'], ['finestra', 'noun', 'a'], ['finestrino', 'noun', 'b'], ['fingere', 'verb', 'a'], ['finimondo', 'noun', 'c'], ['finire', 'verb', 'a'], ['finire', 'noun', 'a'], ['finito', 'past_part', 'b'], ['finito', 'adjective', 'b'], ['finlandese', 'adjective', 'c'], ['finlandese', 'noun', 'c'], ['fino', 'preposition', 'a'], ['fino', 'adverb', 'a'], ['finocchio', 'noun', 'c'], ['finora', 'adverb', 'b'], ['finta', 'noun', 'b'], ['finto', 'past_part', 'a'], ['finto', 'adjective', 'a'], ['fiocco', 'noun', 'c'], ['fionda', 'noun', 'c'], ['fioraio', 'noun', 'c'], ['fiore', 'noun', 'a'], ['fiorentino', 'adjective', 'b'], ['fiorentino', 'noun', 'b'], ['fiorito', 'past_part', 'c'], ['fiorito', 'adjective', 'c'], ['firma', 'noun', 'a'], ['firmare', 'verb', 'a'], ['fiscale', 'adjective', 'b'], ['fiscale', 'noun', 'b'], ['fisicamente', 'adverb', 'b'], ['fisico', 'adjective', 'a'], ['fisico', 'noun', 'a'], ['fissare', 'verb', 'a'], ['fisso', 'adjective', 'a'], ['fisso', 'adverb', 'a'], ['fisso', 'noun', 'a'], ['fitto', 'past_part', 'b'], ['fitto', 'adjective', 'b'], ['fitto', 'adverb', 'b'], ['fitto', 'noun', 'b'], ['fiume', 'noun', 'a'], ['fiuto', 'noun', 'c'], ['flash', 'noun', 'b'], ['flauto', 'noun', 'c'], ['flessibile', 'adjective', 'b'], ['flessibile', 'noun', 'b'], ['flora', 'noun', 'c'], ['fluido', 'adjective', 'b'], ['fluido', 'noun', 'b'], ['fluoro', 'noun', 'c'], ['flusso', 'noun', 'b'], ['foca', 'noun', 'c'], ['focaccia', 'noun', 'c'], ['fodera', 'noun', 'c'], ['foderare', 'verb', 'c'], ['foglia', 'noun', 'b'], ['foglio', 'noun', 'a'], ['fogna', 'noun', 'c'], ['folla', 'noun', 'b'], ['folle', 'adjective', 'b'], ['folle', 'noun', 'b'], ['follia', 'noun', 'b'], ['fondamentale', 'adjective', 'a'], ['fondamentale', 'noun', 'a'], ['fondamentalmente', 'adverb', 'b'], ['fondamento', 'noun', 'b'], ['fondare', 'verb', 'a'], ['fondatore', 'noun', 'b'], ['fondazione', 'noun', 'b'], ['fondere', 'verb', 'b'], ['fondo', 'adjective', 'loc-comando'], ['fondo', 'noun', 'loc-comando'], ['fondo', 'adverb', 'loc-comando'], ['fontana', 'noun', 'b'], ['fontanella', 'noun', 'c'], ['fonte', 'noun', 'a'], ['forare', 'verb', 'b'], ['forbice', 'noun', 'c'], ['forchetta', 'noun', 'c'], ['forcina', 'noun', 'c'], ['foresta', 'noun', 'b'], ['forestale', 'adjective', 'c'], ['forestale', 'noun', 'c'], ['forfora', 'noun', 'c'], ['forma', 'noun', 'a'], ['formaggino', 'noun', 'c'], ['formaggio', 'noun', 'b'], ['formale', 'adjective', 'b'], ['formare', 'verb', 'a'], ['formato', 'past_part', 'b'], ['formato', 'adjective', 'b'], ['formato', 'noun', 'b'], ['formazione', 'noun', 'a'], ['formula', 'noun', 'a'], ['formulare', 'verb', 'b'], ['fornace', 'noun', 'c'], ['fornaio', 'noun', 'c'], ['fornello', 'noun', 'b'], ['fornire', 'verb', 'a'], ['fornitore', 'adjective', 'b'], ['fornitore', 'noun', 'b'], ['forno', 'noun', 'b'], ['foro', 'noun', 'b'], ['forse', 'adverb', 'a'], ['forse', 'noun', 'a'], ['forte', 'adjective', 'a'], ['forte', 'adverb', 'a'], ['forte', 'noun', 'a'], ['fortemente', 'adverb', 'b'], ['fortuna', 'noun', 'a'], ['fortunatamente', 'adverb', 'b'], ['fortunato', 'adjective', 'b'], ['forum', 'noun', 'b'], ['forza', 'noun', 'a'], ['forzare', 'verb', 'b'], ['fosforescente', 'adjective', 'c'], ['fossa', 'noun', 'b'], ['fossetta', 'noun', 'c'], ['fosso', 'noun', 'c'], ['foto', 'noun', 'a'], ['fotografare', 'verb', 'b'], ['fotografia', 'noun', 'a'], ['fotografico', 'adjective', 'b'], ['fotografo', 'noun', 'b'], ['fottere', 'verb', 'b'], ['foulard', 'noun', 'c'], ['fra', 'preposition', 'a'], ['fracasso', 'noun', 'c'], ['fragile', 'adjective', 'b'], ['frammento', 'noun', 'b'], ['francamente', 'adverb', 'b'], ['francese', 'adjective', 'a'], ['francese', 'noun', 'a'], ['francobollo', 'noun', 'c'], ['frangia', 'noun', 'c'], ['frase', 'noun', 'a'], ['fratello', 'noun', 'a'], ['frazione', 'noun', 'b'], ['freccia', 'noun', 'b'], ['freddezza', 'noun', 'c'], ['freddo', 'adjective', 'a'], ['freddo', 'noun', 'a'], ['fregare', 'verb', 'a'], ['frenare', 'verb', 'b'], ['frenetico', 'adjective', 'b'], ['freno', 'noun', 'b'], ['frequentare', 'verb', 'a'], ['frequente', 'adjective', 'b'], ['frequenza', 'noun', 'b'], ['fresco', 'adjective', 'a'], ['fresco', 'noun', 'a'], ['fretta', 'noun', 'a'], ['frigo', 'noun', 'b'], ['frigorifero', 'adjective', 'b'], ['frigorifero', 'noun', 'b'], ['fringuello', 'noun', 'c'], ['frittata', 'noun', 'c'], ['fritto', 'past_part', 'c'], ['fritto', 'adjective', 'c'], ['fritto', 'noun', 'c'], ['friulano', 'adjective', 'c'], ['friulano', 'noun', 'c'], ['fronte', 'noun', 'a'], ['frontiera', 'noun', 'b'], ['frugare', 'verb', 'b'], ['frumento', 'noun', 'c'], ['fruscio', 'noun', 'c'], ['frusta', 'noun', 'c'], ['frutta', 'noun', 'b'], ['fruttivendolo', 'noun', 'c'], ['frutto', 'noun', 'a'], ['fucile', 'noun', 'b'], ['fuga', 'noun', 'a'], ['fuggire', 'verb', 'a'], ['fulmine', 'noun', 'b'], ['fumare', 'verb', 'a'], ['fumetto', 'noun', 'b'], ['fumo', 'noun', 'a'], ['fumo', 'adjective', 'a'], ['fune', 'noun', 'c'], ['funerale', 'noun', 'b'], ['funerale', 'adjective', 'b'], ['fungo', 'noun', 'b'], ['funzionale', 'adjective', 'b'], ['funzionale', 'noun', 'b'], ['funzionamento', 'noun', 'b'], ['funzionare', 'verb', 'a'], ['funzionario', 'noun', 'b'], ['funzione', 'noun', 'a'], ['fuoco', 'noun', 'loc-comando'], ['fuori', 'adverb', 'a'], ['fuori', 'preposition', 'a'], ['fuori', 'noun', 'a'], ['fuori', 'adjective', 'a'], ['furbo', 'adjective', 'b'], ['furbo', 'noun', 'b'], ['furfante', 'noun', 'c'], ['furgone', 'noun', 'b'], ['furia', 'noun', 'b'], ['furioso', 'adjective', 'b'], ['furto', 'noun', 'b'], ['fusione', 'noun', 'b'], ['fuso', 'past_part', 'b'], ['fuso', 'adjective', 'b'], ['fuso', 'noun', 'b'], ['futuro', 'adjective', 'a'], ['futuro', 'noun', 'a'], ['gabbia', 'noun', 'b'], ['galassia', 'noun', 'b'], ['galeotto', 'noun', 'c'], ['galera', 'noun', 'b'], ['galleggiare', 'verb', 'c'], ['galleria', 'noun', 'b'], ['gallese', 'adjective', 'c'], ['gallese', 'noun', 'c'], ['galletta', 'noun', 'c'], ['gallina', 'noun', 'b'], ['gallo', 'noun', 'c'], ['gamba', 'noun', 'a'], ['gambero', 'noun', 'c'], ['gambo', 'noun', 'c'], ['ganascia', 'noun', 'c'], ['gancio', 'noun', 'c'], ['gara', 'noun', 'a'], ['garage', 'noun', 'b'], ['garantire', 'verb', 'a'], ['garanzia', 'noun', 'b'], ['garbo', 'noun', 'c'], ['gargarismo', 'noun', 'c'], ['garofano', 'noun', 'c'], ['garza', 'noun', 'c'], ['gas', 'noun', 'a'], ['gasolio', 'noun', 'c'], ['gassosa', 'noun', 'c'], ['gastronomia', 'noun', 'c'], ['gatto', 'noun', 'a'], ['gavetta', 'noun', 'c'], ['gay', 'adjective', 'b'], ['gay', 'noun', 'b'], ['gazza', 'noun', 'c'], ['gelateria', 'noun', 'c'], ['gelatina', 'noun', 'c'], ['gelato', 'past_part', 'b'], ['gelato', 'adjective', 'b'], ['gelato', 'noun', 'b'], ['gelido', 'adjective', 'b'], ['gelo', 'noun', 'c'], ['gelosia', 'noun', 'b'], ['geloso', 'adjective', 'b'], ['gelsomino', 'noun', 'c'], ['gemello', 'adjective', 'b'], ['gemello', 'noun', 'b'], ['gemma', 'noun', 'c'], ['gene', 'noun', 'b'], ['generale', 'adjective', 'a'], ['generale', 'noun', 'a'], ['generalmente', 'adverb', 'b'], ['generare', 'verb', 'a'], ['generazione', 'noun', 'a'], ['genere', 'noun', 'a'], ['generico', 'adjective', 'b'], ['generico', 'noun', 'b'], ['generosit', 'noun', 'c'], ['generoso', 'adjective', 'b'], ['genetico', 'adjective', 'b'], ['gengiva', 'noun', 'c'], ['geniale', 'adjective', 'b'], ['genio', 'noun', 'b'], ['genitore', 'noun', 'a'], ['gennaio', 'noun', 'a'], ['genovese', 'adjective', 'c'], ['genovese', 'noun', 'c'], ['gente', 'noun', 'a'], ['gentile', 'adjective', 'a'], ['gentile', 'noun', 'a'], ['genuino', 'adjective', 'c'], ['geografico', 'adjective', 'b'], ['geografo', 'noun', 'c'], ['geometra', 'noun', 'c'], ['geometria', 'noun', 'c'], ['geometrico', 'adjective', 'c'], ['gesso', 'noun', 'b'], ['gestione', 'noun', 'a'], ['gestire', 'verb', 'a'], ['gesto', 'noun', 'a'], ['gestore', 'noun', 'b'], ['gettare', 'verb', 'a'], ['gettone', 'noun', 'c'], ['ghiaccio', 'noun', 'b'], ['ghiacciolo', 'noun', 'c'], ['ghianda', 'noun', 'c'], ['ghiro', 'noun', 'c'], ['gi', 'noun', 'c'], ['gi', 'adverb', 'a'], ['giacca', 'noun', 'a'], ['giacere', 'verb', 'b'], ['giaguaro', 'noun', 'c'], ['giallo', 'adjective', 'a'], ['giallo', 'noun', 'a'], ['giapponese', 'adjective', 'a'], ['giapponese', 'noun', 'a'], ['giardinaggio', 'noun', 'c'], ['giardiniera', 'noun', 'c'], ['giardino', 'noun', 'a'], ['gigante', 'noun', 'b'], ['gigante', 'adjective', 'b'], ['gigantesco', 'adjective', 'b'], ['giglio', 'noun', 'b'], ['ginnastica', 'noun', 'b'], ['ginocchio', 'noun', 'a'], ['giocare', 'verb', 'a'], ['giocatore', 'noun', 'a'], ['giocattolo', 'noun', 'b'], ['gioco', 'noun', 'a'], ['gioia', 'noun', 'a'], ['gioiello', 'noun', 'b'], ['gioioso', 'adjective', 'c'], ['giordano', 'adjective', 'c'], ['giordano', 'noun', 'c'], ['giornale', 'noun', 'a'], ['giornale', 'adjective', 'a'], ['giornalino', 'noun', 'c'], ['giornalista', 'noun', 'a'], ['giornata', 'noun', 'a'], ['giorno', 'noun', 'a'], ['giostra', 'noun', 'c'], ['giovane', 'adjective', 'a'], ['giovane', 'noun', 'a'], ['giovanile', 'adjective', 'b'], ['gioved', 'noun', 'b'], ['giovent', 'noun', 'b'], ['giovinezza', 'noun', 'b'], ['giraffa', 'noun', 'c'], ['girare', 'verb', 'a'], ['giravite', 'noun', 'c'], ['giretto', 'noun', 'c'], ['giro', 'noun', 'a'], ['gironzolare', 'verb', 'c'], ['girotondo', 'noun', 'c'], ['gita', 'noun', 'b'], ['gi', 'adverb', 'a'], ['gi', 'adjective', 'a'], ['giubba', 'noun', 'c'], ['giubbotto', 'noun', 'c'], ['giudicare', 'verb', 'a'], ['giudice', 'noun', 'a'], ['giudiziario', 'adjective', 'b'], ['giudizio', 'noun', 'a'], ['giugno', 'noun', 'a'], ['giungere', 'verb', 'a'], ['giungla', 'noun', 'c'], ['giuramento', 'noun', 'b'], ['giurare', 'verb', 'a'], ['giuria', 'noun', 'c'], ['giuridico', 'adjective', 'b'], ['giustamente', 'adverb', 'b'], ['giustificare', 'verb', 'b'], ['giustizia', 'noun', 'a'], ['giusto', 'adjective', 'a'], ['giusto', 'noun', 'a'], ['giusto', 'adverb', 'a'], ['gli', 'pronoun', 'a'], ['glicine', 'noun', 'c'], ['global', 'adjective', 'b'], ['global', 'noun', 'b'], ['globale', 'adjective', 'b'], ['gloria', 'noun', 'b'], ['gnocco', 'noun', 'c'], ['gnomo', 'noun', 'c'], ['goal', 'noun', 'b'], ['gobbo', 'adjective', 'c'], ['gobbo', 'noun', 'c'], ['goccia', 'noun', 'b'], ['godere', 'verb', 'a'], ['gola', 'noun', 'b'], ['goloso', 'adjective', 'c'], ['gomito', 'noun', 'b'], ['gomitolo', 'noun', 'c'], ['gomma', 'noun', 'b'], ['gonfiare', 'verb', 'b'], ['gonfio', 'adjective', 'b'], ['gonfio', 'noun', 'b'], ['gonna', 'noun', 'b'], ['gorgonzola', 'noun', 'c'], ['gorilla', 'noun', 'c'], ['gossip', 'noun', 'b'], ['governare', 'verb', 'b'], ['governatore', 'noun', 'b'], ['governo', 'noun', 'a'], ['gradino', 'noun', 'b'], ['gradire', 'verb', 'b'], ['grado', 'noun', 'a'], ['graffiare', 'verb', 'c'], ['graffio', 'noun', 'c'], ['grafico', 'adjective', 'b'], ['grafico', 'noun', 'b'], ['grammatica', 'noun', 'b'], ['grammo', 'noun', 'b'], ['grana', 'noun', 'c'], ['granaio', 'noun', 'c'], ['granchio', 'noun', 'c'], ['grande', 'adjective', 'a'], ['grande', 'noun', 'a'], ['grandezza', 'noun', 'b'], ['grandine', 'noun', 'c'], ['grandioso', 'adjective', 'b'], ['grano', 'noun', 'b'], ['granturco', 'noun', 'c'], ['grappa', 'noun', 'c'], ['grasso', 'adjective', 'a'], ['grasso', 'noun', 'a'], ['gratis', 'adverb', 'b'], ['gratis', 'adjective', 'b'], ['grattare', 'verb', 'b'], ['grattugiato', 'past_part', 'c'], ['grattugiato', 'adjective', 'c'], ['gratuito', 'adjective', 'b'], ['grave', 'adjective', 'a'], ['grave', 'noun', 'a'], ['grave', 'adverb', 'a'], ['gravidanza', 'noun', 'b'], ['gravit', 'noun', 'b'], ['grazie', 'exclamation', 'a'], ['grazie', 'noun', 'a'], ['grazioso', 'adjective', 'c'], ['greco', 'adjective', 'a'], ['greco', 'noun', 'a'], ['grembiule', 'noun', 'c'], ['gridare', 'verb', 'a'], ['grido', 'noun', 'b'], ['grigio', 'adjective', 'a'], ['grigio', 'noun', 'a'], ['griglia', 'noun', 'c'], ['grinza', 'noun', 'c'], ['grissino', 'noun', 'c'], ['grossista', 'noun', 'c'], ['grosso', 'adjective', 'a'], ['grosso', 'noun', 'a'], ['grotta', 'noun', 'b'], ['gru', 'noun', 'c'], ['gruppo', 'noun', 'a'], ['guadagnare', 'verb', 'a'], ['guadagno', 'noun', 'b'], ['guaio', 'noun', 'b'], ['guaire', 'verb', 'c'], ['guancia', 'noun', 'b'], ['guanciale', 'noun', 'c'], ['guanciale', 'adjective', 'c'], ['guanto', 'noun', 'b'], ['guardare', 'verb', 'a'], ['guardaroba', 'noun', 'c'], ['guardia', 'noun', 'a'], ['guarire', 'verb', 'b'], ['guarnizione', 'noun', 'c'], ['guasto', 'noun', 'c'], ['guerra', 'noun', 'a'], ['guerriero', 'noun', 'b'], ['guerriero', 'adjective', 'b'], ['gufo', 'noun', 'c'], ['guida', 'noun', 'a'], ['guidare', 'verb', 'a'], ['guidatore', 'noun', 'c'], ['guinzaglio', 'noun', 'c'], ['gustare', 'verb', 'b'], ['gusto', 'noun', 'a'], ['gustoso', 'adjective', 'c'], ['hamburger', 'noun', 'c'], ['hobby', 'noun', 'b'], ['home', 'noun', 'b'], ['hotel', 'noun', 'b'], ['hyperlink', 'noun', 'b'], ['i', 'noun', 'c'], ['i', 'determiner', 'b'], ['icona', 'noun', 'b'], ['ics', 'noun', 'c'], ['idea', 'noun', 'a'], ['ideale', 'adjective', 'a'], ['ideale', 'noun', 'a'], ['ideare', 'verb', 'b'], ['identico', 'adjective', 'b'], ['identico', 'noun', 'b'], ['identificare', 'verb', 'a'], ['identificazione', 'noun', 'b'], ['identit', 'noun', 'a'], ['ideologia', 'noun', 'b'], ['ideologico', 'adjective', 'b'], ['idiota', 'adjective', 'a'], ['idiota', 'noun', 'a'], ['idraulico', 'adjective', 'b'], ['idraulico', 'noun', 'b'], ['idrico', 'adjective', 'b'], ['idrogeno', 'noun', 'b'], ['ieri', 'adverb', 'a'], ['ieri', 'noun', 'a'], ['igiene', 'noun', 'c'], ['ignorante', 'pres_part', 'b'], ['ignorante', 'adjective', 'b'], ['ignorante', 'noun', 'b'], ['ignoranza', 'noun', 'b'], ['ignorare', 'verb', 'a'], ['ignoto', 'adjective', 'b'], ['ignoto', 'noun', 'b'], ['il', 'determiner', 'a'], ['il', 'pronoun', 'a'], ['illecito', 'adjective', 'b'], ['illecito', 'noun', 'b'], ['illegale', 'adjective', 'b'], ['illegale', 'noun', 'b'], ['illegittimo', 'adjective', 'c'], ['illegittimo', 'noun', 'c'], ['illudere', 'verb', 'b'], ['illuminare', 'verb', 'b'], ['illuminato', 'past_part', 'b'], ['illuminato', 'adjective', 'b'], ['illuminato', 'noun', 'b'], ['illusione', 'noun', 'b'], ['illustrare', 'verb', 'b'], ['illustre', 'adjective', 'b'], ['imballare', 'verb', 'c'], ['imbarazzante', 'pres_part', 'b'], ['imbarazzante', 'adjective', 'b'], ['imbarazzato', 'past_part', 'b'], ['imbarazzato', 'adjective', 'b'], ['imbarazzo', 'noun', 'b'], ['imbattersi', 'verb', 'b'], ['imbecille', 'adjective', 'b'], ['imbecille', 'noun', 'b'], ['imbiancare', 'verb', 'c'], ['imbianchino', 'noun', 'c'], ['imbottigliare', 'verb', 'c'], ['imbrogliare', 'verb', 'c'], ['imbroglio', 'noun', 'c'], ['imbuto', 'noun', 'c'], ['imitare', 'verb', 'b'], ['immaginare', 'verb', 'a'], ['immaginare', 'noun', 'a'], ['immaginario', 'adjective', 'b'], ['immaginario', 'noun', 'b'], ['immaginazione', 'noun', 'b'], ['immagine', 'noun', 'a'], ['immaturo', 'adjective', 'c'], ['immediatamente', 'adverb', 'a'], ['immediato', 'adjective', 'b'], ['immediato', 'noun', 'b'], ['immenso', 'adjective', 'b'], ['immenso', 'noun', 'b'], ['immergere', 'verb', 'b'], ['immigrato', 'past_part', 'b'], ['immigrato', 'adjective', 'b'], ['immigrato', 'noun', 'b'], ['immobile', 'adjective', 'a'], ['immobile', 'noun', 'a'], ['immobiliare', 'adjective', 'b'], ['immobiliare', 'noun', 'b'], ['immondizia', 'noun', 'c'], ['impallidire', 'verb', 'c'], ['imparare', 'verb', 'a'], ['impastare', 'verb', 'c'], ['impatto', 'noun', 'b'], ['impaziente', 'adjective', 'c'], ['impaziente', 'noun', 'c'], ['impazzire', 'verb', 'b'], ['impedire', 'verb', 'a'], ['impegnare', 'verb', 'a'], ['impegnativo', 'adjective', 'b'], ['impegnato', 'past_part', 'c'], ['impegnato', 'adjective', 'c'], ['impegno', 'noun', 'a'], ['imperare', 'verb', 'b'], ['imperatore', 'noun', 'b'], ['imperiale', 'adjective', 'b'], ['imperiale', 'noun', 'b'], ['impermeabile', 'adjective', 'c'], ['impermeabile', 'noun', 'c'], ['impero', 'noun', 'b'], ['impero', 'adjective', 'b'], ['impianto', 'noun', 'a'], ['impiegare', 'verb', 'a'], ['impiegato', 'past_part', 'b'], ['impiegato', 'adjective', 'b'], ['impiegato', 'noun', 'b'], ['impiego', 'noun', 'b'], ['implicare', 'verb', 'b'], ['imporre', 'verb', 'a'], ['importante', 'pres_part', 'a'], ['importante', 'adjective', 'a'], ['importante', 'noun', 'a'], ['importanza', 'noun', 'a'], ['importare', 'verb', 'a'], ['importo', 'noun', 'b'], ['impossibile', 'adjective', 'a'], ['impossibile', 'noun', 'a'], ['impostare', 'verb', 'b'], ['impostazione', 'noun', 'b'], ['impreciso', 'adjective', 'c'], ['imprenditore', 'noun', 'b'], ['impresa', 'noun', 'a'], ['impressionante', 'pres_part', 'b'], ['impressionante', 'adjective', 'b'], ['impressionare', 'verb', 'b'], ['impressione', 'noun', 'a'], ['imprevisto', 'adjective', 'b'], ['imprevisto', 'noun', 'b'], ['imprigionare', 'verb', 'c'], ['improbabile', 'adjective', 'b'], ['impronta', 'noun', 'b'], ['improvvisamente', 'adverb', 'b'], ['improvvisare', 'verb', 'b'], ['improvviso', 'adjective', 'a'], ['improvviso', 'noun', 'a'], ['imprudente', 'adjective', 'c'], ['imprudente', 'noun', 'c'], ['impulsivo', 'adjective', 'c'], ['impulsivo', 'noun', 'c'], ['impulso', 'noun', 'b'], ['imputata', 'noun', 'b'], ['imputato', 'past_part', 'a'], ['imputato', 'adjective', 'a'], ['imputato', 'noun', 'a'], ['in', 'preposition', 'a'], ['inaspettato', 'adjective', 'b'], ['inaugurare', 'verb', 'b'], ['incamminare', 'verb', 'c'], ['incantare', 'verb', 'c'], ['incapace', 'adjective', 'b'], ['incapace', 'noun', 'b'], ['incapacit', 'noun', 'b'], ['incaricare', 'verb', 'b'], ['incarico', 'noun', 'b'], ['incartare', 'verb', 'c'], ['incassare', 'verb', 'b'], ['incasso', 'noun', 'c'], ['incastrare', 'verb', 'b'], ['incatenare', 'verb', 'c'], ['incazzarsi', 'verb', 'b'], ['incendio', 'noun', 'b'], ['incertezza', 'noun', 'b'], ['incerto', 'adjective', 'b'], ['incerto', 'noun', 'b'], ['inchiesta', 'noun', 'b'], ['inchiodare', 'verb', 'c'], ['incidente', 'noun', 'a'], ['incidere', 'verb', 'b'], ['incinta', 'adjective', 'b'], ['incitare', 'verb', 'c'], ['incivile', 'adjective', 'c'], ['incivile', 'noun', 'c'], ['includere', 'verb', 'b'], ['incluso', 'past_part', 'b'], ['incluso', 'adjective', 'b'], ['incluso', 'noun', 'b'], ['incollare', 'verb', 'b'], ['incominciare', 'verb', 'b'], ['incompleto', 'adjective', 'c'], ['incomprensibile', 'adjective', 'b'], ['inconsolabile', 'adjective', 'c'], ['incontentabile', 'adjective', 'c'], ['incontrare', 'verb', 'a'], ['incontro', 'noun', 'a'], ['incontro', 'adverb', 'b'], ['incoraggiare', 'verb', 'b'], ['incoronare', 'verb', 'c'], ['incorreggibile', 'adjective', 'c'], ['incredibile', 'adjective', 'a'], ['incremento', 'noun', 'b'], ['incrinare', 'verb', 'c'], ['incrociare', 'verb', 'b'], ['incrocio', 'noun', 'c'], ['incubo', 'noun', 'b'], ['incurabile', 'adjective', 'c'], ['incurabile', 'noun', 'c'], ['incuriosire', 'verb', 'b'], ['indagare', 'verb', 'b'], ['indagine', 'noun', 'a'], ['indescrivibile', 'adjective', 'c'], ['indiano', 'adjective', 'b'], ['indiano', 'noun', 'b'], ['indicare', 'verb', 'a'], ['indicazione', 'noun', 'a'], ['indice', 'noun', 'a'], ['indice', 'adjective', 'a'], ['indietreggiare', 'verb', 'c'], ['indietro', 'adverb', 'a'], ['indietro', 'adjective', 'a'], ['indietro', 'loc-comando', 'a'], ['indifeso', 'adjective', 'c'], ['indifferente', 'adjective', 'b'], ['indifferente', 'noun', 'b'], ['indifferenza', 'noun', 'b'], ['indigestione', 'noun', 'c'], ['indimenticabile', 'adjective', 'c'], ['indipendente', 'adjective', 'b'], ['indipendente', 'noun', 'b'], ['indipendentemente', 'adverb', 'b'], ['indipendenza', 'noun', 'b'], ['indiretto', 'adjective', 'b'], ['indirizzare', 'verb', 'b'], ['indirizzo', 'noun', 'a'], ['indisciplinato', 'adjective', 'c'], ['indispensabile', 'adjective', 'b'], ['indispensabile', 'noun', 'b'], ['individuale', 'adjective', 'b'], ['individuare', 'verb', 'a'], ['individuo', 'noun', 'a'], ['individuo', 'adjective', 'a'], ['indizio', 'noun', 'b'], ['indossare', 'verb', 'a'], ['indovinare', 'verb', 'b'], ['indovinello', 'noun', 'c'], ['indubbiamente', 'adverb', 'b'], ['indumento', 'noun', 'c'], ['indurre', 'verb', 'b'], ['industria', 'noun', 'a'], ['industriale', 'adjective', 'a'], ['industriale', 'noun', 'a'], ['inedito', 'adjective', 'b'], ['inefficace', 'adjective', 'c'], ['inerte', 'adjective', 'c'], ['inesistente', 'adjective', 'b'], ['inesperienza', 'noun', 'c'], ['inesperto', 'adjective', 'c'], ['inevitabile', 'adjective', 'b'], ['inevitabile', 'noun', 'b'], ['inevitabilmente', 'adverb', 'b'], ['infame', 'adjective', 'c'], ['infame', 'noun', 'c'], ['infantile', 'adjective', 'b'], ['infanzia', 'noun', 'b'], ['infarto', 'noun', 'b'], ['infatti', 'conjunction', 'a'], ['infatti', 'adverb', 'a'], ['infedele', 'adjective', 'c'], ['infedele', 'noun', 'c'], ['infelice', 'adjective', 'b'], ['infelice', 'noun', 'b'], ['inferiore', 'adjective', 'a'], ['infermiera', 'noun', 'b'], ['infermiere', 'noun', 'c'], ['inferno', 'noun', 'b'], ['inferno', 'adjective', 'b'], ['infezione', 'noun', 'b'], ['infilare', 'verb', 'a'], ['infine', 'adverb', 'a'], ['infinito', 'adjective', 'a'], ['infinito', 'noun', 'a'], ['influenza', 'noun', 'b'], ['influenzare', 'verb', 'b'], ['informare', 'verb', 'a'], ['informatica', 'noun', 'b'], ['informatico', 'adjective', 'b'], ['informatico', 'noun', 'b'], ['informativo', 'adjective', 'b'], ['informazione', 'noun', 'a'], ['infradito', 'adjective', 'c'], ['infradito', 'noun', 'c'], ['infrastruttura', 'noun', 'b'], ['infuriare', 'verb', 'b'], ['infuso', 'past_part', 'c'], ['infuso', 'adjective', 'c'], ['infuso', 'noun', 'c'], ['ingannare', 'verb', 'b'], ['inganno', 'noun', 'b'], ['ingegnere', 'noun', 'b'], ['ingegneria', 'noun', 'b'], ['ingelosire', 'verb', 'c'], ['ingenuo', 'adjective', 'b'], ['ingenuo', 'noun', 'b'], ['ingessare', 'verb', 'c'], ['ingiusto', 'adjective', 'b'], ['ingiusto', 'noun', 'b'], ['inglese', 'adjective', 'a'], ['inglese', 'noun', 'a'], ['ingoiare', 'verb', 'b'], ['ingorgo', 'noun', 'c'], ['ingrandire', 'verb', 'c'], ['ingrassare', 'verb', 'b'], ['ingrediente', 'noun', 'b'], ['ingresso', 'noun', 'a'], ['iniezione', 'noun', 'c'], ['iniziale', 'adjective', 'a'], ['iniziale', 'noun', 'a'], ['inizialmente', 'adverb', 'b'], ['iniziare', 'verb', 'a'], ['iniziativa', 'noun', 'a'], ['inizio', 'noun', 'a'], ['innamorarsi', 'verb', 'a'], ['innamorato', 'past_part', 'b'], ['innamorato', 'adjective', 'b'], ['innamorato', 'noun', 'b'], ['innanzitutto', 'adverb', 'b'], ['innervosire', 'verb', 'c'], ['innocente', 'adjective', 'b'], ['innocente', 'noun', 'b'], ['innocuo', 'adjective', 'b'], ['innovativo', 'adjective', 'b'], ['innovazione', 'noun', 'b'], ['inoltre', 'adverb', 'a'], ['inquadrare', 'verb', 'b'], ['inquietante', 'pres_part', 'b'], ['inquietante', 'adjective', 'b'], ['inquinamento', 'noun', 'b'], ['inquinare', 'verb', 'c'], ['inquinato', 'past_part', 'c'], ['inquinato', 'adjective', 'c'], ['insalata', 'noun', 'b'], ['insegna', 'noun', 'b'], ['insegnamento', 'noun', 'b'], ['insegnante', 'pres_part', 'a'], ['insegnante', 'adjective', 'a'], ['insegnante', 'noun', 'a'], ['insegnare', 'verb', 'a'], ['inseguire', 'verb', 'b'], ['inseparabile', 'adjective', 'c'], ['inseparabile', 'noun', 'c'], ['inserimento', 'noun', 'b'], ['inserire', 'verb', 'a'], ['insetticida', 'adjective', 'c'], ['insetto', 'noun', 'b'], ['insieme', 'adverb', 'a'], ['insieme', 'noun', 'a'], ['insinuare', 'verb', 'b'], ['insistere', 'verb', 'a'], ['insoddisfatto', 'adjective', 'c'], ['insolito', 'adjective', 'b'], ['insolito', 'noun', 'b'], ['insomma', 'adverb', 'a'], ['insopportabile', 'adjective', 'b'], ['insospettire', 'verb', 'c'], ['installare', 'verb', 'b'], ['insuccesso', 'noun', 'c'], ['insultare', 'verb', 'b'], ['insulto', 'noun', 'b'], ['intanto', 'adverb', 'a'], ['intasare', 'verb', 'c'], ['intatto', 'adjective', 'b'], ['integrale', 'adjective', 'b'], ['integrale', 'noun', 'b'], ['integrare', 'verb', 'b'], ['integrazione', 'noun', 'b'], ['intellettuale', 'adjective', 'b'], ['intellettuale', 'noun', 'b'], ['intelligente', 'adjective', 'a'], ['intelligenza', 'noun', 'b'], ['intendere', 'verb', 'a'], ['intensit', 'noun', 'b'], ['intenso', 'adjective', 'a'], ['intento', 'noun', 'b'], ['intenzione', 'noun', 'a'], ['interagire', 'verb', 'b'], ['interamente', 'adverb', 'b'], ['interazione', 'noun', 'b'], ['intercettare', 'verb', 'b'], ['intercettazione', 'noun', 'b'], ['interessante', 'pres_part', 'a'], ['interessante', 'adjective', 'a'], ['interessare', 'verb', 'a'], ['interessato', 'past_part', 'b'], ['interessato', 'adjective', 'b'], ['interessato', 'noun', 'b'], ['interesse', 'noun', 'a'], ['interiore', 'adjective', 'b'], ['interiore', 'noun', 'b'], ['interlocutore', 'noun', 'b'], ['internazionale', 'adjective', 'a'], ['internazionale', 'noun', 'a'], ['internet', 'noun', 'a'], ['interno', 'adjective', 'a'], ['interno', 'noun', 'a'], ['intero', 'adjective', 'a'], ['intero', 'noun', 'a'], ['interpretare', 'verb', 'a'], ['interpretazione', 'noun', 'b'], ['interprete', 'noun', 'b'], ['interrogare', 'verb', 'b'], ['interrogativo', 'adjective', 'b'], ['interrogativo', 'noun', 'b'], ['interrogatorio', 'adjective', 'b'], ['interrogatorio', 'noun', 'b'], ['interrogazione', 'noun', 'c'], ['interrompere', 'verb', 'a'], ['interruttore', 'noun', 'c'], ['interruzione', 'noun', 'b'], ['intervallo', 'noun', 'b'], ['intervenire', 'verb', 'a'], ['intervento', 'noun', 'a'], ['intervista', 'noun', 'a'], ['intesa', 'noun', 'b'], ['intestare', 'verb', 'b'], ['intestino', 'noun', 'c'], ['intimidire', 'verb', 'c'], ['intimit', 'noun', 'b'], ['intimo', 'adjective', 'b'], ['intimo', 'noun', 'b'], ['intitolare', 'verb', 'b'], ['intonaco', 'noun', 'c'], ['intorno', 'adverb', 'a'], ['intorno', 'preposition', 'a'], ['intorno', 'adjective', 'a'], ['intorno', 'noun', 'a'], ['intraprendere', 'verb', 'b'], ['intravedere', 'verb', 'b'], ['intrecciare', 'verb', 'b'], ['introdurre', 'verb', 'a'], ['introduzione', 'noun', 'b'], ['intuire', 'verb', 'b'], ['intuizione', 'noun', 'b'], ['inutile', 'adjective', 'a'], ['invadente', 'pres_part', 'c'], ['invadente', 'adjective', 'c'], ['invadente', 'noun', 'c'], ['invadere', 'verb', 'b'], ['invasione', 'noun', 'b'], ['invecchiare', 'verb', 'b'], ['invece', 'adverb', 'a'], ['inventare', 'verb', 'a'], ['invenzione', 'noun', 'b'], ['invernale', 'adjective', 'b'], ['invernale', 'noun', 'b'], ['inverno', 'noun', 'a'], ['investimento', 'noun', 'b'], ['investire', 'verb', 'a'], ['inviare', 'verb', 'a'], ['inviato', 'past_part', 'b'], ['inviato', 'adjective', 'b'], ['inviato', 'noun', 'b'], ['invidiare', 'verb', 'b'], ['invidioso', 'adjective', 'c'], ['invidioso', 'noun', 'c'], ['invincibile', 'adjective', 'c'], ['invisibile', 'adjective', 'b'], ['invisibile', 'noun', 'b'], ['invitare', 'verb', 'a'], ['invitato', 'past_part', 'b'], ['invitato', 'adjective', 'b'], ['invitato', 'noun', 'b'], ['invito', 'noun', 'b'], ['invocare', 'verb', 'b'], ['inzuppare', 'verb', 'c'], ['io', 'pronoun', 'a'], ['ionico', 'adjective', 'c'], ['ipotesi', 'noun', 'a'], ['ipotizzare', 'verb', 'b'], ['ippopotamo', 'noun', 'c'], ['ipsilon', 'noun', 'c'], ['ira', 'noun', 'b'], ['irlandese', 'adjective', 'b'], ['irlandese', 'noun', 'b'], ['ironia', 'noun', 'b'], ['ironico', 'adjective', 'b'], ['irriconoscibile', 'adjective', 'c'], ['irritare', 'verb', 'b'], ['iscritto', 'past_part', 'b'], ['iscritto', 'adjective', 'b'], ['iscritto', 'noun', 'b'], ['iscrivere', 'verb', 'a'], ['iscrizione', 'noun', 'b'], ['islamico', 'adjective', 'b'], ['islamico', 'noun', 'b'], ['islandese', 'adjective', 'c'], ['islandese', 'noun', 'c'], ['isola', 'noun', 'a'], ['isolare', 'verb', 'b'], ['isolato', 'past_part', 'b'], ['isolato', 'adjective', 'b'], ['isolato', 'noun', 'b'], ['ispettore', 'noun', 'b'], ['ispirare', 'verb', 'a'], ['ispirazione', 'noun', 'b'], ['israeliano', 'adjective', 'c'], ['israeliano', 'noun', 'c'], ['istante', 'noun', 'a'], ['istanza', 'noun', 'b'], ['istintivo', 'adjective', 'c'], ['istinto', 'noun', 'b'], ['istituto', 'noun', 'a'], ['istituzionale', 'adjective', 'b'], ['istituzione', 'noun', 'a'], ['istruttivo', 'adjective', 'c'], ['istruttore', 'noun', 'c'], ['istruzione', 'noun', 'a'], ['italiano', 'adjective', 'a'], ['italiano', 'noun', 'a'], ['iugoslavo', 'adjective', 'c'], ['iugoslavo', 'noun', 'c'], ['jeans', 'noun', 'b'], ['karat', 'noun', 'c'], ['ketchup', 'noun', 'c'], ['killer', 'noun', 'b'], ['killer', 'adjective', 'b'], ['kit', 'noun', 'c'], ['kiwi', 'noun', 'c'], ['l', 'adverb', 'a'], ['la', 'determiner', 'a'], ['la', 'pronoun', 'a'], ['labbro', 'noun', 'a'], ['labirinto', 'noun', 'c'], ['laboratorio', 'noun', 'a'], ['laborioso', 'adjective', 'c'], ['lacca', 'noun', 'c'], ['lacca', 'adjective', 'c'], ['laccio', 'noun', 'c'], ['lacrima', 'noun', 'a'], ['laddove', 'adverb', 'b'], ['laddove', 'conjunction', 'b'], ['ladro', 'noun', 'b'], ['laggi', 'adverb', 'b'], ['lago', 'noun', 'a'], ['laico', 'adjective', 'b'], ['laico', 'noun', 'b'], ['lama', 'noun', 'b'], ['lamentare', 'verb', 'a'], ['lamentela', 'noun', 'c'], ['lametta', 'noun', 'c'], ['lamiera', 'noun', 'c'], ['lampada', 'noun', 'b'], ['lampadario', 'noun', 'c'], ['lampo', 'noun', 'b'], ['lampo', 'adjective', 'b'], ['lampo', 'noun', 'b'], ['lana', 'noun', 'b'], ['lancetta', 'noun', 'c'], ['lanciare', 'verb', 'a'], ['lancio', 'noun', 'b'], ['lanterna', 'noun', 'c'], ['lapis', 'noun', 'c'], ['lardo', 'noun', 'c'], ['larghezza', 'noun', 'c'], ['largo', 'adjective', 'a'], ['largo', 'noun', 'a'], ['largo', 'adverb', 'a'], ['lasagna', 'noun', 'c'], ['lasciare', 'verb', 'a'], ['lass', 'adverb', 'b'], ['lastra', 'noun', 'b'], ['laterale', 'adjective', 'b'], ['laterale', 'noun', 'b'], ['latino', 'adjective', 'b'], ['latino', 'noun', 'b'], ['lato', 'noun', 'a'], ['latta', 'noun', 'c'], ['lattante', 'pres_part', 'c'], ['lattante', 'adjective', 'c'], ['lattante', 'noun', 'c'], ['latte', 'noun', 'a'], ['latte', 'adjective', 'a'], ['latteria', 'noun', 'c'], ['lattina', 'noun', 'c'], ['lattuga', 'noun', 'c'], ['laurea', 'noun', 'b'], ['laureare', 'verb', 'b'], ['laureato', 'past_part', 'b'], ['laureato', 'adjective', 'b'], ['laureato', 'noun', 'b'], ['lava', 'noun', 'c'], ['lavabo', 'noun', 'c'], ['lavagna', 'noun', 'c'], ['lavagna', 'adjective', 'c'], ['lavanda', 'noun', 'c'], ['lavanderia', 'noun', 'c'], ['lavandino', 'noun', 'c'], ['lavapiatti', 'noun', 'c'], ['lavare', 'verb', 'a'], ['lavastoviglie', 'noun', 'c'], ['lavatrice', 'noun', 'b'], ['lavello', 'noun', 'c'], ['lavorare', 'verb', 'a'], ['lavorativo', 'adjective', 'b'], ['lavoratore', 'adjective', 'a'], ['lavoratore', 'noun', 'a'], ['lavorazione', 'noun', 'b'], ['lavoro', 'noun', 'a'], ['laziale', 'adjective', 'c'], ['laziale', 'noun', 'c'], ['le', 'determiner', 'a'], ['le', 'pronoun', 'a'], ['le', 'pronoun', 'a'], ['leader', 'noun', 'b'], ['lealt', 'noun', 'c'], ['lebbra', 'noun', 'c'], ['leccare', 'verb', 'b'], ['leccio', 'noun', 'c'], ['lecito', 'adjective', 'b'], ['lecito', 'noun', 'b'], ['lega', 'noun', 'b'], ['legale', 'adjective', 'a'], ['legale', 'noun', 'a'], ['legame', 'noun', 'b'], ['legare', 'verb', 'a'], ['legato', 'past_part', 'a'], ['legato', 'adjective', 'a'], ['legato', 'noun', 'a'], ['legge', 'noun', 'a'], ['leggenda', 'noun', 'b'], ['leggere', 'verb', 'a'], ['leggermente', 'adverb', 'b'], ['leggero', 'adjective', 'a'], ['leggero', 'adverb', 'a'], ['leggero', 'noun', 'a'], ['legislativo', 'adjective', 'b'], ['legittimo', 'adjective', 'b'], ['legna', 'noun', 'c'], ['legno', 'noun', 'a'], ['legume', 'noun', 'c'], ['lei', 'pronoun', 'a'], ['lentamente', 'adverb', 'a'], ['lente', 'noun', 'c'], ['lenticchia', 'noun', 'c'], ['lentiggine', 'noun', 'c'], ['lento', 'adjective', 'a'], ['lento', 'noun', 'a'], ['lento', 'adverb', 'a'], ['lenza', 'noun', 'c'], ['lenzuolo', 'noun', 'b'], ['leone', 'noun', 'b'], ['leonessa', 'noun', 'c'], ['leopardo', 'noun', 'c'], ['lepre', 'noun', 'c'], ['lesione', 'noun', 'b'], ['lessare', 'verb', 'c'], ['lessema', 'noun', 'b'], ['lettera', 'noun', 'a'], ['letterale', 'adjective', 'c'], ['letteralmente', 'adverb', 'b'], ['letterario', 'adjective', 'b'], ['letteratura', 'noun', 'a'], ['letto', 'noun', 'a'], ['lettone', 'noun', 'c'], ['lettore', 'noun', 'a'], ['lettura', 'noun', 'a'], ['leva', 'noun', 'b'], ['levare', 'verb', 'a'], ['levare', 'noun', 'a'], ['lezione', 'noun', 'a'], ['l', 'adverb', 'a'], ['li', 'pronoun', 'a'], ['libanese', 'adjective', 'b'], ['libanese', 'noun', 'b'], ['liberale', 'adjective', 'b'], ['liberale', 'noun', 'b'], ['liberamente', 'adverb', 'b'], ['liberare', 'verb', 'a'], ['liberazione', 'noun', 'b'], ['libero', 'adjective', 'a'], ['libero', 'noun', 'a'], ['libert', 'noun', 'a'], ['libico', 'adjective', 'c'], ['libico', 'noun', 'c'], ['libraio', 'noun', 'c'], ['libreria', 'noun', 'b'], ['libretto', 'noun', 'b'], ['libro', 'noun', 'a'], ['licenza', 'noun', 'b'], ['licenziamento', 'noun', 'c'], ['licenziare', 'verb', 'b'], ['liceo', 'noun', 'b'], ['lido', 'noun', 'c'], ['lieto', 'adjective', 'b'], ['lieve', 'adjective', 'b'], ['lievito', 'noun', 'c'], ['ligure', 'adjective', 'c'], ['ligure', 'noun', 'c'], ['lima', 'noun', 'c'], ['limare', 'verb', 'c'], ['limitare', 'verb', 'a'], ['limitato', 'past_part', 'b'], ['limitato', 'adjective', 'b'], ['limite', 'noun', 'a'], ['limite', 'adjective', 'a'], ['limonata', 'noun', 'c'], ['limone', 'noun', 'b'], ['limone', 'adjective', 'b'], ['linea', 'noun', 'a'], ['lineare', 'adjective', 'b'], ['lineare', 'noun', 'b'], ['linfa', 'noun', 'b'], ['lingerie', 'noun', 'c'], ['lingua', 'noun', 'a'], ['linguaggio', 'noun', 'a'], ['linguistica', 'noun', 'b'], ['linguistico', 'adjective', 'b'], ['linguistico', 'noun', 'b'], ['link', 'noun', 'b'], ['liquido', 'adjective', 'a'], ['liquido', 'noun', 'a'], ['liquore', 'noun', 'c'], ['lira', 'noun', 'a'], ['lirico', 'adjective', 'b'], ['lisbonese', 'adjective', 'c'], ['lisbonese', 'noun', 'c'], ['liscio', 'adjective', 'b'], ['liscio', 'noun', 'b'], ['lista', 'noun', 'a'], ['lite', 'noun', 'b'], ['litigare', 'verb', 'a'], ['litigio', 'noun', 'b'], ['litro', 'noun', 'b'], ['lituano', 'adjective', 'c'], ['lituano', 'noun', 'c'], ['live', 'adjective', 'b'], ['livello', 'noun', 'a'], ['lo', 'determiner', 'a'], ['lo', 'pronoun', 'a'], ['locale', 'adjective', 'a'], ['locale', 'noun', 'a'], ['locale', 'noun', 'a'], ['localit', 'noun', 'b'], ['locanda', 'noun', 'c'], ['locazione', 'noun', 'b'], ['locomotiva', 'noun', 'c'], ['logica', 'noun', 'b'], ['logico', 'adjective', 'b'], ['logico', 'noun', 'b'], ['logoro', 'past_part', 'c'], ['logoro', 'adjective', 'c'], ['lombardo', 'adjective', 'b'], ['lombardo', 'noun', 'b'], ['londinese', 'adjective', 'c'], ['londinese', 'noun', 'c'], ['lontananza', 'noun', 'b'], ['lontano', 'adjective', 'a'], ['lontano', 'adverb', 'a'], ['lontano', 'noun', 'a'], ['lonza', 'noun', 'c'], ['look', 'noun', 'b'], ['loro', 'pronoun', 'a'], ['loro', 'adjective', 'a'], ['lotta', 'noun', 'a'], ['lottare', 'verb', 'b'], ['lozione', 'noun', 'c'], ['lucano', 'adjective', 'c'], ['lucano', 'noun', 'c'], ['luccicare', 'verb', 'c'], ['lucciola', 'noun', 'c'], ['luce', 'noun', 'a'], ['lucente', 'pres_part', 'c'], ['lucente', 'adjective', 'c'], ['lucente', 'noun', 'c'], ['lucertola', 'noun', 'c'], ['lucidare', 'verb', 'c'], ['lucido', 'adjective', 'b'], ['lucido', 'noun', 'b'], ['luglio', 'noun', 'a'], ['lui', 'pronoun', 'a'], ['lumaca', 'noun', 'c'], ['luminoso', 'adjective', 'b'], ['luna', 'noun', 'a'], ['luned', 'noun', 'a'], ['lunghezza', 'noun', 'b'], ['lungo', 'adjective', 'a'], ['lungo', 'preposition', 'a'], ['lungo', 'noun', 'a'], ['luogo', 'noun', 'a'], ['lupo', 'noun', 'a'], ['lussemburghese', 'adjective', 'c'], ['lussemburghese', 'noun', 'c'], ['lusso', 'noun', 'b'], ['lutto', 'noun', 'b'], ['ma', 'conjunction', 'a'], ['ma', 'noun', 'a'], ['maccherone', 'noun', 'c'], ['macchia', 'noun', 'a'], ['macchina', 'noun', 'a'], ['macchinista', 'noun', 'c'], ['macedone', 'adjective', 'c'], ['macedone', 'noun', 'c'], ['macedonia', 'noun', 'c'], ['maceria', 'noun', 'b'], ['macinare', 'verb', 'c'], ['madonna', 'noun', 'b'], ['madonna', 'exclamation', 'b'], ['madre', 'noun', 'a'], ['madrileno', 'adjective', 'c'], ['madrileno', 'noun', 'c'], ['madrileno', 'adjective', 'c'], ['madrileno', 'noun', 'c'], ['madrina', 'noun', 'c'], ['maestra', 'noun', 'b'], ['maestranza', 'noun', 'c'], ['maestro', 'noun', 'a'], ['maestro', 'adjective', 'a'], ['mafia', 'noun', 'b'], ['mafioso', 'adjective', 'b'], ['mafioso', 'noun', 'b'], ['magari', 'exclamation', 'a'], ['magari', 'conjunction', 'a'], ['magari', 'adverb', 'a'], ['magazzino', 'noun', 'b'], ['maggio', 'noun', 'a'], ['maggioranza', 'noun', 'a'], ['maggiorenne', 'adjective', 'c'], ['maggiorenne', 'noun', 'c'], ['maggiormente', 'adverb', 'b'], ['magia', 'noun', 'b'], ['magico', 'adjective', 'a'], ['magistrato', 'noun', 'b'], ['magistratura', 'noun', 'b'], ['maglia', 'noun', 'a'], ['maglietta', 'noun', 'b'], ['magnetico', 'adjective', 'b'], ['magnifico', 'adjective', 'b'], ['mago', 'noun', 'b'], ['mago', 'adjective', 'b'], ['magro', 'adjective', 'b'], ['magro', 'noun', 'b'], ['mah', 'exclamation', 'b'], ['mai', 'adverb', 'a'], ['maiale', 'noun', 'b'], ['maionese', 'noun', 'c'], ['mais', 'noun', 'c'], ['maiuscola', 'noun', 'c'], ['malato', 'adjective', 'a'], ['malato', 'noun', 'a'], ['malattia', 'noun', 'a'], ['malaugurio', 'noun', 'c'], ['malavita', 'noun', 'c'], ['male', 'adverb', 'a'], ['male', 'exclamation', 'a'], ['male', 'noun', 'a'], ['maledetto', 'past_part', 'b'], ['maledetto', 'adjective', 'b'], ['maledetto', 'noun', 'b'], ['maledizione', 'noun', 'b'], ['maledizione', 'exclamation', 'b'], ['maleducato', 'adjective', 'c'], ['maleducato', 'noun', 'c'], ['maleducazione', 'noun', 'c'], ['malgrado', 'noun', 'b'], ['malgrado', 'adverb', 'b'], ['malgrado', 'conjunction', 'b'], ['malgrado', 'preposition', 'b'], ['malinconia', 'noun', 'b'], ['malinteso', 'adjective', 'c'], ['malinteso', 'noun', 'c'], ['malizia', 'noun', 'c'], ['maltempo', 'noun', 'c'], ['maltese', 'adjective', 'c'], ['maltese', 'noun', 'c'], ['maltrattamento', 'noun', 'c'], ['maltrattare', 'verb', 'c'], ['malva', 'noun', 'c'], ['malvagio', 'adjective', 'b'], ['malvagio', 'noun', 'b'], ['mamma', 'noun', 'a'], ['mammella', 'noun', 'c'], ['mammifero', 'noun', 'c'], ['manager', 'noun', 'b'], ['mancanza', 'noun', 'a'], ['mancare', 'verb', 'a'], ['mancato', 'past_part', 'b'], ['mancato', 'adjective', 'b'], ['mancino', 'adjective', 'c'], ['mancino', 'noun', 'c'], ['manco', 'adjective', 'b'], ['manco', 'adverb', 'b'], ['mandare', 'verb', 'a'], ['mandarino', 'noun', 'c'], ['mandarino', 'adjective', 'c'], ['mandato', 'past_part', 'b'], ['mandato', 'adjective', 'b'], ['mandato', 'noun', 'b'], ['mandorla', 'noun', 'c'], ['mandorlo', 'noun', 'c'], ['manganello', 'noun', 'c'], ['mangiare', 'verb', 'a'], ['mangime', 'noun', 'c'], ['mania', 'noun', 'b'], ['maniaco', 'adjective', 'c'], ['maniaco', 'noun', 'c'], ['manica', 'noun', 'b'], ['manico', 'noun', 'b'], ['maniera', 'noun', 'a'], ['manifestare', 'verb', 'a'], ['manifestazione', 'noun', 'a'], ['manifesto', 'noun', 'b'], ['mano', 'noun', 'a'], ['manodopera', 'noun', 'c'], ['manoscritto', 'adjective', 'b'], ['manoscritto', 'noun', 'b'], ['manovale', 'noun', 'c'], ['manovra', 'noun', 'b'], ['mantello', 'noun', 'b'], ['mantenere', 'verb', 'a'], ['manuale', 'adjective', 'b'], ['manuale', 'noun', 'b'], ['manuale', 'noun', 'b'], ['manutenzione', 'noun', 'b'], ['manzo', 'noun', 'c'], ['mappa', 'noun', 'b'], ['marca', 'noun', 'b'], ['marcare', 'verb', 'b'], ['marchigiano', 'adjective', 'c'], ['marchigiano', 'noun', 'c'], ['marchio', 'noun', 'b'], ['marcia', 'noun', 'b'], ['marciapiede', 'noun', 'b'], ['marcio', 'adjective', 'b'], ['marcio', 'noun', 'b'], ['marcire', 'verb', 'c'], ['marco', 'noun', 'a'], ['mare', 'noun', 'a'], ['marea', 'noun', 'b'], ['maresciallo', 'noun', 'b'], ['margherita', 'noun', 'c'], ['marginale', 'adjective', 'b'], ['marginale', 'noun', 'b'], ['margine', 'noun', 'b'], ['marinaio', 'noun', 'b'], ['marino', 'adjective', 'b'], ['marino', 'noun', 'b'], ['marionetta', 'noun', 'c'], ['marito', 'noun', 'a'], ['marketing', 'noun', 'b'], ['marmellata', 'noun', 'c'], ['marmo', 'noun', 'b'], ['marocchino', 'adjective', 'c'], ['marocchino', 'noun', 'c'], ['marrone', 'noun', 'b'], ['marrone', 'adjective', 'b'], ['marted', 'noun', 'b'], ['marzo', 'noun', 'a'], ['mascarpone', 'noun', 'c'], ['maschera', 'noun', 'b'], ['mascherare', 'verb', 'b'], ['mascherato', 'past_part', 'c'], ['mascherato', 'adjective', 'c'], ['maschile', 'adjective', 'a'], ['maschile', 'noun', 'a'], ['maschio', 'noun', 'a'], ['maschio', 'adjective', 'a'], ['massa', 'noun', 'a'], ['massa', 'adverb', 'a'], ['massacrare', 'verb', 'b'], ['massacro', 'noun', 'c'], ['massaggio', 'noun', 'c'], ['massaia', 'noun', 'c'], ['massiccio', 'adjective', 'b'], ['massiccio', 'noun', 'b'], ['massimo', 'adjective', 'a'], ['massimo', 'noun', 'a'], ['massimo', 'adverb', 'a'], ['master', 'noun', 'b'], ['masticare', 'verb', 'b'], ['masturbare', 'verb', 'b'], ['matematica', 'noun', 'b'], ['matematico', 'adjective', 'b'], ['matematico', 'noun', 'b'], ['materasso', 'noun', 'b'], ['materia', 'noun', 'a'], ['materiale', 'adjective', 'a'], ['materiale', 'noun', 'a'], ['maternit', 'noun', 'b'], ['materno', 'adjective', 'b'], ['matita', 'noun', 'b'], ['matricola', 'noun', 'b'], ['matrimoniale', 'adjective', 'b'], ['matrimoniale', 'noun', 'b'], ['matrimonio', 'noun', 'a'], ['mattina', 'noun', 'a'], ['mattinata', 'noun', 'b'], ['mattino', 'noun', 'a'], ['matto', 'adjective', 'a'], ['matto', 'noun', 'a'], ['mattone', 'noun', 'b'], ['mattone', 'adjective', 'b'], ['mattone', 'noun', 'b'], ['maturare', 'verb', 'b'], ['maturit', 'noun', 'b'], ['maturo', 'adjective', 'b'], ['mazzo', 'noun', 'b'], ['me', 'pronoun', 'a'], ['meccanico', 'adjective', 'a'], ['meccanico', 'noun', 'a'], ['meccanismo', 'noun', 'a'], ['medaglia', 'noun', 'b'], ['medesimo', 'adjective', 'b'], ['medesimo', 'pronoun', 'b'], ['media', 'noun', 'a'], ['media', 'noun', 'b'], ['mediante', 'preposition', 'b'], ['medicare', 'verb', 'c'], ['medicina', 'noun', 'a'], ['medico', 'noun', 'a'], ['medico', 'adjective', 'b'], ['medievale', 'adjective', 'b'], ['medio', 'adjective', 'a'], ['medio', 'noun', 'a'], ['medioevo', 'noun', 'b'], ['meditare', 'verb', 'b'], ['mediterraneo', 'adjective', 'b'], ['mediterraneo', 'noun', 'b'], ['meglio', 'adverb', 'a'], ['meglio', 'adjective', 'a'], ['meglio', 'noun', 'a'], ['mela', 'noun', 'b'], ['melagrana', 'noun', 'c'], ['melanzana', 'noun', 'c'], ['melo', 'noun', 'c'], ['melograno', 'noun', 'c'], ['melone', 'noun', 'c'], ['membrana', 'noun', 'b'], ['membro', 'noun', 'a'], ['memoria', 'noun', 'a'], ['menare', 'verb', 'b'], ['mendicante', 'pres_part', 'c'], ['mendicante', 'adjective', 'c'], ['mendicante', 'noun', 'c'], ['meno', 'adverb', 'a'], ['meno', 'adjective', 'a'], ['meno', 'preposition', 'a'], ['meno', 'noun', 'a'], ['mensa', 'noun', 'b'], ['mensile', 'adjective', 'b'], ['mensile', 'noun', 'b'], ['mensola', 'noun', 'c'], ['menta', 'noun', 'c'], ['mentale', 'adjective', 'a'], ['mentalit', 'noun', 'b'], ['mente', 'noun', 'a'], ['mentire', 'verb', 'a'], ['mento', 'noun', 'b'], ['mentre', 'conjunction', 'a'], ['menu', 'noun', 'b'], ['menzogna', 'noun', 'b'], ['meraviglia', 'noun', 'b'], ['meravigliare', 'verb', 'b'], ['meraviglioso', 'adjective', 'a'], ['meraviglioso', 'noun', 'a'], ['mercante', 'noun', 'b'], ['mercato', 'noun', 'a'], ['merce', 'noun', 'b'], ['merceria', 'noun', 'c'], ['mercoled', 'noun', 'b'], ['merda', 'noun', 'a'], ['merenda', 'noun', 'c'], ['merendina', 'noun', 'c'], ['meridiano', 'adjective', 'c'], ['meridiano', 'noun', 'c'], ['meridionale', 'adjective', 'a'], ['meridionale', 'noun', 'a'], ['meridione', 'noun', 'c'], ['meritare', 'verb', 'a'], ['merito', 'noun', 'a'], ['merlo', 'noun', 'c'], ['merluzzo', 'noun', 'c'], ['mero', 'adjective', 'b'], ['mescolare', 'verb', 'b'], ['mese', 'noun', 'a'], ['messa', 'noun', 'b'], ['messa', 'noun', 'b'], ['messaggio', 'noun', 'a'], ['messe', 'noun', 'c'], ['messicano', 'adjective', 'c'], ['messicano', 'noun', 'c'], ['mestiere', 'noun', 'a'], ['mestolo', 'noun', 'c'], ['mestruazione', 'noun', 'c'], ['met', 'noun', 'a'], ['meta', 'noun', 'b'], ['metafora', 'noun', 'b'], ['metallico', 'adjective', 'b'], ['metallo', 'noun', 'b'], ['metalmeccanico', 'adjective', 'c'], ['metalmeccanico', 'noun', 'c'], ['meteo', 'adjective', 'b'], ['meteo', 'noun', 'b'], ['metodo', 'noun', 'a'], ['metro', 'noun', 'a'], ['metropolitano', 'adjective', 'b'], ['metropolitano', 'noun', 'b'], ['mettere', 'verb', 'a'], ['mezzanotte', 'noun', 'b'], ['mezzo', 'adjective', 'a'], ['mezzo', 'noun', 'a'], ['mezzo', 'adverb', 'a'], ['mezzogiorno', 'noun', 'b'], ['mi', 'pronoun', 'a'], ['miagolare', 'verb', 'c'], ['mica', 'noun', 'a'], ['mica', 'adverb', 'a'], ['micio', 'noun', 'c'], ['microfono', 'noun', 'b'], ['miele', 'noun', 'b'], ['miele', 'adjective', 'b'], ['mietere', 'verb', 'c'], ['migliaio', 'noun', 'c'], ['migliaio', 'noun', 'a'], ['miglioramento', 'noun', 'b'], ['migliorare', 'verb', 'a'], ['migliore', 'adjective', 'a'], ['migliore', 'noun', 'a'], ['migliore', 'adverb', 'a'], ['mignolo', 'noun', 'c'], ['mila', 'adjective', 'a'], ['milanese', 'adjective', 'b'], ['milanese', 'noun', 'b'], ['miliardo', 'noun', 'a'], ['milione', 'noun', 'a'], ['militare', 'adjective', 'a'], ['militare', 'noun', 'a'], ['mille', 'adjective', 'a'], ['mille', 'noun', 'a'], ['millennio', 'noun', 'b'], ['millimetro', 'noun', 'b'], ['mimosa', 'noun', 'c'], ['minaccia', 'noun', 'b'], ['minacciare', 'verb', 'a'], ['minchia', 'noun', 'b'], ['minestra', 'noun', 'c'], ['minestrone', 'noun', 'c'], ['mini', 'adjective', 'c'], ['miniera', 'noun', 'b'], ['minigonna', 'noun', 'c'], ['minimo', 'adjective', 'a'], ['minimo', 'noun', 'a'], ['ministero', 'noun', 'a'], ['ministro', 'noun', 'a'], ['minoranza', 'noun', 'b'], ['minore', 'adjective', 'a'], ['minore', 'noun', 'a'], ['minuscolo', 'adjective', 'b'], ['minuto', 'noun', 'a'], ['mio', 'adjective', 'a'], ['mio', 'pronoun', 'a'], ['miracolo', 'noun', 'a'], ['mirare', 'verb', 'b'], ['mischiare', 'verb', 'b'], ['miscuglio', 'noun', 'c'], ['miseria', 'noun', 'b'], ['misero', 'adjective', 'b'], ['missile', 'adjective', 'c'], ['missile', 'noun', 'c'], ['missione', 'noun', 'a'], ['mister', 'noun', 'c'], ['misterioso', 'adjective', 'b'], ['mistero', 'noun', 'a'], ['misto', 'adjective', 'b'], ['misto', 'noun', 'b'], ['misura', 'noun', 'a'], ['misurare', 'verb', 'b'], ['misurazione', 'noun', 'c'], ['mitico', 'adjective', 'b'], ['mito', 'noun', 'b'], ['mitragliatrice', 'noun', 'c'], ['mobile', 'adjective', 'a'], ['mobile', 'noun', 'a'], ['mobilio', 'noun', 'c'], ['mocassino', 'noun', 'c'], ['moda', 'noun', 'a'], ['modalit', 'noun', 'b'], ['modella', 'noun', 'b'], ['modellare', 'verb', 'c'], ['modello', 'noun', 'a'], ['moderato', 'past_part', 'b'], ['moderato', 'adjective', 'b'], ['moderato', 'adverb', 'b'], ['moderato', 'noun', 'b'], ['moderatore', 'adjective', 'b'], ['moderatore', 'noun', 'b'], ['modernit', 'noun', 'b'], ['moderno', 'adjective', 'a'], ['moderno', 'noun', 'a'], ['modestia', 'noun', 'c'], ['modesto', 'adjective', 'b'], ['modifica', 'noun', 'b'], ['modificare', 'verb', 'a'], ['modificazione', 'noun', 'b'], ['modo', 'noun', 'a'], ['modulo', 'noun', 'b'], ['moglie', 'noun', 'a'], ['molecola', 'noun', 'b'], ['molisano', 'adjective', 'c'], ['molisano', 'noun', 'c'], ['molla', 'noun', 'c'], ['mollare', 'verb', 'b'], ['mollusco', 'noun', 'c'], ['molo', 'noun', 'c'], ['moltiplicare', 'verb', 'b'], ['molto', 'adjective', 'a'], ['molto', 'pronoun', 'a'], ['molto', 'adverb', 'a'], ['molto', 'noun', 'a'], ['momento', 'noun', 'a'], ['monaca', 'noun', 'c'], ['monaco', 'noun', 'c'], ['monarchica', 'noun', 'c'], ['mondiale', 'adjective', 'a'], ['mondiale', 'noun', 'a'], ['mondo', 'noun', 'a'], ['monello', 'noun', 'c'], ['moneta', 'noun', 'a'], ['monetario', 'adjective', 'b'], ['monitor', 'noun', 'b'], ['monologo', 'noun', 'b'], ['montaggio', 'noun', 'b'], ['montagna', 'noun', 'a'], ['montare', 'verb', 'b'], ['monte', 'noun', 'a'], ['montenegrino', 'adjective', 'c'], ['montenegrino', 'noun', 'c'], ['monumento', 'noun', 'b'], ['mora', 'noun', 'b'], ['morale', 'adjective', 'a'], ['morale', 'noun', 'a'], ['morbido', 'adjective', 'b'], ['morbido', 'noun', 'b'], ['mordere', 'verb', 'b'], ['morire', 'verb', 'a'], ['moro', 'adjective', 'b'], ['moro', 'noun', 'b'], ['morsicare', 'verb', 'c'], ['morso', 'noun', 'c'], ['mortadella', 'noun', 'c'], ['mortale', 'adjective', 'b'], ['mortale', 'noun', 'b'], ['morte', 'noun', 'a'], ['morto', 'past_part', 'a'], ['morto', 'adjective', 'a'], ['morto', 'noun', 'a'], ['mosca', 'noun', 'b'], ['moscovita', 'adjective', 'c'], ['moscovita', 'noun', 'c'], ['mossa', 'noun', 'b'], ['mostarda', 'noun', 'c'], ['mostra', 'noun', 'a'], ['mostrare', 'verb', 'a'], ['mostro', 'noun', 'b'], ['motel', 'noun', 'c'], ['motivare', 'verb', 'b'], ['motivazione', 'noun', 'b'], ['motivo', 'noun', 'a'], ['moto', 'noun', 'a'], ['moto', 'noun', 'b'], ['motociclismo', 'noun', 'c'], ['motociclista', 'adjective', 'c'], ['motociclista', 'noun', 'c'], ['motore', 'adjective', 'a'], ['motore', 'noun', 'a'], ['motorino', 'noun', 'b'], ['motoscafo', 'noun', 'c'], ['mousse', 'noun', 'c'], ['movimento', 'noun', 'a'], ['mozzarella', 'noun', 'c'], ['mucca', 'noun', 'b'], ['mucchio', 'noun', 'b'], ['muggire', 'verb', 'c'], ['muggito', 'past_part', 'c'], ['muggito', 'noun', 'c'], ['mugnaio', 'noun', 'c'], ['mugolare', 'verb', 'c'], ['mulino', 'noun', 'c'], ['multa', 'noun', 'b'], ['multare', 'verb', 'c'], ['multinazionale', 'adjective', 'b'], ['multinazionale', 'noun', 'b'], ['multiplo', 'adjective', 'b'], ['multiplo', 'noun', 'b'], ['multipresa', 'noun', 'c'], ['mummia', 'noun', 'c'], ['mungere', 'verb', 'c'], ['municipio', 'noun', 'c'], ['muovere', 'verb', 'a'], ['murare', 'verb', 'c'], ['muratore', 'noun', 'c'], ['muro', 'noun', 'a'], ['muschio', 'noun', 'c'], ['muschio', 'adjective', 'c'], ['muscolare', 'adjective', 'b'], ['muscolare', 'noun', 'b'], ['muscolo', 'noun', 'a'], ['museo', 'noun', 'a'], ['musica', 'noun', 'a'], ['musicale', 'adjective', 'a'], ['musicista', 'noun', 'b'], ['muso', 'noun', 'b'], ['musulmano', 'adjective', 'b'], ['musulmano', 'noun', 'b'], ['muta', 'noun', 'c'], ['mutamento', 'noun', 'b'], ['mutanda', 'noun', 'b'], ['mutandina', 'noun', 'c'], ['mutare', 'verb', 'b'], ['mutazione', 'noun', 'b'], ['mutilato', 'past_part', 'c'], ['mutilato', 'adjective', 'c'], ['mutilato', 'noun', 'c'], ['muto', 'adjective', 'b'], ['muto', 'noun', 'b'], ['mutuo', 'noun', 'b'], ['nanna', 'noun', 'c'], ['nano', 'adjective', 'b'], ['nano', 'noun', 'b'], ['napoletano', 'adjective', 'b'], ['napoletano', 'noun', 'b'], ['narrare', 'verb', 'b'], ['narrativo', 'adjective', 'b'], ['narratore', 'noun', 'b'], ['narrazione', 'noun', 'b'], ['nasale', 'adjective', 'b'], ['nasale', 'noun', 'b'], ['nascere', 'verb', 'a'], ['nascere', 'noun', 'a'], ['nascita', 'noun', 'a'], ['nascondere', 'verb', 'a'], ['nascondiglio', 'noun', 'c'], ['nascondino', 'noun', 'c'], ['nascosto', 'past_part', 'a'], ['nascosto', 'adjective', 'a'], ['nascosto', 'noun', 'a'], ['naso', 'noun', 'a'], ['nastro', 'noun', 'a'], ['natale', 'adjective', 'a'], ['natale', 'noun', 'a'], ['natalizio', 'adjective', 'b'], ['natalizio', 'noun', 'b'], ['nato', 'past_part', 'b'], ['nato', 'adjective', 'b'], ['nato', 'noun', 'b'], ['natura', 'noun', 'a'], ['naturale', 'adjective', 'a'], ['naturale', 'noun', 'a'], ['naturalmente', 'adverb', 'a'], ['naufragio', 'noun', 'c'], ['navale', 'adjective', 'c'], ['nave', 'noun', 'a'], ['navicella', 'noun', 'c'], ['navigare', 'verb', 'b'], ['navigazione', 'noun', 'b'], ['nazionale', 'adjective', 'a'], ['nazionale', 'noun', 'a'], ['nazionalit', 'noun', 'c'], ['nazione', 'noun', 'a'], ['nazista', 'adjective', 'b'], ['nazista', 'noun', 'b'], ['ndrangheta', 'noun', 'c'], ['n', 'conjunction', 'a'], ['ne', 'pronoun', 'a'], ['ne', 'adverb', 'a'], ['neanche', 'adverb', 'a'], ['nebbia', 'noun', 'b'], ['necessariamente', 'adverb', 'b'], ['necessario', 'adjective', 'a'], ['necessario', 'noun', 'a'], ['necessit', 'noun', 'a'], ['necessitare', 'verb', 'b'], ['negare', 'verb', 'a'], ['negativo', 'adjective', 'a'], ['negativo', 'noun', 'a'], ['negativo', 'adverb', 'a'], ['negazione', 'noun', 'c'], ['negoziante', 'pres_part', 'c'], ['negoziante', 'noun', 'c'], ['negozio', 'noun', 'a'], ['negro', 'adjective', 'b'], ['negro', 'noun', 'b'], ['nemico', 'adjective', 'a'], ['nemico', 'noun', 'a'], ['nemmeno', 'adverb', 'a'], ['neo', 'noun', 'c'], ['neonato', 'noun', 'b'], ['neonato', 'adjective', 'b'], ['neppure', 'adverb', 'a'], ['nero', 'adjective', 'a'], ['nero', 'noun', 'a'], ['nervo', 'noun', 'b'], ['nervosismo', 'noun', 'c'], ['nervoso', 'adjective', 'a'], ['nervoso', 'noun', 'a'], ['nessuno', 'adjective', 'a'], ['nessuno', 'pronoun', 'a'], ['nettare', 'noun', 'c'], ['netto', 'adjective', 'b'], ['netto', 'noun', 'b'], ['netto', 'adverb', 'b'], ['network', 'noun', 'b'], ['neutro', 'adjective', 'b'], ['neutro', 'noun', 'b'], ['neve', 'noun', 'a'], ['nevicare', 'verb', 'c'], ['news', 'noun', 'b'], ['newyorkese', 'adjective', 'c'], ['newyorkese', 'noun', 'c'], ['nido', 'noun', 'b'], ['niente', 'pronoun', 'a'], ['niente', 'adjective', 'a'], ['niente', 'adverb', 'a'], ['nipote', 'noun', 'a'], ['no', 'adverb', 'a'], ['no', 'noun', 'a'], ['no', 'adjective', 'a'], ['nobile', 'adjective', 'b'], ['nobile', 'noun', 'b'], ['nocciola', 'noun', 'c'], ['nocciola', 'adjective', 'c'], ['nocciolina', 'noun', 'c'], ['nocivo', 'adjective', 'c'], ['nodo', 'noun', 'b'], ['noi', 'pronoun', 'a'], ['noia', 'noun', 'b'], ['noioso', 'adjective', 'b'], ['noleggiare', 'verb', 'c'], ['nome', 'noun', 'a'], ['nomina', 'noun', 'b'], ['nominare', 'verb', 'a'], ['non', 'adverb', 'a'], ['nonch', 'conjunction', 'b'], ['nonna', 'noun', 'a'], ['nonno', 'noun', 'a'], ['nono', 'adjective', 'b'], ['nono', 'noun', 'b'], ['nonostante', 'preposition', 'a'], ['nonostante', 'conjunction', 'a'], ['nord', 'noun', 'a'], ['nord', 'adjective', 'a'], ['nordamericano', 'adjective', 'c'], ['nordamericano', 'noun', 'c'], ['norma', 'noun', 'a'], ['normale', 'adjective', 'a'], ['normale', 'noun', 'a'], ['normalit', 'noun', 'b'], ['normalmente', 'adverb', 'b'], ['normativa', 'noun', 'b'], ['norvegese', 'adjective', 'c'], ['norvegese', 'noun', 'c'], ['nostalgia', 'noun', 'b'], ['nostro', 'adjective', 'a'], ['nostro', 'pronoun', 'a'], ['nota', 'noun', 'a'], ['notaio', 'noun', 'b'], ['notare', 'verb', 'a'], ['notevole', 'adjective', 'b'], ['notizia', 'noun', 'a'], ['noto', 'adjective', 'a'], ['noto', 'noun', 'a'], ['notte', 'noun', 'a'], ['notturno', 'adjective', 'b'], ['notturno', 'noun', 'b'], ['novanta', 'adjective', 'b'], ['novanta', 'noun', 'b'], ['nove', 'adjective', 'a'], ['nove', 'noun', 'a'], ['novella', 'noun', 'c'], ['novembre', 'noun', 'a'], ['novit', 'noun', 'a'], ['nozione', 'noun', 'b'], ['nozze', 'noun', 'b'], ['nube', 'noun', 'b'], ['nucleare', 'adjective', 'a'], ['nucleare', 'noun', 'a'], ['nucleo', 'noun', 'b'], ['nudo', 'adjective', 'a'], ['nudo', 'noun', 'a'], ['nulla', 'pronoun', 'a'], ['nulla', 'adverb', 'a'], ['numerare', 'verb', 'b'], ['numerazione', 'noun', 'c'], ['numero', 'noun', 'a'], ['numeroso', 'adjective', 'a'], ['nuora', 'noun', 'c'], ['nuotare', 'verb', 'b'], ['nuoto', 'noun', 'b'], ['nuovamente', 'adverb', 'b'], ['nuovo', 'adjective', 'a'], ['nuovo', 'noun', 'a'], ['nutrire', 'verb', 'b'], ['nuvola', 'noun', 'b'], ['nuvoloso', 'adjective', 'c'], ['nylon', 'noun', 'c'], ['o', 'noun', 'c'], ['o', 'conjunction', 'a'], ['obbedire', 'verb', 'b'], ['obbiettivo', 'adjective', 'c'], ['obbiettivo', 'noun', 'c'], ['obbligare', 'verb', 'a'], ['obbligatorio', 'adjective', 'b'], ['obbligazione', 'noun', 'b'], ['obbligo', 'noun', 'b'], ['obiettivo', 'adjective', 'a'], ['obiettivo', 'noun', 'a'], ['obiezione', 'noun', 'b'], ['obl', 'noun', 'c'], ['occasione', 'noun', 'a'], ['occhiaia', 'noun', 'c'], ['occhiale', 'noun', 'a'], ['occhiale', 'adjective', 'a'], ['occhiata', 'noun', 'b'], ['occhiello', 'noun', 'c'], ['occhio', 'noun', 'a'], ['occidentale', 'adjective', 'a'], ['occidentale', 'noun', 'a'], ['occidente', 'noun', 'b'], ['occidente', 'adjective', 'b'], ['occorrere', 'verb', 'a'], ['occupare', 'verb', 'a'], ['occupato', 'past_part', 'c'], ['occupato', 'adjective', 'c'], ['occupato', 'noun', 'c'], ['occupazione', 'noun', 'b'], ['oceano', 'noun', 'b'], ['oculista', 'noun', 'c'], ['oddio', 'exclamation', 'b'], ['odiare', 'verb', 'a'], ['odio', 'noun', 'b'], ['odorare', 'verb', 'c'], ['odore', 'noun', 'a'], ['offendere', 'verb', 'b'], ['offerta', 'noun', 'a'], ['offesa', 'noun', 'b'], ['offeso', 'past_part', 'c'], ['offeso', 'adjective', 'c'], ['offeso', 'noun', 'c'], ['officina', 'noun', 'b'], ['offline', 'adjective', 'b'], ['offline', 'noun', 'b'], ['offrire', 'verb', 'a'], ['oggettivo', 'adjective', 'b'], ['oggetto', 'noun', 'a'], ['oggi', 'adverb', 'a'], ['oggi', 'noun', 'a'], ['ogni', 'adjective', 'a'], ['ognuno', 'pronoun', 'a'], ['ognuno', 'adjective', 'a'], ['ok', 'adverb', 'a'], ['ok', 'noun', 'a'], ['ok', 'adjective', 'a'], ['okay', 'adverb', 'a'], ['okay', 'noun', 'a'], ['okay', 'adjective', 'a'], ['olandese', 'adjective', 'b'], ['olandese', 'noun', 'b'], ['oliare', 'verb', 'c'], ['oliera', 'noun', 'c'], ['olimpico', 'adjective', 'b'], ['olio', 'noun', 'a'], ['oliva', 'noun', 'b'], ['oliva', 'adjective', 'b'], ['oltre', 'adverb', 'a'], ['oltre', 'preposition', 'a'], ['oltrepassare', 'verb', 'c'], ['oltretutto', 'adverb', 'b'], ['omaggio', 'noun', 'b'], ['ombelico', 'noun', 'c'], ['ombra', 'noun', 'a'], ['ombrellone', 'noun', 'c'], ['omicidio', 'noun', 'a'], ['omogeneizzato', 'past_part', 'c'], ['omogeneizzato', 'adjective', 'c'], ['omogeneizzato', 'noun', 'c'], ['omonimo', 'adjective', 'b'], ['omonimo', 'noun', 'b'], ['onda', 'noun', 'a'], ['ondata', 'noun', 'b'], ['ondeggiare', 'verb', 'c'], ['onere', 'noun', 'b'], ['onestamente', 'adverb', 'b'], ['onesto', 'adjective', 'b'], ['onesto', 'noun', 'b'], ['onesto', 'adverb', 'b'], ['online', 'adjective', 'b'], ['online', 'noun', 'b'], ['onorare', 'verb', 'b'], ['onore', 'noun', 'a'], ['opera', 'noun', 'a'], ['operaio', 'noun', 'a'], ['operaio', 'adjective', 'a'], ['operare', 'verb', 'a'], ['operativo', 'adjective', 'b'], ['operativo', 'noun', 'b'], ['operatore', 'adjective', 'b'], ['operatore', 'noun', 'b'], ['operazione', 'noun', 'a'], ['opinione', 'noun', 'a'], ['opporre', 'verb', 'a'], ['opportunit', 'noun', 'b'], ['opportuno', 'adjective', 'b'], ['opposizione', 'noun', 'b'], ['opposto', 'past_part', 'a'], ['opposto', 'adjective', 'a'], ['opposto', 'noun', 'a'], ['oppressivo', 'adjective', 'c'], ['oppresso', 'past_part', 'c'], ['oppresso', 'adjective', 'c'], ['oppresso', 'noun', 'c'], ['oppressore', 'adjective', 'c'], ['oppressore', 'noun', 'c'], ['oppure', 'conjunction', 'a'], ['opzione', 'noun', 'b'], ['ora', 'noun', 'a'], ['ora', 'adverb', 'a'], ['orale', 'adjective', 'b'], ['oramai', 'adverb', 'b'], ['orario', 'adjective', 'a'], ['orario', 'noun', 'a'], ['orbita', 'noun', 'b'], ['orchestra', 'noun', 'b'], ['orco', 'noun', 'b'], ['ordinamento', 'noun', 'b'], ['ordinanza', 'noun', 'b'], ['ordinare', 'verb', 'a'], ['ordinario', 'adjective', 'b'], ['ordinario', 'noun', 'b'], ['ordine', 'noun', 'a'], ['orecchino', 'noun', 'c'], ['orecchio', 'noun', 'a'], ['orefice', 'noun', 'c'], ['organico', 'adjective', 'b'], ['organico', 'noun', 'b'], ['organismo', 'noun', 'a'], ['organizzare', 'verb', 'a'], ['organizzato', 'past_part', 'b'], ['organizzato', 'adjective', 'b'], ['organizzato', 'noun', 'b'], ['organizzazione', 'noun', 'a'], ['organo', 'noun', 'a'], ['orgasmo', 'noun', 'b'], ['orgoglio', 'noun', 'b'], ['orgoglioso', 'adjective', 'b'], ['orientale', 'adjective', 'b'], ['orientale', 'noun', 'b'], ['orientamento', 'noun', 'b'], ['orientare', 'verb', 'b'], ['oriente', 'adjective', 'b'], ['oriente', 'noun', 'b'], ['origano', 'noun', 'c'], ['originale', 'adjective', 'a'], ['originale', 'noun', 'a'], ['originario', 'adjective', 'b'], ['origine', 'noun', 'a'], ['orizzontale', 'adjective', 'b'], ['orizzontale', 'noun', 'b'], ['orizzonte', 'noun', 'b'], ['orlo', 'noun', 'b'], ['orma', 'noun', 'c'], ['ormai', 'adverb', 'a'], ['ormone', 'noun', 'b'], ['oro', 'noun', 'a'], ['orologiaio', 'noun', 'c'], ['orologio', 'noun', 'a'], ['oroscopo', 'noun', 'b'], ['orribile', 'adjective', 'b'], ['orrore', 'noun', 'b'], ['orso', 'noun', 'b'], ['ortaggio', 'noun', 'c'], ['ortensia', 'noun', 'c'], ['ortica', 'noun', 'c'], ['orto', 'noun', 'b'], ['ortolano', 'noun', 'c'], ['ortolano', 'adjective', 'c'], ['orzo', 'noun', 'c'], ['osare', 'verb', 'b'], ['osceno', 'adjective', 'c'], ['oscillare', 'verb', 'b'], ['oscurare', 'verb', 'b'], ['oscuro', 'adjective', 'b'], ['oscuro', 'noun', 'b'], ['oscuro', 'adverb', 'b'], ['ospedale', 'noun', 'a'], ['ospitalit', 'noun', 'c'], ['ospitare', 'verb', 'a'], ['ospite', 'adjective', 'a'], ['ospite', 'noun', 'a'], ['ospizio', 'noun', 'c'], ['osservare', 'verb', 'a'], ['osservazione', 'noun', 'b'], ['ossessione', 'noun', 'b'], ['ossia', 'conjunction', 'b'], ['ossigeno', 'noun', 'b'], ['osso', 'noun', 'a'], ['ostacolare', 'verb', 'b'], ['ostacolo', 'noun', 'b'], ['ostaggio', 'noun', 'c'], ['oste', 'noun', 'c'], ['ostile', 'adjective', 'b'], ['ostinato', 'past_part', 'c'], ['ostinato', 'adjective', 'c'], ['ostrica', 'noun', 'c'], ['ottanta', 'adjective', 'b'], ['ottanta', 'noun', 'b'], ['ottavo', 'adjective', 'b'], ['ottavo', 'noun', 'b'], ['ottenere', 'verb', 'a'], ['ottica', 'noun', 'b'], ['ottimo', 'adjective', 'a'], ['ottimo', 'noun', 'a'], ['otto', 'adjective', 'a'], ['otto', 'noun', 'a'], ['ottobre', 'noun', 'a'], ['ottone', 'noun', 'c'], ['ovale', 'adjective', 'c'], ['ovale', 'noun', 'c'], ['ovatta', 'noun', 'c'], ['ove', 'adverb', 'b'], ['ove', 'conjunction', 'b'], ['ovest', 'noun', 'b'], ['ovest', 'adjective', 'b'], ['ovile', 'noun', 'c'], ['ovino', 'adjective', 'c'], ['ovino', 'noun', 'c'], ['ovunque', 'adverb', 'a'], ['ovunque', 'conjunction', 'a'], ['ovvero', 'conjunction', 'a'], ['ovviamente', 'adverb', 'a'], ['ovviare', 'verb', 'b'], ['ovvio', 'adjective', 'b'], ['ozono', 'noun', 'c'], ['pacchetto', 'noun', 'b'], ['pacco', 'noun', 'b'], ['pace', 'noun', 'a'], ['padella', 'noun', 'c'], ['padre', 'noun', 'a'], ['padrona', 'noun', 'b'], ['padronato', 'noun', 'c'], ['padrone', 'noun', 'a'], ['padroneggiare', 'verb', 'c'], ['paesaggio', 'noun', 'b'], ['paese', 'noun', 'a'], ['paga', 'noun', 'b'], ['pagamento', 'noun', 'a'], ['pagare', 'verb', 'a'], ['pagella', 'noun', 'c'], ['pagina', 'noun', 'a'], ['paglia', 'noun', 'b'], ['paglia', 'adjective', 'b'], ['pagliaio', 'noun', 'c'], ['pago', 'past_part', 'b'], ['pago', 'adjective', 'b'], ['paio', 'noun', 'a'], ['pala', 'noun', 'b'], ['palato', 'noun', 'c'], ['palazzina', 'noun', 'c'], ['palazzo', 'noun', 'a'], ['palco', 'noun', 'b'], ['palcoscenico', 'noun', 'b'], ['palermitano', 'adjective', 'c'], ['palermitano', 'noun', 'c'], ['palestinese', 'adjective', 'c'], ['palestinese', 'noun', 'c'], ['palestra', 'noun', 'b'], ['paletta', 'noun', 'c'], ['palla', 'noun', 'a'], ['pallacanestro', 'noun', 'c'], ['pallanuoto', 'noun', 'c'], ['pallavolo', 'noun', 'c'], ['pallido', 'adjective', 'b'], ['pallina', 'noun', 'b'], ['pallino', 'noun', 'c'], ['palloncino', 'noun', 'c'], ['pallone', 'noun', 'b'], ['pallottola', 'noun', 'c'], ['pallottoliere', 'noun', 'c'], ['palma', 'noun', 'c'], ['palo', 'noun', 'b'], ['palombaro', 'noun', 'c'], ['palpebra', 'noun', 'c'], ['palude', 'noun', 'c'], ['panca', 'noun', 'c'], ['pancarr', 'noun', 'c'], ['pancetta', 'noun', 'c'], ['panchina', 'noun', 'b'], ['pancia', 'noun', 'b'], ['panciotto', 'noun', 'c'], ['panda', 'noun', 'c'], ['pandoro', 'noun', 'c'], ['pane', 'noun', 'a'], ['panetteria', 'noun', 'c'], ['panettiere', 'noun', 'c'], ['panettone', 'noun', 'c'], ['panico', 'adjective', 'b'], ['panico', 'noun', 'b'], ['paniere', 'noun', 'c'], ['panino', 'noun', 'b'], ['panna', 'noun', 'b'], ['pannello', 'noun', 'b'], ['panno', 'noun', 'b'], ['pannocchia', 'noun', 'c'], ['pannolino', 'noun', 'c'], ['pannolone', 'noun', 'c'], ['panorama', 'noun', 'b'], ['pantalone', 'noun', 'a'], ['pantera', 'noun', 'c'], ['pantofola', 'noun', 'c'], ['panzerotto', 'noun', 'c'], ['papa', 'noun', 'a'], ['pap', 'noun', 'a'], ['papavero', 'noun', 'c'], ['papera', 'noun', 'c'], ['papero', 'noun', 'c'], ['pappa', 'noun', 'c'], ['pappagallo', 'noun', 'c'], ['parabola', 'noun', 'c'], ['parabrezza', 'noun', 'c'], ['paracadute', 'noun', 'c'], ['paracadutista', 'noun', 'c'], ['paradiso', 'noun', 'b'], ['paradosso', 'noun', 'b'], ['paradosso', 'adjective', 'b'], ['parafulmine', 'noun', 'c'], ['paragonare', 'verb', 'b'], ['paragone', 'noun', 'b'], ['paralisi', 'noun', 'c'], ['paralizzato', 'past_part', 'c'], ['paralizzato', 'adjective', 'c'], ['parallelepipedo', 'noun', 'c'], ['parallelo', 'adjective', 'b'], ['parallelo', 'noun', 'b'], ['paralume', 'noun', 'c'], ['parametro', 'noun', 'b'], ['paraocchi', 'noun', 'c'], ['parare', 'verb', 'b'], ['paraurti', 'noun', 'c'], ['paravento', 'noun', 'c'], ['parcheggiare', 'verb', 'b'], ['parcheggio', 'noun', 'b'], ['parco', 'noun', 'a'], ['parecchio', 'adjective', 'a'], ['parecchio', 'pronoun', 'a'], ['parecchio', 'adverb', 'a'], ['parecchio', 'adjective', 'a'], ['pareggiare', 'verb', 'c'], ['pareggio', 'noun', 'c'], ['parente', 'noun', 'a'], ['parentesi', 'noun', 'b'], ['parere', 'verb', 'a'], ['parere', 'noun', 'a'], ['parete', 'noun', 'a'], ['pari', 'adjective', 'a'], ['pari', 'adverb', 'a'], ['pari', 'noun', 'a'], ['parigino', 'adjective', 'c'], ['parigino', 'noun', 'c'], ['parit', 'noun', 'c'], ['parlamentare', 'adjective', 'b'], ['parlamentare', 'noun', 'b'], ['parlamento', 'noun', 'b'], ['parlare', 'verb', 'a'], ['parmigiano', 'adjective', 'c'], ['parmigiano', 'noun', 'c'], ['parola', 'noun', 'a'], ['parquet', 'noun', 'c'], ['parroco', 'noun', 'c'], ['parrucca', 'noun', 'c'], ['parrucchiere', 'noun', 'c'], ['parte', 'noun', 'a'], ['parte', 'adverb', 'a'], ['partecipante', 'pres_part', 'b'], ['partecipante', 'adjective', 'b'], ['partecipante', 'noun', 'b'], ['partecipare', 'verb', 'a'], ['partecipazione', 'noun', 'b'], ['parteggiare', 'verb', 'c'], ['partenza', 'noun', 'a'], ['particella', 'noun', 'b'], ['particolare', 'adjective', 'a'], ['particolare', 'noun', 'a'], ['particolarmente', 'adverb', 'a'], ['partigiano', 'noun', 'b'], ['partigiano', 'adjective', 'b'], ['partire', 'verb', 'a'], ['partita', 'noun', 'a'], ['partito', 'noun', 'a'], ['partner', 'noun', 'b'], ['parto', 'noun', 'b'], ['partorire', 'verb', 'b'], ['party', 'noun', 'b'], ['parziale', 'adjective', 'b'], ['parziale', 'noun', 'b'], ['parzialmente', 'adverb', 'b'], ['pascolare', 'verb', 'c'], ['pasqua', 'noun', 'c'], ['pasquale', 'adjective', 'b'], ['passaggio', 'noun', 'a'], ['passare', 'verb', 'a'], ['passata', 'noun', 'c'], ['passatempo', 'noun', 'c'], ['passato', 'past_part', 'a'], ['passato', 'adjective', 'a'], ['passato', 'noun', 'a'], ['passeggero', 'adjective', 'b'], ['passeggero', 'noun', 'b'], ['passeggiare', 'verb', 'b'], ['passeggiata', 'noun', 'b'], ['passeggio', 'noun', 'c'], ['passero', 'noun', 'c'], ['passione', 'noun', 'a'], ['passivo', 'adjective', 'b'], ['passivo', 'noun', 'b'], ['passo', 'noun', 'a'], ['pasta', 'noun', 'a'], ['pasticca', 'noun', 'c'], ['pasticcere', 'noun', 'c'], ['pasticceria', 'noun', 'c'], ['pasticcino', 'noun', 'c'], ['pasticcio', 'noun', 'c'], ['pastiglia', 'noun', 'c'], ['pastina', 'noun', 'c'], ['pasto', 'noun', 'b'], ['pastore', 'noun', 'b'], ['patata', 'noun', 'b'], ['patatina', 'noun', 'c'], ['pat', 'noun', 'c'], ['patente', 'noun', 'b'], ['patetico', 'adjective', 'b'], ['patetico', 'noun', 'b'], ['patologia', 'noun', 'b'], ['patria', 'noun', 'b'], ['patrimonio', 'noun', 'b'], ['pattinaggio', 'noun', 'c'], ['pattinare', 'verb', 'c'], ['pattino', 'noun', 'c'], ['patto', 'noun', 'b'], ['pattumiera', 'noun', 'c'], ['paura', 'noun', 'a'], ['pauroso', 'adjective', 'c'], ['pausa', 'noun', 'a'], ['pavimento', 'noun', 'b'], ['pavone', 'noun', 'c'], ['pavone', 'adjective', 'c'], ['paziente', 'adjective', 'a'], ['paziente', 'noun', 'a'], ['pazienza', 'noun', 'a'], ['pazza', 'noun', 'c'], ['pazzesco', 'adjective', 'b'], ['pazzo', 'adjective', 'a'], ['pazzo', 'noun', 'a'], ['peccato', 'noun', 'b'], ['peccato', 'exclamation', 'b'], ['peccatore', 'noun', 'c'], ['peccatore', 'adjective', 'c'], ['pechinese', 'adjective', 'c'], ['pechinese', 'noun', 'c'], ['pecora', 'noun', 'b'], ['pecorino', 'adjective', 'c'], ['pecorino', 'noun', 'c'], ['pedalare', 'verb', 'c'], ['pedale', 'noun', 'c'], ['pedale', 'adjective', 'c'], ['pedone', 'noun', 'c'], ['pedone', 'adjective', 'c'], ['peggio', 'adverb', 'a'], ['peggio', 'adjective', 'a'], ['peggio', 'noun', 'a'], ['peggioramento', 'noun', 'c'], ['peggiorare', 'verb', 'b'], ['peggiore', 'adjective', 'b'], ['peggiore', 'noun', 'b'], ['peggiore', 'adverb', 'b'], ['pelato', 'past_part', 'c'], ['pelato', 'adjective', 'c'], ['pelato', 'noun', 'c'], ['pelle', 'noun', 'a'], ['pellegrino', 'noun', 'c'], ['pellegrino', 'adjective', 'c'], ['pellerossa', 'adjective', 'c'], ['pellerossa', 'noun', 'c'], ['pelletteria', 'noun', 'c'], ['pellicola', 'noun', 'b'], ['pelo', 'noun', 'b'], ['peloso', 'adjective', 'c'], ['peloso', 'noun', 'c'], ['peluche', 'noun', 'c'], ['pena', 'noun', 'a'], ['penale', 'adjective', 'b'], ['penale', 'noun', 'b'], ['pendere', 'verb', 'b'], ['pendolo', 'noun', 'c'], ['pene', 'noun', 'b'], ['penetrare', 'verb', 'b'], ['penisola', 'noun', 'c'], ['penna', 'noun', 'b'], ['pennarello', 'noun', 'c'], ['pensare', 'verb', 'a'], ['pensiero', 'noun', 'a'], ['pensionato', 'past_part', 'c'], ['pensionato', 'adjective', 'c'], ['pensionato', 'noun', 'c'], ['pensione', 'noun', 'a'], ['pentagono', 'noun', 'c'], ['pentirsi', 'verb', 'b'], ['pentola', 'noun', 'b'], ['penultimo', 'adjective', 'c'], ['pepe', 'noun', 'c'], ['peperoncino', 'noun', 'c'], ['peperone', 'noun', 'c'], ['per', 'preposition', 'a'], ['pera', 'noun', 'c'], ['peraltro', 'adverb', 'b'], ['percentuale', 'adjective', 'b'], ['percentuale', 'noun', 'b'], ['percepire', 'verb', 'a'], ['percezione', 'noun', 'b'], ['perch', 'adverb', 'a'], ['perch', 'conjunction', 'a'], ['perch', 'noun', 'a'], ['perci', 'conjunction', 'a'], ['percorrere', 'verb', 'b'], ['percorso', 'past_part', 'a'], ['percorso', 'adjective', 'a'], ['percorso', 'noun', 'a'], ['perdere', 'verb', 'a'], ['perdita', 'noun', 'a'], ['perdonare', 'verb', 'a'], ['perdono', 'noun', 'b'], ['perduto', 'past_part', 'b'], ['perduto', 'adjective', 'b'], ['perfettamente', 'adverb', 'a'], ['perfetto', 'past_part', 'a'], ['perfetto', 'adjective', 'a'], ['perfetto', 'noun', 'a'], ['perfezione', 'noun', 'b'], ['perfino', 'adverb', 'a'], ['perfino', 'preposition', 'a'], ['pergola', 'noun', 'c'], ['pergolato', 'noun', 'c'], ['pergolato', 'adjective', 'c'], ['pericolo', 'noun', 'a'], ['pericoloso', 'adjective', 'a'], ['periferia', 'noun', 'b'], ['periodico', 'adjective', 'b'], ['periodico', 'noun', 'b'], ['periodo', 'noun', 'a'], ['perito', 'noun', 'b'], ['perito', 'adjective', 'b'], ['perla', 'noun', 'b'], ['perla', 'adjective', 'b'], ['permaloso', 'adjective', 'c'], ['permaloso', 'noun', 'c'], ['permanente', 'pres_part', 'b'], ['permanente', 'adjective', 'b'], ['permanente', 'noun', 'b'], ['permesso', 'past_part', 'b'], ['permesso', 'adjective', 'b'], ['permesso', 'noun', 'b'], ['permettere', 'verb', 'a'], ['pero', 'noun', 'c'], ['per', 'conjunction', 'a'], ['perpendicolare', 'adjective', 'c'], ['perpendicolare', 'noun', 'c'], ['perplesso', 'adjective', 'b'], ['perquisizione', 'noun', 'b'], ['perseguire', 'verb', 'b'], ['persiana', 'noun', 'c'], ['persiano', 'adjective', 'b'], ['persiano', 'noun', 'b'], ['persino', 'adverb', 'a'], ['perso', 'past_part', 'b'], ['perso', 'adjective', 'b'], ['persona', 'noun', 'a'], ['personaggio', 'noun', 'a'], ['personale', 'adjective', 'a'], ['personale', 'noun', 'a'], ['personale', 'noun', 'a'], ['personalit', 'noun', 'b'], ['personalmente', 'adverb', 'a'], ['pertanto', 'conjunction', 'b'], ['perugino', 'adjective', 'c'], ['perugino', 'noun', 'c'], ['peruviano', 'adjective', 'c'], ['peruviano', 'noun', 'c'], ['pervenire', 'verb', 'b'], ['pesante', 'pres_part', 'a'], ['pesante', 'adjective', 'a'], ['pesante', 'adverb', 'a'], ['pesare', 'verb', 'b'], ['pesca', 'noun', 'c'], ['pesca', 'adjective', 'c'], ['pesca', 'noun', 'b'], ['pescare', 'verb', 'b'], ['pescatore', 'noun', 'b'], ['pescatore', 'adjective', 'b'], ['pesce', 'noun', 'a'], ['peschereccio', 'noun', 'c'], ['peschereccio', 'adjective', 'c'], ['pescheria', 'noun', 'c'], ['pesco', 'noun', 'c'], ['peso', 'noun', 'a'], ['pessimo', 'adjective', 'b'], ['pestare', 'verb', 'c'], ['peste', 'noun', 'c'], ['pesto', 'past_part', 'c'], ['pesto', 'adjective', 'c'], ['pesto', 'noun', 'c'], ['petalo', 'noun', 'c'], ['petardo', 'noun', 'c'], ['petroliera', 'noun', 'c'], ['petrolio', 'noun', 'b'], ['pettegolezzo', 'noun', 'c'], ['pettegolo', 'adjective', 'c'], ['pettegolo', 'noun', 'c'], ['pettinare', 'verb', 'c'], ['pettinatura', 'noun', 'c'], ['pettine', 'noun', 'c'], ['pettirosso', 'noun', 'c'], ['petto', 'noun', 'a'], ['pezza', 'noun', 'c'], ['pezzetto', 'noun', 'b'], ['pezzo', 'noun', 'a'], ['pezzuola', 'noun', 'c'], ['pi', 'noun', 'c'], ['piacere', 'verb', 'a'], ['piacere', 'noun', 'a'], ['piacevole', 'adjective', 'b'], ['piadina', 'noun', 'c'], ['piaga', 'noun', 'c'], ['pialla', 'noun', 'c'], ['piallare', 'verb', 'c'], ['pianeggiante', 'pres_part', 'c'], ['pianeggiante', 'adjective', 'c'], ['pianerottolo', 'noun', 'b'], ['pianeta', 'noun', 'a'], ['piangere', 'verb', 'a'], ['piangere', 'noun', 'a'], ['piano', 'noun', 'a'], ['piano', 'noun', 'a'], ['piano', 'adjective', 'a'], ['piano', 'adverb', 'a'], ['pianoforte', 'noun', 'b'], ['pianoterra', 'noun', 'c'], ['pianta', 'noun', 'a'], ['piantare', 'verb', 'b'], ['pianto', 'noun', 'b'], ['pianura', 'noun', 'b'], ['piastra', 'noun', 'c'], ['piattaforma', 'noun', 'b'], ['piatto', 'adjective', 'a'], ['piatto', 'noun', 'a'], ['piazza', 'noun', 'a'], ['piazzale', 'noun', 'b'], ['piazzare', 'verb', 'b'], ['piccante', 'adjective', 'c'], ['picchiare', 'verb', 'b'], ['piccino', 'adjective', 'c'], ['piccino', 'noun', 'c'], ['piccione', 'noun', 'c'], ['picco', 'noun', 'b'], ['piccolo', 'adjective', 'a'], ['piccolo', 'noun', 'a'], ['piccone', 'noun', 'c'], ['picnic', 'noun', 'c'], ['pidocchio', 'noun', 'c'], ['piede', 'noun', 'a'], ['piega', 'noun', 'b'], ['piegare', 'verb', 'b'], ['pieghevole', 'adjective', 'c'], ['pieghevole', 'noun', 'c'], ['piemontese', 'adjective', 'b'], ['piemontese', 'noun', 'b'], ['piena', 'noun', 'c'], ['pienamente', 'adverb', 'b'], ['pieno', 'adjective', 'a'], ['pieno', 'noun', 'a'], ['piet', 'noun', 'b'], ['pietra', 'noun', 'a'], ['pigiama', 'noun', 'c'], ['pigione', 'noun', 'c'], ['pigliare', 'verb', 'b'], ['pigna', 'noun', 'c'], ['pigrizia', 'noun', 'c'], ['pigro', 'adjective', 'c'], ['pigro', 'noun', 'c'], ['pila', 'noun', 'b'], ['pillola', 'noun', 'b'], ['pilota', 'noun', 'b'], ['pineta', 'noun', 'c'], ['ping-pong', 'noun', 'c'], ['pinguino', 'noun', 'c'], ['pinna', 'noun', 'c'], ['pinolo', 'noun', 'c'], ['pinza', 'noun', 'c'], ['pinzetta', 'noun', 'c'], ['pioggia', 'noun', 'a'], ['piombo', 'noun', 'b'], ['piombo', 'adjective', 'b'], ['piombo', 'noun', 'b'], ['pioppo', 'noun', 'c'], ['piovere', 'verb', 'b'], ['piovoso', 'adjective', 'c'], ['piovoso', 'noun', 'c'], ['pip', 'noun', 'c'], ['pipistrello', 'noun', 'c'], ['pirata', 'noun', 'b'], ['piscina', 'noun', 'b'], ['pisello', 'noun', 'c'], ['pisello', 'adjective', 'c'], ['pisolino', 'noun', 'c'], ['pista', 'noun', 'b'], ['pistacchio', 'noun', 'c'], ['pistacchio', 'adjective', 'c'], ['pistola', 'noun', 'a'], ['pittare', 'verb', 'c'], ['pittore', 'noun', 'b'], ['pittore', 'adjective', 'b'], ['pittura', 'noun', 'b'], ['pitturare', 'verb', 'c'], ['pi', 'adverb', 'a'], ['pi', 'adjective', 'a'], ['pi', 'preposition', 'a'], ['pi', 'noun', 'a'], ['piuma', 'noun', 'c'], ['piumino', 'noun', 'c'], ['piuttosto', 'adverb', 'a'], ['pizza', 'noun', 'b'], ['pizzeria', 'noun', 'c'], ['pizzetta', 'noun', 'c'], ['pizzicare', 'verb', 'c'], ['pizzo', 'noun', 'c'], ['plaid', 'noun', 'c'], ['plastica', 'noun', 'b'], ['plastico', 'adjective', 'b'], ['plastico', 'noun', 'b'], ['platano', 'noun', 'c'], ['platino', 'noun', 'c'], ['platino', 'adjective', 'c'], ['plurale', 'noun', 'c'], ['plurale', 'adjective', 'c'], ['pneumatico', 'noun', 'c'], ['pochino', 'noun', 'b'], ['poco', 'adjective', 'a'], ['poco', 'pronoun', 'a'], ['poco', 'adverb', 'a'], ['podere', 'noun', 'c'], ['poema', 'noun', 'b'], ['poesia', 'noun', 'a'], ['poeta', 'noun', 'a'], ['poetico', 'adjective', 'b'], ['poetico', 'noun', 'b'], ['poggiapiedi', 'noun', 'c'], ['poggiare', 'verb', 'c'], ['poi', 'adverb', 'a'], ['poich', 'conjunction', 'a'], ['poker', 'noun', 'b'], ['polacco', 'adjective', 'b'], ['polacco', 'noun', 'b'], ['polemica', 'noun', 'b'], ['polenta', 'noun', 'c'], ['polipo', 'noun', 'c'], ['politica', 'noun', 'a'], ['politico', 'adjective', 'a'], ['politico', 'noun', 'a'], ['polizia', 'noun', 'a'], ['poliziotto', 'noun', 'a'], ['pollaio', 'noun', 'c'], ['pollame', 'noun', 'c'], ['pollice', 'noun', 'b'], ['pollo', 'noun', 'c'], ['polmone', 'noun', 'b'], ['polo', 'noun', 'b'], ['polpa', 'noun', 'c'], ['polpastrello', 'noun', 'c'], ['polpetta', 'noun', 'c'], ['polpo', 'noun', 'c'], ['polsino', 'noun', 'c'], ['polso', 'noun', 'b'], ['poltrona', 'noun', 'b'], ['polvere', 'noun', 'a'], ['polverina', 'noun', 'c'], ['polveroso', 'adjective', 'c'], ['pomata', 'noun', 'c'], ['pomello', 'noun', 'c'], ['pomeriggio', 'noun', 'a'], ['pomodoro', 'noun', 'b'], ['pompa', 'noun', 'b'], ['pompelmo', 'noun', 'c'], ['pompiere', 'noun', 'c'], ['ponte', 'noun', 'a'], ['pony', 'noun', 'c'], ['pop', 'adjective', 'b'], ['pop', 'noun', 'b'], ['popolare', 'adjective', 'a'], ['popolare', 'noun', 'a'], ['popolare', 'verb', 'b'], ['popolarit', 'noun', 'c'], ['popolazione', 'noun', 'a'], ['popolo', 'noun', 'a'], ['porcellana', 'noun', 'c'], ['porcheria', 'noun', 'c'], ['porco', 'noun', 'b'], ['porco', 'adjective', 'b'], ['porgere', 'verb', 'b'], ['porno', 'adjective', 'b'], ['porno', 'noun', 'b'], ['porre', 'verb', 'a'], ['porta', 'noun', 'a'], ['portabagagli', 'noun', 'c'], ['portabagagli', 'adjective', 'c'], ['portacenere', 'noun', 'c'], ['portachiavi', 'noun', 'c'], ['portacipria', 'noun', 'c'], ['portaerei', 'noun', 'c'], ['portafinestra', 'noun', 'c'], ['portafoglio', 'noun', 'b'], ['portafortuna', 'noun', 'c'], ['portale', 'noun', 'b'], ['portamonete', 'noun', 'c'], ['portaombrelli', 'noun', 'c'], ['portare', 'verb', 'a'], ['portata', 'noun', 'b'], ['portatore', 'adjective', 'b'], ['portatore', 'noun', 'b'], ['portiere', 'noun', 'b'], ['portineria', 'noun', 'c'], ['porto', 'noun', 'a'], ['portoghese', 'adjective', 'b'], ['portoghese', 'noun', 'b'], ['portone', 'noun', 'b'], ['porzione', 'noun', 'b'], ['posa', 'noun', 'b'], ['posacenere', 'noun', 'c'], ['posare', 'verb', 'b'], ['posata', 'noun', 'c'], ['positivo', 'adjective', 'a'], ['positivo', 'noun', 'a'], ['positivo', 'adverb', 'a'], ['posizionare', 'verb', 'b'], ['posizione', 'noun', 'a'], ['possedere', 'verb', 'a'], ['possesso', 'noun', 'b'], ['possibile', 'adjective', 'a'], ['possibile', 'noun', 'a'], ['possibilit', 'noun', 'a'], ['post', 'noun', 'b'], ['posta', 'noun', 'a'], ['postale', 'adjective', 'b'], ['postare', 'verb', 'b'], ['posteggiatore', 'noun', 'c'], ['posteriore', 'adjective', 'b'], ['posteriore', 'noun', 'b'], ['postino', 'noun', 'c'], ['postino', 'adjective', 'c'], ['posto', 'noun', 'a'], ['potare', 'verb', 'c'], ['potente', 'pres_part', 'a'], ['potente', 'adjective', 'a'], ['potente', 'noun', 'a'], ['potentino', 'adjective', 'c'], ['potentino', 'noun', 'c'], ['potenza', 'noun', 'b'], ['potenziale', 'adjective', 'b'], ['potenziale', 'noun', 'b'], ['potere', 'verb', 'a'], ['potere', 'noun', 'a'], ['povero', 'adjective', 'a'], ['povert', 'noun', 'b'], ['pozzanghera', 'noun', 'c'], ['pozzo', 'noun', 'b'], ['praghese', 'adjective', 'c'], ['praghese', 'noun', 'c'], ['pranzo', 'noun', 'a'], ['prassi', 'noun', 'b'], ['pratica', 'noun', 'a'], ['praticamente', 'adverb', 'a'], ['praticare', 'verb', 'b'], ['pratico', 'adjective', 'a'], ['prato', 'noun', 'b'], ['precario', 'adjective', 'b'], ['precedente', 'pres_part', 'a'], ['precedente', 'adjective', 'a'], ['precedente', 'noun', 'a'], ['precedentemente', 'adverb', 'b'], ['precedenza', 'noun', 'b'], ['precedere', 'verb', 'b'], ['precipitare', 'verb', 'b'], ['precisamente', 'adverb', 'b'], ['precisare', 'verb', 'a'], ['precisione', 'noun', 'b'], ['preciso', 'adjective', 'a'], ['preciso', 'adverb', 'a'], ['preda', 'noun', 'b'], ['predisporre', 'verb', 'b'], ['preferenza', 'noun', 'b'], ['preferire', 'verb', 'a'], ['preferito', 'past_part', 'b'], ['preferito', 'adjective', 'b'], ['preferito', 'noun', 'b'], ['pregare', 'verb', 'a'], ['preghiera', 'noun', 'b'], ['pregiato', 'past_part', 'c'], ['pregiato', 'adjective', 'c'], ['pregio', 'noun', 'b'], ['pregiudizio', 'noun', 'b'], ['prego', 'exclamation', 'a'], ['prelevare', 'verb', 'b'], ['preliminare', 'adjective', 'b'], ['preliminare', 'noun', 'b'], ['prmaman', 'adjective', 'c'], ['premere', 'verb', 'b'], ['premessa', 'noun', 'b'], ['premiare', 'verb', 'b'], ['premier', 'noun', 'b'], ['premio', 'noun', 'a'], ['premio', 'adjective', 'a'], ['prendere', 'verb', 'a'], ['prenotare', 'verb', 'b'], ['prenotazione', 'noun', 'c'], ['preoccupare', 'verb', 'a'], ['preoccupato', 'past_part', 'b'], ['preoccupato', 'adjective', 'b'], ['preoccupazione', 'noun', 'b'], ['preparare', 'verb', 'a'], ['preparazione', 'noun', 'b'], ['prepotente', 'adjective', 'c'], ['prepotente', 'noun', 'c'], ['presa', 'noun', 'a'], ['prescindere', 'verb', 'b'], ['prescrivere', 'verb', 'b'], ['prescrizione', 'noun', 'b'], ['presentare', 'verb', 'a'], ['presentazione', 'noun', 'b'], ['presente', 'adjective', 'a'], ['presente', 'noun', 'a'], ['presente', 'adverb', 'a'], ['presenza', 'noun', 'a'], ['presepe', 'noun', 'b'], ['preside', 'noun', 'c'], ['presidente', 'noun', 'a'], ['presidente', 'adjective', 'a'], ['presidenza', 'noun', 'b'], ['pressione', 'noun', 'a'], ['presso', 'adverb', 'a'], ['presso', 'preposition', 'a'], ['presso', 'noun', 'a'], ['presso', 'adjective', 'a'], ['prestare', 'verb', 'a'], ['prestazione', 'noun', 'b'], ['prestigio', 'noun', 'b'], ['prestigioso', 'adjective', 'b'], ['prestito', 'noun', 'b'], ['presto', 'adverb', 'a'], ['presto', 'exclamation', 'a'], ['presto', 'adjective', 'a'], ['presumere', 'verb', 'b'], ['presunto', 'past_part', 'b'], ['presunto', 'adjective', 'b'], ['presupposto', 'past_part', 'b'], ['presupposto', 'adjective', 'b'], ['presupposto', 'noun', 'b'], ['prete', 'noun', 'a'], ['pretendere', 'verb', 'a'], ['pretesa', 'noun', 'b'], ['pretesto', 'noun', 'b'], ['prevalentemente', 'adverb', 'b'], ['prevalere', 'verb', 'b'], ['prevedere', 'verb', 'a'], ['prevedibile', 'adjective', 'b'], ['prevenire', 'verb', 'b'], ['preventivo', 'adjective', 'b'], ['preventivo', 'noun', 'b'], ['prevenzione', 'noun', 'b'], ['previdenza', 'noun', 'c'], ['previsione', 'noun', 'b'], ['previsto', 'past_part', 'a'], ['previsto', 'adjective', 'a'], ['previsto', 'noun', 'a'], ['prezioso', 'adjective', 'a'], ['prezioso', 'noun', 'a'], ['prezzemolo', 'noun', 'c'], ['prezzo', 'noun', 'a'], ['prigione', 'noun', 'b'], ['prigioniero', 'adjective', 'b'], ['prigioniero', 'noun', 'b'], ['prima', 'adverb', 'a'], ['prima', 'adjective', 'a'], ['prima', 'noun', 'a'], ['prima', 'noun', 'a'], ['primario', 'adjective', 'b'], ['primario', 'noun', 'b'], ['primavera', 'noun', 'a'], ['primizia', 'noun', 'c'], ['primo', 'adjective', 'a'], ['primo', 'noun', 'a'], ['primo', 'adverb', 'a'], ['primula', 'noun', 'c'], ['principale', 'adjective', 'a'], ['principale', 'noun', 'a'], ['principalmente', 'adverb', 'b'], ['principe', 'noun', 'a'], ['principe', 'adjective', 'a'], ['principessa', 'noun', 'b'], ['principio', 'noun', 'a'], ['priorit', 'noun', 'b'], ['privacy', 'noun', 'b'], ['privare', 'verb', 'b'], ['privato', 'adjective', 'a'], ['privato', 'noun', 'a'], ['privilegio', 'noun', 'b'], ['privo', 'adjective', 'b'], ['privo', 'preposition', 'b'], ['privo', 'noun', 'b'], ['probabile', 'adjective', 'b'], ['probabilit', 'noun', 'b'], ['probabilmente', 'adverb', 'a'], ['problema', 'noun', 'a'], ['problematico', 'adjective', 'b'], ['procedere', 'verb', 'a'], ['procedimento', 'noun', 'b'], ['procedura', 'noun', 'a'], ['processo', 'noun', 'a'], ['proclamare', 'verb', 'b'], ['procura', 'noun', 'b'], ['procurare', 'verb', 'b'], ['procuratore', 'noun', 'b'], ['prodotto', 'past_part', 'a'], ['prodotto', 'adjective', 'a'], ['prodotto', 'noun', 'a'], ['produrre', 'verb', 'a'], ['produttivo', 'adjective', 'b'], ['produttore', 'adjective', 'b'], ['produttore', 'noun', 'b'], ['produzione', 'noun', 'a'], ['prof', 'noun', 'b'], ['professionale', 'adjective', 'a'], ['professione', 'noun', 'b'], ['professionista', 'noun', 'b'], ['professore', 'noun', 'a'], ['professoressa', 'noun', 'b'], ['profeta', 'noun', 'b'], ['profilattico', 'adjective', 'c'], ['profilattico', 'noun', 'c'], ['profilo', 'noun', 'a'], ['profitto', 'noun', 'b'], ['profondamente', 'adverb', 'b'], ['profondit', 'noun', 'b'], ['profondo', 'adjective', 'a'], ['profondo', 'noun', 'a'], ['profondo', 'adverb', 'a'], ['profumare', 'verb', 'b'], ['profumato', 'past_part', 'c'], ['profumato', 'adjective', 'c'], ['profumo', 'noun', 'b'], ['progettare', 'verb', 'b'], ['progettazione', 'noun', 'b'], ['progetto', 'noun', 'a'], ['programma', 'noun', 'a'], ['programmare', 'verb', 'b'], ['programmazione', 'noun', 'b'], ['progressista', 'adjective', 'c'], ['progressista', 'noun', 'c'], ['progressivo', 'adjective', 'b'], ['progresso', 'noun', 'b'], ['proibire', 'verb', 'b'], ['proiettare', 'verb', 'b'], ['proiettile', 'noun', 'b'], ['proiezione', 'noun', 'b'], ['prolunga', 'noun', 'c'], ['promessa', 'noun', 'b'], ['promettere', 'verb', 'a'], ['promozione', 'noun', 'b'], ['promuovere', 'verb', 'b'], ['pronto', 'adjective', 'a'], ['pronuncia', 'noun', 'c'], ['pronunciare', 'verb', 'a'], ['propaganda', 'noun', 'b'], ['propagandare', 'verb', 'c'], ['proporre', 'verb', 'a'], ['proporzione', 'noun', 'b'], ['proposito', 'noun', 'a'], ['proposizione', 'noun', 'c'], ['proposta', 'noun', 'a'], ['propriet', 'noun', 'a'], ['proprietario', 'adjective', 'a'], ['proprietario', 'noun', 'a'], ['proprio', 'adjective', 'a'], ['proprio', 'adverb', 'a'], ['proprio', 'noun', 'a'], ['prosa', 'noun', 'b'], ['prosciugare', 'verb', 'c'], ['prosciutto', 'noun', 'b'], ['prosecco', 'noun', 'c'], ['proseguire', 'verb', 'a'], ['prospettiva', 'noun', 'b'], ['prossimo', 'adjective', 'a'], ['prossimo', 'noun', 'a'], ['prostituta', 'noun', 'b'], ['protagonista', 'adjective', 'a'], ['protagonista', 'noun', 'a'], ['proteggere', 'verb', 'a'], ['proteina', 'noun', 'b'], ['protesta', 'noun', 'b'], ['protestare', 'verb', 'b'], ['protetto', 'past_part', 'b'], ['protetto', 'adjective', 'b'], ['protetto', 'noun', 'b'], ['protezione', 'noun', 'b'], ['protocollo', 'noun', 'b'], ['prova', 'noun', 'a'], ['provare', 'verb', 'a'], ['provenienza', 'noun', 'b'], ['provenire', 'verb', 'a'], ['provincia', 'noun', 'a'], ['provinciale', 'adjective', 'b'], ['provinciale', 'noun', 'b'], ['provocare', 'verb', 'a'], ['provola', 'noun', 'c'], ['provolone', 'noun', 'c'], ['provvedere', 'verb', 'b'], ['provvedimento', 'noun', 'b'], ['provvisorio', 'adjective', 'b'], ['prudere', 'verb', 'c'], ['prugna', 'noun', 'c'], ['prugna', 'adjective', 'c'], ['prurito', 'noun', 'c'], ['pseudonimo', 'noun', 'b'], ['pseudonimo', 'adjective', 'b'], ['psichiatra', 'noun', 'b'], ['psichiatria', 'noun', 'c'], ['psichico', 'adjective', 'b'], ['psicologia', 'noun', 'b'], ['psicologico', 'adjective', 'b'], ['psicologo', 'noun', 'b'], ['pub', 'noun', 'b'], ['pubblicare', 'verb', 'a'], ['pubblicazione', 'noun', 'b'], ['pubblicit', 'noun', 'a'], ['pubblicitario', 'adjective', 'b'], ['pubblicitario', 'noun', 'b'], ['pubblico', 'adjective', 'a'], ['pubblico', 'noun', 'a'], ['pugilato', 'noun', 'c'], ['pugliese', 'adjective', 'c'], ['pugliese', 'noun', 'c'], ['pugno', 'noun', 'a'], ['pulce', 'noun', 'c'], ['pulce', 'adjective', 'c'], ['pulcino', 'noun', 'c'], ['puledro', 'noun', 'c'], ['pulire', 'verb', 'a'], ['pulito', 'past_part', 'b'], ['pulito', 'adjective', 'b'], ['pulito', 'noun', 'b'], ['pulizia', 'noun', 'b'], ['pullman', 'noun', 'b'], ['pullover', 'noun', 'c'], ['pulmino', 'noun', 'c'], ['pulsante', 'pres_part', 'b'], ['pulsante', 'adjective', 'b'], ['pulsante', 'noun', 'b'], ['puma', 'noun', 'c'], ['pungere', 'verb', 'c'], ['punire', 'verb', 'b'], ['punizione', 'noun', 'b'], ['punk', 'adjective', 'c'], ['punk', 'noun', 'c'], ['punta', 'noun', 'a'], ['puntare', 'verb', 'a'], ['puntata', 'noun', 'b'], ['puntato', 'past_part', 'b'], ['puntato', 'adjective', 'b'], ['punteggio', 'noun', 'c'], ['puntiglio', 'noun', 'c'], ['puntino', 'noun', 'b'], ['punto', 'noun', 'a'], ['puntuale', 'adjective', 'b'], ['puntura', 'noun', 'c'], ['pupa', 'noun', 'b'], ['pupazzo', 'noun', 'c'], ['pupo', 'noun', 'c'], ['purch', 'conjunction', 'b'], ['pure', 'adverb', 'a'], ['pure', 'conjunction', 'a'], ['pur', 'noun', 'c'], ['purga', 'noun', 'c'], ['puro', 'adjective', 'a'], ['puro', 'noun', 'a'], ['purtroppo', 'adverb', 'a'], ['puttana', 'noun', 'b'], ['puzza', 'noun', 'b'], ['puzzare', 'verb', 'b'], ['puzzle', 'noun', 'c'], ['qua', 'adverb', 'a'], ['quaderno', 'noun', 'b'], ['quadrato', 'past_part', 'b'], ['quadrato', 'adjective', 'b'], ['quadrato', 'noun', 'b'], ['quadrifoglio', 'noun', 'c'], ['quadro', 'adjective', 'a'], ['quadro', 'noun', 'a'], ['quaglia', 'noun', 'c'], ['qualche', 'adjective', 'a'], ['qualche', 'adverb', 'a'], ['qualcosa', 'pronoun', 'a'], ['qualcuno', 'pronoun', 'a'], ['qualcuno', 'adjective', 'a'], ['qualcuno', 'noun', 'a'], ['quale', 'adjective', 'a'], ['quale', 'pronoun', 'a'], ['quale', 'adverb', 'a'], ['quale', 'noun', 'a'], ['qualificare', 'verb', 'b'], ['qualit', 'noun', 'a'], ['qualora', 'conjunction', 'b'], ['qualsiasi', 'adjective', 'a'], ['qualunque', 'adjective', 'a'], ['qualunque', 'pronoun', 'a'], ['quando', 'conjunction', 'a'], ['quando', 'adverb', 'a'], ['quando', 'noun', 'a'], ['quantit', 'noun', 'a'], ['quantitativo', 'adjective', 'b'], ['quantitativo', 'noun', 'b'], ['quanto', 'adjective', 'a'], ['quanto', 'pronoun', 'a'], ['quanto', 'adverb', 'a'], ['quanto', 'noun', 'a'], ['quaranta', 'adjective', 'a'], ['quaranta', 'noun', 'a'], ['quarta', 'noun', 'b'], ['quartiere', 'noun', 'a'], ['quarto', 'adjective', 'a'], ['quarto', 'noun', 'a'], ['quasi', 'adverb', 'a'], ['quasi', 'conjunction', 'a'], ['quattordici', 'adjective', 'b'], ['quattordici', 'noun', 'b'], ['quattro', 'adjective', 'a'], ['quattro', 'noun', 'a'], ['quello', 'adjective', 'a'], ['quello', 'pronoun', 'a'], ['quercia', 'noun', 'c'], ['questione', 'noun', 'a'], ['questo', 'adjective', 'a'], ['questo', 'pronoun', 'a'], ['questura', 'noun', 'b'], ['qui', 'adverb', 'a'], ['quindi', 'adverb', 'a'], ['quindi', 'conjunction', 'a'], ['quindici', 'adjective', 'a'], ['quindici', 'noun', 'a'], ['quinta', 'noun', 'b'], ['quinto', 'adjective', 'b'], ['quinto', 'noun', 'b'], ['quiz', 'noun', 'a'], ['quota', 'noun', 'a'], ['quotidiano', 'adjective', 'a'], ['quotidiano', 'noun', 'a'], ['rabbia', 'noun', 'a'], ['racchetta', 'noun', 'c'], ['racchiudere', 'verb', 'b'], ['raccogliere', 'verb', 'a'], ['raccolta', 'noun', 'a'], ['raccomandare', 'verb', 'b'], ['raccomandazione', 'noun', 'c'], ['raccontare', 'verb', 'a'], ['racconto', 'noun', 'a'], ['raddoppiare', 'verb', 'b'], ['raddrizzare', 'verb', 'c'], ['radere', 'verb', 'c'], ['radiazione', 'noun', 'b'], ['radicale', 'adjective', 'b'], ['radicale', 'noun', 'b'], ['radicchio', 'noun', 'c'], ['radice', 'noun', 'a'], ['radio', 'noun', 'a'], ['radio', 'adjective', 'a'], ['rado', 'adjective', 'b'], ['rado', 'adverb', 'b'], ['raffigurare', 'verb', 'b'], ['raffinato', 'past_part', 'b'], ['raffinato', 'adjective', 'b'], ['raffinato', 'noun', 'b'], ['rafforzamento', 'noun', 'c'], ['rafforzare', 'verb', 'b'], ['raffreddore', 'noun', 'c'], ['ragazza', 'noun', 'a'], ['ragazzino', 'noun', 'a'], ['ragazzo', 'noun', 'a'], ['raggio', 'noun', 'a'], ['raggiungere', 'verb', 'a'], ['ragionamento', 'noun', 'b'], ['ragionare', 'verb', 'b'], ['ragione', 'noun', 'a'], ['ragionevole', 'adjective', 'b'], ['ragioniere', 'noun', 'b'], ['ragnatela', 'noun', 'c'], ['ragno', 'noun', 'c'], ['rag', 'noun', 'c'], ['rallegrare', 'verb', 'c'], ['rallentare', 'verb', 'b'], ['rame', 'noun', 'b'], ['rammendo', 'noun', 'c'], ['ramo', 'noun', 'b'], ['rampicante', 'pres_part', 'c'], ['rampicante', 'adjective', 'c'], ['rampicante', 'noun', 'c'], ['rana', 'noun', 'c'], ['rancio', 'noun', 'c'], ['rapa', 'noun', 'c'], ['rapidamente', 'adverb', 'b'], ['rapido', 'adjective', 'a'], ['rapido', 'noun', 'a'], ['rapimento', 'noun', 'c'], ['rapina', 'noun', 'b'], ['rapinatore', 'adjective', 'c'], ['rapinatore', 'noun', 'c'], ['rapire', 'verb', 'b'], ['rapporto', 'noun', 'a'], ['rappresentante', 'pres_part', 'b'], ['rappresentante', 'adjective', 'b'], ['rappresentante', 'noun', 'b'], ['rappresentanza', 'noun', 'b'], ['rappresentare', 'verb', 'a'], ['rappresentazione', 'noun', 'b'], ['raramente', 'adverb', 'b'], ['raro', 'adjective', 'a'], ['raro', 'noun', 'a'], ['raro', 'adverb', 'a'], ['rasare', 'verb', 'c'], ['rasoio', 'noun', 'c'], ['rassegna', 'noun', 'b'], ['rassegnare', 'verb', 'b'], ['rassegnazione', 'noun', 'c'], ['rasserenare', 'verb', 'c'], ['rassicurare', 'verb', 'b'], ['rastrello', 'noun', 'c'], ['rata', 'noun', 'c'], ['rateale', 'adjective', 'c'], ['rattristare', 'verb', 'c'], ['rauco', 'adjective', 'c'], ['ravanello', 'noun', 'c'], ['razionale', 'adjective', 'b'], ['razionale', 'noun', 'b'], ['razza', 'noun', 'b'], ['razzo', 'noun', 'c'], ['re', 'noun', 'a'], ['reagire', 'verb', 'a'], ['reale', 'adjective', 'a'], ['reale', 'noun', 'a'], ['realistico', 'adjective', 'b'], ['realizzare', 'verb', 'a'], ['realizzazione', 'noun', 'b'], ['realmente', 'adverb', 'b'], ['realt', 'noun', 'a'], ['reato', 'noun', 'a'], ['reazione', 'noun', 'a'], ['recare', 'verb', 'a'], ['recensione', 'noun', 'b'], ['recente', 'adjective', 'a'], ['recentemente', 'adverb', 'b'], ['recintare', 'verb', 'c'], ['recinto', 'past_part', 'c'], ['recinto', 'adjective', 'c'], ['recinto', 'noun', 'c'], ['recipiente', 'adjective', 'c'], ['recipiente', 'noun', 'c'], ['reciproco', 'adjective', 'b'], ['reciproco', 'noun', 'b'], ['recita', 'noun', 'c'], ['recitare', 'verb', 'a'], ['reclame', 'noun', 'c'], ['reclame', 'adjective', 'c'], ['reclamo', 'noun', 'c'], ['recluta', 'noun', 'c'], ['record', 'noun', 'b'], ['recuperare', 'verb', 'a'], ['recupero', 'noun', 'b'], ['redazione', 'noun', 'b'], ['reddito', 'noun', 'b'], ['redigere', 'verb', 'b'], ['referendum', 'noun', 'b'], ['regalare', 'verb', 'a'], ['regale', 'adjective', 'b'], ['regalo', 'noun', 'a'], ['reggere', 'verb', 'a'], ['reggimento', 'noun', 'c'], ['reggiseno', 'noun', 'b'], ['regia', 'noun', 'b'], ['regime', 'noun', 'a'], ['regina', 'noun', 'a'], ['regionale', 'adjective', 'b'], ['regionale', 'noun', 'b'], ['regione', 'noun', 'a'], ['regista', 'noun', 'a'], ['registrare', 'verb', 'a'], ['registratore', 'adjective', 'c'], ['registratore', 'noun', 'c'], ['registrazione', 'noun', 'a'], ['registro', 'noun', 'b'], ['regnare', 'verb', 'b'], ['regno', 'noun', 'a'], ['regola', 'noun', 'a'], ['regolamento', 'noun', 'b'], ['regolare', 'adjective', 'b'], ['regolare', 'noun', 'b'], ['regolare', 'verb', 'b'], ['regolarmente', 'adverb', 'b'], ['relativamente', 'adverb', 'b'], ['relativo', 'adjective', 'a'], ['relazione', 'noun', 'a'], ['religione', 'noun', 'a'], ['religioso', 'adjective', 'a'], ['religioso', 'noun', 'a'], ['remare', 'verb', 'c'], ['remo', 'noun', 'c'], ['remoto', 'adjective', 'b'], ['rendere', 'verb', 'a'], ['rene', 'noun', 'b'], ['reparto', 'noun', 'b'], ['repertorio', 'noun', 'b'], ['replica', 'noun', 'b'], ['replicare', 'verb', 'b'], ['repressione', 'noun', 'c'], ['reprimere', 'verb', 'c'], ['repubblica', 'noun', 'a'], ['repubblicano', 'adjective', 'b'], ['repubblicano', 'noun', 'b'], ['requisito', 'noun', 'b'], ['resa', 'noun', 'b'], ['residente', 'adjective', 'b'], ['residente', 'noun', 'b'], ['residenza', 'noun', 'b'], ['residuo', 'adjective', 'b'], ['residuo', 'noun', 'b'], ['resistente', 'pres_part', 'b'], ['resistente', 'adjective', 'b'], ['resistente', 'noun', 'b'], ['resistenza', 'noun', 'b'], ['resistere', 'verb', 'a'], ['resoconto', 'noun', 'c'], ['respingere', 'verb', 'b'], ['respirare', 'verb', 'a'], ['respirazione', 'noun', 'c'], ['respiro', 'noun', 'b'], ['responsabile', 'adjective', 'a'], ['responsabile', 'noun', 'a'], ['responsabilit', 'noun', 'a'], ['restare', 'verb', 'a'], ['restituire', 'verb', 'b'], ['resto', 'noun', 'a'], ['restringere', 'verb', 'b'], ['rete', 'noun', 'a'], ['retorica', 'noun', 'b'], ['retro', 'adverb', 'b'], ['retro', 'noun', 'b'], ['retta', 'noun', 'b'], ['rettangolare', 'adjective', 'c'], ['rettile', 'noun', 'c'], ['rettile', 'adjective', 'c'], ['retto', 'adjective', 'b'], ['retto', 'noun', 'b'], ['revisione', 'noun', 'b'], ['rialzare', 'verb', 'b'], ['riaprire', 'verb', 'b'], ['riassumere', 'verb', 'b'], ['ribadire', 'verb', 'b'], ['ribattere', 'verb', 'b'], ['ribellare', 'verb', 'b'], ['ribelle', 'adjective', 'b'], ['ribelle', 'noun', 'b'], ['ricadere', 'verb', 'b'], ['ricaduta', 'noun', 'c'], ['ricalcare', 'verb', 'c'], ['ricamare', 'verb', 'c'], ['ricambiare', 'verb', 'b'], ['ricambio', 'noun', 'c'], ['ricamo', 'noun', 'c'], ['ricarica', 'noun', 'c'], ['ricavare', 'verb', 'b'], ['ricchezza', 'noun', 'b'], ['riccio', 'adjective', 'c'], ['riccio', 'noun', 'c'], ['ricciolo', 'adjective', 'c'], ['ricciolo', 'noun', 'c'], ['ricco', 'adjective', 'a'], ['ricerca', 'noun', 'a'], ['ricercare', 'verb', 'b'], ['ricercatore', 'adjective', 'b'], ['ricercatore', 'noun', 'b'], ['ricetta', 'noun', 'a'], ['ricevere', 'verb', 'a'], ['ricevimento', 'noun', 'c'], ['ricevuta', 'noun', 'b'], ['richiamare', 'verb', 'a'], ['richiamo', 'noun', 'b'], ['richiedere', 'verb', 'a'], ['richiesta', 'noun', 'a'], ['richiudere', 'verb', 'b'], ['ricominciare', 'verb', 'a'], ['ricompensa', 'noun', 'c'], ['ricompensare', 'verb', 'c'], ['riconciliarsi', 'verb', 'c'], ['riconoscere', 'verb', 'a'], ['riconoscimento', 'noun', 'b'], ['ricopiare', 'verb', 'c'], ['ricoprire', 'verb', 'b'], ['ricordare', 'verb', 'a'], ['ricordo', 'noun', 'a'], ['ricorrere', 'verb', 'b'], ['ricorso', 'noun', 'b'], ['ricostruire', 'verb', 'b'], ['ricostruzione', 'noun', 'b'], ['ricotta', 'noun', 'c'], ['ricoverare', 'verb', 'b'], ['ricovero', 'noun', 'c'], ['ricreazione', 'noun', 'c'], ['ridare', 'verb', 'b'], ['ridere', 'verb', 'a'], ['ridere', 'noun', 'a'], ['ridicolo', 'adjective', 'b'], ['ridicolo', 'noun', 'b'], ['ridotto', 'past_part', 'b'], ['ridotto', 'adjective', 'b'], ['ridotto', 'noun', 'b'], ['ridurre', 'verb', 'a'], ['riduzione', 'noun', 'b'], ['riempire', 'verb', 'a'], ['rientrare', 'verb', 'a'], ['rientro', 'noun', 'b'], ['rifare', 'verb', 'a'], ['riferimento', 'noun', 'a'], ['riferire', 'verb', 'a'], ['rifinire', 'verb', 'c'], ['rifiutare', 'verb', 'a'], ['rifiuto', 'noun', 'a'], ['riflessione', 'noun', 'a'], ['riflesso', 'noun', 'b'], ['riflettere', 'verb', 'a'], ['riflettore', 'noun', 'c'], ['riflettore', 'adjective', 'c'], ['riforma', 'noun', 'b'], ['rifornimento', 'noun', 'c'], ['rifugiare', 'verb', 'b'], ['rifugio', 'noun', 'b'], ['riga', 'noun', 'a'], ['rigattiere', 'noun', 'c'], ['rigido', 'adjective', 'b'], ['rigore', 'noun', 'b'], ['rigoroso', 'adjective', 'b'], ['rigovernare', 'verb', 'c'], ['riguardare', 'verb', 'a'], ['riguardo', 'noun', 'a'], ['rilasciare', 'verb', 'b'], ['rilassare', 'verb', 'a'], ['rilegare', 'verb', 'c'], ['rileggere', 'verb', 'b'], ['rilevante', 'pres_part', 'b'], ['rilevante', 'adjective', 'b'], ['rilevare', 'verb', 'b'], ['rilievo', 'noun', 'b'], ['rima', 'noun', 'b'], ['rimandare', 'verb', 'b'], ['rimanenza', 'noun', 'c'], ['rimanere', 'verb', 'a'], ['rimbombare', 'verb', 'c'], ['rimborsare', 'verb', 'c'], ['rimediare', 'verb', 'b'], ['rimedio', 'noun', 'b'], ['rimettere', 'verb', 'a'], ['rimodernare', 'verb', 'c'], ['rimorchio', 'noun', 'c'], ['rimpiangere', 'verb', 'b'], ['rimproverare', 'verb', 'b'], ['rimprovero', 'noun', 'c'], ['rimuovere', 'verb', 'b'], ['rinascere', 'verb', 'b'], ['rinascimento', 'noun', 'b'], ['rinascimento', 'adjective', 'b'], ['rincarare', 'verb', 'c'], ['rinchiudere', 'verb', 'b'], ['rincorsa', 'noun', 'c'], ['rinforzo', 'noun', 'c'], ['rinfresco', 'noun', 'c'], ['ringhiare', 'verb', 'c'], ['ringhiera', 'noun', 'c'], ['ringhio', 'noun', 'c'], ['ringiovanire', 'verb', 'c'], ['ringraziare', 'verb', 'a'], ['rinnegare', 'verb', 'c'], ['rinnovare', 'verb', 'b'], ['rinoceronte', 'noun', 'c'], ['rintracciare', 'verb', 'b'], ['rinuncia', 'noun', 'c'], ['rinunciare', 'verb', 'a'], ['rinvenire', 'verb', 'b'], ['rinviare', 'verb', 'b'], ['rinvio', 'noun', 'c'], ['rione', 'noun', 'c'], ['riordinare', 'verb', 'c'], ['riparare', 'verb', 'b'], ['riparo', 'noun', 'b'], ['ripartire', 'verb', 'b'], ['ripartire', 'verb', 'b'], ['ripensamento', 'noun', 'c'], ['ripensare', 'verb', 'b'], ['ripetente', 'pres_part', 'c'], ['ripetente', 'adjective', 'c'], ['ripetente', 'noun', 'c'], ['ripetere', 'verb', 'a'], ['ripetizione', 'noun', 'b'], ['ripido', 'adjective', 'c'], ['ripiego', 'noun', 'c'], ['ripieno', 'adjective', 'c'], ['ripieno', 'noun', 'c'], ['riportare', 'verb', 'a'], ['riposare', 'verb', 'b'], ['riposo', 'noun', 'b'], ['riposo', 'loc-comando', 'b'], ['riposo', 'noun', 'b'], ['riprendere', 'verb', 'a'], ['ripresa', 'noun', 'b'], ['riprodurre', 'verb', 'b'], ['riproduzione', 'noun', 'a'], ['riproporre', 'verb', 'b'], ['riprovare', 'verb', 'b'], ['ripulire', 'verb', 'b'], ['risaia', 'noun', 'c'], ['risalire', 'verb', 'a'], ['risarcimento', 'noun', 'b'], ['risata', 'noun', 'b'], ['riscaldamento', 'noun', 'b'], ['riscaldare', 'verb', 'b'], ['riscattare', 'verb', 'c'], ['riscatto', 'noun', 'c'], ['rischiare', 'verb', 'a'], ['rischio', 'noun', 'a'], ['rischioso', 'adjective', 'b'], ['risciacquare', 'verb', 'c'], ['riscontrare', 'verb', 'b'], ['riscontro', 'noun', 'b'], ['riscuotere', 'verb', 'b'], ['risentimento', 'noun', 'c'], ['risentire', 'verb', 'b'], ['riserva', 'noun', 'b'], ['riservare', 'verb', 'a'], ['riservato', 'past_part', 'a'], ['riservato', 'adjective', 'a'], ['risiedere', 'verb', 'b'], ['riso', 'noun', 'b'], ['risoluzione', 'noun', 'b'], ['risolvere', 'verb', 'a'], ['risonanza', 'noun', 'b'], ['risorsa', 'noun', 'a'], ['risparmiare', 'verb', 'b'], ['risparmio', 'noun', 'b'], ['rispettare', 'verb', 'a'], ['rispettivamente', 'adverb', 'b'], ['rispettivo', 'adjective', 'b'], ['rispetto', 'noun', 'a'], ['risplendere', 'verb', 'c'], ['rispondere', 'verb', 'a'], ['risposta', 'noun', 'a'], ['rissa', 'noun', 'b'], ['ristampare', 'verb', 'c'], ['ristorante', 'noun', 'a'], ['ristretto', 'past_part', 'b'], ['ristretto', 'adjective', 'b'], ['ristretto', 'noun', 'b'], ['risultare', 'verb', 'a'], ['risultato', 'past_part', 'a'], ['risultato', 'adjective', 'a'], ['risultato', 'noun', 'a'], ['risvegliare', 'verb', 'b'], ['risveglio', 'noun', 'b'], ['ritagliare', 'verb', 'b'], ['ritardare', 'verb', 'b'], ['ritardo', 'noun', 'a'], ['ritenere', 'verb', 'a'], ['ritirare', 'verb', 'a'], ['ritirata', 'noun', 'c'], ['ritiro', 'noun', 'b'], ['ritmo', 'noun', 'a'], ['rito', 'noun', 'b'], ['ritoccare', 'verb', 'c'], ['ritornare', 'verb', 'a'], ['ritornello', 'noun', 'c'], ['ritorno', 'noun', 'a'], ['ritrarre', 'verb', 'b'], ['ritratto', 'past_part', 'b'], ['ritratto', 'adjective', 'b'], ['ritratto', 'noun', 'b'], ['ritrovare', 'verb', 'a'], ['ritrovo', 'noun', 'c'], ['ritto', 'adjective', 'c'], ['ritto', 'noun', 'c'], ['ritto', 'adverb', 'c'], ['ritto', 'preposition', 'c'], ['rituale', 'adjective', 'b'], ['rituale', 'noun', 'b'], ['riunione', 'noun', 'a'], ['riunire', 'verb', 'a'], ['riunito', 'past_part', 'c'], ['riunito', 'adjective', 'c'], ['riunito', 'noun', 'c'], ['riuscire', 'verb', 'a'], ['riuscita', 'noun', 'c'], ['riva', 'noun', 'b'], ['rivale', 'adjective', 'b'], ['rivale', 'noun', 'b'], ['rivedere', 'verb', 'a'], ['rivelare', 'verb', 'a'], ['rivelazione', 'noun', 'b'], ['rivendicare', 'verb', 'b'], ['rivendita', 'noun', 'c'], ['rivestimento', 'noun', 'c'], ['rivestire', 'verb', 'b'], ['rivincita', 'noun', 'c'], ['rivista', 'noun', 'a'], ['rivisto', 'past_part', 'b'], ['rivisto', 'adjective', 'b'], ['rivolgere', 'verb', 'a'], ['rivolta', 'noun', 'b'], ['rivoltare', 'verb', 'c'], ['rivoluzionario', 'adjective', 'b'], ['rivoluzionario', 'noun', 'b'], ['rivoluzione', 'noun', 'a'], ['roba', 'noun', 'a'], ['robot', 'noun', 'b'], ['robusto', 'adjective', 'b'], ['rocca', 'noun', 'c'], ['rocchetto', 'noun', 'c'], ['roccia', 'noun', 'b'], ['roccioso', 'adjective', 'c'], ['rock', 'noun', 'b'], ['rock', 'adjective', 'b'], ['rodaggio', 'noun', 'c'], ['rodere', 'verb', 'c'], ['romagnolo', 'adjective', 'c'], ['romagnolo', 'noun', 'c'], ['romano', 'adjective', 'a'], ['romano', 'noun', 'a'], ['romantico', 'adjective', 'b'], ['romantico', 'noun', 'b'], ['romanzo', 'noun', 'a'], ['rombo', 'noun', 'c'], ['romeno', 'adjective', 'c'], ['romeno', 'noun', 'c'], ['rompere', 'verb', 'a'], ['rondine', 'noun', 'c'], ['ronzare', 'verb', 'c'], ['ronzio', 'noun', 'c'], ['rosa', 'noun', 'a'], ['rosa', 'adjective', 'a'], ['rosario', 'noun', 'c'], ['rosato', 'adjective', 'c'], ['rosato', 'noun', 'c'], ['roseo', 'adjective', 'c'], ['roseo', 'noun', 'c'], ['rosetta', 'noun', 'c'], ['rosmarino', 'noun', 'c'], ['rosolia', 'noun', 'c'], ['rosso', 'adjective', 'a'], ['rosso', 'noun', 'a'], ['rossore', 'noun', 'c'], ['rosticceria', 'noun', 'c'], ['rotaia', 'noun', 'c'], ['rotella', 'noun', 'c'], ['rotolare', 'verb', 'c'], ['rotondo', 'adjective', 'b'], ['rotondo', 'noun', 'b'], ['rotta', 'noun', 'b'], ['rotto', 'past_part', 'b'], ['rotto', 'adjective', 'b'], ['rotto', 'noun', 'b'], ['rottura', 'noun', 'b'], ['roulotte', 'noun', 'c'], ['rovesciare', 'verb', 'b'], ['rovescio', 'adjective', 'b'], ['rovescio', 'noun', 'b'], ['rovina', 'noun', 'b'], ['rovinare', 'verb', 'a'], ['rovo', 'noun', 'c'], ['rozzo', 'adjective', 'c'], ['rubare', 'verb', 'a'], ['rubinetto', 'noun', 'c'], ['rubrica', 'noun', 'b'], ['rude', 'adjective', 'c'], ['ruga', 'noun', 'c'], ['ruggine', 'noun', 'c'], ['ruggine', 'adjective', 'c'], ['ruggire', 'verb', 'c'], ['ruggito', 'past_part', 'c'], ['ruggito', 'noun', 'c'], ['rullo', 'noun', 'c'], ['rumeno', 'adjective', 'c'], ['rumeno', 'noun', 'c'], ['ruminante', 'pres_part', 'c'], ['ruminante', 'adjective', 'c'], ['ruminante', 'noun', 'c'], ['rumore', 'noun', 'a'], ['ruolo', 'noun', 'a'], ['ruota', 'noun', 'b'], ['ruotare', 'verb', 'b'], ['ruscello', 'noun', 'c'], ['ruspa', 'noun', 'c'], ['russare', 'verb', 'c'], ['russo', 'adjective', 'a'], ['russo', 'noun', 'a'], ['rustico', 'adjective', 'c'], ['rustico', 'noun', 'c'], ['ruttare', 'verb', 'c'], ['rutto', 'noun', 'c'], ['sabato', 'noun', 'a'], ['sabbia', 'noun', 'b'], ['sabbia', 'adjective', 'b'], ['sabotare', 'verb', 'c'], ['saccheggiare', 'verb', 'c'], ['sacchetto', 'noun', 'b'], ['sacco', 'noun', 'a'], ['sacerdote', 'noun', 'b'], ['sacrificare', 'verb', 'b'], ['sacrificio', 'noun', 'b'], ['sacro', 'adjective', 'b'], ['sacro', 'noun', 'b'], ['safari', 'noun', 'c'], ['saga', 'noun', 'b'], ['saggezza', 'noun', 'b'], ['saggio', 'adjective', 'b'], ['saggio', 'noun', 'b'], ['saggio', 'noun', 'b'], ['sagra', 'noun', 'c'], ['sagrestano', 'noun', 'c'], ['sagrestano', 'adjective', 'c'], ['sala', 'noun', 'a'], ['salame', 'noun', 'c'], ['salare', 'verb', 'c'], ['salario', 'adjective', 'b'], ['salario', 'noun', 'b'], ['salatino', 'noun', 'c'], ['salato', 'past_part', 'b'], ['salato', 'adjective', 'b'], ['salato', 'noun', 'b'], ['saldatura', 'noun', 'c'], ['sale', 'noun', 'b'], ['salice', 'noun', 'c'], ['saliera', 'noun', 'c'], ['salire', 'verb', 'a'], ['salita', 'noun', 'b'], ['saliva', 'noun', 'c'], ['salmone', 'noun', 'c'], ['salmone', 'adjective', 'c'], ['salone', 'noun', 'b'], ['salotto', 'noun', 'b'], ['salsa', 'noun', 'b'], ['salsiccia', 'noun', 'c'], ['saltare', 'verb', 'a'], ['saltellare', 'verb', 'c'], ['salto', 'noun', 'b'], ['salume', 'noun', 'c'], ['salutare', 'verb', 'a'], ['salutare', 'noun', 'a'], ['salute', 'noun', 'a'], ['salute', 'exclamation', 'a'], ['saluto', 'noun', 'a'], ['salvadanaio', 'noun', 'c'], ['salvagente', 'noun', 'c'], ['salvare', 'verb', 'a'], ['salvaslip', 'noun', 'c'], ['salvatore', 'adjective', 'b'], ['salvatore', 'noun', 'b'], ['salve', 'exclamation', 'b'], ['salvezza', 'noun', 'b'], ['salvia', 'noun', 'c'], ['salvietta', 'noun', 'c'], ['salvo', 'adjective', 'a'], ['salvo', 'preposition', 'a'], ['sandalo', 'noun', 'c'], ['sangue', 'noun', 'a'], ['sangue', 'adjective', 'a'], ['sanguinare', 'verb', 'c'], ['sanguisuga', 'noun', 'c'], ['sanit', 'noun', 'b'], ['sanitaria', 'noun', 'c'], ['sanitario', 'adjective', 'b'], ['sanitario', 'noun', 'b'], ['sano', 'adjective', 'a'], ['santo', 'adjective', 'a'], ['santo', 'noun', 'a'], ['sanzione', 'noun', 'b'], ['sapere', 'verb', 'a'], ['sapere', 'noun', 'b'], ['sapiente', 'adjective', 'c'], ['sapiente', 'noun', 'c'], ['sapone', 'noun', 'b'], ['saponetta', 'noun', 'c'], ['sapore', 'noun', 'b'], ['saporito', 'past_part', 'c'], ['saporito', 'adjective', 'c'], ['sardina', 'noun', 'c'], ['sardo', 'adjective', 'b'], ['sardo', 'noun', 'b'], ['sarto', 'noun', 'c'], ['sasso', 'noun', 'b'], ['satellite', 'noun', 'b'], ['sazio', 'past_part', 'c'], ['sazio', 'adjective', 'c'], ['sbadato', 'adjective', 'c'], ['sbadato', 'noun', 'c'], ['sbadigliare', 'verb', 'c'], ['sbadiglio', 'noun', 'c'], ['sbagliare', 'verb', 'a'], ['sbagliato', 'past_part', 'a'], ['sbagliato', 'adjective', 'a'], ['sbaglio', 'noun', 'b'], ['sbarbare', 'verb', 'c'], ['sbarcare', 'verb', 'b'], ['sbarra', 'noun', 'c'], ['sbarramento', 'noun', 'c'], ['sbattere', 'verb', 'a'], ['sberla', 'noun', 'c'], ['sbiadire', 'verb', 'c'], ['sbiancare', 'verb', 'c'], ['sbigottire', 'verb', 'c'], ['sbloccare', 'verb', 'c'], ['sboccare', 'verb', 'c'], ['sbocciare', 'verb', 'c'], ['sbocco', 'noun', 'c'], ['sbornia', 'noun', 'c'], ['sbottonare', 'verb', 'c'], ['sbriciolare', 'verb', 'c'], ['sbrigare', 'verb', 'b'], ['sbronza', 'noun', 'c'], ['sbronzo', 'adjective', 'c'], ['sbucciare', 'verb', 'c'], ['sbuffare', 'verb', 'c'], ['scacchiera', 'noun', 'c'], ['scadenza', 'noun', 'b'], ['scadere', 'verb', 'b'], ['scaffale', 'noun', 'b'], ['scafo', 'noun', 'c'], ['scala', 'noun', 'a'], ['scalare', 'verb', 'b'], ['scalata', 'noun', 'c'], ['scaldabagno', 'noun', 'c'], ['scaldare', 'verb', 'b'], ['scalinata', 'noun', 'c'], ['scalino', 'noun', 'c'], ['scalpello', 'noun', 'c'], ['scalzo', 'adjective', 'c'], ['scambiare', 'verb', 'a'], ['scambio', 'noun', 'a'], ['scamorza', 'noun', 'c'], ['scampagnata', 'noun', 'c'], ['scampo', 'noun', 'c'], ['scandalizzare', 'verb', 'c'], ['scandalo', 'noun', 'b'], ['scandire', 'verb', 'b'], ['scansare', 'verb', 'c'], ['scapito', 'noun', 'c'], ['scappamento', 'noun', 'c'], ['scappare', 'verb', 'a'], ['scappatoia', 'noun', 'c'], ['scarabocchiare', 'verb', 'c'], ['scarabocchio', 'noun', 'c'], ['scarafaggio', 'noun', 'c'], ['scarcerare', 'verb', 'c'], ['scaricare', 'verb', 'a'], ['scaricatore', 'noun', 'c'], ['scarico', 'noun', 'b'], ['scarlattina', 'noun', 'c'], ['scarpa', 'noun', 'a'], ['scarpiera', 'noun', 'c'], ['scarpone', 'noun', 'c'], ['scarso', 'adjective', 'b'], ['scartare', 'verb', 'b'], ['scatenare', 'verb', 'b'], ['scatola', 'noun', 'a'], ['scattare', 'verb', 'a'], ['scatto', 'noun', 'b'], ['scavalcare', 'verb', 'c'], ['scavare', 'verb', 'b'], ['scavo', 'noun', 'c'], ['scegliere', 'verb', 'a'], ['scelta', 'noun', 'a'], ['scemo', 'past_part', 'b'], ['scemo', 'adjective', 'b'], ['scemo', 'noun', 'b'], ['scena', 'noun', 'a'], ['scenario', 'noun', 'b'], ['scendere', 'verb', 'a'], ['sceneggiatura', 'noun', 'b'], ['sceriffo', 'noun', 'c'], ['scheda', 'noun', 'b'], ['schedario', 'noun', 'c'], ['scheggia', 'noun', 'c'], ['scheletro', 'noun', 'c'], ['schema', 'noun', 'b'], ['schermo', 'noun', 'a'], ['scherzare', 'verb', 'a'], ['scherzo', 'noun', 'b'], ['scherzoso', 'adjective', 'c'], ['schiacciare', 'verb', 'b'], ['schiacciato', 'past_part', 'c'], ['schiacciato', 'adjective', 'c'], ['schiaffo', 'noun', 'b'], ['schiavo', 'adjective', 'b'], ['schiavo', 'noun', 'b'], ['schiena', 'noun', 'a'], ['schierare', 'verb', 'b'], ['schietto', 'adjective', 'c'], ['schifo', 'noun', 'a'], ['schifo', 'adjective', 'a'], ['schiuma', 'noun', 'c'], ['schizzare', 'verb', 'b'], ['schizzo', 'noun', 'b'], ['sci', 'noun', 'b'], ['scia', 'noun', 'b'], ['sciacquare', 'verb', 'c'], ['scialle', 'noun', 'c'], ['sciame', 'noun', 'c'], ['sciare', 'verb', 'c'], ['sciarpa', 'noun', 'c'], ['sciatore', 'noun', 'c'], ['scientifico', 'adjective', 'a'], ['scientifico', 'noun', 'a'], ['scienza', 'noun', 'a'], ['scienziato', 'noun', 'b'], ['scienziato', 'adjective', 'b'], ['scimmia', 'noun', 'b'], ['scintilla', 'noun', 'b'], ['sciocchezza', 'noun', 'b'], ['sciocco', 'adjective', 'b'], ['sciocco', 'noun', 'b'], ['sciogliere', 'verb', 'b'], ['scioperare', 'verb', 'c'], ['sciopero', 'noun', 'b'], ['scirocco', 'noun', 'c'], ['sciroppo', 'noun', 'c'], ['scivolare', 'verb', 'b'], ['scivolata', 'noun', 'c'], ['scivolo', 'noun', 'c'], ['scocciare', 'verb', 'c'], ['scodella', 'noun', 'c'], ['scodinzolare', 'verb', 'c'], ['scoglio', 'noun', 'c'], ['scoiattolo', 'noun', 'c'], ['scolapiatti', 'noun', 'c'], ['scolaro', 'noun', 'c'], ['scolastico', 'adjective', 'b'], ['scolastico', 'noun', 'b'], ['scolpire', 'verb', 'c'], ['scommessa', 'noun', 'b'], ['scommettere', 'verb', 'b'], ['scomodo', 'adjective', 'c'], ['scomparire', 'verb', 'a'], ['scomparsa', 'noun', 'b'], ['scompartimento', 'noun', 'c'], ['sconfiggere', 'verb', 'b'], ['sconfitta', 'noun', 'b'], ['scongelare', 'verb', 'c'], ['sconosciuto', 'past_part', 'a'], ['sconosciuto', 'adjective', 'a'], ['sconsigliare', 'verb', 'c'], ['scontato', 'past_part', 'b'], ['scontato', 'adjective', 'b'], ['scontento', 'adjective', 'c'], ['sconto', 'noun', 'b'], ['scontrare', 'verb', 'b'], ['scontro', 'noun', 'b'], ['sconvolgere', 'verb', 'b'], ['scopa', 'noun', 'c'], ['scopare', 'verb', 'b'], ['scoperta', 'noun', 'a'], ['scopo', 'noun', 'a'], ['scoppiare', 'verb', 'a'], ['scoprire', 'verb', 'a'], ['scordare', 'verb', 'b'], ['scorgere', 'verb', 'b'], ['scorpione', 'noun', 'c'], ['scorrere', 'verb', 'a'], ['scorretto', 'adjective', 'c'], ['scorso', 'past_part', 'a'], ['scorso', 'adjective', 'a'], ['scorso', 'noun', 'a'], ['scorta', 'noun', 'b'], ['scortese', 'adjective', 'c'], ['scossa', 'noun', 'c'], ['scout', 'noun', 'c'], ['scout', 'adjective', 'c'], ['scozzese', 'adjective', 'c'], ['scozzese', 'noun', 'c'], ['screpolare', 'verb', 'c'], ['scricchiolare', 'verb', 'c'], ['scritta', 'noun', 'b'], ['scritto', 'past_part', 'b'], ['scritto', 'adjective', 'b'], ['scritto', 'noun', 'b'], ['scrittore', 'noun', 'a'], ['scrittura', 'noun', 'a'], ['scrivania', 'noun', 'b'], ['scrivere', 'verb', 'a'], ['scrofa', 'noun', 'c'], ['scrupolo', 'noun', 'c'], ['scudetto', 'noun', 'c'], ['scudo', 'noun', 'b'], ['scultore', 'noun', 'c'], ['scultura', 'noun', 'b'], ['scuola', 'noun', 'a'], ['scuotere', 'verb', 'b'], ['scure', 'noun', 'c'], ['scurire', 'verb', 'c'], ['scuro', 'adjective', 'b'], ['scuro', 'noun', 'b'], ['scuro', 'adverb', 'b'], ['scusa', 'noun', 'a'], ['scusare', 'verb', 'a'], ['sdebitarsi', 'verb', 'c'], ['sdegnare', 'verb', 'c'], ['sdraiare', 'verb', 'b'], ['sdraiato', 'past_part', 'c'], ['sdraiato', 'adjective', 'c'], ['se', 'pronoun', 'a'], ['se', 'conjunction', 'a'], ['se', 'noun', 'a'], ['sebbene', 'conjunction', 'b'], ['seccare', 'verb', 'b'], ['seccatura', 'noun', 'c'], ['secchio', 'noun', 'b'], ['secchione', 'noun', 'b'], ['secco', 'adjective', 'a'], ['secco', 'noun', 'a'], ['secolo', 'noun', 'a'], ['seconda', 'noun', 'b'], ['secondario', 'adjective', 'b'], ['secondario', 'noun', 'b'], ['secondo', 'adjective', 'a'], ['secondo', 'noun', 'a'], ['secondo', 'adverb', 'a'], ['secondo', 'preposition', 'a'], ['secondo', 'conjunction', 'a'], ['sedano', 'noun', 'c'], ['sede', 'noun', 'a'], ['sedere', 'verb', 'a'], ['sedia', 'noun', 'a'], ['sedici', 'adjective', 'b'], ['sedici', 'noun', 'b'], ['sedile', 'noun', 'b'], ['sedurre', 'verb', 'b'], ['seduta', 'noun', 'b'], ['seduttore', 'adjective', 'c'], ['seduttore', 'noun', 'c'], ['seggiolino', 'noun', 'c'], ['seggiovia', 'noun', 'c'], ['segheria', 'noun', 'c'], ['segmento', 'noun', 'b'], ['segnalare', 'verb', 'a'], ['segnalazione', 'noun', 'b'], ['segnale', 'noun', 'a'], ['segnare', 'verb', 'a'], ['segno', 'noun', 'a'], ['segretaria', 'noun', 'b'], ['segretario', 'noun', 'b'], ['segreteria', 'noun', 'b'], ['segreto', 'noun', 'a'], ['segreto', 'adjective', 'a'], ['segreto', 'noun', 'a'], ['segreto', 'adverb', 'a'], ['seguente', 'pres_part', 'a'], ['seguente', 'adjective', 'a'], ['seguente', 'noun', 'a'], ['seguire', 'verb', 'a'], ['seguito', 'noun', 'a'], ['sei', 'adjective', 'a'], ['sei', 'noun', 'a'], ['selezionare', 'verb', 'b'], ['selezione', 'noun', 'b'], ['selva', 'noun', 'c'], ['selvaggina', 'noun', 'c'], ['selvaggio', 'adjective', 'b'], ['selvaggio', 'noun', 'b'], ['semaforo', 'noun', 'c'], ['semantico', 'adjective', 'b'], ['sembrare', 'verb', 'a'], ['seme', 'noun', 'b'], ['semestre', 'noun', 'c'], ['semifreddo', 'adjective', 'c'], ['semifreddo', 'noun', 'c'], ['seminare', 'verb', 'b'], ['semmai', 'conjunction', 'b'], ['semmai', 'adverb', 'b'], ['semolino', 'noun', 'c'], ['semplice', 'adjective', 'a'], ['semplice', 'noun', 'a'], ['semplicemente', 'adverb', 'a'], ['semplicit', 'noun', 'b'], ['semplificare', 'verb', 'b'], ['sempre', 'adverb', 'a'], ['senape', 'noun', 'c'], ['senape', 'adjective', 'c'], ['senato', 'noun', 'b'], ['senatore', 'noun', 'b'], ['senn', 'adverb', 'b'], ['seno', 'noun', 'a'], ['sensazione', 'noun', 'a'], ['sensibile', 'adjective', 'b'], ['sensibile', 'noun', 'b'], ['sensibilit', 'noun', 'b'], ['senso', 'noun', 'a'], ['sensuale', 'adjective', 'b'], ['sentenza', 'noun', 'a'], ['sentiero', 'noun', 'b'], ['sentimentale', 'adjective', 'b'], ['sentimentale', 'noun', 'b'], ['sentimento', 'noun', 'a'], ['sentire', 'verb', 'a'], ['sentito', 'past_part', 'b'], ['sentito', 'adjective', 'b'], ['senza', 'preposition', 'a'], ['senza', 'conjunction', 'a'], ['separare', 'verb', 'a'], ['separato', 'past_part', 'b'], ['separato', 'adjective', 'b'], ['separato', 'noun', 'b'], ['separazione', 'noun', 'b'], ['sepolto', 'past_part', 'b'], ['sepolto', 'adjective', 'b'], ['sepolto', 'noun', 'b'], ['seppellire', 'verb', 'b'], ['seppia', 'noun', 'c'], ['seppia', 'adjective', 'c'], ['seppia', 'noun', 'c'], ['sequenza', 'noun', 'b'], ['sequestrare', 'verb', 'b'], ['sequestro', 'noun', 'b'], ['sera', 'noun', 'a'], ['serata', 'noun', 'a'], ['serbo', 'adjective', 'c'], ['serbo', 'noun', 'c'], ['serenata', 'noun', 'c'], ['serenit', 'noun', 'b'], ['sereno', 'adjective', 'a'], ['sereno', 'noun', 'a'], ['sergente', 'noun', 'b'], ['seriamente', 'adverb', 'b'], ['serie', 'noun', 'a'], ['seriet', 'noun', 'c'], ['serio', 'adjective', 'a'], ['serio', 'noun', 'a'], ['serpente', 'noun', 'b'], ['serra', 'noun', 'b'], ['servire', 'verb', 'a'], ['servizio', 'noun', 'a'], ['servo', 'noun', 'b'], ['servo', 'adjective', 'b'], ['sessanta', 'adjective', 'b'], ['sessanta', 'noun', 'b'], ['sesso', 'noun', 'a'], ['sessuale', 'adjective', 'a'], ['sesto', 'adjective', 'b'], ['sesto', 'noun', 'b'], ['set', 'noun', 'b'], ['seta', 'noun', 'b'], ['sete', 'noun', 'b'], ['setta', 'noun', 'b'], ['settanta', 'adjective', 'b'], ['settanta', 'noun', 'b'], ['sette', 'adjective', 'a'], ['sette', 'noun', 'a'], ['settembre', 'noun', 'a'], ['settentrione', 'noun', 'c'], ['settimana', 'noun', 'a'], ['settimanale', 'adjective', 'b'], ['settimanale', 'noun', 'b'], ['settimo', 'adjective', 'b'], ['settimo', 'noun', 'b'], ['settore', 'noun', 'a'], ['severo', 'adjective', 'b'], ['sexy', 'adjective', 'b'], ['sezione', 'noun', 'a'], ['sfera', 'noun', 'b'], ['sfida', 'noun', 'a'], ['sfidare', 'verb', 'b'], ['sfiducia', 'noun', 'c'], ['sfigato', 'adjective', 'b'], ['sfigato', 'noun', 'b'], ['sfilare', 'verb', 'b'], ['sfilata', 'noun', 'b'], ['sfinire', 'verb', 'c'], ['sfiorare', 'verb', 'b'], ['sfociare', 'verb', 'c'], ['sfogare', 'verb', 'b'], ['sfoglia', 'noun', 'c'], ['sfogliare', 'verb', 'b'], ['sfogo', 'noun', 'b'], ['sfollamento', 'noun', 'c'], ['sfollare', 'verb', 'c'], ['sfondare', 'verb', 'b'], ['sfondo', 'noun', 'b'], ['sfortunato', 'adjective', 'c'], ['sforzare', 'verb', 'b'], ['sforzo', 'noun', 'a'], ['sfrenato', 'past_part', 'c'], ['sfrenato', 'adjective', 'c'], ['sfruttare', 'verb', 'a'], ['sfuggire', 'verb', 'a'], ['sgabello', 'noun', 'c'], ['sganciare', 'verb', 'c'], ['sgarbato', 'adjective', 'c'], ['sgarbato', 'noun', 'c'], ['sgarbo', 'noun', 'c'], ['sgombro', 'noun', 'c'], ['sgomento', 'noun', 'c'], ['sgonfiare', 'verb', 'c'], ['sgozzare', 'verb', 'c'], ['sgrassare', 'verb', 'c'], ['sgrassatore', 'noun', 'c'], ['sgridare', 'verb', 'c'], ['sguardo', 'noun', 'a'], ['shampoo', 'noun', 'c'], ['share', 'noun', 'b'], ['shopping', 'noun', 'b'], ['shorts', 'noun', 'c'], ['show', 'noun', 'b'], ['s', 'adverb', 'a'], ['s', 'noun', 'a'], ['s', 'adjective', 'a'], ['si', 'pronoun', 'a'], ['sia', 'conjunction', 'a'], ['siamese', 'adjective', 'c'], ['siamese', 'noun', 'c'], ['sicch', 'conjunction', 'b'], ['siccit', 'noun', 'c'], ['siccome', 'conjunction', 'a'], ['siccome', 'adverb', 'a'], ['siciliano', 'adjective', 'b'], ['siciliano', 'noun', 'b'], ['sicuramente', 'adverb', 'a'], ['sicurezza', 'noun', 'a'], ['sicuro', 'adjective', 'a'], ['sicuro', 'noun', 'a'], ['sicuro', 'adverb', 'a'], ['siepe', 'noun', 'c'], ['sigaretta', 'noun', 'a'], ['sigaro', 'noun', 'c'], ['sigla', 'noun', 'b'], ['significare', 'verb', 'a'], ['significativo', 'adjective', 'b'], ['significato', 'past_part', 'a'], ['significato', 'noun', 'a'], ['signora', 'noun', 'a'], ['signore', 'noun', 'a'], ['signorina', 'noun', 'a'], ['silenzio', 'noun', 'a'], ['silenzioso', 'adjective', 'b'], ['sillaba', 'noun', 'c'], ['simbolico', 'adjective', 'b'], ['simbolo', 'noun', 'a'], ['simile', 'adjective', 'a'], ['simile', 'adjective', 'a'], ['simile', 'noun', 'a'], ['simile', 'adverb', 'a'], ['simpatia', 'noun', 'b'], ['simpatico', 'adjective', 'a'], ['simulare', 'verb', 'b'], ['sinceramente', 'adverb', 'b'], ['sincero', 'adjective', 'b'], ['sindacale', 'adjective', 'b'], ['sindacato', 'noun', 'b'], ['sindaco', 'noun', 'b'], ['sindrome', 'noun', 'b'], ['single', 'noun', 'b'], ['singolare', 'adjective', 'b'], ['singolare', 'noun', 'b'], ['singolo', 'adjective', 'a'], ['singolo', 'noun', 'a'], ['sinistra', 'noun', 'a'], ['sinistro', 'adjective', 'a'], ['sinistro', 'noun', 'a'], ['sino', 'preposition', 'a'], ['sino', 'adverb', 'a'], ['sinonimo', 'noun', 'b'], ['sintesi', 'noun', 'b'], ['sintetico', 'adjective', 'b'], ['sintetizzare', 'verb', 'b'], ['sintomo', 'noun', 'b'], ['sir', 'noun', 'b'], ['siriano', 'adjective', 'c'], ['siriano', 'noun', 'c'], ['siringa', 'noun', 'c'], ['sistema', 'noun', 'a'], ['sistemare', 'verb', 'a'], ['sito', 'noun', 'a'], ['sito', 'adjective', 'a'], ['situare', 'verb', 'b'], ['situazione', 'noun', 'a'], ['slacciare', 'verb', 'c'], ['slanciato', 'past_part', 'c'], ['slanciato', 'adjective', 'c'], ['slavo', 'adjective', 'c'], ['slavo', 'noun', 'c'], ['slegare', 'verb', 'c'], ['slip', 'noun', 'c'], ['slitta', 'noun', 'c'], ['slogan', 'noun', 'b'], ['slogare', 'verb', 'c'], ['slogatura', 'noun', 'c'], ['slovacco', 'adjective', 'c'], ['slovacco', 'noun', 'c'], ['sloveno', 'adjective', 'c'], ['sloveno', 'noun', 'c'], ['smacchiare', 'verb', 'c'], ['smacchiatore', 'adjective', 'c'], ['smacchiatore', 'noun', 'c'], ['smaltimento', 'noun', 'b'], ['smalto', 'noun', 'c'], ['smascherare', 'verb', 'c'], ['smentire', 'verb', 'b'], ['smettere', 'verb', 'a'], ['smisurato', 'past_part', 'c'], ['smisurato', 'adjective', 'c'], ['smog', 'noun', 'c'], ['smontare', 'verb', 'b'], ['smorfia', 'noun', 'c'], ['smuovere', 'verb', 'c'], ['snack', 'noun', 'c'], ['sneaker', 'noun', 'c'], ['snello', 'adjective', 'c'], ['soccorrere', 'verb', 'c'], ['soccorso', 'noun', 'b'], ['socialdemocratico', 'adjective', 'c'], ['socialdemocratico', 'noun', 'c'], ['sociale', 'adjective', 'a'], ['sociale', 'noun', 'a'], ['socialista', 'adjective', 'b'], ['socialista', 'noun', 'b'], ['societ', 'noun', 'a'], ['socievole', 'adjective', 'c'], ['socio', 'noun', 'b'], ['soddisfare', 'verb', 'a'], ['soddisfatto', 'past_part', 'b'], ['soddisfatto', 'adjective', 'b'], ['soddisfazione', 'noun', 'a'], ['sodo', 'adjective', 'b'], ['sodo', 'noun', 'b'], ['sodo', 'adverb', 'b'], ['sof', 'noun', 'c'], ['sofferenza', 'noun', 'a'], ['soffermare', 'verb', 'b'], ['soffiare', 'verb', 'b'], ['soffice', 'adjective', 'c'], ['soffitta', 'noun', 'c'], ['soffitto', 'noun', 'b'], ['soffocare', 'verb', 'b'], ['soffriggere', 'verb', 'c'], ['soffrire', 'verb', 'a'], ['sofisticato', 'past_part', 'b'], ['sofisticato', 'adjective', 'b'], ['software', 'noun', 'b'], ['soggettivo', 'adjective', 'b'], ['soggetto', 'noun', 'a'], ['soggetto', 'adjective', 'b'], ['soggezione', 'noun', 'c'], ['soggiorno', 'noun', 'a'], ['soglia', 'noun', 'b'], ['sogliola', 'noun', 'c'], ['sognare', 'verb', 'a'], ['sogno', 'noun', 'a'], ['sol', 'noun', 'c'], ['solaio', 'noun', 'c'], ['solamente', 'adverb', 'a'], ['solamente', 'conjunction', 'a'], ['solare', 'adjective', 'b'], ['solare', 'noun', 'b'], ['solco', 'noun', 'b'], ['soldato', 'noun', 'a'], ['soldo', 'noun', 'a'], ['sole', 'noun', 'a'], ['solenne', 'adjective', 'b'], ['solidariet', 'noun', 'b'], ['solido', 'adjective', 'b'], ['solido', 'noun', 'b'], ['solitamente', 'adverb', 'b'], ['solitario', 'adjective', 'b'], ['solitario', 'noun', 'b'], ['solito', 'adjective', 'a'], ['solito', 'noun', 'a'], ['solitudine', 'noun', 'b'], ['solletico', 'noun', 'c'], ['sollevare', 'verb', 'a'], ['sollievo', 'noun', 'b'], ['solo', 'adjective', 'a'], ['solo', 'noun', 'a'], ['solo', 'adverb', 'a'], ['solo', 'conjunction', 'a'], ['soltanto', 'adverb', 'a'], ['soltanto', 'conjunction', 'a'], ['soluzione', 'noun', 'a'], ['somigliare', 'verb', 'b'], ['somma', 'noun', 'a'], ['sommare', 'verb', 'b'], ['sondaggio', 'noun', 'a'], ['sonno', 'noun', 'a'], ['sonoro', 'adjective', 'b'], ['sonoro', 'noun', 'b'], ['soppalco', 'noun', 'c'], ['sopportare', 'verb', 'a'], ['sopra', 'preposition', 'a'], ['sopra', 'adverb', 'a'], ['sopra', 'adjective', 'a'], ['sopra', 'noun', 'a'], ['soprabito', 'noun', 'c'], ['sopracciglio', 'noun', 'c'], ['soprammobile', 'noun', 'c'], ['soprannome', 'noun', 'c'], ['soprattutto', 'adverb', 'a'], ['sopravvalutare', 'verb', 'c'], ['sopravvivenza', 'noun', 'b'], ['sopravvivere', 'verb', 'a'], ['sorcio', 'noun', 'c'], ['sordo', 'adjective', 'b'], ['sordo', 'noun', 'b'], ['sorella', 'noun', 'a'], ['sorgente', 'pres_part', 'b'], ['sorgente', 'adjective', 'b'], ['sorgente', 'noun', 'b'], ['sorgere', 'verb', 'b'], ['sorpassare', 'verb', 'c'], ['sorpasso', 'noun', 'c'], ['sorprendente', 'pres_part', 'b'], ['sorprendente', 'adjective', 'b'], ['sorprendere', 'verb', 'b'], ['sorpresa', 'noun', 'a'], ['sorridente', 'pres_part', 'c'], ['sorridente', 'adjective', 'c'], ['sorridere', 'verb', 'a'], ['sorriso', 'noun', 'a'], ['sorso', 'noun', 'c'], ['sorta', 'noun', 'a'], ['sorte', 'noun', 'b'], ['sorteggiare', 'verb', 'c'], ['sorteggio', 'noun', 'c'], ['sorvegliare', 'verb', 'b'], ['sospendere', 'verb', 'b'], ['sospensione', 'noun', 'b'], ['sospeso', 'past_part', 'b'], ['sospeso', 'adjective', 'b'], ['sospeso', 'noun', 'b'], ['sospettare', 'verb', 'b'], ['sospetto', 'noun', 'a'], ['sospetto', 'adjective', 'a'], ['sospetto', 'noun', 'a'], ['sospirare', 'verb', 'b'], ['sospiro', 'noun', 'b'], ['sosta', 'noun', 'b'], ['sostanza', 'noun', 'a'], ['sostanzialmente', 'adverb', 'b'], ['sostare', 'verb', 'c'], ['sostegno', 'noun', 'b'], ['sostenere', 'verb', 'a'], ['sostenitore', 'adjective', 'b'], ['sostenitore', 'noun', 'b'], ['sostituire', 'verb', 'a'], ['sostituzione', 'noun', 'b'], ['sottaceto', 'adjective', 'c'], ['sottaceto', 'adverb', 'c'], ['sottaceto', 'noun', 'c'], ['sotterraneo', 'adjective', 'b'], ['sotterraneo', 'noun', 'b'], ['sottile', 'adjective', 'a'], ['sottile', 'noun', 'a'], ['sottile', 'adverb', 'a'], ['sottinteso', 'past_part', 'c'], ['sottinteso', 'adjective', 'c'], ['sottinteso', 'noun', 'c'], ['sotto', 'preposition', 'a'], ['sotto', 'adverb', 'a'], ['sotto', 'adjective', 'a'], ['sotto', 'noun', 'a'], ['sottofondo', 'noun', 'b'], ['sottolineare', 'verb', 'a'], ['sottolio', 'adverb', 'c'], ['sottolio', 'adjective', 'c'], ['sottomarino', 'adjective', 'c'], ['sottomarino', 'noun', 'c'], ['sottopassaggio', 'noun', 'c'], ['sottoporre', 'verb', 'a'], ['sottoscrivere', 'verb', 'b'], ['sottovalutare', 'verb', 'b'], ['sottrarre', 'verb', 'b'], ['sovietico', 'adjective', 'b'], ['sovietico', 'noun', 'b'], ['sovrano', 'adjective', 'b'], ['sovrano', 'noun', 'b'], ['sovrapporre', 'verb', 'b'], ['spaccare', 'verb', 'b'], ['spaccatura', 'noun', 'c'], ['spacciare', 'verb', 'b'], ['spacciatore', 'noun', 'c'], ['spaccio', 'noun', 'c'], ['spada', 'noun', 'b'], ['spaghetto', 'noun', 'b'], ['spagnolo', 'adjective', 'a'], ['spagnolo', 'noun', 'a'], ['spago', 'noun', 'c'], ['spalancare', 'verb', 'b'], ['spalla', 'noun', 'a'], ['spalmabile', 'adjective', 'c'], ['spalmare', 'verb', 'c'], ['spam', 'noun', 'b'], ['sparare', 'verb', 'a'], ['sparecchiare', 'verb', 'c'], ['spargere', 'verb', 'b'], ['sparire', 'verb', 'a'], ['sparo', 'noun', 'b'], ['sparso', 'past_part', 'b'], ['sparso', 'adjective', 'b'], ['spassare', 'verb', 'b'], ['spasso', 'noun', 'c'], ['spavaldo', 'adjective', 'c'], ['spaventare', 'verb', 'a'], ['spaventato', 'past_part', 'b'], ['spaventato', 'adjective', 'b'], ['spaventoso', 'adjective', 'b'], ['spaziale', 'adjective', 'b'], ['spazio', 'noun', 'a'], ['spazioso', 'adjective', 'c'], ['spazzare', 'verb', 'b'], ['spazzatura', 'noun', 'b'], ['spazzino', 'noun', 'c'], ['spazzola', 'noun', 'c'], ['spazzolare', 'verb', 'c'], ['spazzolino', 'noun', 'c'], ['spazzolone', 'noun', 'c'], ['specchiarsi', 'verb', 'c'], ['specchio', 'noun', 'a'], ['speciale', 'adjective', 'a'], ['speciale', 'noun', 'a'], ['specialista', 'noun', 'b'], ['specializzato', 'past_part', 'b'], ['specializzato', 'adjective', 'b'], ['specializzato', 'noun', 'b'], ['specialmente', 'adverb', 'b'], ['specie', 'noun', 'a'], ['specie', 'adverb', 'a'], ['specificare', 'verb', 'b'], ['specifico', 'adjective', 'a'], ['specifico', 'noun', 'a'], ['speck', 'noun', 'c'], ['spedire', 'verb', 'b'], ['spedizione', 'noun', 'b'], ['spegnere', 'verb', 'a'], ['spellare', 'verb', 'c'], ['spendere', 'verb', 'a'], ['spennare', 'verb', 'c'], ['spensierato', 'adjective', 'c'], ['spento', 'past_part', 'b'], ['spento', 'adjective', 'b'], ['speranza', 'noun', 'a'], ['sperare', 'verb', 'a'], ['sperimentale', 'adjective', 'b'], ['sperimentare', 'verb', 'b'], ['sperimentazione', 'noun', 'b'], ['sperone', 'noun', 'c'], ['spesa', 'noun', 'a'], ['spesso', 'adjective', 'b'], ['spesso', 'adverb', 'a'], ['spessore', 'noun', 'b'], ['spettacolare', 'adjective', 'b'], ['spettacolo', 'noun', 'a'], ['spettare', 'verb', 'b'], ['spettatore', 'noun', 'b'], ['spettinare', 'verb', 'c'], ['spettro', 'noun', 'b'], ['spezia', 'noun', 'c'], ['spezzare', 'verb', 'b'], ['spia', 'noun', 'b'], ['spiacere', 'verb', 'b'], ['spiaggia', 'noun', 'a'], ['spianare', 'verb', 'c'], ['spiare', 'verb', 'b'], ['spiazzo', 'noun', 'c'], ['spiccare', 'verb', 'b'], ['spicciolo', 'adjective', 'c'], ['spicciolo', 'noun', 'c'], ['spiedino', 'noun', 'c'], ['spiedo', 'noun', 'c'], ['spiegare', 'verb', 'a'], ['spiegazione', 'noun', 'a'], ['spietato', 'adjective', 'b'], ['spiga', 'noun', 'c'], ['spigolo', 'noun', 'c'], ['spillo', 'noun', 'c'], ['spina', 'noun', 'b'], ['spinacio', 'noun', 'c'], ['spingere', 'verb', 'a'], ['spinta', 'noun', 'b'], ['spionaggio', 'noun', 'c'], ['spirito', 'noun', 'a'], ['spiritoso', 'adjective', 'c'], ['spirituale', 'adjective', 'b'], ['spirituale', 'noun', 'b'], ['splendente', 'pres_part', 'c'], ['splendente', 'adjective', 'c'], ['splendere', 'verb', 'b'], ['splendido', 'adjective', 'b'], ['splendore', 'noun', 'b'], ['spogliare', 'verb', 'b'], ['spogliatoio', 'noun', 'c'], ['spoglio', 'noun', 'c'], ['spolverare', 'verb', 'c'], ['sponda', 'noun', 'b'], ['spontaneo', 'adjective', 'b'], ['sporcare', 'verb', 'b'], ['sporcizia', 'noun', 'c'], ['sporco', 'adjective', 'a'], ['sporco', 'noun', 'a'], ['sporgente', 'pres_part', 'c'], ['sporgente', 'adjective', 'c'], ['sporgente', 'noun', 'c'], ['sporgere', 'verb', 'b'], ['sport', 'noun', 'a'], ['sport', 'adjective', 'a'], ['sportello', 'noun', 'b'], ['sportivo', 'adjective', 'a'], ['sportivo', 'noun', 'a'], ['sposare', 'verb', 'a'], ['sposato', 'past_part', 'b'], ['sposato', 'adjective', 'b'], ['sposato', 'noun', 'b'], ['sposo', 'noun', 'b'], ['spostamento', 'noun', 'b'], ['spostare', 'verb', 'a'], ['spot', 'noun', 'b'], ['spranga', 'noun', 'c'], ['spray', 'adjective', 'c'], ['spray', 'noun', 'c'], ['sprecare', 'verb', 'b'], ['spreco', 'noun', 'c'], ['spremere', 'verb', 'c'], ['spremuta', 'noun', 'c'], ['sprofondare', 'verb', 'b'], ['sproposito', 'noun', 'c'], ['spruzzare', 'verb', 'c'], ['spuma', 'noun', 'c'], ['spumante', 'pres_part', 'c'], ['spumante', 'adjective', 'c'], ['spumante', 'noun', 'c'], ['spuntare', 'verb', 'b'], ['spuntino', 'noun', 'c'], ['spunto', 'noun', 'b'], ['sputare', 'verb', 'b'], ['sputo', 'noun', 'c'], ['squadra', 'noun', 'a'], ['squallido', 'adjective', 'c'], ['squalo', 'noun', 'c'], ['squarcio', 'noun', 'c'], ['squillare', 'verb', 'b'], ['squisito', 'adjective', 'c'], ['stabile', 'adjective', 'b'], ['stabile', 'noun', 'b'], ['stabilire', 'verb', 'a'], ['stabilit', 'noun', 'b'], ['staccare', 'verb', 'a'], ['stacco', 'noun', 'c'], ['stadio', 'noun', 'b'], ['staffa', 'noun', 'c'], ['stagione', 'noun', 'a'], ['stagno', 'noun', 'c'], ['stalla', 'noun', 'b'], ['stallone', 'noun', 'c'], ['stamattina', 'adverb', 'b'], ['stampa', 'noun', 'a'], ['stampare', 'verb', 'b'], ['stampatello', 'noun', 'c'], ['stampato', 'past_part', 'b'], ['stampato', 'adjective', 'b'], ['stampato', 'noun', 'b'], ['stampella', 'noun', 'c'], ['stampo', 'noun', 'c'], ['stancare', 'verb', 'b'], ['stanchezza', 'noun', 'b'], ['stanco', 'adjective', 'a'], ['standard', 'noun', 'b'], ['standard', 'adjective', 'b'], ['stanga', 'noun', 'c'], ['stanotte', 'adverb', 'b'], ['stanza', 'noun', 'a'], ['star', 'noun', 'b'], ['stare', 'verb', 'a'], ['stasera', 'adverb', 'a'], ['statale', 'adjective', 'b'], ['statale', 'noun', 'b'], ['statistica', 'noun', 'b'], ['statistico', 'adjective', 'b'], ['statistico', 'noun', 'b'], ['stato', 'noun', 'a'], ['stato', 'noun', 'a'], ['statua', 'noun', 'b'], ['statunitense', 'adjective', 'b'], ['statunitense', 'noun', 'b'], ['status', 'noun', 'b'], ['stavolta', 'adverb', 'b'], ['stazione', 'noun', 'a'], ['stella', 'noun', 'a'], ['stellare', 'adjective', 'b'], ['stendere', 'verb', 'b'], ['stendibiancheria', 'noun', 'c'], ['stereo', 'adjective', 'c'], ['stereo', 'noun', 'c'], ['sterlina', 'noun', 'b'], ['sterzare', 'verb', 'c'], ['sterzo', 'noun', 'c'], ['stesso', 'adjective', 'a'], ['stesso', 'pronoun', 'a'], ['stile', 'noun', 'a'], ['stima', 'noun', 'b'], ['stimare', 'verb', 'b'], ['stimolare', 'verb', 'b'], ['stimolo', 'noun', 'b'], ['stinco', 'noun', 'c'], ['stipendiare', 'verb', 'c'], ['stipendio', 'noun', 'a'], ['stirare', 'verb', 'b'], ['stivaletto', 'noun', 'c'], ['stoffa', 'noun', 'b'], ['stomaco', 'noun', 'b'], ['stonare', 'verb', 'c'], ['stop', 'loc-comando', 'c'], ['stop', 'noun', 'c'], ['stoppa', 'noun', 'c'], ['storcere', 'verb', 'c'], ['storia', 'noun', 'a'], ['storico', 'adjective', 'a'], ['storico', 'noun', 'a'], ['stornello', 'noun', 'c'], ['storta', 'noun', 'c'], ['storto', 'past_part', 'b'], ['storto', 'adjective', 'b'], ['storto', 'adverb', 'b'], ['storto', 'noun', 'b'], ['stoviglia', 'noun', 'c'], ['stracchino', 'noun', 'c'], ['straccio', 'noun', 'b'], ['strada', 'noun', 'a'], ['stradale', 'adjective', 'b'], ['stradale', 'noun', 'b'], ['strage', 'noun', 'b'], ['strangolare', 'verb', 'c'], ['straniero', 'adjective', 'a'], ['straniero', 'noun', 'a'], ['strano', 'adjective', 'a'], ['straordinario', 'adjective', 'a'], ['straordinario', 'noun', 'a'], ['strappare', 'verb', 'b'], ['strategia', 'noun', 'a'], ['strategico', 'adjective', 'b'], ['strato', 'noun', 'b'], ['strega', 'noun', 'a'], ['stregare', 'verb', 'b'], ['stregone', 'noun', 'c'], ['stress', 'noun', 'b'], ['stretta', 'noun', 'b'], ['strettamente', 'adverb', 'b'], ['stretto', 'past_part', 'a'], ['stretto', 'adjective', 'a'], ['stretto', 'noun', 'a'], ['strillare', 'verb', 'b'], ['strillo', 'noun', 'c'], ['stringa', 'noun', 'c'], ['stringere', 'verb', 'a'], ['striscia', 'noun', 'b'], ['strisciare', 'verb', 'b'], ['strofinaccio', 'noun', 'c'], ['stronzata', 'noun', 'b'], ['stronzo', 'noun', 'a'], ['stronzo', 'adjective', 'a'], ['strumento', 'noun', 'a'], ['strutto', 'past_part', 'c'], ['strutto', 'adjective', 'c'], ['strutto', 'noun', 'c'], ['struttura', 'noun', 'a'], ['strutturale', 'adjective', 'b'], ['struzzo', 'noun', 'c'], ['studente', 'noun', 'a'], ['studiare', 'verb', 'a'], ['studio', 'noun', 'a'], ['studioso', 'adjective', 'b'], ['studioso', 'noun', 'b'], ['stufa', 'noun', 'c'], ['stuoia', 'noun', 'c'], ['stupefacente', 'pres_part', 'b'], ['stupefacente', 'adjective', 'b'], ['stupefacente', 'noun', 'b'], ['stupendo', 'adjective', 'b'], ['stupido', 'adjective', 'a'], ['stupido', 'noun', 'a'], ['stupire', 'verb', 'b'], ['stupito', 'past_part', 'b'], ['stupito', 'adjective', 'b'], ['stupore', 'noun', 'b'], ['stuzzicadenti', 'noun', 'c'], ['stuzzicare', 'verb', 'c'], ['style', 'noun', 'b'], ['su', 'preposition', 'a'], ['su', 'adverb', 'a'], ['su', 'exclamation', 'a'], ['su', 'noun', 'a'], ['subire', 'verb', 'a'], ['subito', 'adverb', 'a'], ['succedere', 'verb', 'a'], ['successione', 'noun', 'b'], ['successivamente', 'adverb', 'b'], ['successivo', 'adjective', 'a'], ['successo', 'noun', 'a'], ['succhiare', 'verb', 'b'], ['succo', 'noun', 'b'], ['sud', 'noun', 'a'], ['sud', 'adjective', 'a'], ['sudamericano', 'adjective', 'c'], ['sudamericano', 'noun', 'c'], ['sudare', 'verb', 'b'], ['sudato', 'past_part', 'c'], ['sudato', 'adjective', 'c'], ['suddito', 'noun', 'b'], ['suddito', 'adjective', 'b'], ['suddividere', 'verb', 'b'], ['sudicio', 'adjective', 'c'], ['sudicio', 'noun', 'c'], ['sudore', 'noun', 'b'], ['sudtirolese', 'adjective', 'c'], ['sudtirolese', 'noun', 'c'], ['sufficiente', 'adjective', 'a'], ['suggerimento', 'noun', 'b'], ['suggerire', 'verb', 'a'], ['suggestivo', 'adjective', 'b'], ['sughero', 'noun', 'c'], ['sugo', 'noun', 'b'], ['suicidio', 'noun', 'b'], ['suino', 'noun', 'c'], ['suino', 'adjective', 'c'], ['suo', 'adjective', 'a'], ['suo', 'pronoun', 'a'], ['suocera', 'noun', 'c'], ['suocero', 'noun', 'c'], ['suola', 'noun', 'c'], ['suolo', 'noun', 'b'], ['suonare', 'verb', 'a'], ['suono', 'noun', 'a'], ['suora', 'noun', 'a'], ['super', 'adjective', 'b'], ['super', 'noun', 'b'], ['superare', 'verb', 'a'], ['superbia', 'noun', 'c'], ['superficiale', 'adjective', 'b'], ['superficie', 'noun', 'a'], ['superiore', 'adjective', 'a'], ['superiore', 'noun', 'a'], ['supermercato', 'noun', 'b'], ['supporre', 'verb', 'b'], ['supportare', 'verb', 'b'], ['supporto', 'noun', 'a'], ['supremo', 'adjective', 'b'], ['surgelato', 'past_part', 'c'], ['surgelato', 'adjective', 'c'], ['surgelato', 'noun', 'c'], ['suscitare', 'verb', 'b'], ['susina', 'noun', 'c'], ['susino', 'noun', 'c'], ['susseguirsi', 'verb', 'c'], ['sussurrare', 'verb', 'b'], ['svanire', 'verb', 'b'], ['svedese', 'adjective', 'c'], ['svedese', 'noun', 'c'], ['sveglia', 'noun', 'c'], ['svegliare', 'verb', 'a'], ['svegliarsi', 'verb', 'c'], ['sveglio', 'past_part', 'b'], ['sveglio', 'adjective', 'b'], ['svelare', 'verb', 'b'], ['svelto', 'adjective', 'c'], ['svenire', 'verb', 'b'], ['sventola', 'noun', 'c'], ['sviluppare', 'verb', 'a'], ['sviluppato', 'past_part', 'b'], ['sviluppato', 'adjective', 'b'], ['sviluppo', 'noun', 'a'], ['svizzero', 'adjective', 'b'], ['svizzero', 'noun', 'b'], ['svolazzare', 'verb', 'c'], ['svolgere', 'verb', 'a'], ['svolgimento', 'noun', 'c'], ['svolta', 'noun', 'b'], ['svuotare', 'verb', 'b'], ['tabaccaio', 'noun', 'c'], ['tabella', 'noun', 'b'], ['tacca', 'noun', 'c'], ['tacchino', 'noun', 'c'], ['tacco', 'noun', 'b'], ['tacere', 'verb', 'a'], ['tacere', 'noun', 'a'], ['tag', 'noun', 'b'], ['taglia', 'noun', 'b'], ['tagliare', 'verb', 'a'], ['tagliatella', 'noun', 'c'], ['tagliato', 'past_part', 'b'], ['tagliato', 'adjective', 'b'], ['tagliere', 'noun', 'c'], ['taglio', 'noun', 'a'], ['tagliola', 'noun', 'c'], ['talco', 'noun', 'c'], ['tale', 'adjective', 'a'], ['tale', 'pronoun', 'a'], ['tale', 'adverb', 'a'], ['taleggio', 'noun', 'c'], ['talento', 'noun', 'b'], ['talmente', 'adverb', 'a'], ['talpa', 'noun', 'c'], ['talpa', 'adjective', 'c'], ['talpa', 'noun', 'c'], ['talvolta', 'adverb', 'b'], ['tamburo', 'noun', 'c'], ['tamponare', 'verb', 'c'], ['tangente', 'pres_part', 'b'], ['tangente', 'adjective', 'b'], ['tangente', 'noun', 'b'], ['tanto', 'adjective', 'a'], ['tanto', 'pronoun', 'a'], ['tanto', 'noun', 'a'], ['tanto', 'adverb', 'a'], ['tanto', 'conjunction', 'a'], ['tappa', 'noun', 'b'], ['tappare', 'verb', 'b'], ['tappetino', 'noun', 'c'], ['tappeto', 'noun', 'b'], ['tappezzare', 'verb', 'c'], ['tappo', 'noun', 'c'], ['tarallo', 'noun', 'c'], ['tarantella', 'noun', 'c'], ['tardi', 'adverb', 'a'], ['tardo', 'adjective', 'a'], ['tardo', 'adverb', 'a'], ['targa', 'noun', 'b'], ['tariffa', 'noun', 'b'], ['tarlo', 'noun', 'c'], ['tartaruga', 'noun', 'c'], ['tartufo', 'noun', 'c'], ['tasca', 'noun', 'a'], ['tassa', 'noun', 'a'], ['tassare', 'verb', 'c'], ['tassello', 'noun', 'c'], ['tasso', 'noun', 'b'], ['tastiera', 'noun', 'b'], ['tasto', 'noun', 'b'], ['tatto', 'noun', 'c'], ['tatuaggio', 'noun', 'b'], ['taverna', 'noun', 'c'], ['tavola', 'noun', 'a'], ['tavoletta', 'noun', 'c'], ['tavolino', 'noun', 'b'], ['tavolo', 'noun', 'a'], ['taxi', 'noun', 'b'], ['tazza', 'noun', 'b'], ['t', 'noun', 'b'], ['te', 'pronoun', 'noun'], ['te', 'team', 'noun'], ['teatrale', 'adjective', 'b'], ['teatro', 'noun', 'a'], ['tecnica', 'noun', 'a'], ['tecnicamente', 'adverb', 'b'], ['tecnico', 'adjective', 'a'], ['tecnico', 'noun', 'a'], ['tecnologia', 'noun', 'a'], ['tecnologico', 'adjective', 'b'], ['tedesco', 'adjective', 'a'], ['tedesco', 'noun', 'a'], ['tegame', 'noun', 'c'], ['teglia', 'noun', 'c'], ['tegola', 'noun', 'c'], ['tela', 'noun', 'b'], ['telaio', 'noun', 'c'], ['telecamera', 'noun', 'b'], ['telecomandato', 'past_part', 'c'], ['telecomandato', 'adjective', 'c'], ['telecronaca', 'noun', 'c'], ['telecronista', 'noun', 'c'], ['telefilm', 'noun', 'b'], ['telefonare', 'verb', 'a'], ['telefonata', 'noun', 'a'], ['telefonico', 'adjective', 'a'], ['telefonino', 'noun', 'b'], ['telefono', 'noun', 'a'], ['telegiornale', 'noun', 'b'], ['telegrafico', 'adjective', 'c'], ['telegrafo', 'noun', 'c'], ['telegramma', 'noun', 'c'], ['telescopio', 'noun', 'b'], ['televisione', 'noun', 'a'], ['televisivo', 'adjective', 'a'], ['televisore', 'noun', 'b'], ['tema', 'noun', 'a'], ['temere', 'verb', 'a'], ['temperatura', 'noun', 'a'], ['tempesta', 'noun', 'b'], ['tempio', 'noun', 'b'], ['tempo', 'noun', 'a'], ['temporale', 'noun', 'b'], ['temporaneo', 'adjective', 'b'], ['tenaglia', 'noun', 'c'], ['tenda', 'noun', 'a'], ['tendenza', 'noun', 'a'], ['tendere', 'verb', 'a'], ['tenebra', 'noun', 'c'], ['tenente', 'noun', 'b'], ['tenere', 'verb', 'a'], ['tenerezza', 'noun', 'b'], ['tenero', 'adjective', 'b'], ['tenero', 'noun', 'b'], ['tennis', 'noun', 'b'], ['tensione', 'noun', 'a'], ['tentare', 'verb', 'a'], ['tentativo', 'noun', 'a'], ['tentazione', 'noun', 'b'], ['tenuta', 'noun', 'b'], ['teologia', 'noun', 'b'], ['teologo', 'noun', 'b'], ['teoria', 'noun', 'a'], ['teorico', 'adjective', 'b'], ['teorico', 'noun', 'b'], ['terapia', 'noun', 'a'], ['tergicristallo', 'noun', 'c'], ['terminale', 'adjective', 'b'], ['terminale', 'noun', 'b'], ['terminare', 'verb', 'a'], ['termine', 'noun', 'a'], ['termosifone', 'noun', 'c'], ['terra', 'noun', 'a'], ['terrazzo', 'noun', 'b'], ['terremoto', 'noun', 'b'], ['terreno', 'noun', 'a'], ['terrestre', 'adjective', 'b'], ['terrestre', 'noun', 'b'], ['terribile', 'adjective', 'a'], ['terriccio', 'noun', 'c'], ['territoriale', 'adjective', 'b'], ['territoriale', 'noun', 'b'], ['territorio', 'noun', 'a'], ['terrore', 'noun', 'b'], ['terrorismo', 'noun', 'b'], ['terrorista', 'adjective', 'b'], ['terrorista', 'noun', 'b'], ['terrorizzare', 'verb', 'b'], ['terzo', 'adjective', 'a'], ['terzo', 'noun', 'a'], ['teschio', 'noun', 'b'], ['tesi', 'noun', 'a'], ['teso', 'past_part', 'b'], ['teso', 'adjective', 'b'], ['tesoro', 'noun', 'a'], ['tessera', 'noun', 'b'], ['tessile', 'adjective', 'c'], ['tessile', 'noun', 'c'], ['tessuto', 'past_part', 'b'], ['tessuto', 'adjective', 'b'], ['tessuto', 'noun', 'b'], ['test', 'noun', 'a'], ['testa', 'noun', 'a'], ['testamento', 'noun', 'b'], ['testare', 'verb', 'b'], ['testimone', 'noun', 'a'], ['testimonianza', 'noun', 'b'], ['testimoniare', 'verb', 'b'], ['testo', 'noun', 'a'], ['tetta', 'noun', 'b'], ['tetto', 'noun', 'a'], ['tettoia', 'noun', 'c'], ['tg', 'sigla', 'b'], ['thermos', 'noun', 'c'], ['ti', 'noun', 'c'], ['ti', 'pronoun', 'a'], ['tic', 'noun', 'c'], ['ticchettio', 'noun', 'c'], ['tifare', 'verb', 'b'], ['tifo', 'noun', 'c'], ['tifoso', 'adjective', 'b'], ['tifoso', 'noun', 'b'], ['tigre', 'noun', 'b'], ['timbro', 'noun', 'c'], ['timidezza', 'noun', 'c'], ['timido', 'adjective', 'b'], ['timido', 'noun', 'b'], ['timone', 'noun', 'c'], ['timoniere', 'noun', 'c'], ['timore', 'noun', 'b'], ['tinello', 'noun', 'c'], ['tino', 'noun', 'c'], ['tipico', 'adjective', 'a'], ['tipo', 'noun', 'a'], ['tipologia', 'noun', 'b'], ['tiramis', 'noun', 'c'], ['tiranno', 'noun', 'c'], ['tiranno', 'adjective', 'c'], ['tirare', 'verb', 'a'], ['tiro', 'noun', 'b'], ['tirocinio', 'noun', 'b'], ['tirrenico', 'adjective', 'c'], ['tisana', 'noun', 'c'], ['titolare', 'adjective', 'b'], ['titolare', 'noun', 'b'], ['titolo', 'noun', 'a'], ['tiv', 'noun', 'a'], ['tizio', 'noun', 'b'], ['toast', 'noun', 'c'], ['toccare', 'verb', 'a'], ['tocco', 'noun', 'b'], ['togliere', 'verb', 'a'], ['toilette', 'noun', 'c'], ['toletta', 'noun', 'c'], ['tolleranza', 'noun', 'b'], ['tollerare', 'verb', 'b'], ['tomba', 'noun', 'b'], ['tombola', 'noun', 'c'], ['tonaca', 'noun', 'c'], ['tondo', 'adjective', 'b'], ['tondo', 'noun', 'b'], ['tonnellata', 'noun', 'b'], ['tonno', 'noun', 'c'], ['tono', 'noun', 'a'], ['tonsilla', 'noun', 'c'], ['top', 'noun', 'b'], ['topo', 'noun', 'b'], ['topo', 'adjective', 'b'], ['toppa', 'noun', 'c'], ['torbido', 'adjective', 'c'], ['torbido', 'noun', 'c'], ['torcere', 'verb', 'b'], ['torcia', 'noun', 'c'], ['torcicollo', 'noun', 'c'], ['tordo', 'noun', 'c'], ['torero', 'noun', 'c'], ['torinese', 'adjective', 'c'], ['torinese', 'noun', 'c'], ['tormentare', 'verb', 'b'], ['tornaconto', 'noun', 'c'], ['tornare', 'verb', 'a'], ['torneo', 'noun', 'b'], ['tornio', 'noun', 'c'], ['toro', 'noun', 'b'], ['torre', 'noun', 'b'], ['torrone', 'noun', 'c'], ['torta', 'noun', 'b'], ['tortellino', 'noun', 'c'], ['torto', 'noun', 'b'], ['tortora', 'noun', 'c'], ['tortora', 'adjective', 'c'], ['tortora', 'noun', 'c'], ['tosare', 'verb', 'c'], ['toscano', 'adjective', 'b'], ['toscano', 'noun', 'b'], ['tosse', 'noun', 'b'], ['tossico', 'adjective', 'b'], ['tossico', 'noun', 'b'], ['tossire', 'verb', 'c'], ['tostapane', 'noun', 'c'], ['totale', 'adjective', 'a'], ['totale', 'noun', 'a'], ['totalmente', 'adverb', 'b'], ['tour', 'noun', 'b'], ['tovaglia', 'noun', 'b'], ['tovaglietta', 'noun', 'c'], ['tovagliolo', 'noun', 'c'], ['tra', 'preposition', 'a'], ['traballare', 'verb', 'c'], ['traboccare', 'verb', 'c'], ['trabocchetto', 'noun', 'c'], ['traccia', 'noun', 'a'], ['tracciare', 'verb', 'b'], ['tradimento', 'noun', 'b'], ['tradire', 'verb', 'b'], ['tradizionale', 'adjective', 'a'], ['tradizione', 'noun', 'a'], ['tradurre', 'verb', 'a'], ['traduzione', 'noun', 'a'], ['traffico', 'noun', 'a'], ['trafila', 'noun', 'c'], ['traforo', 'noun', 'c'], ['tragedia', 'noun', 'b'], ['traghetto', 'noun', 'c'], ['tragico', 'adjective', 'b'], ['tragico', 'noun', 'b'], ['trainare', 'verb', 'c'], ['trama', 'noun', 'b'], ['tramezzino', 'noun', 'c'], ['tramite', 'noun', 'preposition'], ['tramontare', 'verb', 'c'], ['tramonto', 'noun', 'b'], ['trampolino', 'noun', 'c'], ['trancio', 'noun', 'c'], ['tranne', 'preposition', 'a'], ['tranquillamente', 'adverb', 'b'], ['tranquillit', 'noun', 'b'], ['tranquillizzare', 'verb', 'c'], ['tranquillo', 'adjective', 'a'], ['tranquillo', 'adverb', 'a'], ['tranquillo', 'noun', 'a'], ['transito', 'noun', 'c'], ['trapano', 'noun', 'c'], ['trapezio', 'noun', 'c'], ['trapezio', 'adjective', 'c'], ['trapianto', 'noun', 'c'], ['trappola', 'noun', 'b'], ['trapunta', 'noun', 'c'], ['trarre', 'verb', 'a'], ['trascinare', 'verb', 'a'], ['trascorrere', 'verb', 'a'], ['trascrizione', 'noun', 'b'], ['trascurare', 'verb', 'b'], ['trasferimento', 'noun', 'b'], ['trasferire', 'verb', 'a'], ['trasformare', 'verb', 'a'], ['trasformazione', 'noun', 'b'], ['trasfusione', 'noun', 'c'], ['traslocare', 'verb', 'c'], ['trasloco', 'noun', 'c'], ['trasmettere', 'verb', 'a'], ['trasmissione', 'noun', 'a'], ['trasparente', 'adjective', 'b'], ['trasparente', 'noun', 'b'], ['trasparenza', 'noun', 'b'], ['trasportare', 'verb', 'b'], ['trasporto', 'noun', 'a'], ['trattamento', 'noun', 'a'], ['trattare', 'verb', 'a'], ['trattativa', 'noun', 'b'], ['trattato', 'noun', 'b'], ['trattenere', 'verb', 'a'], ['trattenuta', 'noun', 'c'], ['tratto', 'noun', 'a'], ['trattore', 'noun', 'c'], ['trauma', 'noun', 'b'], ['travasare', 'verb', 'c'], ['travestire', 'verb', 'c'], ['travolgere', 'verb', 'b'], ['tre', 'adjective', 'a'], ['tre', 'noun', 'a'], ['trebbiare', 'verb', 'c'], ['trecento', 'adjective', 'b'], ['trecento', 'noun', 'b'], ['tredici', 'adjective', 'b'], ['tredici', 'noun', 'b'], ['tremare', 'verb', 'b'], ['tremendo', 'adjective', 'b'], ['trend', 'noun', 'b'], ['treno', 'noun', 'a'], ['trenta', 'adjective', 'a'], ['trenta', 'noun', 'a'], ['trentino', 'adjective', 'c'], ['trentino', 'noun', 'c'], ['triangolo', 'noun', 'b'], ['trib', 'noun', 'c'], ['tribunale', 'noun', 'a'], ['triestino', 'adjective', 'c'], ['triestino', 'noun', 'c'], ['trifoglio', 'noun', 'c'], ['trina', 'noun', 'c'], ['trincea', 'noun', 'c'], ['trionfo', 'noun', 'b'], ['triste', 'adjective', 'a'], ['tristezza', 'noun', 'b'], ['tritare', 'verb', 'c'], ['trofeo', 'noun', 'c'], ['tronco', 'noun', 'b'], ['trono', 'noun', 'b'], ['troppo', 'adjective', 'a'], ['troppo', 'pronoun', 'a'], ['troppo', 'adverb', 'a'], ['troppo', 'noun', 'a'], ['trota', 'noun', 'c'], ['trottare', 'verb', 'c'], ['trottola', 'noun', 'c'], ['trovare', 'verb', 'a'], ['truccare', 'verb', 'c'], ['trucco', 'noun', 'b'], ['trucco', 'noun', 'b'], ['truffa', 'noun', 'b'], ['truffare', 'verb', 'c'], ['truppa', 'noun', 'b'], ['t-shirt', 'noun', 'c'], ['tu', 'pronoun', 'a'], ['tubo', 'noun', 'b'], ['tuffare', 'verb', 'b'], ['tuffo', 'noun', 'c'], ['tulipano', 'noun', 'c'], ['tumore', 'noun', 'b'], ['tunica', 'noun', 'c'], ['tunisino', 'adjective', 'c'], ['tunisino', 'noun', 'c'], ['tunnel', 'noun', 'c'], ['tuo', 'adjective', 'a'], ['tuo', 'pronoun', 'a'], ['tuono', 'noun', 'c'], ['turbare', 'verb', 'b'], ['turco', 'adjective', 'b'], ['turco', 'noun', 'b'], ['turismo', 'noun', 'b'], ['turista', 'noun', 'b'], ['turistico', 'adjective', 'b'], ['turno', 'noun', 'a'], ['tuta', 'noun', 'b'], ['tutela', 'noun', 'b'], ['tutelare', 'verb', 'b'], ['tutore', 'noun', 'c'], ['tuttavia', 'conjunction', 'a'], ['tuttavia', 'adverb', 'a'], ['tutto', 'adjective', 'a'], ['tutto', 'pronoun', 'a'], ['tuttora', 'adverb', 'b'], ['u', 'noun', 'c'], ['ubriaco', 'adjective', 'b'], ['ubriaco', 'noun', 'b'], ['uccello', 'noun', 'a'], ['uccidere', 'verb', 'a'], ['ucraino', 'adjective', 'c'], ['ucraino', 'noun', 'c'], ['udienza', 'noun', 'b'], ['udinese', 'adjective', 'c'], ['udinese', 'noun', 'c'], ['udire', 'verb', 'b'], ['udire', 'noun', 'b'], ['ufficiale', 'noun', 'b'], ['ufficiale', 'adjective', 'a'], ['ufficialmente', 'adverb', 'b'], ['ufficio', 'noun', 'a'], ['uguale', 'adjective', 'a'], ['uguale', 'adverb', 'a'], ['uguale', 'noun', 'a'], ['ugualmente', 'adverb', 'b'], ['ulcera', 'noun', 'c'], ['ulteriore', 'adjective', 'a'], ['ulteriormente', 'adverb', 'b'], ['ultimamente', 'adverb', 'b'], ['ultimo', 'adjective', 'a'], ['ultimo', 'noun', 'a'], ['ultravioletto', 'noun', 'c'], ['ultravioletto', 'adjective', 'c'], ['umanit', 'noun', 'a'], ['umano', 'adjective', 'a'], ['umano', 'noun', 'a'], ['umbro', 'adjective', 'c'], ['umbro', 'noun', 'c'], ['umido', 'adjective', 'b'], ['umido', 'noun', 'b'], ['umile', 'adjective', 'b'], ['umile', 'noun', 'b'], ['umiliare', 'verb', 'b'], ['umore', 'noun', 'b'], ['umorismo', 'noun', 'c'], ['una', 'determiner', 'a'], ['una', 'pronoun', 'a'], ['undici', 'adjective', 'b'], ['undici', 'noun', 'b'], ['ungherese', 'adjective', 'c'], ['ungherese', 'noun', 'c'], ['unghia', 'noun', 'b'], ['unguento', 'noun', 'c'], ['unico', 'adjective', 'a'], ['unico', 'noun', 'a'], ['uniforme', 'adjective', 'b'], ['unione', 'noun', 'b'], ['unire', 'verb', 'a'], ['unit', 'noun', 'a'], ['unito', 'past_part', 'a'], ['unito', 'adjective', 'a'], ['unito', 'noun', 'a'], ['universale', 'adjective', 'b'], ['universale', 'noun', 'b'], ['universit', 'noun', 'a'], ['universitario', 'adjective', 'b'], ['universitario', 'noun', 'b'], ['universo', 'noun', 'a'], ['uno', 'adjective', 'a'], ['uno', 'noun', 'a'], ['uno', 'determiner', 'a'], ['uno', 'pronoun', 'a'], ['uomo', 'noun', 'a'], ['uovo', 'noun', 'a'], ['uragano', 'noun', 'c'], ['urbanistico', 'adjective', 'b'], ['urbano', 'adjective', 'b'], ['urgente', 'adjective', 'b'], ['urgenza', 'noun', 'b'], ['urlare', 'verb', 'a'], ['urlo', 'noun', 'b'], ['urna', 'noun', 'c'], ['urtare', 'verb', 'b'], ['usare', 'verb', 'a'], ['usato', 'past_part', 'b'], ['usato', 'adjective', 'b'], ['usato', 'noun', 'b'], ['uscire', 'verb', 'a'], ['uscita', 'noun', 'a'], ['usignolo', 'noun', 'c'], ['uso', 'noun', 'a'], ['utensile', 'noun', 'c'], ['utente', 'noun', 'a'], ['utenza', 'noun', 'b'], ['utile', 'adjective', 'a'], ['utile', 'noun', 'a'], ['utilit', 'noun', 'b'], ['utilizzare', 'verb', 'a'], ['utilizzo', 'noun', 'b'], ['vabb', 'exclamation', 'b'], ['vacanza', 'noun', 'a'], ['vacca', 'noun', 'b'], ['vaccino', 'noun', 'c'], ['vaffanculo', 'exclamation', 'b'], ['vagare', 'verb', 'b'], ['vagire', 'verb', 'c'], ['vago', 'adjective', 'b'], ['vago', 'noun', 'b'], ['valanga', 'noun', 'c'], ['valdostano', 'adjective', 'c'], ['valdostano', 'noun', 'c'], ['valere', 'verb', 'a'], ['valido', 'adjective', 'b'], ['valigia', 'noun', 'b'], ['valle', 'noun', 'b'], ['valore', 'noun', 'a'], ['valorizzare', 'verb', 'b'], ['valoroso', 'adjective', 'c'], ['valoroso', 'noun', 'c'], ['valutare', 'verb', 'a'], ['valutazione', 'noun', 'b'], ['valvola', 'noun', 'c'], ['vampata', 'noun', 'c'], ['vampiro', 'noun', 'b'], ['vandalo', 'adjective', 'c'], ['vandalo', 'noun', 'c'], ['vanga', 'noun', 'c'], ['vangelo', 'noun', 'b'], ['vanitoso', 'adjective', 'c'], ['vanitoso', 'noun', 'c'], ['vano', 'adjective', 'b'], ['vano', 'noun', 'b'], ['vantaggio', 'noun', 'a'], ['vantaggioso', 'adjective', 'c'], ['vantare', 'verb', 'b'], ['vanto', 'noun', 'c'], ['vapore', 'noun', 'b'], ['variabile', 'adjective', 'b'], ['variabile', 'noun', 'b'], ['variante', 'pres_part', 'b'], ['variante', 'adjective', 'b'], ['variante', 'noun', 'b'], ['variare', 'verb', 'b'], ['variazione', 'noun', 'b'], ['variet', 'noun', 'b'], ['vario', 'adjective', 'a'], ['vario', 'adjective', 'a'], ['vario', 'pronoun', 'a'], ['variopinto', 'adjective', 'c'], ['vasca', 'noun', 'b'], ['vaso', 'noun', 'b'], ['vasto', 'adjective', 'b'], ['vasto', 'noun', 'b'], ['ve', 'pronoun', 'a'], ['ve', 'adverb', 'a'], ['vecchio', 'adjective', 'a'], ['vecchio', 'noun', 'a'], ['vedere', 'verb', 'a'], ['vedere', 'noun', 'a'], ['vedova', 'noun', 'b'], ['vegetale', 'adjective', 'b'], ['vegetale', 'noun', 'b'], ['veglia', 'noun', 'c'], ['veglione', 'noun', 'c'], ['veicolo', 'noun', 'b'], ['vela', 'noun', 'b'], ['veleno', 'noun', 'b'], ['velenoso', 'adjective', 'c'], ['vellutato', 'past_part', 'c'], ['vellutato', 'adjective', 'c'], ['velluto', 'noun', 'c'], ['velo', 'noun', 'b'], ['veloce', 'adjective', 'a'], ['veloce', 'adverb', 'a'], ['veloce', 'noun', 'a'], ['velocemente', 'adverb', 'b'], ['velocit', 'noun', 'a'], ['vena', 'noun', 'b'], ['vendemmiare', 'verb', 'c'], ['vendere', 'verb', 'a'], ['vendetta', 'noun', 'b'], ['vendicare', 'verb', 'b'], ['vendita', 'noun', 'a'], ['venditore', 'adjective', 'b'], ['venditore', 'noun', 'b'], ['venerd', 'noun', 'a'], ['veneto', 'adjective', 'b'], ['veneto', 'noun', 'b'], ['veneziano', 'adjective', 'c'], ['veneziano', 'noun', 'c'], ['venire', 'verb', 'a'], ['ventaglio', 'noun', 'c'], ['ventata', 'noun', 'c'], ['venti', 'adjective', 'a'], ['venti', 'noun', 'a'], ['venticinque', 'adjective', 'b'], ['venticinque', 'noun', 'b'], ['ventilatore', 'adjective', 'c'], ['ventilatore', 'noun', 'c'], ['ventina', 'noun', 'b'], ['ventiquattro', 'adjective', 'b'], ['ventiquattro', 'noun', 'b'], ['vento', 'noun', 'a'], ['ventre', 'noun', 'b'], ['venuta', 'noun', 'c'], ['veramente', 'adverb', 'a'], ['verbale', 'adjective', 'a'], ['verbale', 'noun', 'a'], ['verbo', 'noun', 'b'], ['verde', 'adjective', 'a'], ['verde', 'noun', 'a'], ['verdura', 'noun', 'b'], ['vergine', 'adjective', 'b'], ['vergine', 'noun', 'b'], ['vergogna', 'noun', 'b'], ['vergognarsi', 'verb', 'b'], ['verifica', 'noun', 'b'], ['verificare', 'verb', 'a'], ['verit', 'noun', 'a'], ['verme', 'noun', 'b'], ['vernice', 'noun', 'b'], ['vero', 'adjective', 'a'], ['vero', 'noun', 'a'], ['versare', 'verb', 'a'], ['versione', 'noun', 'a'], ['verso', 'noun', 'a'], ['verso', 'preposition', 'a'], ['vertebra', 'noun', 'c'], ['verticale', 'adjective', 'b'], ['verticale', 'noun', 'b'], ['vertice', 'noun', 'b'], ['vertigine', 'noun', 'c'], ['vescovo', 'noun', 'b'], ['vescovo', 'adjective', 'b'], ['vespa', 'noun', 'c'], ['veste', 'noun', 'b'], ['vestire', 'verb', 'a'], ['vestito', 'noun', 'a'], ['vestito', 'past_part', 'b'], ['vestito', 'adjective', 'b'], ['veterinario', 'adjective', 'c'], ['veterinario', 'noun', 'c'], ['vetrina', 'noun', 'b'], ['vetro', 'noun', 'a'], ['vettura', 'noun', 'b'], ['vi', 'pronoun', 'a'], ['vi', 'adverb', 'a'], ['via', 'noun', 'a'], ['via', 'adverb', 'a'], ['via', 'exclamation', 'a'], ['via', 'noun', 'a'], ['viaggiare', 'verb', 'a'], ['viaggiatore', 'noun', 'b'], ['viaggiatrice', 'noun', 'c'], ['viaggio', 'noun', 'a'], ['viale', 'noun', 'b'], ['vibrare', 'verb', 'b'], ['vice', 'noun', 'b'], ['vicenda', 'noun', 'a'], ['viceversa', 'adverb', 'b'], ['vicinanza', 'noun', 'b'], ['vicino', 'adjective', 'a'], ['vicino', 'noun', 'a'], ['vicino', 'adverb', 'a'], ['vicolo', 'noun', 'b'], ['video', 'adjective', 'a'], ['video', 'noun', 'a'], ['videogioco', 'noun', 'b'], ['viennese', 'adjective', 'c'], ['viennese', 'noun', 'c'], ['vietare', 'verb', 'b'], ['vigile', 'adjective', 'b'], ['vigile', 'noun', 'b'], ['vigilia', 'noun', 'b'], ['vigna', 'noun', 'c'], ['vigore', 'noun', 'b'], ['villa', 'noun', 'a'], ['villaggio', 'noun', 'a'], ['vincente', 'pres_part', 'b'], ['vincente', 'adjective', 'b'], ['vincente', 'noun', 'b'], ['vincere', 'verb', 'a'], ['vincitore', 'adjective', 'b'], ['vincitore', 'noun', 'b'], ['vincolo', 'noun', 'b'], ['vino', 'noun', 'a'], ['vino', 'adjective', 'a'], ['viola', 'noun', 'b'], ['viola', 'adjective', 'b'], ['violare', 'verb', 'b'], ['violazione', 'noun', 'b'], ['violentare', 'verb', 'c'], ['violento', 'adjective', 'a'], ['violento', 'noun', 'a'], ['violenza', 'noun', 'a'], ['violetta', 'noun', 'c'], ['violetto', 'adjective', 'c'], ['violetto', 'noun', 'c'], ['violino', 'noun', 'b'], ['vipera', 'noun', 'c'], ['virgola', 'noun', 'b'], ['virt', 'noun', 'b'], ['virtuale', 'adjective', 'b'], ['virus', 'noun', 'b'], ['visibile', 'adjective', 'b'], ['visibile', 'noun', 'b'], ['visione', 'noun', 'a'], ['visita', 'noun', 'a'], ['visitare', 'verb', 'a'], ['visitatore', 'noun', 'b'], ['visivo', 'adjective', 'b'], ['viso', 'noun', 'a'], ['vissuto', 'past_part', 'b'], ['vissuto', 'adjective', 'b'], ['vissuto', 'noun', 'b'], ['vista', 'noun', 'a'], ['vita', 'noun', 'a'], ['vitale', 'adjective', 'b'], ['vitale', 'noun', 'b'], ['vitamina', 'noun', 'c'], ['vite', 'noun', 'c'], ['vitello', 'noun', 'c'], ['vittima', 'noun', 'a'], ['vittoria', 'noun', 'a'], ['vivace', 'adjective', 'b'], ['vivace', 'adverb', 'b'], ['vivace', 'noun', 'b'], ['vivente', 'pres_part', 'b'], ['vivente', 'adjective', 'b'], ['vivente', 'noun', 'b'], ['vivere', 'verb', 'a'], ['vivere', 'noun', 'a'], ['vivo', 'adjective', 'a'], ['vivo', 'noun', 'a'], ['viziare', 'verb', 'c'], ['viziato', 'past_part', 'c'], ['viziato', 'adjective', 'c'], ['vizio', 'noun', 'b'], ['vocabolario', 'noun', 'b'], ['vocale', 'noun', 'b'], ['vocale', 'adjective', 'b'], ['vocazione', 'noun', 'b'], ['voce', 'noun', 'a'], ['vodka', 'noun', 'c'], ['voglia', 'noun', 'a'], ['voi', 'pronoun', 'a'], ['volantino', 'noun', 'c'], ['volare', 'verb', 'a'], ['volata', 'noun', 'c'], ['volenteroso', 'adjective', 'c'], ['volentieri', 'adverb', 'b'], ['volere', 'verb', 'a'], ['volgare', 'adjective', 'b'], ['volgare', 'noun', 'b'], ['volgere', 'verb', 'b'], ['volo', 'noun', 'a'], ['volont', 'noun', 'a'], ['volontariato', 'noun', 'b'], ['volontario', 'adjective', 'b'], ['volontario', 'noun', 'b'], ['volta', 'noun', 'a'], ['voltare', 'verb', 'a'], ['volto', 'noun', 'a'], ['volume', 'noun', 'a'], ['vomitare', 'verb', 'b'], ['vomito', 'noun', 'c'], ['vongola', 'noun', 'c'], ['vostro', 'adjective', 'a'], ['vostro', 'pronoun', 'a'], ['votare', 'verb', 'a'], ['votazione', 'noun', 'c'], ['voto', 'noun', 'a'], ['vu', 'noun', 'c'], ['vuotare', 'verb', 'c'], ['vuoto', 'adjective', 'a'], ['vuoto', 'noun', 'a'], ['wafer', 'noun', 'c'], ['web', 'noun', 'a'], ['weekend', 'noun', 'b'], ['whisky', 'noun', 'c'], ['wurstel', 'noun', 'c'], ['yogurt', 'noun', 'c'], ['zaino', 'noun', 'b'], ['zampa', 'noun', 'b'], ['zampogna', 'noun', 'c'], ['zanna', 'noun', 'c'], ['zanzara', 'noun', 'c'], ['zattera', 'noun', 'c'], ['zebra', 'noun', 'c'], ['zero', 'adjective', 'a'], ['zero', 'noun', 'a'], ['zero', 'symbol', 'a'], ['zeta', 'noun', 'c'], ['zia', 'noun', 'a'], ['zingaro', 'adjective', 'c'], ['zingaro', 'noun', 'c'], ['zio', 'noun', 'a'], ['zitella', 'noun', 'c'], ['zitto', 'adjective', 'a'], ['zitto', 'noun', 'a'], ['zoccolo', 'noun', 'c'], ['zolla', 'noun', 'c'], ['zona', 'noun', 'a'], ['zoo', 'noun', 'c'], ['zoppicare', 'verb', 'c'], ['zoppo', 'adjective', 'c'], ['zoppo', 'noun', 'c'], ['zucca', 'noun', 'b'], ['zucchero', 'noun', 'b'], ['zucchina', 'noun', 'c'], ['zuffa', 'noun', 'c'], ['zuppa', 'noun', 'c'], ]
[ 18893, 62, 270, 796, 685, 198, 17816, 64, 3256, 705, 77, 977, 3256, 705, 66, 6, 4357, 198, 17816, 64, 3256, 705, 3866, 9150, 3256, 705, 64, 6, 4357, 198, 17816, 6485, 363, 75, 3014, 68, 3256, 705, 18302, 62, 3911, 3256, 705, 66, ...
2.008447
126,431
from __future__ import absolute_import, division, print_function, unicode_literals import os # this arbitrary-looking assortment of functionality is provided here # to have a central place for overrideable behavior. The motivating # use is the FB build environment, where this source file is replaced # by an equivalent. if os.path.basename(os.path.dirname(__file__)) == 'shared': torch_parent = os.path.dirname(os.path.dirname(os.path.dirname(__file__))) else: torch_parent = os.path.dirname(os.path.dirname(__file__)) TEST_MASTER_ADDR = '127.0.0.1' TEST_MASTER_PORT = 29500
[ 6738, 11593, 37443, 834, 1330, 4112, 62, 11748, 11, 7297, 11, 3601, 62, 8818, 11, 28000, 1098, 62, 17201, 874, 198, 198, 11748, 28686, 198, 198, 2, 428, 14977, 12, 11534, 36168, 286, 11244, 318, 2810, 994, 198, 2, 284, 423, 257, 431...
3.061856
194
import json from collections import OrderedDict from typing import Union, List import clip import torch import torch.nn as nn import torch.nn.functional as F from libs.definitions import ROOT label_file = ROOT / 'imagenet_class_index.json' with open(label_file, 'r') as f: labels = json.load(f) _DEFAULT_CLASSNAMES = [value[1] for value in labels.values()] # templates are copied from https://github.com/openai/CLIP/blob/main/notebooks/Prompt_Engineering_for_ImageNet.ipynb _DEFAULT_TEMPLATES = [ 'a bad photo of a {}.', 'a photo of many {}.', 'a sculpture of a {}.', 'a photo of the hard to see {}.', 'a low resolution photo of the {}.', 'a rendering of a {}.', 'graffiti of a {}.', 'a bad photo of the {}.', 'a cropped photo of the {}.', 'a tattoo of a {}.', 'the embroidered {}.', 'a photo of a hard to see {}.', 'a bright photo of a {}.', 'a photo of a clean {}.', 'a photo of a dirty {}.', 'a dark photo of the {}.', 'a drawing of a {}.', 'a photo of my {}.', 'the plastic {}.', 'a photo of the cool {}.', 'a close-up photo of a {}.', 'a black and white photo of the {}.', 'a painting of the {}.', 'a painting of a {}.', 'a pixelated photo of the {}.', 'a sculpture of the {}.', 'a bright photo of the {}.', 'a cropped photo of a {}.', 'a plastic {}.', 'a photo of the dirty {}.', 'a jpeg corrupted photo of a {}.', 'a blurry photo of the {}.', 'a photo of the {}.', 'a good photo of the {}.', 'a rendering of the {}.', 'a {} in a video game.', 'a photo of one {}.', 'a doodle of a {}.', 'a close-up photo of the {}.', 'a photo of a {}.', 'the origami {}.', 'the {} in a video game.', 'a sketch of a {}.', 'a doodle of the {}.', 'a origami {}.', 'a low resolution photo of a {}.', 'the toy {}.', 'a rendition of the {}.', 'a photo of the clean {}.', 'a photo of a large {}.', 'a rendition of a {}.', 'a photo of a nice {}.', 'a photo of a weird {}.', 'a blurry photo of a {}.', 'a cartoon {}.', 'art of a {}.', 'a sketch of the {}.', 'a embroidered {}.', 'a pixelated photo of a {}.', 'itap of the {}.', 'a jpeg corrupted photo of the {}.', 'a good photo of a {}.', 'a plushie {}.', 'a photo of the nice {}.', 'a photo of the small {}.', 'a photo of the weird {}.', 'the cartoon {}.', 'art of the {}.', 'a drawing of the {}.', 'a photo of the large {}.', 'a black and white photo of a {}.', 'the plushie {}.', 'a dark photo of a {}.', 'itap of a {}.', 'graffiti of the {}.', 'a toy {}.', 'itap of my {}.', 'a photo of a cool {}.', 'a photo of a small {}.', 'a tattoo of the {}.', ]
[ 11748, 33918, 198, 6738, 17268, 1330, 14230, 1068, 35, 713, 198, 6738, 19720, 1330, 4479, 11, 7343, 198, 198, 11748, 10651, 198, 11748, 28034, 198, 11748, 28034, 13, 20471, 355, 299, 77, 198, 11748, 28034, 13, 20471, 13, 45124, 355, 376...
2.387288
1,180
#!/usr/bin/env python from __future__ import print_function from optparse import OptionParser import os import random import subprocess import matplotlib matplotlib.use('Agg') import numpy as np import matplotlib.pyplot as plt import pysam from scipy.stats import binom from scipy.stats.mstats import mquantiles import seaborn as sns import stats ################################################################################ # basset_sick_loss.py # # Shuffle SNPs that overlap DNase sites within their sites and compare the SAD # distributions. # # Todo: # -Control for GC% changes introduced by mutation shuffles. # -Control for positional changes within the DHS regions. # -Properly handle indels. ################################################################################ ################################################################################ # main ################################################################################ def compute_sad(sample_vcf_file, model_file, si, out_dir, seq_len, gpu, replot): ''' Run basset_sad.py to compute scores. ''' cuda_str = '' if gpu: cuda_str = '--cudnn' cmd = 'basset_sad.py %s -l %d -o %s %s %s' % (cuda_str, seq_len, out_dir, model_file, sample_vcf_file) if not replot: subprocess.call(cmd, shell=True) sad = [] for line in open('%s/sad_table.txt' % out_dir): a = line.split() if a[3] == 't%d'%si: sad.append(float(a[-1])) return np.array(sad) def filter_vcf(vcf_file, bed_file, sample_vcf_file): ''' Filter the VCF file for SNPs that overlap the BED file, removing indels. ''' # open filtered file sample_vcf_out = open(sample_vcf_file, 'w') # intersect p = subprocess.Popen('bedtools intersect -wo -a %s -b %s' % (vcf_file, bed_file), stdout=subprocess.PIPE, shell=True) for line in p.stdout: a = line.split() if len(a[3]) == len(a[4]) == 1: print(line, file=sample_vcf_out, end='') sample_vcf_out.close() def retrieve_sad(sample_vcf_file, sad_table_file, si): ''' Retrieve SAD scores from a pre-computed table. Note that I'm assuming here the table has all SAD scores in one row for each SNP so I can pull out the score I want as column si+1. ''' snp_indexes = {} vi = 0 for line in open(sample_vcf_file): a = line.split() snp_indexes[a[2]] = vi vi += 1 sad = np.zeros(len(snp_indexes)) for line in open(sad_table_file): a = line.split() print(a) if a[0] in snp_indexes: sad[snp_indexes[a[0]]] = float(a[si+1]) return sad def shuffle_snps(in_vcf_file, out_vcf_file, genome): ''' Shuffle the SNPs within their overlapping DHS. ''' out_vcf_open = open(out_vcf_file, 'w') for line in open(in_vcf_file): a = line.split() # read SNP info snp_chrom = a[0] snp_pos = int(a[1]) snp_nt = a[3] # determine BED start bi = 5 while a[bi] != snp_chrom: bi += 1 # read BED info bed_chrom = a[bi] bed_start = int(a[bi+1]) bed_end = int(a[bi+2]) # get sequence bed_seq = genome.fetch(bed_chrom, bed_start, bed_end) # determine matching positions bed_nt_matches = [i for i in range(len(bed_seq)) if bed_seq[i] == snp_nt] while len(bed_nt_matches) == 0: # expand segment by 10 nt bed_start = max(0, bed_start-10) bed_end += 10 bed_seq = genome.fetch(bed_chrom, bed_start, bed_end) # sample new SNP position shuf_pos = bed_start + 1 + random.choice(bed_nt_matches) # write into columns a[1] = str(shuf_pos) print('\t'.join(a), file=out_vcf_open) out_vcf_open.close() def shuffle_snps_old(in_vcf_file, out_vcf_file, genome): ''' Shuffle the SNPs within their overlapping DHS. ''' out_vcf_open = open(out_vcf_file, 'w') for line in open(in_vcf_file): a = line.split() # read SNP info snp_chrom = a[0] snp_pos = int(a[1]) # determine BED start bi = 5 while a[bi] != snp_chrom: bi += 1 # read BED info bed_chrom = a[bi] bed_start = int(a[bi+1]) bed_end = int(a[bi+2]) # sample new SNP position shuf_pos = random.randint(bed_start, bed_end-1) while shuf_pos == snp_pos: shuf_pos = random.randint(bed_start, bed_end-1) # set reference allele ref_nt = genome.fetch(snp_chrom, shuf_pos-1, shuf_pos) # sample alternate allele alt_nt = random.choice('ACGT') while alt_nt == ref_nt: alt_nt = random.choice('ACGT') # write into columns a[1] = str(shuf_pos) a[3] = ref_nt a[4] = alt_nt print('\t'.join(a), file=out_vcf_open) ################################################################################ # __main__ ################################################################################ if __name__ == '__main__': main()
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 6738, 11593, 37443, 834, 1330, 3601, 62, 8818, 198, 6738, 2172, 29572, 1330, 16018, 46677, 198, 11748, 28686, 198, 11748, 4738, 198, 11748, 850, 14681, 198, 198, 11748, 2603, 29487, 8019, ...
2.300133
2,249
#!/usr/bin/env python3 import codecs import re def trans_addr(addr): """Traduit une position de fichier en une adresse du programme""" if addr < 0x1000: return 0 if 0x0000000000001000 <= addr < 0x0000000000001000 + 0x00001ef000000000: return 0x00002b0000000000 + addr - 0x0000000000001000 if 0x00002afffffe1000 <= addr < 0x00002afffffe1000 + 0x0000161000000000: return 0x000049f000000000 + addr - 0x00002afffffe1000 if 0x000049effffe1000 <= addr < 0x000049effffe1000 + 0x00002afffffe0000: return 0x0000000000020000 + addr - 0x000049effffe1000 raise Exception("Invalid addr {:#x}".format(addr)) blobs = {} with open('strace_tar_output.log', 'r') as f: curseek = 0 for line in f: m = re.match(r'lseek\(4, ([^,]*), SEEK_SET\)', line) if m is not None: curseek = int(m.group(1)) continue if line.startswith('write(4, "'): m = re.match(r'write\(4, "(.*)", ([0-9]*)\) = ([0-9]*)', line) assert m is not None: rawdata, count1, count2 = m.groups() assert count1 == count2 addr = curseek curseek += int(count1) data = codecs.escape_decode(rawdata.encode('ascii'))[0] # Trouve le premier octet non-nul dans le bloc de donnes i = 0 while i < len(data) and not data[i]: i += 1 if i >= len(data): continue addr = trans_addr(addr + i) data = data[i:].rstrip(b'\0') with open('out/blob-{:016x}.bin'.format(addr), 'wb') as f: f.write(data)
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 18, 198, 11748, 40481, 82, 198, 11748, 302, 198, 198, 4299, 1007, 62, 29851, 7, 29851, 2599, 198, 220, 220, 220, 37227, 2898, 324, 5013, 17809, 2292, 390, 277, 488, 959, 551, 17809, 512, ...
2.02445
818
import pygame_gui import pygame TITLE = "TUNAPRO1234" BACKGROUND = colors.darkslategray WIDTH, HEIGHT = 1920, 1080 """ Hzlca bir plan yapacam Neural alar kontrol edebileceimiz kk bir framework Alarn geliimini grebileceiz deitirebileceiz ve kaydedebileceiz Bunun iin -select box yaps -balatmak iin buton -kaydetme olaylar iin stteki eyden Pencereler -tkladmz nronun bilgilerini gsteren ve deitirebilen bir pencere -tkladmz weightin deimesini salayan bir pencere Norn, katman ve a iin pygame wrapperlar yazacam Weigth iin de bir class olur Kaydetme olayna daha var """ # elements: dict: {"buttons": butonlar, "entries", entryler} if __name__ == "__main__": main()
[ 11748, 12972, 6057, 62, 48317, 198, 11748, 12972, 6057, 628, 198, 198, 49560, 2538, 796, 366, 51, 4944, 2969, 13252, 1065, 2682, 1, 198, 31098, 46025, 796, 7577, 13, 67, 5558, 17660, 44605, 198, 54, 2389, 4221, 11, 11179, 9947, 796, 1...
2.480144
277
import time import pytest from flask import g from flask import session import paho.mqtt.client as paho from SmartSleep.db import get_db from flask import json import runpy msg_nr = 0 messages = [""] broker = 'broker.emqx.io' port = 1883
[ 11748, 640, 198, 198, 11748, 12972, 9288, 198, 6738, 42903, 1330, 308, 198, 6738, 42903, 1330, 6246, 198, 11748, 279, 17108, 13, 76, 80, 926, 13, 16366, 355, 279, 17108, 198, 6738, 10880, 40555, 13, 9945, 1330, 651, 62, 9945, 198, 673...
2.939759
83
import theano.tensor as tt
[ 11748, 262, 5733, 13, 83, 22854, 355, 256, 83, 628, 628 ]
2.727273
11
from pyEtherCAT import MasterEtherCAT # nic = "eth0" # cat = MasterEtherCAT.MasterEtherCAT(nic) ADP = 0x0000 #1 ADDR = 0x0E00 # cat.APRD(IDX=0x00, ADP=ADP, ADO=ADDR, DATA=[0,0,0,0,0,0,0,0]) #DATA(64bit) (DATA, WKC) = cat.socket_read() # print("[0x{:04X}]= 0x{:02x}{:02x},0x{:02x}{:02x},0x{:02x}{:02x},0x{:02x}{:02x}".format(ADDR, DATA[7],DATA[6],DATA[5],DATA[4],DATA[3],DATA[2],DATA[1],DATA[0])) #
[ 6738, 12972, 36, 490, 34, 1404, 1330, 5599, 36, 490, 34, 1404, 1303, 198, 6988, 796, 366, 2788, 15, 1, 1303, 220, 198, 9246, 796, 5599, 36, 490, 34, 1404, 13, 18254, 36, 490, 34, 1404, 7, 6988, 8, 220, 220, 198, 198, 2885, 47, ...
1.806306
222
#!/usr/bin/env python # -*- coding: UTF-8 -*- import pprint from base import BaseObject from base import FileIO
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 2, 532, 9, 12, 19617, 25, 41002, 12, 23, 532, 9, 12, 628, 198, 11748, 279, 4798, 198, 198, 6738, 2779, 1330, 7308, 10267, 198, 6738, 2779, 1330, 9220, 9399, 628 ]
2.9
40
import pandas as pd import time df = pd.read_csv("data/sample.csv") for num in range(1000): argx = str(df["x"][num:num+1].get_values()) argy = str(df["y"][num:num+1].get_values()) print("x:{0} / y:{1}".format(argx,argy)) time.sleep(0.1)
[ 11748, 19798, 292, 355, 279, 67, 198, 11748, 640, 198, 198, 7568, 796, 279, 67, 13, 961, 62, 40664, 7203, 7890, 14, 39873, 13, 40664, 4943, 198, 198, 1640, 997, 287, 2837, 7, 12825, 2599, 198, 220, 220, 220, 1822, 87, 796, 965, 7,...
2.125
120
from django.urls import path from . import views app_name = "graphs" urlpatterns = [ path("graphs/", views.GraphListView.as_view(), name="graph.list"), path("graphs/create/", views.GraphCreateView.as_view(), name="graph.create"), path( "graphs/<str:pk>/", views.GraphDetailView.as_view(), name="graph.detail", ), path( "graphs/<str:pk>/update/", views.GraphUpdateView.as_view(), name="graph.update", ), path( "graphs/<str:pk>/delete/", views.GraphDeleteView.as_view(), name="graph.delete", ), ]
[ 6738, 42625, 14208, 13, 6371, 82, 1330, 3108, 198, 198, 6738, 764, 1330, 5009, 198, 198, 1324, 62, 3672, 796, 366, 34960, 82, 1, 198, 6371, 33279, 82, 796, 685, 198, 220, 220, 220, 3108, 7203, 34960, 82, 14, 1600, 5009, 13, 37065, ...
2.13879
281
# -------------------------------------------------------- # Fast R-CNN # Copyright (c) 2015 Microsoft # Licensed under The MIT License [see LICENSE for details] # Written by Ross Girshick - extended by Lukas Tuggener # -------------------------------------------------------- """Blob helper functions.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np import cv2 import random def im_list_to_blob(ims): """Convert a list of images into a network input. Assumes images are already prepared (means subtracted, BGR order, ...). """ max_shape = np.array([im.shape for im in ims]).max(axis=0) num_images = len(ims) blob = np.zeros((num_images, max_shape[0], max_shape[1], 3), dtype=np.float32) for i in range(num_images): im = ims[i] blob[i, 0:im.shape[0], 0:im.shape[1], :] = im return blob def prep_im_for_blob(im, pixel_means, global_scale, args): """Mean subtract and scale an image for use in a blob.""" im = im.astype(np.float32, copy=False) # substract mean if args.substract_mean == "True": im -= pixel_means # do global scaling im = cv2.resize(im, None, None, fx=global_scale, fy=global_scale, interpolation=cv2.INTER_LINEAR) im_size_max = np.max(im.shape[0:2]) # Prevent the biggest axis from being more than MAX_SIZE if im_size_max > args.max_edge: if not args.crop == "True": # scale down if bigger than max size re_scale = (float(args.max_edge) / float(im_size_max)) im = cv2.resize(im, None, None, fx=re_scale, fy=re_scale, interpolation=cv2.INTER_LINEAR) global_scale = global_scale*re_scale crop_box = [0,0,im.shape[0],im.shape[1]] else: # Crop image topleft = random.uniform(0,1)<args.crop_top_left_bias # crop to max size if necessary if im.shape[0] <= args.max_edge or topleft: crop_0 = 0 else: crop_0 = random.randint(0,im.shape[0]-args.max_edge) if im.shape[1] <= args.max_edge or topleft: crop_1 = 0 else: crop_1 = random.randint(0,im.shape[1]-args.max_edge) crop_box = [crop_0, crop_1, min(crop_0+args.max_edge,im.shape[0]), min(crop_1+args.max_edge,im.shape[1])] im = im[crop_box[0]:crop_box[2],crop_box[1]:crop_box[3]] else: crop_box = [0, 0, im.shape[0], im.shape[1]] if not args.pad_to == 0: # pad to fit RefineNet #TODO fix refinenet padding problem y_mulity = int(np.ceil(im.shape[0] / float(args.pad_to))) x_mulity = int(np.ceil(im.shape[1] / float(args.pad_to))) canv = np.ones([y_mulity * args.pad_to, x_mulity * args.pad_to,3], dtype=np.uint8) * 255 canv[0:im.shape[0], 0:im.shape[1]] = im im = canv return im, global_scale, crop_box
[ 2, 20368, 22369, 198, 2, 12549, 371, 12, 18474, 198, 2, 15069, 357, 66, 8, 1853, 5413, 198, 2, 49962, 739, 383, 17168, 13789, 685, 3826, 38559, 24290, 329, 3307, 60, 198, 2, 22503, 416, 9847, 23837, 1477, 624, 532, 7083, 416, 28102,...
2.390347
1,181
# coding=utf-8 """ Gathers the media folders """ import elib from ._context import Context def get_media_folders(ctx: Context): """ Gathers the media folders """ ctx.info('gathering media folders') media_folders = [] this_folder = ctx.source_folder while True: ctx.debug(f'traversing: "{this_folder}"') media_folder_candidate = elib.path.ensure_path(this_folder, 'media', must_exist=False).absolute() if media_folder_candidate.exists() and media_folder_candidate.is_dir(): ctx.debug(f'media folder found: "{media_folder_candidate}"') media_folders.append(media_folder_candidate) if len(this_folder.parents) is 1: ctx.debug(f'reach mount point at: "{this_folder}"') break this_folder = this_folder.parent # if not media_folders: # raise ConvertError('no media folder found', ctx) ctx.info(f'media folders:\n{elib.pretty_format(media_folders)}') ctx.media_folders = media_folders
[ 2, 19617, 28, 40477, 12, 23, 198, 37811, 198, 38, 1032, 82, 262, 2056, 24512, 198, 37811, 198, 198, 11748, 1288, 571, 198, 198, 6738, 47540, 22866, 1330, 30532, 628, 198, 4299, 651, 62, 11431, 62, 11379, 364, 7, 49464, 25, 30532, 25...
2.427553
421
from inspect import signature import twister2.tset.TSet as ts from twister2.utils import function_wrapper
[ 6738, 10104, 1330, 9877, 198, 198, 11748, 665, 1694, 17, 13, 912, 316, 13, 4694, 316, 355, 40379, 198, 6738, 665, 1694, 17, 13, 26791, 1330, 2163, 62, 48553, 628 ]
3.6
30
from e2cnn.gspaces import * from e2cnn.nn import FieldType from e2cnn.nn import GeometricTensor from e2cnn.group import Representation from e2cnn.group.representation import build_from_discrete_group_representation from ..equivariant_module import EquivariantModule import torch from typing import List, Tuple, Any import numpy as np import math __all__ = ["ConcatenatedNonLinearity"]
[ 6738, 304, 17, 66, 20471, 13, 70, 2777, 2114, 1330, 1635, 198, 6738, 304, 17, 66, 20471, 13, 20471, 1330, 7663, 6030, 198, 6738, 304, 17, 66, 20471, 13, 20471, 1330, 2269, 16996, 51, 22854, 198, 6738, 304, 17, 66, 20471, 13, 8094, ...
3.128
125
import requests import re import socket from common.colors import bad,que, info, good,run,W,end from common.uriParser import parsing_url as hostd
[ 11748, 7007, 198, 11748, 302, 198, 11748, 17802, 198, 6738, 2219, 13, 4033, 669, 1330, 2089, 11, 4188, 11, 7508, 11, 922, 11, 5143, 11, 54, 11, 437, 198, 6738, 2219, 13, 9900, 46677, 1330, 32096, 62, 6371, 355, 2583, 67 ]
3.536585
41
from key import Key from query import Cursor ''' Hello Tiered Access >>> import pymongo >>> import datastore.core >>> >>> from datastore.impl.mongo import MongoDatastore >>> from datastore.impl.lrucache import LRUCache >>> from datastore.impl.filesystem import FileSystemDatastore >>> >>> conn = pymongo.Connection() >>> mongo = MongoDatastore(conn.test_db) >>> >>> cache = LRUCache(1000) >>> fs = FileSystemDatastore('/tmp/.test_db') >>> >>> ds = datastore.TieredDatastore([cache, mongo, fs]) >>> >>> hello = datastore.Key('hello') >>> ds.put(hello, 'world') >>> ds.contains(hello) True >>> ds.get(hello) 'world' >>> ds.delete(hello) >>> ds.get(hello) None Hello Sharding >>> import datastore.core >>> >>> shards = [datastore.DictDatastore() for i in range(0, 10)] >>> >>> ds = datastore.ShardedDatastore(shards) >>> >>> hello = datastore.Key('hello') >>> ds.put(hello, 'world') >>> ds.contains(hello) True >>> ds.get(hello) 'world' >>> ds.delete(hello) >>> ds.get(hello) None '''
[ 198, 6738, 1994, 1330, 7383, 198, 6738, 12405, 1330, 327, 21471, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 198, 7061, 6, 198, 198, 15496, 3628...
2.317829
516
from django.contrib import admin from django.contrib.admin import widgets from .models import Person, Publication, Position, Talk, Project, Poster, Keyword, News, Banner, Video, Project_header, Photo, Project_umbrella, Project_Role, Sponsor from website.admin_list_filters import PositionRoleListFilter, PositionTitleListFilter, PubVenueTypeListFilter, PubVenueListFilter from sortedm2m_filter_horizontal_widget.forms import SortedFilteredSelectMultiple import django from django import forms from django.http import HttpResponse from datetime import datetime from django.template import loader from django.template import RequestContext from django.shortcuts import redirect from django import forms import urllib import bibtexparser from image_cropping import ImageCroppingMixin # Uses format as per https://github.com/jonasundderwolf/django-image-cropping to add cropping to the admin page #from https://stackoverflow.com/questions/9602217/define-an-order-for-manytomanyfield-with-django #display items inline admin.site.register(Person, PersonAdmin) admin.site.register(Publication, PublicationAdmin) admin.site.register(Talk, TalkAdmin) admin.site.register(Project, ProjectAdmin) admin.site.register(Poster, PosterAdmin) admin.site.register(Keyword) admin.site.register(News, NewsAdmin) admin.site.register(Banner, BannerAdmin) admin.site.register(Video, VideoAdmin) admin.site.register(Photo, PhotoAdmin) admin.site.register(Project_umbrella, ProjectUmbrellaAdmin) admin.site.register(Sponsor) # For modifying more on the front admin landing page, see https://medium.com/django-musings/customizing-the-django-admin-site-b82c7d325510 admin.site.index_title = "Makeability Lab Admin. Django version: " + django.get_version() + " ML Version: 0.5.7a"
[ 6738, 42625, 14208, 13, 3642, 822, 1330, 13169, 198, 6738, 42625, 14208, 13, 3642, 822, 13, 28482, 1330, 40803, 198, 6738, 764, 27530, 1330, 7755, 11, 45065, 11, 23158, 11, 12167, 11, 4935, 11, 44996, 11, 7383, 4775, 11, 3000, 11, 274...
3.439453
512
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jun 26 21:39:04 2019 @author: fmk """ import argparse,subprocess,string,random import pandas as pd ''' positional and optional argument parser''' parser = argparse.ArgumentParser(formatter_class=argparse.RawDescriptionHelpFormatter, description='''\ Infer orthologs across two or more prokka-based annotations, and returns overview table for all genes. Homology is inferred using CD-HIT and annotations need to be in fasta format (nucleotide (*.ffn) or amino acid (*.faa)) CD-HIT: %identity optional. Fixed: -s 0.9, ie. shorter sequences need to be at least 90% length of the representative of the cluster. ''', epilog="Questions or comments? --> fkey@mit.edu") parser.add_argument("-f", dest="file_sample_annotation", help="2-col TSV file with subject-identifier and annotation file path.",type=argparse.FileType('r'),required=True) parser.add_argument('-p', dest="percentIdentity", action="store", default='0.98', help="Percent identity cd-hit. Default: 0.98") parser.add_argument('-o', dest="outpath", action="store", help="Output path.",required=True) parser.add_argument("-c", dest="cdhit", help="Path to CD-HIT executable", action="store",default="cd-hit") args = parser.parse_args() ''' FUNCTIONS''' ''' MAIN ''' # TEST Vars #file_sample_annotation = "/Users/fmk/Documents/mit/stapAD/tmp/pycode/prokka_ffn/subject_4_9_16.list" ##annopath = "/Users/fmk/Documents/mit/stapAD/mlst" ##filetype = "txt" #outpath = "/Users/fmk/Documents/mit/stapAD/tmp/pycode" #percentIdentity=0.95 #cdhit_executable = '/usr/local/bin/cd-hit' if __name__ == "__main__": # assign argparse arguments file_sample_annotation = args.file_sample_annotation # annopath = fix_path(args.annopath) # fix path to annotation has trailing "/" outpath = fix_path(args.outpath) # filetype = args.filetype cdhit_executable = args.cdhit percentIdentity = args.percentIdentity # get concatenated annotation file (output: merged_annotation.fa) and dict[subject]=prokka-tag subj_tag_dict = read_merge_sample_annotation_file(file_sample_annotation) subject_list_ord = list(subj_tag_dict.keys()) prokkaTag_list_ord = [ subj_tag_dict[k] for k in subject_list_ord ] # cd-hit command_cdhit = cdhit_executable + " -s 0.9 -c " + percentIdentity + " -i " + outpath + "merged_annotation.fa" + " -o " + outpath+"cdhit_results" subprocess.run(command_cdhit,shell=True) # read-in cdhit results: dict[SAAB_XXXXX_pidZZZ_YYY]=[geneX,geneY,geneZ] cdhit_res_dict = read_cdhit_cluster(outpath+"cdhit_results.clstr",percentIdentity,prokkaTag_list_ord) # build table of gene annotation cdhit_res_df = pd.DataFrame.from_dict(cdhit_res_dict,orient='index',columns=subject_list_ord) # write cdhit res cdhit_res_df.to_csv(outpath+'annotation_orthologs.tsv',sep="\t")
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 18, 198, 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 37811, 198, 41972, 319, 3300, 7653, 2608, 2310, 25, 2670, 25, 3023, 13130, 198, 198, 31, 9800, 25, 277, 28015, 19...
2.531826
1,194
import os import angr test_location = os.path.join(os.path.dirname(os.path.realpath(__file__)), '..', '..', 'binaries', 'tests') if __name__ == "__main__": test_vtable_extraction_x86_64()
[ 11748, 28686, 201, 198, 11748, 281, 2164, 201, 198, 201, 198, 9288, 62, 24886, 796, 28686, 13, 6978, 13, 22179, 7, 418, 13, 6978, 13, 15908, 3672, 7, 418, 13, 6978, 13, 5305, 6978, 7, 834, 7753, 834, 36911, 705, 492, 3256, 705, 49...
2.277778
90
from zeep import Client if __name__ == "__main__": test2()
[ 6738, 41271, 538, 1330, 20985, 198, 198, 361, 11593, 3672, 834, 6624, 366, 834, 12417, 834, 1298, 198, 220, 220, 220, 1332, 17, 3419 ]
2.625
24
'''The HTTP agent - provides observation using http push to remote agent and expects action in the reply''' import json import time import os import threading import requests from . import BaseAgent from .. import utility from .. import characters
[ 7061, 6, 464, 14626, 5797, 532, 3769, 13432, 1262, 2638, 4574, 284, 6569, 198, 220, 220, 5797, 290, 13423, 2223, 287, 262, 10971, 7061, 6, 198, 11748, 33918, 198, 11748, 640, 198, 11748, 28686, 198, 11748, 4704, 278, 198, 11748, 7007, ...
4.288136
59
# -*- coding: utf-8 -*- from rest_framework import serializers from django_countries.serializer_fields import CountryField from .models import Event, CountryEvent
[ 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 198, 6738, 1334, 62, 30604, 1330, 11389, 11341, 198, 6738, 42625, 14208, 62, 9127, 1678, 13, 46911, 7509, 62, 25747, 1330, 12946, 15878, 198, 198, 6738, 764, 27530, 1330, ...
3.408163
49
#!/usr/bin/env python """ Prints a map of the entire world. """ import os, sys import math from struct import pack # local module try: import nbt except ImportError: # nbt not in search path. Let's see if it can be found in the parent folder extrasearchpath = os.path.realpath(os.path.join(__file__,os.pardir,os.pardir)) if not os.path.exists(os.path.join(extrasearchpath,'nbt')): raise sys.path.append(extrasearchpath) from nbt.region import RegionFile from nbt.chunk import Chunk from nbt.world import WorldFolder,McRegionWorldFolder # PIL module (not build-in) try: from PIL import Image except ImportError: # PIL not in search path. Let's see if it can be found in the parent folder sys.stderr.write("Module PIL/Image not found. Pillow (a PIL fork) can be found at http://python-imaging.github.io/\n") # Note: it may also be possible that PIL is installed, but JPEG support is disabled or broken sys.exit(70) # EX_SOFTWARE # List of blocks to ignore # Uncomment all the lines to show underground structures # TODO: move this list into a separate config file block_ignore = [ 'air', # At least this one # 'cave_air', 'water', 'lava', 'snow', 'ice', # 'grass', 'tall_grass', 'dead_bush', # 'seagrass', 'tall_seagrass', 'kelp', 'kelp_plant', # 'dandelion', 'poppy', 'oxeye_daisy', 'white_tulip', # 'azure_bluet', 'lilac', 'rose_bush', 'peony', 'blue_orchid', # 'lily_pad', 'sugar_cane', 'vine', 'pumpkin', 'cactus', # 'wheat', 'potatoes', 'beetroots', 'carrots', # 'oak_leaves', 'dark_oak_leaves', 'birch_leaves', # 'acacia_leaves', 'spruce_leaves', # 'oak_log', 'dark_oak_log', 'birch_log', # 'acacia_log', 'spruce_log', # 'brown_mushroom', 'red_mushroom', # 'brown_mushroom_block', 'red_mushroom_block', 'mushroom_stem', # 'grass_block', 'grass_path', 'farmland', 'dirt', # 'stone', 'sand', 'gravel', 'clay', # 'sandstone', 'diorite', 'andesite', 'granite', 'obsidian', # 'coal_ore', 'iron_ore', 'gold_ore', 'diamond_ore', # 'redstone_ore', 'lapis_ore', 'emerald_ore', # 'cobweb', ] # Map of block colors from names # Legacy block numeric identifiers are now hidden by Block class # and mapped to alpha identifiers in best effort # TODO: move this map into a separate config file block_colors = { 'acacia_leaves': {'h':114, 's':64, 'l':22 }, 'acacia_log': {'h':35, 's':93, 'l':30 }, 'air': {'h':0, 's':0, 'l':0 }, 'andesite': {'h':0, 's':0, 'l':32 }, 'azure_bluet': {'h':0, 's':0, 'l':100}, 'bedrock': {'h':0, 's':0, 'l':10 }, 'birch_leaves': {'h':114, 's':64, 'l':22 }, 'birch_log': {'h':35, 's':93, 'l':30 }, 'blue_orchid': {'h':0, 's':0, 'l':100}, 'bookshelf': {'h':0, 's':0, 'l':100}, 'brown_mushroom': {'h':0, 's':0, 'l':100}, 'brown_mushroom_block': {'h':0, 's':0, 'l':100}, 'cactus': {'h':126, 's':61, 'l':20 }, 'cave_air': {'h':0, 's':0, 'l':0 }, 'chest': {'h':0, 's':100, 'l':50 }, 'clay': {'h':7, 's':62, 'l':23 }, 'coal_ore': {'h':0, 's':0, 'l':10 }, 'cobblestone': {'h':0, 's':0, 'l':25 }, 'cobblestone_stairs': {'h':0, 's':0, 'l':25 }, 'crafting_table': {'h':0, 's':0, 'l':100}, 'dandelion': {'h':60, 's':100, 'l':60 }, 'dark_oak_leaves': {'h':114, 's':64, 'l':22 }, 'dark_oak_log': {'h':35, 's':93, 'l':30 }, 'dark_oak_planks': {'h':35, 's':93, 'l':30 }, 'dead_bush': {'h':0, 's':0, 'l':100}, 'diorite': {'h':0, 's':0, 'l':32 }, 'dirt': {'h':27, 's':51, 'l':15 }, 'end_portal_frame': {'h':0, 's':100, 'l':50 }, 'farmland': {'h':35, 's':93, 'l':15 }, 'fire': {'h':55, 's':100, 'l':50 }, 'flowing_lava': {'h':16, 's':100, 'l':48 }, 'flowing_water': {'h':228, 's':50, 'l':23 }, 'glass_pane': {'h':0, 's':0, 'l':100}, 'granite': {'h':0, 's':0, 'l':32 }, 'grass': {'h':94, 's':42, 'l':25 }, 'grass_block': {'h':94, 's':42, 'l':32 }, 'gravel': {'h':21, 's':18, 'l':20 }, 'ice': {'h':240, 's':10, 'l':95 }, 'infested_stone': {'h':320, 's':100, 'l':50 }, 'iron_ore': {'h':22, 's':65, 'l':61 }, 'iron_bars': {'h':22, 's':65, 'l':61 }, 'ladder': {'h':35, 's':93, 'l':30 }, 'lava': {'h':16, 's':100, 'l':48 }, 'lilac': {'h':0, 's':0, 'l':100}, 'lily_pad': {'h':114, 's':64, 'l':18 }, 'lit_pumpkin': {'h':24, 's':100, 'l':45 }, 'mossy_cobblestone': {'h':115, 's':30, 'l':50 }, 'mushroom_stem': {'h':0, 's':0, 'l':100}, 'oak_door': {'h':35, 's':93, 'l':30 }, 'oak_fence': {'h':35, 's':93, 'l':30 }, 'oak_fence_gate': {'h':35, 's':93, 'l':30 }, 'oak_leaves': {'h':114, 's':64, 'l':22 }, 'oak_log': {'h':35, 's':93, 'l':30 }, 'oak_planks': {'h':35, 's':93, 'l':30 }, 'oak_pressure_plate': {'h':35, 's':93, 'l':30 }, 'oak_stairs': {'h':114, 's':64, 'l':22 }, 'peony': {'h':0, 's':0, 'l':100}, 'pink_tulip': {'h':0, 's':0, 'l':0 }, 'poppy': {'h':0, 's':100, 'l':50 }, 'pumpkin': {'h':24, 's':100, 'l':45 }, 'rail': {'h':33, 's':81, 'l':50 }, 'red_mushroom': {'h':0, 's':50, 'l':20 }, 'red_mushroom_block': {'h':0, 's':50, 'l':20 }, 'rose_bush': {'h':0, 's':0, 'l':100}, 'sugar_cane': {'h':123, 's':70, 'l':50 }, 'sand': {'h':53, 's':22, 'l':58 }, 'sandstone': {'h':48, 's':31, 'l':40 }, 'seagrass': {'h':94, 's':42, 'l':25 }, 'sign': {'h':114, 's':64, 'l':22 }, 'spruce_leaves': {'h':114, 's':64, 'l':22 }, 'spruce_log': {'h':35, 's':93, 'l':30 }, 'stone': {'h':0, 's':0, 'l':32 }, 'stone_slab': {'h':0, 's':0, 'l':32 }, 'tall_grass': {'h':94, 's':42, 'l':25 }, 'tall_seagrass': {'h':94, 's':42, 'l':25 }, 'torch': {'h':60, 's':100, 'l':50 }, 'snow': {'h':240, 's':10, 'l':85 }, 'spawner': {'h':180, 's':100, 'l':50 }, 'vine': {'h':114, 's':64, 'l':18 }, 'wall_torch': {'h':60, 's':100, 'l':50 }, 'water': {'h':228, 's':50, 'l':23 }, 'wheat': {'h':123, 's':60, 'l':50 }, 'white_wool': {'h':0, 's':0, 'l':100}, } ## Color functions for map generation ## # Hue given in degrees, # saturation and lightness given either in range 0-1 or 0-100 and returned in kind # From http://www.easyrgb.com/index.php?X=MATH&H=19#text19 if __name__ == '__main__': if (len(sys.argv) == 1): print("No world folder specified!") sys.exit(64) # EX_USAGE if sys.argv[1] == '--noshow' and len(sys.argv) > 2: show = False world_folder = sys.argv[2] else: show = True world_folder = sys.argv[1] # clean path name, eliminate trailing slashes. required for os.path.basename() world_folder = os.path.normpath(world_folder) if (not os.path.exists(world_folder)): print("No such folder as "+world_folder) sys.exit(72) # EX_IOERR sys.exit(main(world_folder, show))
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 37811, 198, 18557, 82, 257, 3975, 286, 262, 2104, 995, 13, 198, 37811, 198, 198, 11748, 28686, 11, 25064, 198, 11748, 10688, 198, 6738, 2878, 1330, 2353, 198, 2, 1957, 8265, 198, 28311...
1.769697
4,455
from flaskcbv.url import Url, make_urls from .views import mainView namespases = make_urls( Url('', mainView(), name="main"), )
[ 6738, 42903, 21101, 85, 13, 6371, 1330, 8799, 75, 11, 787, 62, 6371, 82, 198, 6738, 764, 33571, 1330, 1388, 7680, 198, 198, 14933, 79, 1386, 796, 787, 62, 6371, 82, 7, 198, 220, 220, 220, 8799, 75, 10786, 3256, 1388, 7680, 22784, ...
2.666667
51
# Rest framework from rest_framework import permissions
[ 2, 8324, 9355, 198, 6738, 1334, 62, 30604, 1330, 21627, 628 ]
5.181818
11
""" femagtools.vtu ~~~~~~~~~~~~~~ Read FEMAG vtu files """ import vtk import pathlib import numpy as np def read(filename): """ Read vtu file and return Reader object. Arguments: filename: name of vtu file to be read """ return Reader(filename)
[ 37811, 198, 220, 220, 220, 2796, 363, 31391, 13, 85, 28047, 198, 220, 220, 220, 220, 15116, 8728, 4907, 628, 220, 220, 220, 4149, 376, 3620, 4760, 410, 28047, 3696, 198, 37811, 198, 11748, 410, 30488, 198, 11748, 3108, 8019, 198, 1174...
2.589286
112
#!/usr/bin/python #-------------------------------IMPORT--------------------------------# from lib import * #-------------------------------EXPORT--------------------------------# __all__ = ['<#PREFIX#>_app','<#PREFIX#>_index']
[ 2, 48443, 14629, 14, 8800, 14, 29412, 628, 198, 2, 1783, 24305, 3955, 15490, 3880, 2, 198, 6738, 9195, 1330, 1635, 628, 198, 2, 1783, 24305, 6369, 15490, 3880, 2, 198, 834, 439, 834, 796, 37250, 27, 2, 47, 31688, 10426, 2, 29, 62,...
4.070175
57
from collections import Counter import numpy as np
[ 6738, 17268, 1330, 15034, 198, 11748, 299, 32152, 355, 45941, 628, 628, 198 ]
4.230769
13
#!/usr/bin/python ''' Tests for MPS and MPO ''' from numpy import * import matplotlib.pyplot as plt from numpy.testing import dec, assert_, assert_raises, assert_almost_equal, assert_allclose import pdb from ..sweep import * if __name__ == '__main__': test_iterator()
[ 2, 48443, 14629, 14, 8800, 14, 29412, 198, 7061, 6, 198, 51, 3558, 329, 337, 3705, 290, 4904, 46, 198, 7061, 6, 198, 6738, 299, 32152, 1330, 1635, 198, 11748, 2603, 29487, 8019, 13, 9078, 29487, 355, 458, 83, 198, 6738, 299, 32152, ...
2.76
100
from threading import Lock from time import time from ui import Menu from ui.utils import clamp, check_value_lock, to_be_foreground
[ 6738, 4704, 278, 1330, 13656, 198, 6738, 640, 1330, 640, 198, 198, 6738, 334, 72, 1330, 21860, 198, 6738, 334, 72, 13, 26791, 1330, 29405, 11, 2198, 62, 8367, 62, 5354, 11, 284, 62, 1350, 62, 754, 2833, 628 ]
3.435897
39
# coding=utf-8 """Russian Peasant Multiplication algorithm Python implementation.""" if __name__ == '__main__': for x in range(10): for y in range(10): print(x, y, x * y, russ_peasant(x, y))
[ 2, 19617, 28, 40477, 12, 23, 198, 37811, 16220, 2631, 8775, 15237, 489, 3299, 11862, 11361, 7822, 526, 15931, 628, 198, 198, 361, 11593, 3672, 834, 6624, 705, 834, 12417, 834, 10354, 198, 220, 220, 220, 329, 2124, 287, 2837, 7, 940, ...
2.395604
91
from urllib.parse import urlparse, parse_qs from io import BytesIO import json
[ 6738, 2956, 297, 571, 13, 29572, 1330, 19016, 29572, 11, 21136, 62, 48382, 198, 6738, 33245, 1330, 2750, 4879, 9399, 198, 11748, 33918, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628, 628 ]
2.941176
34
import math import torch from bisect import bisect_right
[ 11748, 10688, 198, 11748, 28034, 198, 6738, 47457, 478, 1330, 47457, 478, 62, 3506, 628, 628, 628 ]
3.647059
17
# -*- coding: utf-8 -*- from .._globals import IDENTITY from ..connection import ConnectionPool from .base import BaseAdapter
[ 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 198, 6738, 11485, 62, 4743, 672, 874, 1330, 4522, 3525, 9050, 198, 6738, 11485, 38659, 1330, 26923, 27201, 198, 6738, 764, 8692, 1330, 7308, 47307, 628 ]
3.2
40
import os import platform import unittest import nose from conans import tools from conans.errors import ConanException from conans.model.version import Version from conans import __version__ as client_version from conans.model import settings from conans.test.utils.tools import TestClient from conans.test.assets.visual_project_files import get_vs_project_files
[ 11748, 28686, 198, 11748, 3859, 198, 11748, 555, 715, 395, 198, 198, 11748, 9686, 198, 198, 6738, 369, 504, 1330, 4899, 198, 6738, 369, 504, 13, 48277, 1330, 31634, 16922, 198, 6738, 369, 504, 13, 19849, 13, 9641, 1330, 10628, 198, 67...
3.822917
96
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import argparse import json import os from pinliner import __version__ import sys TEMPLATE_FILE = 'importer.template' TEMPLATE_PATTERN = '${CONTENTS}' if __name__ == '__main__': main()
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 6738, 11593, 37443, 834, 1330, 4112, 62, 11748, 198, 198, 11748, 1822, 29572, 198, 11748, 33918, 198, 11748, 28686, 198, ...
2.675926
108
from dotenv import load_dotenv load_dotenv() import os import boto3 #s3 = boto3.resource('s3') s3 = boto3.resource('s3', aws_access_key_id=os.environ.get("AWS_KEY_ID"), aws_secret_access_key=os.environ.get("AWS_SECRET_KEY")) for bucket in s3.buckets.all(): print(bucket.name)
[ 6738, 16605, 24330, 1330, 3440, 62, 26518, 24330, 198, 2220, 62, 26518, 24330, 3419, 198, 198, 11748, 28686, 198, 11748, 275, 2069, 18, 198, 198, 2, 82, 18, 796, 275, 2069, 18, 13, 31092, 10786, 82, 18, 11537, 198, 82, 18, 796, 275,...
2.103448
145
import os from sonar.testbench import Testbench, Module, TestVector, Thread from sonar.interfaces import AXIS from sonar_strToInt import strToInt hold_buffer = Testbench.default('hold_buffer') filepath = os.path.join(os.path.dirname(__file__), 'build/hold_buffer/') dut = Module.default("DUT") dut.add_clock_port('ap_clk', '20ns') dut.add_reset_port('ap_rst_n') dut.add_port('dataRelease_V', 'input', 16) axis_input = AXIS('axis_input', 'slave', 'ap_clk', c_struct='axis_word', c_stream='uaxis_l') axis_input.port.init_channels('tkeep', 64, True) dut.add_interface(axis_input) axis_output = AXIS('axis_output', 'master', 'ap_clk', c_struct='axis_word', c_stream='uaxis_l') axis_output.port.init_channels('tkeep', 64, True) dut.add_interface(axis_output) hold_buffer.add_module(dut) ################################################################################ # Test Vectors ################################################################################ # Initialization thread (added to each test vector to reset everything) initT = Thread() initT.init_signals() initT.wait_negedge('ap_clk') initT.add_delay('40ns') initT.set_signal('ap_rst_n', 1) initT.set_signal('axis_output_tready', 1) #------------------------------------------------------------------------------- # #------------------------------------------------------------------------------- Release_A = TestVector() Release_A.add_thread(initT) rA_t1 = Thread() rA_t1.add_delay('100ns') rA_t1.init_timer() rA_t1.set_signal('dataRelease_V', 1) axis_input.writes(rA_t1, [ {"tdata": 0xDEF, "callTB": 1}, {"tdata": 0xFED, "callTB": 1}, ]) Release_A.add_thread(rA_t1) rA_t2 = Thread() axis_output.read(rA_t2, 0xDEF) axis_output.read(rA_t2, 0xFED) rA_t2.print_elapsed_time("Release_A") rA_t2.end_vector() Release_A.add_thread(rA_t2) #------------------------------------------------------------------------------- # Medium Message A # # #------------------------------------------------------------------------------- Release_B = TestVector() Release_B.add_thread(initT) rB_t1 = Thread() rB_t1.add_delay('100ns') rB_t1.init_timer() axis_input.writes(rB_t1, [ {"tdata": 0xDEF, "callTB": 1}, {"tdata": 0xFED, "callTB": 1}, ]) rB_t1.set_signal('dataRelease_V', 1) Release_B.add_thread(rB_t1) rB_t2 = Thread() axis_output.read(rB_t2, 0xDEF) axis_output.read(rB_t2, 0xFED) rB_t2.print_elapsed_time("Release_B") rB_t2.end_vector() Release_B.add_thread(rB_t2) hold_buffer.add_test_vector(Release_A) hold_buffer.add_test_vector(Release_B) hold_buffer.generateTB(filepath, 'all')
[ 11748, 28686, 198, 198, 6738, 3367, 283, 13, 9288, 26968, 1330, 6208, 26968, 11, 19937, 11, 6208, 38469, 11, 14122, 198, 6738, 3367, 283, 13, 3849, 32186, 1330, 43051, 1797, 198, 198, 6738, 3367, 283, 62, 2536, 2514, 5317, 1330, 965, ...
2.679875
959
#!/usr/bin/env python # filename: pair.py # # Copyright (c) 2015 Bryan Briney # License: The MIT license (http://opensource.org/licenses/MIT) # # Permission is hereby granted, free of charge, to any person obtaining a copy of this software # and associated documentation files (the "Software"), to deal in the Software without restriction, # including without limitation the rights to use, copy, modify, merge, publish, distribute, # sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all copies or # substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING # BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, # DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # import copy import sys import traceback from Bio.Seq import Seq from Bio.Alphabet import generic_dna from abtools import germlines from abtools.alignment import global_alignment from abtools.sequence import Sequence def fasta(self, key='vdj_nt', append_chain=True): ''' Returns the sequence pair as a fasta string. If the Pair object contains both heavy and light chain sequences, both will be returned as a single string. By default, the fasta string contains the 'vdj_nt' sequence for each chain. To change, use the <key> option to select an alternate sequence. By default, the chain (heavy or light) will be appended to the sequence name: >MySequence_heavy To just use the pair name (which will result in duplicate sequence names for Pair objects with both heavy and light chains), set <append_chain> to False. ''' fastas = [] for s, chain in [(self.heavy, 'heavy'), (self.light, 'light')]: if s is not None: c = '_{}'.format(chain) if append_chain else '' fastas.append('>{}{}\n{}'.format(s['seq_id'], c, s[key])) return '\n'.join(fastas) def get_pairs(db, collection, experiment=None, subject=None, group=None, name='seq_id', delim=None, delim_occurance=1, pairs_only=False): ''' Gets sequences and assigns them to the appropriate mAb pair, based on the sequence name. Inputs: ::db:: is a pymongo database connection object ::collection:: is the collection name, as a string If ::subject:: is provided, only sequences with a 'subject' field matching ::subject:: will be included. ::subject:: can be either a single subject (as a string) or an iterable (list or tuple) of subject strings. If ::group:: is provided, only sequences with a 'group' field matching ::group:: will be included. ::group:: can be either a single group (as a string) or an iterable (list or tuple) of group strings. ::name:: is the dict key of the field to be used to group the sequences into pairs. Default is 'seq_id' ::delim:: is an optional delimiter used to truncate the contents of the ::name:: field. Default is None, which results in no name truncation. ::delim_occurance:: is the occurance of the delimiter at which to trim. Trimming is performed as delim.join(name.split(delim)[:delim_occurance]), so setting delim_occurance to -1 will trucate after the last occurance of delim. Default is 1. ::pairs_only:: setting to True results in only truly paired sequences (pair.is_pair == True) will be returned. Default is False. Returns a list of Pair objects, one for each mAb pair. ''' match = {} if subject is not None: if type(subject) in (list, tuple): match['subject'] = {'$in': subject} elif type(subject) in (str, str): match['subject'] = subject if group is not None: if type(group) in (list, tuple): match['group'] = {'$in': group} elif type(group) in (str, str): match['group'] = group if experiment is not None: if type(experiment) in (list, tuple): match['experiment'] = {'$in': experiment} elif type(experiment) in (str, str): match['experiment'] = experiment seqs = list(db[collection].find(match)) return assign_pairs(seqs, name=name, delim=delim, delim_occurance=delim_occurance, pairs_only=pairs_only) def assign_pairs(seqs, name='seq_id', delim=None, delim_occurance=1, pairs_only=False): ''' Assigns sequences to the appropriate mAb pair, based on the sequence name. Inputs: ::seqs:: is a list of dicts, of the format returned by querying a MongoDB containing Abstar output. ::name:: is the dict key of the field to be used to group the sequences into pairs. Default is 'seq_id' ::delim:: is an optional delimiter used to truncate the contents of the ::name:: field. Default is None, which results in no name truncation. ::delim_occurance:: is the occurance of the delimiter at which to trim. Trimming is performed as delim.join(name.split(delim)[:delim_occurance]), so setting delim_occurance to -1 will trucate after the last occurance of delim. Default is 1. ::pairs_only:: setting to True results in only truly paired sequences (pair.is_pair == True) will be returned. Default is False. Returns a list of Pair objects, one for each mAb pair. ''' pdict = {} for s in seqs: if delim is not None: pname = delim.join(s[name].split(delim)[:delim_occurance]) else: pname = s[name] if pname not in pdict: pdict[pname] = [s, ] else: pdict[pname].append(s) pairs = [Pair(pdict[n], name=n) for n in list(pdict.keys())] if pairs_only: pairs = [p for p in pairs if p.is_pair] return pairs def deduplicate(pairs, aa=False, ignore_primer_regions=False): ''' Removes duplicate sequences from a list of Pair objects. If a Pair has heavy and light chains, both chains must identically match heavy and light chains from another Pair to be considered a duplicate. If a Pair has only a single chain, identical matches to that chain will cause the single chain Pair to be considered a duplicate, even if the comparison Pair has both chains. Note that identical sequences are identified by simple string comparison, so sequences of different length that are identical over the entirety of the shorter sequence are not considered duplicates. By default, comparison is made on the nucleotide sequence. To use the amino acid sequence instead, set aa=True. ''' nr_pairs = [] just_pairs = [p for p in pairs if p.is_pair] single_chains = [p for p in pairs if not p.is_pair] _pairs = just_pairs + single_chains for p in _pairs: duplicates = [] for nr in nr_pairs: identical = True vdj = 'vdj_aa' if aa else 'vdj_nt' offset = 4 if aa else 12 if p.heavy is not None: if nr.heavy is None: identical = False else: heavy = p.heavy[vdj][offset:-offset] if ignore_primer_regions else p.heavy[vdj] nr_heavy = nr.heavy[vdj][offset:-offset] if ignore_primer_regions else nr.heavy[vdj] if heavy != nr_heavy: identical = False if p.light is not None: if nr.light is None: identical = False else: light = p.light[vdj][offset:-offset] if ignore_primer_regions else p.light[vdj] nr_light = nr.light[vdj][offset:-offset] if ignore_primer_regions else nr.light[vdj] if light != nr_light: identical = False duplicates.append(identical) if any(duplicates): continue else: nr_pairs.append(p) return nr_pairs def refine(pairs, heavy=True, light=True, species='human'): refined_pairs = copy.deepcopy(pairs) for p in refined_pairs: p.refine(heavy, light, species) return refined_pairs
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 2, 29472, 25, 5166, 13, 9078, 628, 198, 2, 198, 2, 15069, 357, 66, 8, 1853, 17857, 1709, 500, 88, 198, 2, 13789, 25, 383, 17168, 5964, 357, 4023, 1378, 44813, 1668, 13, 2398, 14, ...
2.650201
3,239
import tensorflow as tf from t3f.tensor_train import TensorTrain from t3f.tensor_train_batch import TensorTrainBatch from t3f import shapes from t3f import decompositions def project_sum(what, where, weights=None): """Project sum of `what` TTs on the tangent space of `where` TT. project_sum(what, x) = P_x(what) project_sum(batch_what, x) = P_x(\sum_i batch_what[i]) project_sum(batch_what, x, weights) = P_x(\sum_j weights[j] * batch_what[j]) This function implements the algorithm from the paper [1], theorem 3.1. [1] C. Lubich, I. Oseledets and B. Vandereycken, Time integration of Tensor Trains. Args: what: TensorTrain or TensorTrainBatch. In the case of batch returns projection of the sum of elements in the batch. where: TensorTrain, TT-tensor or TT-matrix on which tangent space to project weights: python list or tf.Tensor of numbers or None, weights of the sum Returns: a TensorTrain with the TT-ranks equal 2 * tangent_space_tens.get_tt_ranks() Complexity: O(d r_where^3 m) for orthogonalizing the TT-cores of where +O(batch_size d r_what r_where n (r_what + r_where)) d is the number of TT-cores (what.ndims()); r_what is the largest TT-rank of what max(what.get_tt_rank()) r_where is the largest TT-rank of where n is the size of the axis dimension of what and where e.g. for a tensor of size 4 x 4 x 4, n is 4; for a 9 x 64 matrix of raw shape (3, 3, 3) x (4, 4, 4) n is 12 """ # Always work with batch of TT objects for simplicity. what = shapes.expand_batch_dim(what) if weights is not None: weights = tf.convert_to_tensor(weights, dtype=where.dtype) if not isinstance(where, TensorTrain): raise ValueError('The first argument should be a TensorTrain object, got ' '"%s".' % where) if where.get_raw_shape() != what.get_raw_shape(): raise ValueError('The shapes of the tensor we want to project and of the ' 'tensor on which tangent space we want to project should ' 'match, got %s and %s.' % (where.get_raw_shape(), what.get_raw_shape())) dtypes_compatible = (where.dtype.is_compatible_with(what.dtype) or what.dtype.is_compatible_with(where.dtype)) if not dtypes_compatible: raise ValueError('Dtypes of the arguments should coincide, got %s and %s.' % (where.dtype, what.dtype)) left_tangent_space_tens = decompositions.orthogonalize_tt_cores( where) right_tangent_space_tens = decompositions.orthogonalize_tt_cores( left_tangent_space_tens, left_to_right=False) ndims = where.ndims() dtype = where.dtype raw_shape = shapes.lazy_raw_shape(where) batch_size = shapes.lazy_batch_size(what) right_tangent_tt_ranks = shapes.lazy_tt_ranks(right_tangent_space_tens) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left_tangent_space_tens) # For einsum notation. mode_str = 'ij' if where.is_tt_matrix() else 'i' right_rank_dim = where.right_tt_rank_dim left_rank_dim = where.left_tt_rank_dim if weights is not None: weights_shape = weights.get_shape() output_is_batch = len(weights_shape) > 1 and weights_shape[1] > 1 else: output_is_batch = False output_batch_str = 'o' if output_is_batch else '' if output_is_batch: right_rank_dim += 1 left_rank_dim += 1 output_batch_size = weights.get_shape()[1].value # Prepare rhs vectors. # rhs[core_idx] is of size # batch_size x tensor_tt_ranks[core_idx] x tangent_tt_ranks[core_idx] rhs = [None] * (ndims + 1) rhs[ndims] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1, 0, -1): tens_core = what.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sa{0}b,sbd,c{0}d->sac'.format(mode_str) rhs[core_idx] = tf.einsum(einsum_str, tens_core, rhs[core_idx + 1], right_tang_core) # Prepare lhs vectors. # lhs[core_idx] is of size # batch_size x tangent_tt_ranks[core_idx] x tensor_tt_ranks[core_idx] lhs = [None] * (ndims + 1) lhs[0] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sab,a{0}c,sb{0}d->scd'.format(mode_str) lhs[core_idx + 1] = tf.einsum(einsum_str, lhs[core_idx], left_tang_core, tens_core) # Left to right sweep. res_cores_list = [] for core_idx in range(ndims): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] if core_idx < ndims - 1: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) einsum_str = 'a{0}b,sbc->sa{0}c'.format(mode_str) proj_core -= tf.einsum(einsum_str, left_tang_core, lhs[core_idx + 1]) if weights is None: einsum_str = 'sa{0}b,sbc->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, proj_core, rhs[core_idx + 1]) else: einsum_str = 'sa{0}b,sbc->sa{0}c'.format(mode_str, output_batch_str) proj_core_s = tf.einsum(einsum_str, proj_core, rhs[core_idx + 1]) einsum_str = 's{1},sa{0}c->{1}a{0}c'.format(mode_str, output_batch_str) proj_core = tf.einsum(einsum_str, weights, proj_core_s) if core_idx == ndims - 1: if weights is None: einsum_str = 'sab,sb{0}c->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) else: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str, output_batch_str) proj_core_s = tf.einsum(einsum_str, lhs[core_idx], tens_core) einsum_str = 's{1},sa{0}c->{1}a{0}c'.format(mode_str, output_batch_str) proj_core = tf.einsum(einsum_str, weights, proj_core_s) if output_is_batch: # Add batch dimension of size output_batch_size to left_tang_core and # right_tang_core extended_left_tang_core = tf.expand_dims(left_tang_core, 0) extended_right_tang_core = tf.expand_dims(right_tang_core, 0) if where.is_tt_matrix(): extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1, 1]) else: extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1]) else: extended_left_tang_core = left_tang_core extended_right_tang_core = right_tang_core if core_idx == 0: res_core = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) elif core_idx == ndims - 1: res_core = tf.concat((extended_right_tang_core, proj_core), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[core_idx] rank_2 = left_tangent_tt_ranks[core_idx + 1] if where.is_tt_matrix(): mode_size_n = raw_shape[0][core_idx] mode_size_m = raw_shape[1][core_idx] shape = [rank_1, mode_size_n, mode_size_m, rank_2] else: mode_size = raw_shape[0][core_idx] shape = [rank_1, mode_size, rank_2] if output_is_batch: shape = [output_batch_size] + shape zeros = tf.zeros(shape, dtype) upper = tf.concat((extended_right_tang_core, zeros), axis=right_rank_dim) lower = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) res_core = tf.concat((upper, lower), axis=left_rank_dim) res_cores_list.append(res_core) # TODO: TT-ranks. if output_is_batch: res = TensorTrainBatch(res_cores_list, where.get_raw_shape(), batch_size=output_batch_size) else: res = TensorTrain(res_cores_list, where.get_raw_shape()) res.projection_on = where return res def project(what, where): """Project `what` TTs on the tangent space of `where` TT. project(what, x) = P_x(what) project(batch_what, x) = batch(P_x(batch_what[0]), ..., P_x(batch_what[N])) This function implements the algorithm from the paper [1], theorem 3.1. [1] C. Lubich, I. Oseledets and B. Vandereycken, Time integration of Tensor Trains. Args: what: TensorTrain or TensorTrainBatch. In the case of batch returns batch with projection of each individual tensor. where: TensorTrain, TT-tensor or TT-matrix on which tangent space to project Returns: a TensorTrain with the TT-ranks equal 2 * tangent_space_tens.get_tt_ranks() Complexity: O(d r_where^3 m) for orthogonalizing the TT-cores of where +O(batch_size d r_what r_where n (r_what + r_where)) d is the number of TT-cores (what.ndims()); r_what is the largest TT-rank of what max(what.get_tt_rank()) r_where is the largest TT-rank of where n is the size of the axis dimension of what and where e.g. for a tensor of size 4 x 4 x 4, n is 4; for a 9 x 64 matrix of raw shape (3, 3, 3) x (4, 4, 4) n is 12 """ if not isinstance(where, TensorTrain): raise ValueError('The first argument should be a TensorTrain object, got ' '"%s".' % where) if where.get_raw_shape() != what.get_raw_shape(): raise ValueError('The shapes of the tensor we want to project and of the ' 'tensor on which tangent space we want to project should ' 'match, got %s and %s.' % (where.get_raw_shape(), what.get_raw_shape())) dtypes_compatible = (where.dtype.is_compatible_with(what.dtype) or what.dtype.is_compatible_with(where.dtype)) if not dtypes_compatible: raise ValueError('Dtypes of the arguments should coincide, got %s and %s.' % (where.dtype, what.dtype)) left_tangent_space_tens = decompositions.orthogonalize_tt_cores( where) right_tangent_space_tens = decompositions.orthogonalize_tt_cores( left_tangent_space_tens, left_to_right=False) ndims = where.ndims() dtype = where.dtype raw_shape = shapes.lazy_raw_shape(where) right_tangent_tt_ranks = shapes.lazy_tt_ranks(right_tangent_space_tens) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left_tangent_space_tens) # For einsum notation. mode_str = 'ij' if where.is_tt_matrix() else 'i' right_rank_dim = what.right_tt_rank_dim left_rank_dim = what.left_tt_rank_dim output_is_batch = isinstance(what, TensorTrainBatch) if output_is_batch: output_batch_size = what.batch_size # Always work with batch of TT objects for simplicity. what = shapes.expand_batch_dim(what) batch_size = shapes.lazy_batch_size(what) # Prepare rhs vectors. # rhs[core_idx] is of size # batch_size x tensor_tt_ranks[core_idx] x tangent_tt_ranks[core_idx] rhs = [None] * (ndims + 1) rhs[ndims] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1, 0, -1): tens_core = what.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sa{0}b,sbd,c{0}d->sac'.format(mode_str) rhs[core_idx] = tf.einsum(einsum_str, tens_core, rhs[core_idx + 1], right_tang_core) # Prepare lhs vectors. # lhs[core_idx] is of size # batch_size x tangent_tt_ranks[core_idx] x tensor_tt_ranks[core_idx] lhs = [None] * (ndims + 1) lhs[0] = tf.ones((batch_size, 1, 1), dtype=dtype) for core_idx in range(ndims - 1): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] einsum_str = 'sab,a{0}c,sb{0}d->scd'.format(mode_str) lhs[core_idx + 1] = tf.einsum(einsum_str, lhs[core_idx], left_tang_core, tens_core) # Left to right sweep. res_cores_list = [] for core_idx in range(ndims): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] if core_idx < ndims - 1: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) einsum_str = 'a{0}b,sbc->sa{0}c'.format(mode_str) proj_core -= tf.einsum(einsum_str, left_tang_core, lhs[core_idx + 1]) if output_is_batch: einsum_str = 'sa{0}b,sbc->sa{0}c'.format(mode_str) else: einsum_str = 'sa{0}b,sbc->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, proj_core, rhs[core_idx + 1]) if core_idx == ndims - 1: if output_is_batch: einsum_str = 'sab,sb{0}c->sa{0}c'.format(mode_str) else: einsum_str = 'sab,sb{0}c->a{0}c'.format(mode_str) proj_core = tf.einsum(einsum_str, lhs[core_idx], tens_core) if output_is_batch: # Add batch dimension of size output_batch_size to left_tang_core and # right_tang_core extended_left_tang_core = tf.expand_dims(left_tang_core, 0) extended_right_tang_core = tf.expand_dims(right_tang_core, 0) if where.is_tt_matrix(): extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1, 1]) else: extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1]) else: extended_left_tang_core = left_tang_core extended_right_tang_core = right_tang_core if core_idx == 0: res_core = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) elif core_idx == ndims - 1: res_core = tf.concat((extended_right_tang_core, proj_core), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[core_idx] rank_2 = left_tangent_tt_ranks[core_idx + 1] if where.is_tt_matrix(): mode_size_n = raw_shape[0][core_idx] mode_size_m = raw_shape[1][core_idx] shape = [rank_1, mode_size_n, mode_size_m, rank_2] else: mode_size = raw_shape[0][core_idx] shape = [rank_1, mode_size, rank_2] if output_is_batch: shape = [output_batch_size] + shape zeros = tf.zeros(shape, dtype) upper = tf.concat((extended_right_tang_core, zeros), axis=right_rank_dim) lower = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) res_core = tf.concat((upper, lower), axis=left_rank_dim) res_cores_list.append(res_core) # TODO: TT-ranks. if output_is_batch: res = TensorTrainBatch(res_cores_list, where.get_raw_shape(), batch_size=output_batch_size) else: res = TensorTrain(res_cores_list, where.get_raw_shape()) res.projection_on = where return res def project_matmul(what, where, matrix): """Project `matrix` * `what` TTs on the tangent space of `where` TT. project(what, x) = P_x(what) project(batch_what, x) = batch(P_x(batch_what[0]), ..., P_x(batch_what[N])) This function implements the algorithm from the paper [1], theorem 3.1. [1] C. Lubich, I. Oseledets and B. Vandereycken, Time integration of Tensor Trains. Args: what: TensorTrain or TensorTrainBatch. In the case of batch returns batch with projection of each individual tensor. where: TensorTrain, TT-tensor or TT-matrix on which tangent space to project matrix: TensorTrain, TT-matrix to multiply by what Returns: a TensorTrain with the TT-ranks equal 2 * tangent_space_tens.get_tt_ranks() Complexity: O(d r_where^3 m) for orthogonalizing the TT-cores of where +O(batch_size d R r_what r_where (n r_what + n m R + m r_where)) d is the number of TT-cores (what.ndims()); r_what is the largest TT-rank of what max(what.get_tt_rank()) r_where is the largest TT-rank of where matrix is of TT-rank R and of raw-shape (m, m, ..., m) x (n, n, ..., n). """ if not isinstance(where, TensorTrain): raise ValueError('The first argument should be a TensorTrain object, got ' '"%s".' % where) if where.get_raw_shape() != what.get_raw_shape(): raise ValueError('The shapes of the tensor we want to project and of the ' 'tensor on which tangent space we want to project should ' 'match, got %s and %s.' % (where.get_raw_shape(), what.get_raw_shape())) dtypes_compatible = (where.dtype.is_compatible_with(what.dtype) or what.dtype.is_compatible_with(where.dtype)) if not dtypes_compatible: raise ValueError('Dtypes of the arguments should coincide, got %s and %s.' % (where.dtype, what.dtype)) left_tangent_space_tens = decompositions.orthogonalize_tt_cores( where) right_tangent_space_tens = decompositions.orthogonalize_tt_cores( left_tangent_space_tens, left_to_right=False) ndims = where.ndims() dtype = where.dtype raw_shape = shapes.lazy_raw_shape(where) batch_size = shapes.lazy_batch_size(what) right_tangent_tt_ranks = shapes.lazy_tt_ranks(right_tangent_space_tens) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left_tangent_space_tens) # For einsum notation. right_rank_dim = what.right_tt_rank_dim left_rank_dim = what.left_tt_rank_dim output_is_batch = isinstance(what, TensorTrainBatch) if output_is_batch: output_batch_size = what.batch_size # Always work with batch of TT objects for simplicity. what = shapes.expand_batch_dim(what) # Prepare rhs vectors. # rhs[core_idx] is of size # batch_size x tensor_tt_ranks[core_idx] x matrix_tt_ranks[core_idx] x tangent_tt_ranks[core_idx] rhs = [None] * (ndims + 1) rhs[ndims] = tf.ones((batch_size, 1, 1, 1), dtype=dtype) for core_idx in range(ndims - 1, 0, -1): tens_core = what.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] matrix_core = matrix.tt_cores[core_idx] rhs[core_idx] = tf.einsum('bije,cikf,sdef,sajkd->sabc', matrix_core, right_tang_core, rhs[core_idx + 1], tens_core) # Prepare lhs vectors. # lhs[core_idx] is of size # batch_size x tangent_tt_ranks[core_idx] x matrix_tt_ranks[core_idx] x tensor_tt_ranks[core_idx] lhs = [None] * (ndims + 1) lhs[0] = tf.ones((batch_size, 1, 1, 1), dtype=dtype) for core_idx in range(ndims - 1): tens_core = what.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] matrix_core = matrix.tt_cores[core_idx] # TODO: brutforce order of indices in lhs?? lhs[core_idx + 1] = tf.einsum('bije,aikd,sabc,scjkf->sdef', matrix_core, left_tang_core, lhs[core_idx], tens_core) # Left to right sweep. res_cores_list = [] for core_idx in range(ndims): tens_core = what.tt_cores[core_idx] matrix_core = matrix.tt_cores[core_idx] left_tang_core = left_tangent_space_tens.tt_cores[core_idx] right_tang_core = right_tangent_space_tens.tt_cores[core_idx] if core_idx < ndims - 1: proj_core = tf.einsum('scjke,sabc,bijd->saikde', tens_core, lhs[core_idx], matrix_core) proj_core -= tf.einsum('aikb,sbcd->saikcd', left_tang_core, lhs[core_idx + 1]) proj_core = tf.einsum('saikcb,sbcd->saikd', proj_core, rhs[core_idx + 1]) if core_idx == ndims - 1: # d and e dimensions take 1 value, since its the last rank. # To make the result shape (?, ?, ?, 1), we are summing d and leaving e, # but we could have done the opposite -- sum e and leave d. proj_core = tf.einsum('sabc,bijd,scjke->saike', lhs[core_idx], matrix_core, tens_core) if output_is_batch: # Add batch dimension of size output_batch_size to left_tang_core and # right_tang_core extended_left_tang_core = tf.expand_dims(left_tang_core, 0) extended_right_tang_core = tf.expand_dims(right_tang_core, 0) extended_left_tang_core = tf.tile(extended_left_tang_core, [output_batch_size, 1, 1, 1, 1]) extended_right_tang_core = tf.tile(extended_right_tang_core, [output_batch_size, 1, 1, 1, 1]) else: extended_left_tang_core = left_tang_core extended_right_tang_core = right_tang_core if core_idx == 0: res_core = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) elif core_idx == ndims - 1: res_core = tf.concat((extended_right_tang_core, proj_core), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[core_idx] rank_2 = left_tangent_tt_ranks[core_idx + 1] mode_size_n = raw_shape[0][core_idx] mode_size_m = raw_shape[1][core_idx] shape = [rank_1, mode_size_n, mode_size_m, rank_2] if output_is_batch: shape = [output_batch_size] + shape zeros = tf.zeros(shape, dtype) upper = tf.concat((extended_right_tang_core, zeros), axis=right_rank_dim) lower = tf.concat((proj_core, extended_left_tang_core), axis=right_rank_dim) res_core = tf.concat((upper, lower), axis=left_rank_dim) res_cores_list.append(res_core) # TODO: TT-ranks. if output_is_batch: res = TensorTrainBatch(res_cores_list, where.get_raw_shape(), batch_size=output_batch_size) else: res = TensorTrain(res_cores_list, where.get_raw_shape()) res.projection_on = where return res def pairwise_flat_inner_projected(projected_tt_vectors_1, projected_tt_vectors_2): """Scalar products between two batches of TTs from the same tangent space. res[i, j] = t3f.flat_inner(projected_tt_vectors_1[i], projected_tt_vectors_1[j]). pairwise_flat_inner_projected(projected_tt_vectors_1, projected_tt_vectors_2) is equivalent to pairwise_flat_inner(projected_tt_vectors_1, projected_tt_vectors_2) , but works only on objects from the same tangent space and is much faster than general pairwise_flat_inner. Args: projected_tt_vectors_1: TensorTrainBatch of tensors projected on the same tangent space as projected_tt_vectors_2. projected_tt_vectors_2: TensorTrainBatch. Returns: tf.tensor with the scalar product matrix. Complexity: O(batch_size^2 d r^2 n), where d is the number of TT-cores (projected_tt_vectors_1.ndims()); r is the largest TT-rank max(projected_tt_vectors_1.get_tt_rank()) (i.e. 2 * {the TT-rank of the object we projected vectors onto}. and n is the size of the axis dimension, e.g. for a tensor of size 4 x 4 x 4, n is 4; for a 9 x 64 matrix of raw shape (3, 3, 3) x (4, 4, 4) n is 12. """ if not hasattr(projected_tt_vectors_1, 'projection_on') or \ not hasattr(projected_tt_vectors_2, 'projection_on'): raise ValueError('Both arguments should be projections on the tangent ' 'space of some other TT-object. All projection* functions ' 'leave .projection_on field in the resulting TT-object ' 'which is not present in the arguments you\'ve provided') if projected_tt_vectors_1.projection_on != projected_tt_vectors_2.projection_on: raise ValueError('Both arguments should be projections on the tangent ' 'space of the same TT-object. The provided arguments are ' 'projections on different TT-objects (%s and %s). Or at ' 'least the pointers are different.' % (projected_tt_vectors_1.projection_on, projected_tt_vectors_2.projection_on)) # Always work with batches of objects for simplicity. projected_tt_vectors_1 = shapes.expand_batch_dim(projected_tt_vectors_1) projected_tt_vectors_2 = shapes.expand_batch_dim(projected_tt_vectors_2) ndims = projected_tt_vectors_1.ndims() tt_ranks = shapes.lazy_tt_ranks(projected_tt_vectors_1) if projected_tt_vectors_1.is_tt_matrix(): right_size = tt_ranks[1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[0] curr_core_2 = projected_tt_vectors_2.tt_cores[0] curr_du_1 = curr_core_1[:, :, :, :, :right_size] curr_du_2 = curr_core_2[:, :, :, :, :right_size] res = tf.einsum('paijb,qaijb->pq', curr_du_1, curr_du_2) for core_idx in range(1, ndims): left_size = tt_ranks[core_idx] // 2 right_size = tt_ranks[core_idx + 1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[core_idx] curr_core_2 = projected_tt_vectors_2.tt_cores[core_idx] curr_du_1 = curr_core_1[:, left_size:, :, :, :right_size] curr_du_2 = curr_core_2[:, left_size:, :, :, :right_size] res += tf.einsum('paijb,qaijb->pq', curr_du_1, curr_du_2) left_size = tt_ranks[-2] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[-1] curr_core_2 = projected_tt_vectors_2.tt_cores[-1] curr_du_1 = curr_core_1[:, left_size:, :, :, :] curr_du_2 = curr_core_2[:, left_size:, :, :, :] res += tf.einsum('paijb,qaijb->pq', curr_du_1, curr_du_2) else: # Working with TT-tensor, not TT-matrix. right_size = tt_ranks[1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[0] curr_core_2 = projected_tt_vectors_2.tt_cores[0] curr_du_1 = curr_core_1[:, :, :, :right_size] curr_du_2 = curr_core_2[:, :, :, :right_size] res = tf.einsum('paib,qaib->pq', curr_du_1, curr_du_2) for core_idx in range(1, ndims): left_size = tt_ranks[core_idx] // 2 right_size = tt_ranks[core_idx + 1] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[core_idx] curr_core_2 = projected_tt_vectors_2.tt_cores[core_idx] curr_du_1 = curr_core_1[:, left_size:, :, :right_size] curr_du_2 = curr_core_2[:, left_size:, :, :right_size] res += tf.einsum('paib,qaib->pq', curr_du_1, curr_du_2) left_size = tt_ranks[-2] // 2 curr_core_1 = projected_tt_vectors_1.tt_cores[-1] curr_core_2 = projected_tt_vectors_2.tt_cores[-1] curr_du_1 = curr_core_1[:, left_size:, :, :] curr_du_2 = curr_core_2[:, left_size:, :, :] res += tf.einsum('paib,qaib->pq', curr_du_1, curr_du_2) return res def add_n_projected(tt_objects, coef=None): """Adds all input TT-objects that are projections on the same tangent space. add_projected((a, b)) is equivalent add(a, b) for a and b that are from the same tangent space, but doesn't increase the TT-ranks. Args: tt_objects: a list of TT-objects that are projections on the same tangent space. coef: a list of numbers or anything else convertable to tf.Tensor. If provided, computes weighted sum. The size of this array should be len(tt_objects) x tt_objects[0].batch_size Returns: TT-objects representing the sum of the tt_objects (weighted sum if coef is provided). The TT-rank of the result equals to the TT-ranks of the arguments. """ for tt in tt_objects: if not hasattr(tt, 'projection_on'): raise ValueError('Both arguments should be projections on the tangent ' 'space of some other TT-object. All projection* functions ' 'leave .projection_on field in the resulting TT-object ' 'which is not present in the argument you\'ve provided.') projection_on = tt_objects[0].projection_on for tt in tt_objects[1:]: if tt.projection_on != projection_on: raise ValueError('All tt_objects should be projections on the tangent ' 'space of the same TT-object. The provided arguments are ' 'projections on different TT-objects (%s and %s). Or at ' 'least the pointers are different.' % (tt.projection_on, projection_on)) if coef is not None: coef = tf.convert_to_tensor(coef, dtype=tt_objects[0].dtype) if coef.get_shape().ndims > 1: # In batch case we will need to multiply each core by this coefficients # along the first axis. To do it need to reshape the coefs to match # the TT-cores number of dimensions. some_core = tt_objects[0].tt_cores[0] dim_array = [1] * (some_core.get_shape().ndims + 1) dim_array[0] = coef.get_shape()[0].value dim_array[1] = coef.get_shape()[1].value coef = tf.reshape(coef, dim_array) ndims = tt_objects[0].ndims() tt_ranks = shapes.lazy_tt_ranks(tt_objects[0]) left_rank_dim = tt_objects[0].left_tt_rank_dim right_rank_dim = tt_objects[0].right_tt_rank_dim res_cores = [] right_half_rank = tt_ranks[1] // 2 left_chunks = [] for obj_idx, tt in enumerate(tt_objects): curr_core = slice_tt_core(tt.tt_cores[0], slice(None), slice(0, right_half_rank)) if coef is not None: curr_core *= coef[obj_idx] left_chunks.append(curr_core) left_part = tf.add_n(left_chunks) first_obj_core = tt_objects[0].tt_cores[0] right_part = slice_tt_core(first_obj_core, slice(None), slice(right_half_rank, None)) first_core = tf.concat((left_part, right_part), axis=right_rank_dim) res_cores.append(first_core) for core_idx in range(1, ndims - 1): first_obj_core = tt_objects[0].tt_cores[core_idx] left_half_rank = tt_ranks[core_idx] // 2 right_half_rank = tt_ranks[core_idx + 1] // 2 upper_part = slice_tt_core(tt.tt_cores[core_idx], slice(0, left_half_rank), slice(None)) lower_right_part = slice_tt_core(first_obj_core, slice(left_half_rank, None), slice(right_half_rank, None)) lower_left_chunks = [] for obj_idx, tt in enumerate(tt_objects): curr_core = slice_tt_core(tt.tt_cores[core_idx], slice(left_half_rank, None), slice(0, right_half_rank)) if coef is not None: curr_core *= coef[obj_idx] lower_left_chunks.append(curr_core) lower_left_part = tf.add_n(lower_left_chunks) lower_part = tf.concat((lower_left_part, lower_right_part), axis=right_rank_dim) curr_core = tf.concat((upper_part, lower_part), axis=left_rank_dim) res_cores.append(curr_core) left_half_rank = tt_ranks[ndims - 1] // 2 upper_part = slice_tt_core(tt.tt_cores[-1], slice(0, left_half_rank), slice(None)) lower_chunks = [] for obj_idx, tt in enumerate(tt_objects): curr_core = slice_tt_core(tt.tt_cores[-1], slice(left_half_rank, None), slice(None)) if coef is not None: curr_core *= coef[obj_idx] lower_chunks.append(curr_core) lower_part = tf.add_n(lower_chunks) last_core = tf.concat((upper_part, lower_part), axis=left_rank_dim) res_cores.append(last_core) raw_shape = tt_objects[0].get_raw_shape() static_tt_ranks = tt_objects[0].get_tt_ranks() if isinstance(tt_objects[0], TensorTrain): res = TensorTrain(res_cores, raw_shape, static_tt_ranks) elif isinstance(tt_objects[0], TensorTrainBatch): res = TensorTrainBatch(res_cores, raw_shape, static_tt_ranks, tt_objects[0].batch_size) # Maintain the projection_on property. res.projection_on = tt_objects[0].projection_on return res def tangent_space_to_deltas(tt, name='t3f_tangent_space_to_deltas'): """Convert an element of the tangent space to deltas representation. Tangent space elements (outputs of t3f.project) look like: dP1 V2 ... Vd + U1 dP2 V3 ... Vd + ... + U1 ... Ud-1 dPd. This function takes as input an element of the tangent space and converts it to the list of deltas [dP1, ..., dPd]. Args: tt: `TensorTrain` or `TensorTrainBatch` that is a result of t3f.project, t3f.project_matmul, or other similar functions. name: string, name of the Op. Returns: A list of delta-cores (tf.Tensors). """ if not hasattr(tt, 'projection_on') or tt.projection_on is None: raise ValueError('tt argument is supposed to be a projection, but it ' 'lacks projection_on field') num_dims = tt.ndims() left_tt_rank_dim = tt.left_tt_rank_dim right_tt_rank_dim = tt.right_tt_rank_dim deltas = [None] * num_dims tt_ranks = shapes.lazy_tt_ranks(tt) for i in range(1, num_dims - 1): if int(tt_ranks[i] / 2) != tt_ranks[i] / 2: raise ValueError('tt argument is supposed to be a projection, but its ' 'ranks are not even.') with tf.compat.v1.name_scope(name, values=tt.tt_cores): for i in range(1, num_dims - 1): r1, r2 = tt_ranks[i], tt_ranks[i + 1] curr_core = tt.tt_cores[i] slc = [slice(None)] * len(curr_core.shape) slc[left_tt_rank_dim] = slice(int(r1 / 2), None) slc[right_tt_rank_dim] = slice(0, int(r2 / 2)) deltas[i] = curr_core[slc] slc = [slice(None)] * len(tt.tt_cores[0].shape) slc[right_tt_rank_dim] = slice(0, int(tt_ranks[1] / 2)) deltas[0] = tt.tt_cores[0][slc] slc = [slice(None)] * len(tt.tt_cores[0].shape) slc[left_tt_rank_dim] = slice(int(tt_ranks[-2] / 2), None) deltas[num_dims - 1] = tt.tt_cores[num_dims - 1][slc] return deltas def deltas_to_tangent_space(deltas, tt, left=None, right=None, name='t3f_deltas_to_tangent_space'): """Converts deltas representation of tangent space vector to TT object. Takes as input a list of [dP1, ..., dPd] and returns dP1 V2 ... Vd + U1 dP2 V3 ... Vd + ... + U1 ... Ud-1 dPd. This function is hard to use correctly because deltas should abey the so called gauge conditions. If the don't, the function will silently return incorrect result. This is why this function is not imported in __init__. Args: deltas: a list of deltas (essentially TT-cores) obeying the gauge conditions. tt: `TensorTrain` object on which the tangent space tensor represented by delta is projected. left: t3f.orthogonilize_tt_cores(tt). If you have it already compute, you may pass it as argument to avoid recomputing. right: t3f.orthogonilize_tt_cores(left, left_to_right=False). If you have it already compute, you may pass it as argument to avoid recomputing. name: string, name of the Op. Returns: `TensorTrain` object constructed from deltas, that is from the tangent space at point `tt`. """ cores = [] dtype = tt.dtype num_dims = tt.ndims() # TODO: add cache instead of mannually pasisng precomputed stuff? input_tensors = list(tt.tt_cores) + list(deltas) if left is not None: input_tensors += list(left.tt_cores) if right is not None: input_tensors += list(right.tt_cores) with tf.compat.v1.name_scope(name, values=input_tensors): if left is None: left = decompositions.orthogonalize_tt_cores(tt) if right is None: right = decompositions.orthogonalize_tt_cores(left, left_to_right=False) left_tangent_tt_ranks = shapes.lazy_tt_ranks(left) right_tangent_tt_ranks = shapes.lazy_tt_ranks(left) raw_shape = shapes.lazy_raw_shape(left) right_rank_dim = left.right_tt_rank_dim left_rank_dim = left.left_tt_rank_dim is_batch_case = len(deltas[0].shape) > len(tt.tt_cores[0].shape) if is_batch_case: right_rank_dim += 1 left_rank_dim += 1 batch_size = deltas[0].shape.as_list()[0] for i in range(num_dims): left_tt_core = left.tt_cores[i] right_tt_core = right.tt_cores[i] if is_batch_case: tile = [1] * len(left_tt_core.shape) tile = [batch_size] + tile left_tt_core = tf.tile(left_tt_core[None, ...], tile) right_tt_core = tf.tile(right_tt_core[None, ...], tile) if i == 0: tangent_core = tf.concat((deltas[i], left_tt_core), axis=right_rank_dim) elif i == num_dims - 1: tangent_core = tf.concat((right_tt_core, deltas[i]), axis=left_rank_dim) else: rank_1 = right_tangent_tt_ranks[i] rank_2 = left_tangent_tt_ranks[i + 1] if tt.is_tt_matrix(): mode_size_n = raw_shape[0][i] mode_size_m = raw_shape[1][i] shape = [rank_1, mode_size_n, mode_size_m, rank_2] else: mode_size_n = raw_shape[0][i] shape = [rank_1, mode_size_n, rank_2] if is_batch_case: shape = [batch_size] + shape zeros = tf.zeros(shape, dtype=dtype) upper = tf.concat((right_tt_core, zeros), axis=right_rank_dim) lower = tf.concat((deltas[i], left_tt_core), axis=right_rank_dim) tangent_core = tf.concat((upper, lower), axis=left_rank_dim) cores.append(tangent_core) if is_batch_case: tangent = TensorTrainBatch(cores, batch_size=batch_size) else: tangent = TensorTrain(cores) tangent.projection_on = tt return tangent
[ 11748, 11192, 273, 11125, 355, 48700, 198, 198, 6738, 256, 18, 69, 13, 83, 22854, 62, 27432, 1330, 309, 22854, 44077, 198, 6738, 256, 18, 69, 13, 83, 22854, 62, 27432, 62, 43501, 1330, 309, 22854, 44077, 33, 963, 198, 6738, 256, 18,...
2.122114
17,844
from django.contrib import admin from . import models
[ 6738, 42625, 14208, 13, 3642, 822, 1330, 13169, 198, 198, 6738, 764, 1330, 4981, 628 ]
3.733333
15
import logging from iptv_proxy.providers.beast.constants import BeastConstants from iptv_proxy.providers.iptv_provider.json_api import ProviderConfigurationJSONAPI logger = logging.getLogger(__name__)
[ 11748, 18931, 198, 198, 6738, 1312, 457, 85, 62, 36436, 13, 15234, 4157, 13, 1350, 459, 13, 9979, 1187, 1330, 10984, 34184, 1187, 198, 6738, 1312, 457, 85, 62, 36436, 13, 15234, 4157, 13, 10257, 85, 62, 15234, 1304, 13, 17752, 62, 1...
3.1875
64
#!/usr/bin/env python # Copyright 2015 Wieger Wesselink. # Distributed under the Boost Software License, Version 1.0. # (See accompanying file LICENSE_1_0.txt or http://www.boost.org/LICENSE_1_0.txt) import os import os.path import random import re import sys import traceback sys.path += [os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'python'))] import random_state_formula_generator from random_bes_generator import make_bes from random_pbes_generator import make_pbes import random_process_expression from testing import YmlTest from text_utility import write_text MCRL2_ROOT = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', '..')) MCRL2_INSTALL_DIR = os.path.join(MCRL2_ROOT, 'install', 'bin') # generates stochastic random processes # generates random process with higher probability of tau transitions # N.B. does not work yet due to unusable abstraction map # N.B. This test has been disabled, since the tool has been deprecated. # N.B does not work due to unknown expressions (F_or) available_tests = { 'alphabet-reduce' : lambda name, settings: AlphabetReduceTest(name, settings) , 'lpssuminst' : lambda name, settings: LpsSuminstTest(name, settings) , 'lpssumelm' : lambda name, settings: LpsSumelmTest(name, settings) , 'lpsparelm' : lambda name, settings: LpsParelmTest(name, settings) , 'lps-quantifier-one-point' : lambda name, settings: LpsOnePointRuleRewriteTest(name, settings) , 'lpsconfcheck-commutative' : lambda name, settings: LpsConfcheckTest(name, 'commutative', settings) , 'lpsconfcheck-commutative-disjoint' : lambda name, settings: LpsConfcheckTest(name, 'commutative-disjoint', settings) , 'lpsconfcheck-disjoint' : lambda name, settings: LpsConfcheckTest(name, 'disjoint', settings) , 'lpsconfcheck-triangular' : lambda name, settings: LpsConfcheckTest(name, 'triangular', settings) , 'lpsconfcheck-trivial' : lambda name, settings: LpsConfcheckTest(name, 'trivial', settings) , 'lpsconstelm' : lambda name, settings: LpsConstelmTest(name, settings) , 'lpsbinary' : lambda name, settings: LpsBinaryTest(name, settings) , 'lps2lts-algorithms' : lambda name, settings: Lps2ltsAlgorithmsTest(name, settings) , 'lps2pbes' : lambda name, settings: Lps2pbesTest(name, settings) , 'lpsstategraph' : lambda name, settings: LpsstategraphTest(name, settings) , 'lts2pbes' : lambda name, settings: Lts2pbesTest(name, settings) , 'ltscompare-bisim' : lambda name, settings: LtscompareTest(name, 'bisim', settings) , 'ltscompare-bisim-gv' : lambda name, settings: LtscompareTest(name, 'bisim-gv', settings) , 'ltscompare-bisim-gjkw' : lambda name, settings: LtscompareTest(name, 'bisim-gjkw', settings) , 'ltscompare-branching-bisim' : lambda name, settings: LtscompareTest(name, 'branching-bisim', settings) , 'ltscompare-branching-bisim-gv' : lambda name, settings: LtscompareTest(name, 'branching-bisim-gv', settings) , 'ltscompare-branching-bisim-gjkw' : lambda name, settings: LtscompareTest(name, 'branching-bisim-gjkw', settings) , 'ltscompare-dpbranching-bisim' : lambda name, settings: LtscompareTest(name, 'dpbranching-bisim', settings) , 'ltscompare-dpbranching-bisim-gv' : lambda name, settings: LtscompareTest(name, 'dpbranching-bisim-gv', settings) , 'ltscompare-dpbranching-bisim-gjkw' : lambda name, settings: LtscompareTest(name, 'dpbranching-bisim-gjkw', settings) , 'ltscompare-weak-bisim' : lambda name, settings: LtscompareTest(name, 'weak-bisim', settings) , 'ltscompare-dpweak-bisim' : lambda name, settings: LtscompareTest(name, 'dpweak-bisim', settings) , 'ltscompare-sim' : lambda name, settings: LtscompareTest(name, 'sim', settings) , 'ltscompare-ready-sim' : lambda name, settings: LtscompareTest(name, 'ready-sim', settings) , 'ltscompare-trace' : lambda name, settings: LtscompareTest(name, 'trace', settings) , 'ltscompare-weak-trace' : lambda name, settings: LtscompareTest(name, 'weak-trace', settings) , 'bisimulation-bisim' : lambda name, settings: BisimulationTest(name, 'bisim', settings) , 'bisimulation-bisim-gv' : lambda name, settings: BisimulationTest(name, 'bisim-gv', settings) , 'bisimulation-bisim-gjkw' : lambda name, settings: BisimulationTest(name, 'bisim-gjkw', settings) , 'bisimulation-branching-bisim' : lambda name, settings: BisimulationTest(name, 'branching-bisim', settings) , 'bisimulation-branching-bisim-gv' : lambda name, settings: BisimulationTest(name, 'branching-bisim-gv', settings) , 'bisimulation-branching-bisim-gjkw' : lambda name, settings: BisimulationTest(name, 'branching-bisim-gjkw', settings) , 'bisimulation-weak-bisim' : lambda name, settings: BisimulationTest(name, 'weak-bisim', settings) , 'pbesconstelm' : lambda name, settings: PbesconstelmTest(name, settings) , 'pbesparelm' : lambda name, settings: PbesparelmTest(name, settings) , 'pbespareqelm' : lambda name, settings: PbespareqelmTest(name, settings) , 'pbespor2' : lambda name, settings: Pbespor2Test(name, settings) , 'pbesrewr-simplify' : lambda name, settings: PbesrewrTest(name, 'simplify', settings) , 'pbesrewr-pfnf' : lambda name, settings: PbesrewrTest(name, 'pfnf', settings) , 'pbesrewr-quantifier-all' : lambda name, settings: PbesrewrTest(name, 'quantifier-all', settings) , 'pbesrewr-quantifier-finite' : lambda name, settings: PbesrewrTest(name, 'quantifier-finite', settings) , 'pbesrewr-quantifier-inside' : lambda name, settings: PbesrewrTest(name, 'quantifier-inside', settings) , 'pbesrewr-quantifier-one-point' : lambda name, settings: PbesrewrTest(name, 'quantifier-one-point', settings) , 'pbesrewr-data-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-data-rewriter', settings) , 'pbesrewr-simplify-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-rewriter', settings) , 'pbesrewr-simplify-data-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-data-rewriter', settings) , 'pbesrewr-simplify-quantifiers-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-quantifiers-rewriter', settings) , 'pbesrewr-simplify-quantifiers-data-rewriter' : lambda name, settings: PbestransformTest(name, 'pbes-simplify-quantifiers-data-rewriter', settings), 'pbesinst-lazy' : lambda name, settings: PbesinstTest(name, ['-slazy'], settings) , 'pbesinst-alternative_lazy' : lambda name, settings: PbesinstTest(name, ['-salternative-lazy'], settings) , 'pbesinst-finite' : lambda name, settings: PbesinstTest(name, ['-sfinite', '-f*(*:Bool)'], settings) , 'pbespgsolve' : lambda name, settings: PbespgsolveTest(name, settings) , 'pbessolve' : lambda name, settings: Pbes2boolTest(name, settings) , 'pbessolve-depth-first' : lambda name, settings: Pbes2boolDepthFirstTest(name, settings) , 'pbessolve-counter-example-optimization-0' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 0, settings) , 'pbessolve-counter-example-optimization-1' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 1, settings) , 'pbessolve-counter-example-optimization-2' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 2, settings) , 'pbessolve-counter-example-optimization-3' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 3, settings) , 'pbessolve-counter-example-optimization-4' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 4, settings) , 'pbessolve-counter-example-optimization-5' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 5, settings) , 'pbessolve-counter-example-optimization-6' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 6, settings) , 'pbessolve-counter-example-optimization-7' : lambda name, settings: Pbes2bool_counter_exampleTest(name, 7, settings) , 'pbesstategraph' : lambda name, settings: PbesstategraphTest(name, settings) , 'pbes-unify-parameters' : lambda name, settings: Pbes_unify_parametersTest(name, settings) , 'pbes-srf' : lambda name, settings: Pbes_srfTest(name, settings) , # 'pbessymbolicbisim' : lambda name, settings: PbessymbolicbisimTest(name, settings) , # excluded from the tests because of Z3 dependency 'bessolve' : lambda name, settings: BessolveTest(name, settings) , #'stochastic-ltscompare' : lambda name, settings: StochasticLtscompareTest(name, settings) , } # These test do not work on Windows due to dependencies. if os.name != 'nt': available_tests.update({'pbessolvesymbolic' : lambda name, settings: PbessolvesymbolicTest(name, settings) }) # available_tests.update({ 'pbesbddsolve' : lambda name, settings: PbesbddsolveTest(name, settings) }) # Return all tests that match with pattern. In case of an exact match, only this exact match is returned. if __name__ == '__main__': main(available_tests)
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 198, 2, 15069, 1853, 370, 494, 1362, 370, 7878, 676, 13, 198, 2, 4307, 6169, 739, 262, 19835, 10442, 13789, 11, 10628, 352, 13, 15, 13, 198, 2, 357, 6214, 19249, 2393, 38559, 24290, ...
1.810519
7,035
""" Main program to launch proc/hdfs.py """ import argparse import logging from pars import addargs import sys import logging logging.basicConfig(format='%(levelname)s:%(message)s', level=logging.INFO) from proc.hdfs import DIRHDFS if __name__ == "__main__": # execute only if run as a script main()
[ 37811, 8774, 1430, 284, 4219, 13834, 14, 71, 7568, 82, 13, 9078, 198, 37811, 198, 11748, 1822, 29572, 198, 11748, 18931, 198, 6738, 13544, 1330, 751, 22046, 198, 198, 11748, 25064, 198, 11748, 18931, 198, 6404, 2667, 13, 35487, 16934, 7...
2.888889
108
"""JSON (de)serialization framework. The framework presented here is somewhat based on `Go's "json" package`_ (especially the ``omitempty`` functionality). .. _`Go's "json" package`: http://golang.org/pkg/encoding/json/ """ import abc import binascii import logging import OpenSSL import six from josepy import b64, errors, interfaces, util logger = logging.getLogger(__name__) def encode_b64jose(data): """Encode JOSE Base-64 field. :param bytes data: :rtype: `unicode` """ # b64encode produces ASCII characters only return b64.b64encode(data).decode('ascii') def decode_b64jose(data, size=None, minimum=False): """Decode JOSE Base-64 field. :param unicode data: :param int size: Required length (after decoding). :param bool minimum: If ``True``, then `size` will be treated as minimum required length, as opposed to exact equality. :rtype: bytes """ error_cls = TypeError if six.PY2 else binascii.Error try: decoded = b64.b64decode(data.encode()) except error_cls as error: raise errors.DeserializationError(error) if size is not None and ((not minimum and len(decoded) != size) or (minimum and len(decoded) < size)): raise errors.DeserializationError( "Expected at least or exactly {0} bytes".format(size)) return decoded def encode_hex16(value): """Hexlify. :param bytes value: :rtype: unicode """ return binascii.hexlify(value).decode() def decode_hex16(value, size=None, minimum=False): """Decode hexlified field. :param unicode value: :param int size: Required length (after decoding). :param bool minimum: If ``True``, then `size` will be treated as minimum required length, as opposed to exact equality. :rtype: bytes """ value = value.encode() if size is not None and ((not minimum and len(value) != size * 2) or (minimum and len(value) < size * 2)): raise errors.DeserializationError() error_cls = TypeError if six.PY2 else binascii.Error try: return binascii.unhexlify(value) except error_cls as error: raise errors.DeserializationError(error) def encode_cert(cert): """Encode certificate as JOSE Base-64 DER. :type cert: `OpenSSL.crypto.X509` wrapped in `.ComparableX509` :rtype: unicode """ return encode_b64jose(OpenSSL.crypto.dump_certificate( OpenSSL.crypto.FILETYPE_ASN1, cert.wrapped)) def decode_cert(b64der): """Decode JOSE Base-64 DER-encoded certificate. :param unicode b64der: :rtype: `OpenSSL.crypto.X509` wrapped in `.ComparableX509` """ try: return util.ComparableX509(OpenSSL.crypto.load_certificate( OpenSSL.crypto.FILETYPE_ASN1, decode_b64jose(b64der))) except OpenSSL.crypto.Error as error: raise errors.DeserializationError(error) def encode_csr(csr): """Encode CSR as JOSE Base-64 DER. :type csr: `OpenSSL.crypto.X509Req` wrapped in `.ComparableX509` :rtype: unicode """ return encode_b64jose(OpenSSL.crypto.dump_certificate_request( OpenSSL.crypto.FILETYPE_ASN1, csr.wrapped)) def decode_csr(b64der): """Decode JOSE Base-64 DER-encoded CSR. :param unicode b64der: :rtype: `OpenSSL.crypto.X509Req` wrapped in `.ComparableX509` """ try: return util.ComparableX509(OpenSSL.crypto.load_certificate_request( OpenSSL.crypto.FILETYPE_ASN1, decode_b64jose(b64der))) except OpenSSL.crypto.Error as error: raise errors.DeserializationError(error)
[ 37811, 40386, 357, 2934, 8, 46911, 1634, 9355, 13, 198, 198, 464, 9355, 5545, 994, 318, 6454, 1912, 319, 4600, 5247, 338, 366, 17752, 1, 5301, 63, 62, 198, 7, 16480, 262, 7559, 296, 270, 28920, 15506, 11244, 737, 198, 198, 492, 4808...
2.478912
1,470
from norm.models.norm import Status, Lambda from norm.executable import NormExecutable from typing import Union, List import logging logger = logging.getLogger(__name__)
[ 6738, 2593, 13, 27530, 13, 27237, 1330, 12678, 11, 21114, 6814, 198, 6738, 2593, 13, 18558, 18187, 1330, 11220, 23002, 18187, 198, 198, 6738, 19720, 1330, 4479, 11, 7343, 198, 11748, 18931, 198, 6404, 1362, 796, 18931, 13, 1136, 11187, ...
3.591837
49
from __future__ import print_function import os import sys import json import tika from tqdm import tqdm from utils import LogUtil from parser import Parser from ioutils import read_lines from tika import parser as tk_parser if __name__ == '__main__': main()
[ 6738, 11593, 37443, 834, 1330, 3601, 62, 8818, 198, 198, 11748, 28686, 198, 11748, 25064, 198, 11748, 33918, 198, 11748, 256, 9232, 198, 6738, 256, 80, 36020, 1330, 256, 80, 36020, 198, 6738, 3384, 4487, 1330, 5972, 18274, 346, 198, 673...
3.127907
86
from checkov.cloudformation.checks.resource.base_resource_check import BaseResourceCheck from checkov.common.parsers.node import DictNode from checkov.common.models.enums import CheckResult, CheckCategories check = DocDBAuditLogs()
[ 6738, 2198, 709, 13, 17721, 1161, 13, 42116, 13, 31092, 13, 8692, 62, 31092, 62, 9122, 1330, 7308, 26198, 9787, 198, 6738, 2198, 709, 13, 11321, 13, 79, 945, 364, 13, 17440, 1330, 360, 713, 19667, 198, 6738, 2198, 709, 13, 11321, 13...
3.455882
68
# # Copyright (c) 2016-2022 Deephaven Data Labs and Patent Pending # import jpy _JCallbackAdapter = jpy.get_type('io.deephaven.server.plugin.python.CallbackAdapter')
[ 2, 198, 2, 15069, 357, 66, 8, 1584, 12, 1238, 1828, 10766, 39487, 6060, 23500, 290, 34830, 350, 1571, 198, 2, 198, 198, 11748, 474, 9078, 198, 198, 62, 41, 47258, 47307, 796, 474, 9078, 13, 1136, 62, 4906, 10786, 952, 13, 22089, 3...
3.054545
55
"""Tests for parse tree pretty printing that preserves formatting Test case descriptions are in file test/data/output.test. """ import os.path import re from typing import Undefined, Any from mypy import build from mypy.myunit import Suite, run_test from mypy.test.helpers import assert_string_arrays_equal from mypy.test.data import parse_test_cases from mypy.test.config import test_data_prefix, test_temp_dir from mypy.parse import parse from mypy.output import OutputVisitor from mypy.errors import CompileError # Files which contain test case descriptions. output_files = ['output.test'] def test_output(testcase): """Perform an identity source code transformation test case.""" expected = testcase.output if expected == []: expected = testcase.input try: src = '\n'.join(testcase.input) # Parse and semantically analyze the source program. # Test case names with a special suffix get semantically analyzed. This # lets us test that semantic analysis does not break source code pretty # printing. if testcase.name.endswith('_SemanticAnalyzer'): result = build.build('main', target=build.SEMANTIC_ANALYSIS, program_text=src, flags=[build.TEST_BUILTINS], alt_lib_path=test_temp_dir) files = result.files else: files = {'main': parse(src, 'main')} a = [] first = True # Produce an output containing the pretty-printed forms (with original # formatting) of all the relevant source files. for fnam in sorted(files.keys()): f = files[fnam] # Omit the builtins and files marked for omission. if (not f.path.endswith(os.sep + 'builtins.py') and '-skip.' not in f.path): # Add file name + colon for files other than the first. if not first: a.append('{}:'.format(fix_path(remove_prefix( f.path, test_temp_dir)))) v = OutputVisitor() f.accept(v) s = v.output() if s != '': a += s.split('\n') first = False except CompileError as e: a = e.messages assert_string_arrays_equal( expected, a, 'Invalid source code output ({}, line {})'.format( testcase.file, testcase.line)) if __name__ == '__main__': import sys run_test(OutputSuite(), sys.argv[1:])
[ 37811, 51, 3558, 329, 21136, 5509, 2495, 13570, 326, 43759, 33313, 198, 198, 14402, 1339, 16969, 389, 287, 2393, 1332, 14, 7890, 14, 22915, 13, 9288, 13, 198, 37811, 198, 198, 11748, 28686, 13, 6978, 198, 11748, 302, 198, 198, 6738, 1...
2.205662
1,201
# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # -------------------------------------------------------------------------- import unittest from sklearn.tree import DecisionTreeClassifier from sklearn.tree import DecisionTreeRegressor from skl2onnx.common.data_types import onnx_built_with_ml from test_utils import ( dump_one_class_classification, dump_binary_classification, dump_multiple_classification, ) from test_utils import dump_multiple_regression, dump_single_regression if __name__ == "__main__": unittest.main()
[ 2, 16529, 45537, 198, 2, 15069, 357, 66, 8, 5413, 10501, 13, 1439, 2489, 10395, 13, 198, 2, 49962, 739, 262, 17168, 13789, 13, 4091, 13789, 13, 14116, 287, 262, 1628, 6808, 329, 198, 2, 5964, 1321, 13, 198, 2, 16529, 35937, 198, 1...
4.033149
181
import logging import os import tempfile import PIL.Image import numpy import tensorflow import microscopeimagequality.constants import microscopeimagequality.data_provider import microscopeimagequality.evaluation import microscopeimagequality.prediction
[ 11748, 18931, 198, 11748, 28686, 198, 11748, 20218, 7753, 198, 198, 11748, 350, 4146, 13, 5159, 198, 11748, 299, 32152, 198, 11748, 11192, 273, 11125, 198, 198, 11748, 36396, 9060, 13237, 13, 9979, 1187, 198, 11748, 36396, 9060, 13237, 13...
4.229508
61
""" QuickBot wiring config. Specifies which pins are used for motor control, IR sensors and wheel encoders. """ # Motor pins: (dir1_pin, dir2_pin, pwd_pin) RIGHT_MOTOR_PINS = 'P8_12', 'P8_10', 'P9_14' LEFT_MOTOR_PINS = 'P8_14', 'P8_16', 'P9_16' # IR sensors (clock-wise, starting with the rear left sensor): # rear-left, front-left, front, front-right, rear-right IR_PINS = ('P9_38', 'P9_40', 'P9_36', 'P9_35', 'P9_33') # Wheel encoder sensors: (left, right) ENC_PINS = ('P9_39', 'P9_37')
[ 37811, 198, 21063, 20630, 29477, 4566, 13, 198, 198, 22882, 6945, 543, 20567, 389, 973, 329, 5584, 1630, 11, 14826, 15736, 290, 7825, 2207, 375, 364, 13, 198, 37811, 198, 198, 2, 12533, 20567, 25, 357, 15908, 16, 62, 11635, 11, 26672,...
2.305556
216
#!/usr/bin/env python # Copyright (c) 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. # Self-test for skimage. import filecmp import os import subprocess import sys import tempfile # Find a path to the binary to use. Iterates through a list of possible # locations the binary may be. def PickBinaryPath(base_dir): POSSIBLE_BINARY_PATHS = [ 'out/Debug/skimage', 'out/Release/skimage', 'xcodebuild/Debug/skimage', 'xcodebuild/Release/skimage', ] for binary in POSSIBLE_BINARY_PATHS: binary_full_path = os.path.join(base_dir, binary) if (os.path.exists(binary_full_path)): return binary_full_path raise BinaryNotFoundException # Quit early if two files have different content. def test_invalid_file(file_dir, skimage_binary): """ Test the return value of skimage when an invalid file is decoded. If there is no expectation file, or the file expects a particular result, skimage should return nonzero indicating failure. If the file has no expectation, or ignore-failure is set to true, skimage should return zero indicating success. """ invalid_file = os.path.join(file_dir, "skimage", "input", "bad-images", "invalid.png") # No expectations file: args = [skimage_binary, "--readPath", invalid_file] result = subprocess.call(args) if 0 == result: print "'%s' should have reported failure!" % " ".join(args) exit(1) # Directory holding all expectations files expectations_dir = os.path.join(file_dir, "skimage", "input", "bad-images") # Expectations file expecting a valid decode: incorrect_expectations = os.path.join(expectations_dir, "incorrect-results.json") args = [skimage_binary, "--readPath", invalid_file, "--readExpectationsPath", incorrect_expectations] result = subprocess.call(args) if 0 == result: print "'%s' should have reported failure!" % " ".join(args) exit(1) # Empty expectations: empty_expectations = os.path.join(expectations_dir, "empty-results.json") output = subprocess.check_output([skimage_binary, "--readPath", invalid_file, "--readExpectationsPath", empty_expectations], stderr=subprocess.STDOUT) if not "Missing" in output: # Another test (in main()) tests to ensure that "Missing" does not appear # in the output. That test could be passed if the output changed so # "Missing" never appears. This ensures that an error is not missed if # that happens. print "skimage output changed! This may cause other self tests to fail!" exit(1) # Ignore failure: ignore_expectations = os.path.join(expectations_dir, "ignore-results.json") output = subprocess.check_output([skimage_binary, "--readPath", invalid_file, "--readExpectationsPath", ignore_expectations], stderr=subprocess.STDOUT) if not "failures" in output: # Another test (in main()) tests to ensure that "failures" does not # appear in the output. That test could be passed if the output changed # so "failures" never appears. This ensures that an error is not missed # if that happens. print "skimage output changed! This may cause other self tests to fail!" exit(1) def test_incorrect_expectations(file_dir, skimage_binary): """ Test that comparing to incorrect expectations fails, unless ignore-failures is set to true. """ valid_file = os.path.join(file_dir, "skimage", "input", "images-with-known-hashes", "1209453360120438698.png") expectations_dir = os.path.join(file_dir, "skimage", "input", "images-with-known-hashes") incorrect_results = os.path.join(expectations_dir, "incorrect-results.json") args = [skimage_binary, "--readPath", valid_file, "--readExpectationsPath", incorrect_results] result = subprocess.call(args) if 0 == result: print "'%s' should have reported failure!" % " ".join(args) exit(1) ignore_results = os.path.join(expectations_dir, "ignore-failures.json") subprocess.check_call([skimage_binary, "--readPath", valid_file, "--readExpectationsPath", ignore_results]) if __name__ == "__main__": main()
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 2, 15069, 357, 66, 8, 2211, 383, 18255, 1505, 46665, 13, 1439, 2489, 10395, 13, 198, 2, 5765, 286, 428, 2723, 2438, 318, 21825, 416, 257, 347, 10305, 12, 7635, 5964, 326, 460, 307, ...
2.408081
1,980
from collections import namedtuple Condition = namedtuple( 'Condition', ['current_state', 'input_character'], )
[ 6738, 17268, 1330, 3706, 83, 29291, 198, 198, 48362, 796, 3706, 83, 29291, 7, 198, 220, 220, 220, 705, 48362, 3256, 198, 220, 220, 220, 37250, 14421, 62, 5219, 3256, 705, 15414, 62, 22769, 6, 4357, 198, 8, 198 ]
3.102564
39
"""Asset definitions for the simple_lakehouse example.""" import pandas as pd from lakehouse import Column, computed_table, source_table from pyarrow import date32, float64, string sfo_q2_weather_sample_table = source_table( path="data", columns=[Column("tmpf", float64()), Column("valid_date", string())], )
[ 37811, 45869, 17336, 329, 262, 2829, 62, 27180, 4803, 1672, 526, 15931, 198, 11748, 19798, 292, 355, 279, 67, 198, 6738, 13546, 4803, 1330, 29201, 11, 29231, 62, 11487, 11, 2723, 62, 11487, 198, 6738, 12972, 6018, 1330, 3128, 2624, 11, ...
3.351064
94
#!/usr/bin/env python # -*- coding: utf-8 -*- # #@created: 07.01.2018 #@author: Aleksey Komissarov #@contact: ad3002@gmail.com from aindex import * settings = { "index_prefix": "tests/kmers.23", "aindex_prefix": "tests/kmers.23", "reads_file": "tests/reads.reads", } index = load_aindex(settings) k = 23 sequence = "TAAGTTATTATTTAGTTAATACTTTTAACAATATTATTAAGGTATTTAAAAAATACTATTATAGTATTTAACATAGTTAAATACCTTCCTTAATACTGTTAAATTATATTCAATCAATACATATATAATATTATTAAAATACTTGATAAGTATTATTTAGATATTAGACAAATACTAATTTTATATTGCTTTAATACTTAATAAATACTACTTATGTATTAAGTAAATATTACTGTAATACTAATAACAATATTATTACAATATGCTAGAATAATATTGCTAGTATCAATAATTACTAATATAGTATTAGGAAAATACCATAATAATATTTCTACATAATACTAAGTTAATACTATGTGTAGAATAATAAATAATCAGATTAAAAAAATTTTATTTATCTGAAACATATTTAATCAATTGAACTGATTATTTTCAGCAGTAATAATTACATATGTACATAGTACATATGTAAAATATCATTAATTTCTGTTATATATAATAGTATCTATTTTAGAGAGTATTAATTATTACTATAATTAAGCATTTATGCTTAATTATAAGCTTTTTATGAACAAAATTATAGACATTTTAGTTCTTATAATAAATAATAGATATTAAAGAAAATAAAAAAATAGAAATAAATATCATAACCCTTGATAACCCAGAAATTAATACTTAATCAAAAATGAAAATATTAATTAATAAAAGTGAATTGAATAAAATTTTGAAAAAAATGAATAACGTTATTATTTCCAATAACAAAATAAAACCACATCATTCATATTTTTTAATAGAGGCAAAAGAAAAAGAAATAAACTTTTATGCTAACAATGAATACTTTTCTGTCAAATGTAATTTAAATAAAAATATTGATATTCTTGAACAAGGCTCCTTAATTGTTAAAGGAAAAATTTTTAACGATCTTATTAATGGCATAAAAGAAGAGATTATTACTATTCAAGAAAAAGATCAAACACTTTTGGTTAAAACAAAAAAAACAAGTATTAATTTAAACACAATTAATGTGAATGAATTTCCAAGAATAAGGTTTAATGAAAAAAACGATTTAAGTGAATTTAATCAATTCAAAATAAATTATTCACTTTTAGTAAAAGGCATTAAAAAAATTTTTCACTCAGTTTCAAATAATCGTGAAATATCTTCTAAATTTAATGGAGTAAATTTCAATGGATCCAATGGAAAAGAAATATTTTTAGAAGCTTCTGACACTTATAAACTATCTGTTTTTGAGATAAAGCAAGAAACAGAACCATTTGATTTCATTTTGGAGAGTAATTTACTTAGTTTCATTAATTCTTTTAATCCTGAAGAAGATAAATCTATTGTTTTTTATTACAGAAAAGATAATAAAGATAGCTTTAGTACAGAAATGTTGATTTCAATGGATAACTTTATGATTAGTTACACATCGGTTAATGAAAAATTTCCAGAGGTAAACTACTTTTTTGAATTTGAACCTGAAACTAAAATAGTTGTTCAAAAAAATGAATTAAAAGATGCACTTCAAAGAATTCAAACTTTGGCTCAAAATGAAAGAACTTTTTTATGCGATATGCAAATTAACAGTTCTGAATTAAAAATAAGAGCTATTGTTAATAATATCGGAAATTCTCTTGAGGAAATTTCTTGTCTTAAATTTGAAGGTTATAAACTTAATATTTCTTTTAACCCAAGTTCTCTATTAGATCACATAGAGTCTTTTGAATCAAATGAAATAAATTTTGATTTCCAAGGAAATAGTAAGTATTTTTTGATAACCTCTAAAAGTGAACCTGAACTTAAGCAAATATTGGTTCCTTCAAGATAATGAATCTTTACGATCTTTTAGAACTACCAACTACAGCATCAATAAAAGAAATAAAAATTGCTTATAAAAGATTAGCAAAGCGTTATCACCCTGATGTAAATAAATTAGGTTCGCAAACTTTTGTTGAAATTAATAATGCTTATTCAATATTAAGTGATCCTAACCAAAAGGAAAAATATGATTCAATGCTGAAAGTTAATGATTTTCAAAATCGCATCAAAAATTTAGATATTAGTGTTAGATGACATGAAAATTTCATGGAAGAACTCGAACTTCGTAAGAACTGAGAATTTGATTTTTTTTCATCTGATGAAGATTTCTTTTATTCTCCATTTACAAAAA" test_kmer = "TAAGTTATTATTTAGTTAATACT" right_kmer = "AGTTAATACTTTTAACAATATTA" print("Task 1. Get kmer frequency") # raw_input("\nReady?") for i in range(len(sequence)-k+1): kmer = sequence[i:i+k] print("Position %s kmer %s freq = %s" % (i, kmer, index[kmer])) print("Task 2. Iter read by read, print the first 20 reads") # raw_input("\nReady?") for i, read in enumerate(index.iter_reads()): if i == 20: break print(i, read) print("Task 3. Iter reads by kmer, returs (start, next_read_start, read, pos_if_uniq|None, all_poses)") # raw_input("\nReady?") for read in iter_reads_by_kmer(test_kmer, index): print(read) print("Task 4. Get distances in reads for two kmers, returns a list of (rid, left_kmer_pos, right_kmer_pos) tuples.") # raw_input("\nReady?") print(get_left_right_distances(test_kmer, right_kmer, index)) print("Task 5. Get layout for kmer, returns (max_pos, reads, lefts, rights, rids, starts), for details see source code") # raw_input("\nReady?") max_pos, reads, lefts, rights, rids, starts = get_layout_for_kmer(right_kmer, index) print("Central layout:") for read in reads: print(read) print("Left flanks:") print(lefts) print("Right flanks:") print(rights) print("Task 6. Iter reads by sequence, returs (start, next_read_start, read, pos_if_uniq|None, all_poses)") # raw_input("\nReady?") sequence = "AATATTATTAAGGTATTTAAAAAATACTATTATAGTATTTAACATA" for read in iter_reads_by_sequence(sequence, index): print(read) print("Task 7. Iter reads by kmer with reads as SE, returns (start, next_read_start, subread, kmere_pos, -1|0|1 for spring_pos, was_reversed, poses_in_read)") # raw_input("\nReady?") user_reads = set() sequence = "AATATTATTAAGGTATTTAAAAAATACTATTATAGTATTTAACATA" for rid, nextrid, read, pos, spring_pos, was_reversed, poses in get_reads_se_by_kmer(kmer, index, user_reads, k=23): print(rid, read, pos)
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 2, 198, 2, 31, 25598, 25, 8753, 13, 486, 13, 7908, 198, 2, 31, 9800, 25, 9300, 74, 4397, 32364, 747, 42737, 198, ...
2.261475
1,939
"""Implements test for pytest-resource-path Fixtures with pytest.ini.""" from pathlib import Path import pytest def test_resource_path_ini(resource_path, request): """Fixture resource_path should be following absolute path.""" assert resource_path == Path(str(request.fspath)).parents[1] / Path( "data/test_package/test_module_something/test_resource_path_ini" ) def test_resource_path_root_ini(resource_path_root, request): """Fixture resource_path_root should be following absolute path.""" assert resource_path_root == Path(str(request.fspath)).parents[1] / Path("data") # Reason: To define fixture in same module. pylint: disable=redefined-outer-name def test_resource_path_root_scope_package_ini(resource_path_root_scope_package_ini, request): assert resource_path_root_scope_package_ini == Path(str(request.fspath)).parents[1] / Path("data")
[ 37811, 3546, 1154, 902, 1332, 329, 12972, 9288, 12, 31092, 12, 6978, 376, 25506, 351, 12972, 9288, 13, 5362, 526, 15931, 198, 6738, 3108, 8019, 1330, 10644, 198, 198, 11748, 12972, 9288, 628, 198, 4299, 1332, 62, 31092, 62, 6978, 62, ...
3.083333
288
# Illustrates combining exception / error handling # with file access print('Start') try: with open('myfile2.txt', 'r') as f: lines = f.readlines() for line in lines: print(line, end='') except FileNotFoundError as err: print('oops') print(err) print('Done')
[ 2, 23279, 9700, 19771, 6631, 1220, 4049, 9041, 198, 2, 351, 2393, 1895, 198, 198, 4798, 10786, 10434, 11537, 198, 28311, 25, 198, 220, 220, 220, 351, 1280, 10786, 1820, 7753, 17, 13, 14116, 3256, 705, 81, 11537, 355, 277, 25, 198, 2...
2.487603
121
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Mon Sep 11 13:30:53 2017 @author: laoj """ import numpy as np import pymc3 as pm import theano.tensor as tt from pymc3.distributions.distribution import Discrete, draw_values, generate_samples, infer_shape from pymc3.distributions.dist_math import bound, logpow, factln, Cholesky from pymc3.math import tround #%% n scaler, p 1D #n = 183 n = np.array([[106], [143], [102], [116], [183], [150]]) p = np.array([[ 0.21245365, 0.41223126, 0.37531509], [ 0.13221011, 0.50537169, 0.3624182 ], [ 0.08813779, 0.54447146, 0.36739075], [ 0.18932804, 0.4630365, 0.34763546], [ 0.11006472, 0.49227755, 0.39765773], [ 0.17886852, 0.41098834, 0.41014314]]) # p = np.array([ 0.21245365, 0.41223126, 0.37531509]) n = tt.as_tensor_variable(n) p = tt.as_tensor_variable(p) n = np.squeeze(n) n = tt.shape_padright(n) if n.ndim == 1 else tt.as_tensor_variable(n) n.ndim n * p #%% n = np.array([[106], [143], [102], [116], [183], [150]]) #n = 183 p = np.array([[ 0.21245365, 0.41223126, 0.37531509], [ 0.13221011, 0.50537169, 0.3624182 ], [ 0.08813779, 0.54447146, 0.36739075], [ 0.18932804, 0.4630365, 0.34763546], [ 0.11006472, 0.49227755, 0.39765773], [ 0.17886852, 0.41098834, 0.41014314]]) #p = np.array([[ 0.21245365, 0.41223126, 0.37531509]]) #n = tt.as_tensor_variable(n) p = tt.as_tensor_variable(p) #%% Multinomial.dist(1,np.ones(3)/3,shape=(6, 3)).mode.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).p.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).n.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).mean.eval() #%% Multinomial.dist(n,p,shape=(6, 3)).random() #%% counts =np.asarray([[19, 50, 37], [21, 67, 55], [11, 53, 38], [17, 54, 45], [24, 93, 66], [27, 53, 70]]) Multinomial.dist(n,p,shape=(6, 3)).logp(x=counts).eval() #%% with pm.Model() as model: like = Multinomial('obs_ABC', n, p, observed=counts, shape=counts.shape) #%% paramall = ( [[.25, .25, .25, .25], 4, 2], [[.25, .25, .25, .25], (1, 4), 3], # 3: expect to fail # [[.25, .25, .25, .25], (10, 4)], [[.25, .25, .25, .25], (10, 1, 4), 5], # 5: expect to fail # [[[.25, .25, .25, .25]], (2, 4), [7, 11]], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (2, 4), 13], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (2, 4), [17, 19]], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (1, 2, 4), [23, 29]], [[[.25, .25, .25, .25], [.25, .25, .25, .25]], (10, 2, 4), [31, 37]], ) for p, shape, n in paramall: with pm.Model() as model: m = Multinomial('m', n=n, p=np.asarray(p), shape=shape) print(m.random().shape) #%% counts =np.asarray([[19, 50, 37], [21, 67, 55], [11, 53, 38], [17, 54, 45], [24, 93, 66], [27, 53, 70]]) n = np.array([[106], [143], [102], [116], [183], [150]]) sparsity=1 #not zero beta=np.ones(counts.shape) #input for dirichlet with pm.Model() as model: theta=pm.Dirichlet('theta',beta/sparsity, shape = counts.shape) transition=pm.Multinomial('transition',n,theta,observed=counts) trace=pm.sample(1000) #%% import numpy as np import pymc3 as pm import theano.tensor as tt def norm_simplex(p): """Sum-to-zero transformation.""" return (p.T / p.sum(axis=-1)).T def ccmodel(beta, x): """Community composition model.""" return norm_simplex(tt.exp(tt.dot(x, tt.log(beta)))) n, k, r = 25, 10, 2 x = np.random.randint(0, 1000, size=(n, k)) y = np.random.randint(0, 1000, size=n) design = np.vstack((np.ones(25), np.random.randint(2, size=n))).T with pm.Model() as model: # Community composition pi = pm.Dirichlet('pi', np.ones(k), shape=(r, k)) comp = pm.Deterministic('comp', ccmodel(pi, design)) # Inferred population density of observed taxa (hierarchical model) rho = pm.Normal('rho', shape=r) tau = pm.Lognormal('tau') dens = pm.Lognormal('dens', tt.dot(design, rho), tau=tau, shape=n) # Community composition *with* the spike expected_recovery = as_col(1 / dens) _comp = norm_simplex(tt.concatenate((comp, expected_recovery), axis=1)) # Variability mu = pm.Lognormal('mu') # Data obs = DirichletMultinomial('obs', _comp * mu, observed=tt.concatenate((x, as_col(y)), axis=1)) pm.sample(1000)
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 18, 198, 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 37811, 198, 41972, 319, 2892, 8621, 1367, 1511, 25, 1270, 25, 4310, 2177, 198, 198, 31, 9800, 25, 300, 5488, 73, ...
2.003599
2,223
from useful import * from os import system if __name__ == "__main__": system("cls") #system("clear") while(True): formula = input("Frmula: ") if formula == 'q': break print(formula) print("Removendo implicaes: ") A1 = remove_implication(formula) print(A1) print("Aplicando Lei de Morgan: ") A2 = morgan_law(A1) print(A2) print("Removendo dupla negao: ") A3 = remove_double_negation(A2) print(A3) print("Aplicando distributividade: ") A4 = distributivity(A3) print(A4) print("Aplicando distributividade com novo tomo: ") A5 = distributivity_new_aton(A3) print(A5) system("pause")
[ 6738, 4465, 1330, 1635, 198, 6738, 28686, 1330, 1080, 198, 198, 361, 11593, 3672, 834, 6624, 366, 834, 12417, 834, 1298, 198, 220, 220, 220, 1080, 7203, 565, 82, 4943, 198, 220, 220, 220, 1303, 10057, 7203, 20063, 4943, 198, 220, 220,...
2.002695
371
# Copyright 2018 Oliver Berger # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Core features """ import asyncio import concurrent import logging import os import time from implant import core log = logging.getLogger(__name__)
[ 2, 15069, 2864, 15416, 33337, 198, 2, 198, 2, 49962, 739, 262, 24843, 13789, 11, 10628, 362, 13, 15, 357, 1169, 366, 34156, 15341, 198, 2, 345, 743, 407, 779, 428, 2393, 2845, 287, 11846, 351, 262, 13789, 13, 198, 2, 921, 743, 733...
3.80829
193
from trainer.utils.losses import * from trainer.utils import custom_ssim
[ 6738, 21997, 13, 26791, 13, 22462, 274, 1330, 1635, 198, 6738, 21997, 13, 26791, 1330, 2183, 62, 824, 320, 198 ]
3.65
20
############################################################################## # # Copyright (c) 2005 Zope Foundation and Contributors. # # This software is subject to the provisions of the Zope Public License, # Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution. # THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED # WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS # FOR A PARTICULAR PURPOSE. # ############################################################################## """Names exported by the ZopeTestCase package """ import ZopeLite as Zope2 import utils import layer from ZopeLite import hasProduct from ZopeLite import installProduct from ZopeLite import hasPackage from ZopeLite import installPackage from ZopeLite import _print from ZopeTestCase import folder_name from ZopeTestCase import user_name from ZopeTestCase import user_password from ZopeTestCase import user_role from ZopeTestCase import standard_permissions from ZopeTestCase import ZopeTestCase from ZopeTestCase import FunctionalTestCase from PortalTestCase import portal_name from PortalTestCase import PortalTestCase from sandbox import Sandboxed from functional import Functional from base import TestCase from base import app from base import close from warnhook import WarningsHook from unittest import main from zopedoctest import ZopeDocTestSuite from zopedoctest import ZopeDocFileSuite from zopedoctest import FunctionalDocTestSuite from zopedoctest import FunctionalDocFileSuite import zopedoctest as doctest import transaction import placeless Zope = Zope2
[ 29113, 29113, 7804, 4242, 2235, 198, 2, 198, 2, 15069, 357, 66, 8, 5075, 1168, 3008, 5693, 290, 25767, 669, 13, 198, 2, 198, 2, 770, 3788, 318, 2426, 284, 262, 8617, 286, 262, 1168, 3008, 5094, 13789, 11, 198, 2, 10628, 362, 13, ...
3.874142
437
# Copyright (C) 2010-2011 Richard Lincoln # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from CIM14.CPSM.Equipment.LoadModel.EnergyArea import EnergyArea
[ 2, 15069, 357, 34, 8, 3050, 12, 9804, 6219, 12406, 198, 2, 198, 2, 2448, 3411, 318, 29376, 7520, 11, 1479, 286, 3877, 11, 284, 597, 1048, 16727, 257, 4866, 198, 2, 286, 428, 3788, 290, 3917, 10314, 3696, 357, 1169, 366, 25423, 123...
3.822951
305
#!/usr/bin/env python import sys HATOMS = ["HG", "HD", "HE", "HH"] lines = open(sys.argv[1]).readlines() for line in lines: if line[:6] == "ATOM " or line[:6] == "HETATM": if line[17:20] == "WAT": continue if line[13] == "H": continue if line[12:14] in HATOMS: continue print(line.strip("\n"))
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 11748, 25064, 198, 198, 39, 1404, 2662, 50, 796, 14631, 39, 38, 1600, 366, 10227, 1600, 366, 13909, 1600, 366, 16768, 8973, 198, 198, 6615, 796, 1280, 7, 17597, 13, 853, 85, 58, 16, ...
1.873096
197
import Pmw import os import re from FastIndex import FastIndex, timer from PrologFrame import PrologFrame from TerminalFrame import TerminalFrame regexp = re.compile('[a-zA-z0-9_]+__[0-9]+')
[ 11748, 350, 76, 86, 198, 11748, 28686, 198, 11748, 302, 198, 6738, 12549, 15732, 1330, 12549, 15732, 11, 19781, 198, 6738, 1041, 6404, 19778, 1330, 1041, 6404, 19778, 198, 6738, 24523, 19778, 1330, 24523, 19778, 198, 198, 260, 25636, 79, ...
2.909091
66
import sqlite3 with sqlite3.connect('storage.db') as conn: cursor = conn.cursor() cursor.execute( """ CREATE TABLE IF NOT EXISTS Produtos ( id INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT, name TEXT NOT NULL, price REAL, compra_id INTEGER, FOREIGN KEY (compra_id) REFERENCES Compras(id) ); """ ) cursor.execute( """ CREATE TABLE IF NOT EXISTS Compras ( id INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT, date TEXT NOT NULL ); """ )
[ 11748, 44161, 578, 18, 198, 198, 4480, 44161, 578, 18, 13, 8443, 10786, 35350, 13, 9945, 11537, 355, 48260, 25, 198, 220, 220, 220, 23493, 796, 48260, 13, 66, 21471, 3419, 628, 220, 220, 220, 23493, 13, 41049, 7, 198, 220, 220, 220,...
1.931818
308
disable_gtk_binaries = True
[ 40223, 62, 13655, 74, 62, 8800, 3166, 796, 6407, 198 ]
2.8
10
import sys from gensim import models from gensim.models.doc2vec import LabeledSentence import pickle with open('corpus_text', 'rb') as f: corpus = pickle.load(f) sentences = corpus_to_sentences(corpus) model = models.Doc2Vec(vector_size=400, window=15, alpha=.025, min_alpha=.025, min_count=1, sample=1e-6) model.build_vocab(sentences) print(len(corpus)) model.train(sentences, total_examples=len(corpus), epochs=20) model.save('doc2vec.model')
[ 11748, 25064, 198, 6738, 308, 641, 320, 1330, 4981, 198, 6738, 308, 641, 320, 13, 27530, 13, 15390, 17, 35138, 1330, 3498, 18449, 31837, 594, 198, 11748, 2298, 293, 628, 198, 198, 4480, 1280, 10786, 10215, 79, 385, 62, 5239, 3256, 705...
2.559322
177
# -*- encoding: utf-8 -*- from collections import UserDict from itertools import count import shutil import winreg import uuid PATH = "Software\\Microsoft\\Windows\\CurrentVersion\\Lxss" KEY = winreg.HKEY_CURRENT_USER def __exit__(self, *exc): self._state = 1 return False class _Lxss(object): default_distribution = RegistryDescriptor("DefaultDistribution") def __iter__(self): for i in count(): with self._key() as k: try: name = winreg.EnumKey(k, i) except OSError as e: if e.winerror != 259: raise break yield Distribution(name) Lxss = _Lxss()
[ 2, 532, 9, 12, 21004, 25, 3384, 69, 12, 23, 532, 9, 12, 201, 198, 201, 198, 6738, 17268, 1330, 11787, 35, 713, 201, 198, 6738, 340, 861, 10141, 1330, 954, 201, 198, 11748, 4423, 346, 201, 198, 11748, 1592, 2301, 201, 198, 11748, ...
1.920398
402
# changelog.py -- Python module for Debian changelogs # Copyright (C) 2006-7 James Westby <jw+debian@jameswestby.net> # Copyright (C) 2008 Canonical Ltd. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA # The parsing code is based on that from dpkg which is: # Copyright 1996 Ian Jackson # Copyright 2005 Frank Lichtenheld <frank@lichtenheld.de> # and licensed under the same license as above. """This module implements facilities to deal with Debian changelogs.""" from __future__ import absolute_import import os import pwd import re import socket import warnings import sys import six from debian import debian_support # Python 3 doesn't have StandardError, but let's avoid changing our # exception inheritance hierarchy for Python 2. try: _base_exception_class = StandardError except NameError: _base_exception_class = Exception # TODO(jsw): Remove this in favor of using debian_support.Version directly. I # don't think we gain anything by using this empty subclass. if sys.version >= '3': __str__ = _format else: __unicode__ = _format topline = re.compile(r'^(\w%(name_chars)s*) \(([^\(\) \t]+)\)' r'((\s+%(name_chars)s+)+)\;' % {'name_chars': '[-+0-9a-z.]'}, re.IGNORECASE) blankline = re.compile(r'^\s*$') change = re.compile(r'^\s\s+.*$') endline = re.compile(r'^ -- (.*) <(.*)>( ?)((\w+\,\s*)?\d{1,2}\s+\w+\s+' r'\d{4}\s+\d{1,2}:\d\d:\d\d\s+[-+]\d{4}\s*)$') endline_nodetails = re.compile(r'^ --(?: (.*) <(.*)>( ?)((\w+\,\s*)?\d{1,2}' r'\s+\w+\s+\d{4}\s+\d{1,2}:\d\d:\d\d\s+[-+]\d{4}' r'))?\s*$') keyvalue= re.compile(r'^([-0-9a-z]+)=\s*(.*\S)$', re.IGNORECASE) value_re = re.compile(r'^([-0-9a-z]+)((\s+.*)?)$', re.IGNORECASE) xbcs_re = re.compile('^X[BCS]+-', re.IGNORECASE) emacs_variables = re.compile(r'^(;;\s*)?Local variables:', re.IGNORECASE) vim_variables = re.compile('^vim:', re.IGNORECASE) cvs_keyword = re.compile(r'^\$\w+:.*\$') comments = re.compile(r'^\# ') more_comments = re.compile(r'^/\*.*\*/') closes = re.compile(r'closes:\s*(?:bug)?\#?\s?\d+(?:,\s*(?:bug)?\#?\s?\d+)*', re.IGNORECASE) closeslp = re.compile(r'lp:\s+\#\d+(?:,\s*\#\d+)*', re.IGNORECASE) old_format_re1 = re.compile(r'^(\w+\s+\w+\s+\d{1,2} \d{1,2}:\d{1,2}:\d{1,2}' r'\s+[\w\s]*\d{4})\s+(.*)\s+(<|\()(.*)(\)|>)') old_format_re2 = re.compile(r'^(\w+\s+\w+\s+\d{1,2},?\s*\d{4})\s+(.*)' r'\s+(<|\()(.*)(\)|>)') old_format_re3 = re.compile(r'^(\w[-+0-9a-z.]*) \(([^\(\) \t]+)\)\;?', re.IGNORECASE) old_format_re4 = re.compile(r'^([\w.+-]+)(-| )(\S+) Debian (\S+)', re.IGNORECASE) old_format_re5 = re.compile('^Changes from version (.*) to (.*):', re.IGNORECASE) old_format_re6 = re.compile(r'^Changes for [\w.+-]+-[\w.+-]+:?\s*$', re.IGNORECASE) old_format_re7 = re.compile(r'^Old Changelog:\s*$', re.IGNORECASE) old_format_re8 = re.compile(r'^(?:\d+:)?\w[\w.+~-]*:?\s*$') def get_maintainer(): """Get the maintainer information in the same manner as dch. This function gets the information about the current user for the maintainer field using environment variables of gecos informations as approriate. It uses the same methods as dch to get the information, namely DEBEMAIL, DEBFULLNAME, EMAIL, NAME, /etc/mailname and gecos. :returns: a tuple of the full name, email pair as strings. Either of the pair may be None if that value couldn't be determined. """ env = os.environ regex = re.compile(r"^(.*)\s+<(.*)>$") # Split email and name if 'DEBEMAIL' in env: match_obj = regex.match(env['DEBEMAIL']) if match_obj: if not 'DEBFULLNAME' in env: env['DEBFULLNAME'] = match_obj.group(1) env['DEBEMAIL'] = match_obj.group(2) if 'DEBEMAIL' not in env or 'DEBFULLNAME' not in env: if 'EMAIL' in env: match_obj = regex.match(env['EMAIL']) if match_obj: if not 'DEBFULLNAME' in env: env['DEBFULLNAME'] = match_obj.group(1) env['EMAIL'] = match_obj.group(2) # Get maintainer's name if 'DEBFULLNAME' in env: maintainer = env['DEBFULLNAME'] elif 'NAME' in env: maintainer = env['NAME'] else: # Use password database if no data in environment variables try: maintainer = re.sub(r',.*', '', pwd.getpwuid(os.getuid()).pw_gecos) except (KeyError, AttributeError): maintainer = None # Get maintainer's mail address if 'DEBEMAIL' in env: email = env['DEBEMAIL'] elif 'EMAIL' in env: email = env['EMAIL'] else: addr = None if os.path.exists('/etc/mailname'): f = open('/etc/mailname') try: addr = f.readline().strip() finally: f.close() if not addr: addr = socket.getfqdn() if addr: user = pwd.getpwuid(os.getuid()).pw_name if not user: addr = None else: addr = "%s@%s" % (user, addr) if addr: email = addr else: email = None return (maintainer, email)
[ 2, 1488, 417, 519, 13, 9078, 1377, 11361, 8265, 329, 26062, 1488, 417, 18463, 198, 2, 15069, 357, 34, 8, 4793, 12, 22, 3700, 2688, 1525, 1279, 73, 86, 10, 24689, 31, 73, 1047, 7038, 1525, 13, 3262, 29, 198, 2, 15069, 357, 34, 8,...
2.122507
2,808
# Define a procedure, median, that takes three # numbers as its inputs, and returns the median # of the three numbers. # Make sure your procedure has a return statement. print(median(1,2,3)) #>>> 2 print(median(9,3,6)) #>>> 6 print(median(7,8,7)) #>>> 7
[ 2, 2896, 500, 257, 8771, 11, 14288, 11, 326, 2753, 1115, 201, 198, 2, 3146, 355, 663, 17311, 11, 290, 5860, 262, 14288, 201, 198, 2, 286, 262, 1115, 3146, 13, 201, 198, 201, 198, 2, 6889, 1654, 534, 8771, 468, 257, 1441, 2643, 1...
2.27907
129
import requests import os import subprocess import gidgethub from gidgethub import sansio AUTOMERGE_LABEL = ":robot: automerge" def comment_on_pr(issue_number, message): """ Leave a comment on a PR/Issue """ request_headers = sansio.create_headers( "miss-islington", oauth_token=os.getenv("GH_AUTH") ) issue_comment_url = ( f"https://api.github.com/repos/python/cpython/issues/{issue_number}/comments" ) data = {"body": message} response = requests.post(issue_comment_url, headers=request_headers, json=data) if response.status_code == requests.codes.created: print(f"Commented at {response.json()['html_url']}, message: {message}") else: print(response.status_code) print(response.text) return response def assign_pr_to_core_dev(issue_number, coredev_login): """ Assign the PR to a core dev. Should be done when miss-islington failed to backport. """ request_headers = sansio.create_headers( "miss-islington", oauth_token=os.getenv("GH_AUTH") ) edit_issue_url = ( f"https://api.github.com/repos/python/cpython/issues/{issue_number}" ) data = {"assignees": [coredev_login]} response = requests.patch(edit_issue_url, headers=request_headers, json=data) if response.status_code == requests.codes.created: print(f"Assigned PR {issue_number} to {coredev_login}") else: print(response.status_code) print(response.text) return response def normalize_title(title, body): """Normalize the title if it spills over into the PR's body.""" if not (title.endswith("") and body.startswith("")): return title else: # Being paranoid in case \r\n is used. return title[:-1] + body[1:].partition("\r\n")[0] def normalize_message(body): """Normalize the message body to make it commit-worthy. Mostly this just means removing HTML comments, but also removes unwanted leading or trailing whitespace. Returns the normalized body. """ while "<!--" in body: body = body[: body.index("<!--")] + body[body.index("-->") + 3 :] return "\n\n" + body.strip() # Copied over from https://github.com/python/bedevere
[ 11748, 7007, 198, 11748, 28686, 198, 11748, 850, 14681, 198, 198, 11748, 308, 17484, 40140, 198, 198, 6738, 308, 17484, 40140, 1330, 38078, 952, 628, 198, 39371, 2662, 1137, 8264, 62, 48780, 3698, 796, 366, 25, 305, 13645, 25, 3557, 263...
2.584668
874
import aiosqlite import sqlite3 import asyncio import nonebot from nonebot.log import logger driver: nonebot.Driver = nonebot.get_driver() config: nonebot.config.Config = driver.config
[ 11748, 257, 4267, 13976, 578, 198, 11748, 44161, 578, 18, 198, 11748, 30351, 952, 198, 11748, 4844, 13645, 198, 6738, 4844, 13645, 13, 6404, 1330, 49706, 198, 198, 26230, 25, 4844, 13645, 13, 32103, 796, 4844, 13645, 13, 1136, 62, 26230...
3.321429
56
from typing import Callable, Optional, Sequence, Tuple, Union import numpy from dexp.processing.utils.nd_slice import nd_split_slices, remove_margin_slice from dexp.processing.utils.normalise import Normalise from dexp.utils import xpArray from dexp.utils.backends import Backend def scatter_gather_i2i( function: Callable, image: xpArray, tiles: Union[int, Tuple[int, ...]], margins: Optional[Union[int, Tuple[int, ...]]] = None, normalise: bool = False, clip: bool = False, to_numpy: bool = True, internal_dtype: Optional[numpy.dtype] = None, ) -> xpArray: """ Image-2-image scatter-gather. 'Scatters' computation of a given unary function by splitting the input array into tiles, computing using a given backend, and reassembling the tiles into a single array of same shape as the inpout that is either backed by the same backend than that of the input image, or that is backed by numpy -- usefull when the compute backend cannot hold the whole input and output images in memory. Parameters ---------- function : unary function image : input image (can be any backend, numpy ) tiles : tile sizes to cut input image into, can be a single integer or a tuple of integers. margins : margins to add to each tile, can be a single integer or a tuple of integers. if None, no margins are added. normalise : normalises the input image. clip : clip after normalisation/denormalisation to_numpy : should the result be a numpy array? Very usefull when the compute backend cannot hold the whole input and output images in memory. internal_dtype : internal dtype for computation Returns ------- Result of applying the unary function to the input image, if to_numpy==True then the image is """ if internal_dtype is None: internal_dtype = image.dtype if type(tiles) == int: tiles = (tiles,) * image.ndim # If None is passed for a tile that means that we don't tile along that axis, we als clip the tile size: tiles = tuple((length if tile is None else min(length, tile)) for tile, length in zip(tiles, image.shape)) if margins is None: margins = (0,) * image.ndim if type(margins) == int: margins = (margins,) * image.ndim if to_numpy: result = numpy.empty(shape=image.shape, dtype=internal_dtype) else: result = Backend.get_xp_module(image).empty_like(image, dtype=internal_dtype) # Normalise: norm = Normalise(Backend.to_backend(image), do_normalise=normalise, clip=clip, quantile=0.005) # image shape: shape = image.shape # We compute the slices objects to cut the input and target images into batches: tile_slices = list(nd_split_slices(shape, chunks=tiles, margins=margins)) tile_slices_no_margins = list(nd_split_slices(shape, chunks=tiles)) # Zipping together slices with and without margins: slices = zip(tile_slices, tile_slices_no_margins) # Number of tiles: number_of_tiles = len(tile_slices) if number_of_tiles == 1: # If there is only one tile, let's not be complicated about it: result = norm.backward(function(norm.forward(image))) if to_numpy: result = Backend.to_numpy(result, dtype=internal_dtype) else: result = Backend.to_backend(result, dtype=internal_dtype) else: _scatter_gather_loop( norm.backward, function, image, internal_dtype, norm.forward, result, shape, slices, to_numpy ) return result # Dask turned out not too work great here, HUGE overhead compared to the light approach above. # def scatter_gather_dask(backend: Backend, # function, # image, # chunks, # margins=None): # boundary=None # trim=True # align_arrays=True # # image_d = from_array(image, chunks=chunks, asarray=False) # # def function_numpy(_image): # print(_image.shape) # return backend.to_numpy(function(_image)) # # #func, *args, depth=None, boundary=None, trim=True, align_arrays=True, **kwargs # computation= map_overlap(function_numpy, # image_d, # depth=margins, # boundary=boundary, # trim=trim, # align_arrays=align_arrays, # dtype=image.dtype # ) # # #computation.visualize(filename='transpose.png') # result = computation.compute() # # return result
[ 6738, 19720, 1330, 4889, 540, 11, 32233, 11, 45835, 11, 309, 29291, 11, 4479, 198, 198, 11748, 299, 32152, 198, 198, 6738, 36017, 79, 13, 36948, 13, 26791, 13, 358, 62, 48369, 1330, 299, 67, 62, 35312, 62, 82, 677, 274, 11, 4781, ...
2.546363
1,801
from .backend import Backend from .thread import HttpPool
[ 6738, 764, 1891, 437, 1330, 5157, 437, 198, 6738, 764, 16663, 1330, 367, 29281, 27201, 198 ]
3.625
16
#!/usr/local/bin/python3 if __name__ == '__main__': main()
[ 2, 48443, 14629, 14, 12001, 14, 8800, 14, 29412, 18, 628, 198, 198, 361, 11593, 3672, 834, 6624, 705, 834, 12417, 834, 10354, 198, 220, 220, 220, 1388, 3419, 198 ]
2.2
30
import fnmatch import glob import os import sys import setuptools.command.build_ext APPLE = sys.platform == 'darwin' BASE_DIR = os.path.dirname(__file__) ABOUT = {} _read_about() EXCLUDED_STATIC_FILE_PATHS = [ '*.py', '*/__pycache__/*', '*/tests/*', '*/_ext/cc/*', '*/_ext/cy/*', '*/_ext/rs/*', ] PACKAGE_DATA = [ '.revision', ] + _get_static_files('omnibus') INSTALL_REQUIRES = [ 'toolz>=0.9.0', ] EXTRAS_REQUIRE = { 'bintrees': ['bintrees>=0.2.7'], 'cytoolz': ['cytoolz>=0.9.0'], 'docker': ['docker>=3.7.0'], 'sortedcontainers': ['sortedcontainers>=2.1.0'], } DEBUG = 'DEBUG' in os.environ EXT_MODULES = [] try: import Cython except ImportError: pass else: import Cython.Build import Cython.Compiler.Options EXT_MODULES.extend([ *[ setuptools.Extension( 'omnibus._ext.cc.' + os.path.basename(fpath).rpartition('.')[0], sources=[fpath] ) for fpath in glob.glob('omnibus/_ext/cc/*.cc') ], *Cython.Build.cythonize( [ setuptools.Extension( 'omnibus._ext.cy.' + os.path.basename(fpath).rpartition('.')[0], sources=[fpath], language='c++', ) for fpath in glob.glob('omnibus/_ext/cy/**/*.pyx', recursive=True) ], language_level=3, gdb_debug=DEBUG, compiler_directives={ **Cython.Compiler.Options.get_directive_defaults(), 'embedsignature': True, 'binding': True, }, ), ]) if APPLE: EXT_MODULES.extend([ setuptools.Extension( 'omnibus._ext.m.' + os.path.basename(fpath).rpartition('.')[0], sources=[fpath], extra_link_args=[ '-framework', 'AppKit', '-framework', 'CoreFoundation', ] ) for fpath in glob.glob('omnibus/_ext/m/*.m') ]) if __name__ == '__main__': setuptools.setup( name=ABOUT['__title__'], version=ABOUT['__version__'], description=ABOUT['__description__'], author=ABOUT['__author__'], url=ABOUT['__url__'], python_requires='>=3.7', classifiers=[ 'Intended Audience :: Developers', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: ' + '.'.join(map(str, sys.version_info[:2])), 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python', ], # zip_safe=True, setup_requires=['setuptools'], packages=setuptools.find_packages( include=['omnibus', 'omnibus.*'], exclude=['tests', '*.tests', '*.tests.*'], ), py_modules=['omnibus'], package_data={'omnibus': PACKAGE_DATA}, include_package_data=True, entry_points={}, install_requires=INSTALL_REQUIRES, extras_require=EXTRAS_REQUIRE, ext_modules=EXT_MODULES, )
[ 11748, 24714, 15699, 198, 11748, 15095, 198, 11748, 28686, 198, 11748, 25064, 198, 198, 11748, 900, 37623, 10141, 13, 21812, 13, 11249, 62, 2302, 628, 198, 2969, 16437, 796, 25064, 13, 24254, 6624, 705, 27455, 5404, 6, 628, 198, 33, 111...
1.891112
1,699
from day1 import * import unittest
[ 6738, 1110, 16, 1330, 1635, 198, 11748, 555, 715, 395, 198 ]
3.181818
11
import AppKit from PyObjCTools.TestSupport import TestCase, min_os_level
[ 11748, 2034, 20827, 198, 6738, 9485, 49201, 4177, 10141, 13, 14402, 15514, 1330, 6208, 20448, 11, 949, 62, 418, 62, 5715, 628 ]
3.363636
22
from collections import OrderedDict, defaultdict import numpy as np ''' generate a id to length dic '''
[ 6738, 17268, 1330, 14230, 1068, 35, 713, 11, 4277, 11600, 198, 11748, 299, 32152, 355, 45941, 198, 198, 7061, 6, 198, 8612, 378, 257, 4686, 284, 4129, 288, 291, 198, 7061, 6, 198 ]
3.181818
33
from ._schema import DefaultSchema from ._utils import get_minutes, get_yields, normalize_string
[ 6738, 47540, 15952, 2611, 1330, 15161, 27054, 2611, 198, 6738, 47540, 26791, 1330, 651, 62, 1084, 1769, 11, 651, 62, 88, 1164, 82, 11, 3487, 1096, 62, 8841, 628 ]
3.37931
29
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright [2017] Tatarnikov Viktor [viktor@tatarnikov.org] # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. try: from urllib.parse import unquote except ImportError: from urllib import unquote from flask import url_for, current_app from werkzeug.utils import import_string
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 198, 2, 198, 2, 15069, 685, 5539, 60, 29070, 1501, 1134, 709, 31096, 685, 28930, 13165, 31, 83, 265, 1501, 1134, 709, 13, ...
3.378049
246
# -*- coding: utf-8 -*- # !/usr/bin/python __author__ = 'ma_keling' # Version : 1.0.0 # Start Time : 2018-11-29 # Update Time : # Change Log : ## 1. ## 2. ## 3. import time import arcpy import math # Description: Loop layers and calculate lod for every feature in the layer. # Description: Compute the total records for a featureclass # Description: get the start level based on layer extent # Description: Add a new field with name lod to a table # Description: Compute get area and length per pixel based on dpi and scale # Description: Compute get length per pixel based on dpi and scale # Description: Calculate lod for every polygon # Description: Calculate lod for every point # Description: Calculate lod for every polyline # execute()
[ 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 201, 198, 2, 5145, 14, 14629, 14, 8800, 14, 29412, 201, 198, 834, 9800, 834, 796, 705, 2611, 62, 365, 1359, 6, 201, 198, 2, 10628, 220, 220, 220, 220, 1058, 352, 13, 15...
2.800676
296
# -*- coding: utf-8 -*- """ Created on Sat Feb 8 07:38:05 2020 @author: Ham Self Challenge: Count Cluster of Non-0s Given a 1-dimension array of integers, determine how many 'clusters' of non-0 in the array. A 'cluster' is a group of consecutive non-0 values. Scoring: a solution needs to be a 1-liner; i.e. NO point if implementing with a traditional 'for' loop! Sample Input (see STDIN_SIO) A : [ 9, 0, 0, 22, 0, 0, 39, 11, 3, 0, \ 0, 24, 1, 0, 50, 23, 3, 44, 0, 23, \ 25, 6, 36, 19, 10, 23, 0, 37, 4, 1, \ 7, 12, 0, 0, 49 ] Expected Output: 8 """ import itertools STDIN_SIO = """ 9, 0, 0, 22, 0, 0, 39, 11, 3, 0, \ 0, 24, 1, 0, 50, 23, 3, 44, 0, 23, \ 2, 8, 20, 35, 0, 40, 34, 26, 36, 0, \ 35, 19, 20, 18, 11, 43, 19, 21, 40, 0, \ 14, 0, 14, 0, 0, 25, 35, 24, 49, 15, \ 13, 3, 0, 10, 31, 25, 27, 37, 27, 43, \ 44, 27, 8, 43, 0, 0, 33, 25, 19, 47, \ 0, 29, 5, 2, 12, 8, 7, 0, 16, 36, \ 0, 6, 17, 35, 36, 21, 0, 9, 1, 0, \ 43, 29, 39, 15, 18, 0, 34, 26, 48, 0, \ 34, 35, 7, 10, 0, 0, 15, 5, 12, 26, \ 0, 37, 30, 33, 27, 34, 9, 37, 22, 0, \ 0, 24, 30, 0, 0, 38, 23, 25, 0, 30, \ 39, 24, 31, 0, 6, 19, 25, 0, 28, 15, \ 8, 0, 48, 0, 35, 41, 0, 24, 1, 41, \ 31, 0, 35, 21, 15, 26, 15, 27, 4, 0, \ 8, 4, 0, 0, 2, 42, 18, 0, 28, 18, \ 49, 34, 5, 10, 41, 48, 26, 14, 45, 44, \ 9, 0, 49, 50, 24, 0, 0, 0, 23, 0, \ 17, 0, 47, 31, 0, 42, 0, 0, 0, 40, \ 46, 22, 50, 32, 20, 3, 44, 22, 0, 37, \ 25, 0, 19, 26, 14, 23, 27, 41, 0, 1, \ 13, 0, 48, 20, 37, 8, 0, 18, 0, 26, \ 12, 19, 32, 19, 22, 0, 0, 0, 0, 0, \ 16, 0, 0, 43, 0, 10, 5, 0, 6, 26, \ 0, 24, 40, 29, 0, 43, 18, 27, 0, 0, \ 37, 0, 46, 35, 17, 0, 20, 44, 29, 29, \ 40, 33, 22, 27, 0, 0, 38, 21, 4, 0, \ 0, 15, 31, 48, 36, 10, 0, 41, 0, 45, \ 39, 0, 11, 9, 3, 38, 16, 0, 11, 22, \ 37, 0, 3, 44, 10, 12, 47, 22, 32, 7, \ 24, 1, 0, 22, 25, 0, 14, 0, 0, 0, \ 23, 0, 36, 1, 42, 46, 0, 48, 0, 33, \ 5, 27, 45, 0, 15, 29, 0, 50, 2, 31, \ 25, 6, 36, 19, 10, 23, 0, 37, 4, 1, \ 7, 12, 0, 0, 49 """.strip() def count_non_0_clusters_1(arr): """Translate each non-0 to an 'A' char, and 0 to a space. Then join together to become a string. Then split(), then return number of tokens. """ return len("".join(["A" if e else " " for e in arr]).split()) def count_non_0_clusters_2(arr): """groupby() partitions into groups as: [[True , [list of non-0]], [False, [list of 0s]], [True , [list of non-0]], [False, [list of 0s]], ... [True , [list of non-0]]] (Old) Next, the list comprenhension iterates thru each tuple, then collects the 1st element in each tuple if True. Finally, return the len/count of Trues: return len([t[0] for t in itertools.groupby(...) if t[0]]) Next, the list comprenhension iterates thru each tuple, then collects the 1st element in each tuple. Then return the count() of True elements. """ return [t[0] for t in itertools.groupby(arr, lambda e: bool(e))].count(True) if __name__ == '__main__': a = list(map(int, STDIN_SIO.split(","))) # Nicely print it, 10 entries per line, with continuation # so can copy-n-paste back into STDIN_SIO #print(len(a)) #for i in range(0, (len(a) // 10) * 10, 10): # print("%3u," * 10 % tuple(a[i:i+10]), end=" \\\n") #j = a[(len(a) // 10) * 10:] #print("%3u," * (len(j) - 1) % tuple(j[:-1]), end="") #print("%3u" % j[-1]) print("count_*_1() returns", count_non_0_clusters_1(a), "clusters of non-0") print("count_*_2() returns", count_non_0_clusters_2(a), "clusters of non-0")
[ 2, 532, 9, 12, 19617, 25, 3384, 69, 12, 23, 532, 9, 12, 201, 198, 37811, 201, 198, 41972, 319, 7031, 3158, 220, 807, 8753, 25, 2548, 25, 2713, 12131, 201, 198, 201, 198, 31, 9800, 25, 4345, 201, 198, 201, 198, 24704, 13879, 25, ...
1.93059
2,017
"""Tests for node configuration.""" import json import logging import time from pathlib import Path import allure import pytest from _pytest.tmpdir import TempdirFactory from cardano_clusterlib import clusterlib from cardano_node_tests.utils import cluster_management from cardano_node_tests.utils import cluster_nodes from cardano_node_tests.utils import configuration from cardano_node_tests.utils import helpers LOGGER = logging.getLogger(__name__) # use the "temp_dir" fixture for all tests automatically pytestmark = pytest.mark.usefixtures("temp_dir") def check_epoch_length(cluster_obj: clusterlib.ClusterLib) -> None: end_sec = 15 end_sec_padded = end_sec + 15 # padded to make sure tip got updated cluster_obj.wait_for_new_epoch() epoch = cluster_obj.get_epoch() sleep_time = cluster_obj.epoch_length_sec - end_sec time.sleep(sleep_time) assert epoch == cluster_obj.get_epoch() time.sleep(end_sec_padded) assert epoch + 1 == cluster_obj.get_epoch()
[ 37811, 51, 3558, 329, 10139, 8398, 526, 15931, 198, 11748, 33918, 198, 11748, 18931, 198, 11748, 640, 198, 6738, 3108, 8019, 1330, 10644, 198, 198, 11748, 477, 495, 198, 11748, 12972, 9288, 198, 6738, 4808, 9078, 9288, 13, 22065, 15908, ...
3.01791
335
#!/usr/bin/env python from __future__ import print_function import matplotlib matplotlib.use('Agg') import os import heapq import numpy as np import matplotlib.pyplot as plt from mpl_toolkits.mplot3d import Axes3D import matplotlib.colors as colors import matplotlib.cm as cm from numpy import * from random import sample, seed, randint from os.path import getsize as getFileSize import math import random import csv from cycler import cycler from io import StringIO #np.set_printoptions(threshold=np.nan) from collections import Counter from matplotlib.colors import LogNorm from mpl_toolkits.axes_grid1 import AxesGrid from astropy import units as u from astropy import cosmology import matplotlib.ticker as mtick import PlotScripts import ReadScripts import AllVars import GalaxyPhotoion as photo import ObservationalData as Obs import gnedin_analytic as ga from mpi4py import MPI import sys comm = MPI.COMM_WORLD rank = comm.Get_rank() size = comm.Get_size() AllVars.Set_Params_Kali() AllVars.Set_Constants() PlotScripts.Set_Params_Plot() output_format = ".png" # For the Tiamat extended results there is a weird hump when calculating the escape fraction. # This hump occurs at a halo mass of approximately 10.3. # The calculation of fesc skips this hump range (defined from kink_low to kink_high) kink_low = 10.3 kink_high = 10.30000001 m_low = 7.0 # We only sum the photons coming from halos within the mass range m_low < Halo Mass < m_high m_high = 15.0 m_gal_low = 3.0 m_gal_high = 12.0 m_low_SAGE = pow(10, m_low)/1.0e10 * AllVars.Hubble_h m_high_SAGE = pow(10, m_high)/1.0e10 * AllVars.Hubble_h bin_width = 0.2 NB = int((m_high - m_low) / bin_width) NB_gal = int((m_gal_high - m_gal_low) / bin_width) fej_low = 0.0 fej_high = 1.0 fej_bin_width = 0.05 NB_fej = int((fej_high - fej_low) / fej_bin_width) def calculate_beta(MUV, z): ''' Calculation of the dust attenuation parameter Beta. Fit values are from Bouwens (2015) ApJ 793, 115. For z = 5 and 6, Bouwens uses a piece-wise linear relationship and a linear relationship for higher redshift. ## Parameters ---------- MUV : `float' A value of the absolute magnitude in the UV (generally M1600) in the AB magnitude system. z : `float' Redshift the attenuation is calculated at. Returns ------ beta : `float' Value of the UV continuum paramaeter beta. ''' if (z >= 4.5 and z < 5.5): # z = 5 fits. if (MUV > -18.8): dB = -0.08 else: dB = -0.17 B = -2.05 offset = 18.8 elif (z >= 5.5 and z < 6.5): # z = 6 fits. if (MUV > -18.8): dB = -0.08 else: dB = -0.24 B = -2.22 offset = 18.8 elif (z >= 6.5 and z < 7.5): # z = 7 fits. dB = -0.20 B = -2.05 offset = 19.5 elif (z >= 7.5 and z < 8.5): # z = 8 fits. dB = -0.15 B = -2.13 offset = 19.5 elif (z >= 8.5 and z < 9.5): # z = 9 fits. dB = -0.16 B = -2.19 offset = 19.5 elif (z >= 9.5 and z < 10.5): # z = 10 fits. dB = -0.16 B = -2.16 offset = 19.5 beta = dB * (MUV + offset) + B return beta def multiply(array): ''' Performs element wise multiplication. Parameters ---------- array : `~numpy.darray' The array to be multiplied. Returns ------- total : `float' Total of the elements multiplied together. ''' total = 1 for i in range(0, len(array)): total *= array[i] return total ## def Sum_Log(array): ''' Performs an element wise sum of an array who's elements are in log-space. Parameters ---------- array : array Array with elements in log-space. Returns ------ sum_total : float Value of the elements taken to the power of 10 and summed. Units ----- All units are kept the same as the inputs. ''' sum_total = 0.0 for i in range(0, len(array)): sum_total += 10**array[i] return sum_total ## def Std_Log(array, mean): ''' Calculates the standard deviation of an array with elements in log-space. Parameters ---------- array : array Array with elements in log-space. mean : float Mean of the array (not in log). Returns ------ std : float Standard deviation of the input array taken to the power of 10. Units ----- All units are kept the same as the inputs. ''' sum_total = 0.0 for i in range(0, len(array)): sum_total += (10**array[i] - mean)**2 sum_total *= 1.0/len(array) std = np.sqrt(sum_total) return std ### def collect_across_tasks(mean_per_task, std_per_task, N_per_task, SnapList, BinSnapList=[], binned=False, m_bin_low=0.0, m_bin_high=0.0, my_bin_width=bin_width): """ Reduces arrays that are unique to each task onto the master task. The dimensions of the input arrays will change slightly if we are collecting a statistics that is binned across e.g., halo mass or galaxy stellar mass. Parameters ---------- mean_per_task, std_per_task, N_per_task: Nested 2D (or 3D if binned == True) arrays of floats. Outer length is equal to the number of models. Inner length is equal to the number of snapshots the data has been calculated for. Most inner length is equal to the number of bins. Contains the mean/standard deviation/number of objects unique for each task. SnapList: Nested 2D arrays of integers. Outer length is equal to the number of models. Contains the snapshot numbers the data has been calculated for each model. BinSnapList: Nested 2D arrays of integers. Outer length is equal to the number of models. Often statistics are calculated for ALL snapshots but we only wish to plot for a subset of snapshots. This variable allows the binned data to be collected for only a subset of the snapshots. binned: Boolean. Dictates whether the collected data is a 2D or 3D array with the inner-most array being binned across e.g., halo mass. Returns ---------- master_mean, master_std, master_N: Nested 2D (or 3D if binned == True) arrays of floats. Shape is identical to the input mean_per_task etc. If rank == 0 these contain the collected statistics. Otherwise these will be none. master_bin_middle: Array of floats. Contains the location of the middle of the bins for the data. """ master_mean = [] master_std = [] master_N = [] master_bin_middle = [] for model_number in range(0, len(SnapList)): master_mean.append([]) master_std.append([]) master_N.append([]) master_bin_middle.append([]) # If we're collecting a binned statistic (e.g., binned across halo mass), then we need to perform the collecting per snapshot. if binned: count = 0 for snapshot_idx in range(len(SnapList[model_number])): if SnapList[model_number][snapshot_idx] == BinSnapList[model_number][count]: master_mean[model_number], master_std[model_number], master_N[model_number] = calculate_pooled_stats(master_mean[model_number], master_std[model_number], master_N[model_number], mean_per_task[model_number][snapshot_idx], std_per_task[model_number][snapshot_idx], N_per_task[model_number][snapshot_idx]) master_bin_middle[model_number].append(np.arange(m_bin_low, m_bin_high+my_bin_width, my_bin_width)[:-1] + my_bin_width* 0.5) count += 1 if count == len(BinSnapList[model_number]): break else: master_mean[model_number], master_std[model_number], master_N[model_number] = calculate_pooled_stats(master_mean[model_number], master_std[model_number], master_N[model_number], mean_per_task[model_number], std_per_task[model_number], N_per_task[model_number]) if rank == 0: master_mean[model_number] = master_mean[model_number][0] master_std[model_number] = master_std[model_number][0] master_N[model_number] = master_N[model_number][0] return master_mean, master_std, master_N, master_bin_middle ### def calculate_pooled_stats(mean_pool, std_pool, N_pool, mean_local, std_local, N_local): ''' Calculates the pooled mean and standard deviation from multiple processors and appends it to an input array. Formulae taken from https://en.wikipedia.org/wiki/Pooled_variance As we only care about these stats on the rank 0 process, we make use of junk inputs/outputs for other ranks. NOTE: Since the input data may be an array (e.g. pooling the mean/std for a stellar mass function). Parameters ---------- mean_pool, std_pool, N_pool : array of floats. Arrays that contain the current pooled means/standard deviation/number of data points (for rank 0) or just a junk input (for other ranks). mean_local, mean_std : float or array of floats. The non-pooled mean and standard deviation unique for each process. N_local : floating point number or array of floating point numbers. Number of data points used to calculate the mean/standard deviation that is going to be added to the pool. NOTE: Use floating point here so we can use MPI.DOUBLE for all MPI functions. Returns ------- mean_pool, std_pool : array of floats. Original array with the new pooled mean/standard deviation appended (for rank 0) or the new pooled mean/standard deviation only (for other ranks). Units ----- All units are the same as the input. All inputs MUST BE real-space (not log-space). ''' if isinstance(mean_local, list) == True: if len(mean_local) != len(std_local): print("len(mean_local) = {0} \t len(std_local) = {1}".format(len(mean_local), len(std_local))) raise ValueError("Lengths of mean_local and std_local should be equal") if ((type(mean_local).__module__ == np.__name__) == True or (isinstance(mean_local, list) == True)): # Checks to see if we are dealing with arrays. N_times_mean_local = np.multiply(N_local, mean_local) N_times_var_local = np.multiply(N_local, np.multiply(std_local, std_local)) N_local = np.array(N_local).astype(float) N_times_mean_local = np.array(N_times_mean_local).astype(np.float32) if rank == 0: # Only rank 0 holds the final arrays so only it requires proper definitions. N_times_mean_pool = np.zeros_like(N_times_mean_local) N_pool_function = np.zeros_like(N_local) N_times_var_pool = np.zeros_like(N_times_var_local) N_times_mean_pool = N_times_mean_pool.astype(np.float64) # Recast everything to double precision then use MPI.DOUBLE. N_pool_function = N_pool_function.astype(np.float64) N_times_var_pool = N_times_var_pool.astype(np.float64) else: N_times_mean_pool = None N_pool_function = None N_times_var_pool = None comm.Barrier() N_times_mean_local = N_times_mean_local.astype(np.float64) N_local = N_local.astype(np.float64) N_times_var_local = N_times_var_local.astype(np.float64) comm.Reduce([N_times_mean_local, MPI.DOUBLE], [N_times_mean_pool, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the arrays across processors. comm.Reduce([N_local, MPI.DOUBLE],[N_pool_function, MPI.DOUBLE], op = MPI.SUM, root = 0) comm.Reduce([N_times_var_local, MPI.DOUBLE], [N_times_var_pool, MPI.DOUBLE], op = MPI.SUM, root = 0) else: N_times_mean_local = N_local * mean_local N_times_var_local = N_local * std_local * std_local N_times_mean_pool = comm.reduce(N_times_mean_local, op = MPI.SUM, root = 0) N_pool_function = comm.reduce(N_local, op = MPI.SUM, root = 0) N_times_var_pool = comm.reduce(N_times_var_local, op = MPI.SUM, root = 0) if rank == 0: mean_pool_function = np.zeros((len(N_pool_function))) std_pool_function = np.zeros((len(N_pool_function))) for i in range(0, len(N_pool_function)): if N_pool_function[i] == 0: mean_pool_function[i] = 0.0 else: mean_pool_function[i] = np.divide(N_times_mean_pool[i], N_pool_function[i]) if N_pool_function[i] < 3: std_pool_function[i] = 0.0 else: std_pool_function[i] = np.sqrt(np.divide(N_times_var_pool[i], N_pool_function[i])) mean_pool.append(mean_pool_function) std_pool.append(std_pool_function) N_pool.append(N_pool_function) return mean_pool, std_pool, N_pool else: return mean_pool, std_pool, N_pool_function # Junk return because non-rank 0 doesn't care. ## def StellarMassFunction(SnapList, SMF, simulation_norm, FirstFile, LastFile, NumFile, ResolutionLimit_mean, model_tags, observations, paper_plot, output_tag): ''' Calculates the stellar mass function for given galaxies with the option to overplot observations by Song et al. (2013) at z = 6, 7, 8 and/or Baldry et al. (2008) at z = 0.1. Parallel compatible. NOTE: The plotting assumes the redshifts we are plotting at are (roughly) the same for each model. Parameters --------- SnapList : Nested 'array-like`, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots that we plot the stellar mass function at for each model. SMF : Nested 2-dimensional array, SMF[model_number0][snapshot0] = [bin0galaxies, ..., binNgalaxies], with length equal to the number of bins (NB_gal). The count of galaxies within each stellar mass bin. Bounds are given by 'm_gal_low' and 'm_gal_high' in bins given by 'bin_width'. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali FirstFile, LastFile, NumFile : array of integers with length equal to the number of models. The file numbers for each model that were read in (defined by the range between [FirstFile, LastFile] inclusive) and the TOTAL number of files for this model (we may only be plotting a subset of the volume). ResolutionLimit_mean : array of floats with the same shape as SMF. This is the mean stellar mass for a halo with len (number of N-body simulation particles) between 'stellar_mass_halolen_lower' and 'stellar_mass_halolen_upper'. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. observations : int Denotes whether we want to overplot observational results. 0 : Don't plot anything. 1 : Plot Song et al. (2016) at z = 6, 7, 8. 2 : Plot Baldry et al. (2008) at z = 0.1. 3 : Plot both of these. paper_plot : int Denotes whether we want to split the plotting over three panels (z = 6, 7, 8) for the paper or keep it all to one figure. output_tag : string Name of the file that will be generated. File will be saved in the current directory with the output format defined by the 'output_format' variable at the beggining of the file. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Stellar Mass is in units of log10(Msun). ''' ## Empty array initialization ## title = [] normalization_array = [] redshift_labels = [] counts_array = [] bin_middle_array = [] for model_number in range(0, len(SnapList)): counts_array.append([]) bin_middle_array.append([]) redshift_labels.append([]) #### for model_number in range(0, len(SnapList)): # Does this for each of the models. ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() box_factor = (LastFile[model_number] - FirstFile[model_number] + 1.0)/(NumFile[model_number]) # This factor allows us to take a sub-volume of the box and scale the results to represent the entire box. print("We are creating the stellar mass function using {0:.4f} of the box's volume.".format(box_factor)) norm = pow(AllVars.BoxSize,3) / pow(AllVars.Hubble_h, 3) * bin_width * box_factor normalization_array.append(norm) #### for snapshot_idx in range(0, len(SnapList[model_number])): # Loops for each snapshot in each model. tmp = 'z = %.2f' %(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) # Assigns a redshift label. redshift_labels[model_number].append(tmp) ## We perform the plotting on Rank 0 so only this rank requires the final counts array. ## if rank == 0: counts_total = np.zeros_like(SMF[model_number][snapshot_idx]) else: counts_total = None comm.Reduce([SMF[model_number][snapshot_idx], MPI.FLOAT], [counts_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. if rank == 0: counts_array[model_number].append(counts_total) bin_middle_array[model_number].append(np.arange(m_gal_low, m_gal_high+bin_width, bin_width)[:-1] + bin_width * 0.5) #### ## Plotting ## if rank == 0: # Plot only on rank 0. if paper_plot == 0: f = plt.figure() ax = plt.subplot(111) for model_number in range(0, len(SnapList)): for snapshot_idx in range(0, len(SnapList[model_number])): if model_number == 0: # We assume the redshifts for each model are the same, we only want to put a legend label for each redshift once. title = redshift_labels[model_number][snapshot_idx] else: title = '' plt.plot(bin_middle_array[model_number][snapshot_idx], counts_array[model_number][snapshot_idx] / normalization_array[model_number], color = PlotScripts.colors[snapshot_idx], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = title, linewidth = PlotScripts.global_linewidth) #print(np.min(np.log10(ResolutionLimit_mean))) #ax.axvline(np.max(np.log10(ResolutionLimit_mean)), color = 'k', linewidth = PlotScripts.global_linewidth, linestyle = '--') #ax.text(np.max(np.log10(ResolutionLimit_mean)) + 0.1, 1e-3, "Resolution Limit", color = 'k') for model_number in range(0, len(SnapList)): # Place legend labels for each of the models. NOTE: Placed after previous loop for proper formatting of labels. plt.plot(1e100, 1e100, color = 'k', linestyle = PlotScripts.linestyles[model_number], label = model_tags[model_number], rasterized=True, linewidth = PlotScripts.global_linewidth) ## Adjusting axis labels/limits. ## plt.yscale('log', nonposy='clip') plt.axis([6, 11.5, 1e-6, 1e-0]) ax.set_xlabel(r'$\log_{10}\ m_{\mathrm{*}} \:[M_{\odot}]$', fontsize = PlotScripts.global_fontsize) ax.set_ylabel(r'$\Phi\ [\mathrm{Mpc}^{-3}\: \mathrm{dex}^{-1}]$', fontsize = PlotScripts.global_fontsize) ax.xaxis.set_minor_locator(plt.MultipleLocator(0.25)) ax.set_xticks(np.arange(6.0, 12.0)) if (observations == 1 or observations == 3): # If we wanted to plot Song. Obs.Get_Data_SMF() delta = 0.05 caps = 5 ## Song (2016) Plotting ## plt.errorbar(Obs.Song_SMF_z6[:,0], 10**Obs.Song_SMF_z6[:,1], yerr= (10**Obs.Song_SMF_z6[:,1] - 10**Obs.Song_SMF_z6[:,3], 10**Obs.Song_SMF_z6[:,2] - 10**Obs.Song_SMF_z6[:,1]), xerr = 0.25, capsize = caps, elinewidth = PlotScripts.global_errorwidth, alpha = 1.0, lw=2.0, marker='o', ls='none', label = 'Song 2015, z = 6', color = PlotScripts.colors[0], rasterized=True) plt.errorbar(Obs.Song_SMF_z7[:,0], 10**Obs.Song_SMF_z7[:,1], yerr= (10**Obs.Song_SMF_z7[:,1] - 10**Obs.Song_SMF_z7[:,3], 10**Obs.Song_SMF_z7[:,2] - 10**Obs.Song_SMF_z7[:,1]), xerr = 0.25, capsize = caps, alpha=0.75, elinewidth = PlotScripts.global_errorwidth, lw=1.0, marker='o', ls='none', label = 'Song 2015, z = 7', color = PlotScripts.colors[1], rasterized=True) plt.errorbar(Obs.Song_SMF_z8[:,0], 10**Obs.Song_SMF_z8[:,1], yerr= (10**Obs.Song_SMF_z8[:,1] - 10**Obs.Song_SMF_z8[:,3], 10**Obs.Song_SMF_z8[:,2] - 10**Obs.Song_SMF_z8[:,1]), xerr = 0.25, capsize = caps, alpha=0.75, elinewidth = PlotScripts.global_errorwidth, lw=1.0, marker='o', ls='none', label = 'Song 2015, z = 8', color = PlotScripts.colors[2], rasterized=True) #### if ((observations == 2 or observations == 3) and rank == 0): # If we wanted to plot Baldry. Baldry_xval = np.log10(10 ** Obs.Baldry_SMF_z0[:, 0] /AllVars.Hubble_h/AllVars.Hubble_h) Baldry_xval = Baldry_xval - 0.26 # convert back to Chabrier IMF Baldry_yvalU = (Obs.Baldry_SMF_z0[:, 1]+Obs.Baldry_SMF_z0[:, 2]) * AllVars.Hubble_h*AllVars.Hubble_h*AllVars.Hubble_h Baldry_yvalL = (Obs.Baldry_SMF_z0[:, 1]-Obs.Baldry_SMF_z0[:, 2]) * AllVars.Hubble_h*AllVars.Hubble_h*AllVars.Hubble_h plt.fill_between(Baldry_xval, Baldry_yvalU, Baldry_yvalL, facecolor='purple', alpha=0.25, label='Baldry et al. 2008 (z=0.1)') #### leg = plt.legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile = './%s%s' %(output_tag, output_format) plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() if (paper_plot == 1): fig, ax = plt.subplots(nrows=1, ncols=3, sharex=False, sharey=True, figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 for model_number in range(0, len(SnapList)): for count in range(len(SnapList[model_number])): w = np.where((counts_array[model_number][count] > 0))[0] ax[count].plot(bin_middle_array[model_number][count][w], counts_array[model_number][count][w] / normalization_array[model_number], color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = r"$\mathbf{SAGE}$", linewidth = PlotScripts.global_linewidth) tick_locs = np.arange(6.0, 12.0) ax[count].set_xticklabels([r"$\mathbf{%d}$" % x for x in tick_locs], fontsize = PlotScripts.global_fontsize) ax[count].set_xlim([6.8, 10.3]) ax[count].tick_params(which = 'both', direction='in', width = PlotScripts.global_tickwidth) ax[count].tick_params(which = 'major', length = PlotScripts.global_ticklength) ax[count].tick_params(which = 'minor', length = PlotScripts.global_ticklength-2) ax[count].set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize - delta_fontsize) ax[count].xaxis.set_minor_locator(plt.MultipleLocator(0.25)) #ax[count].set_xticks(np.arange(6.0, 12.0)) for axis in ['top','bottom','left','right']: # Adjust axis thickness. ax[count].spines[axis].set_linewidth(PlotScripts.global_axiswidth) # Since y-axis is shared, only need to do this once. ax[0].set_yscale('log', nonposy='clip') ax[0].set_yticklabels([r"$\mathbf{10^{-5}}$",r"$\mathbf{10^{-5}}$",r"$\mathbf{10^{-4}}$", r"$\mathbf{10^{-3}}$", r"$\mathbf{10^{-2}}$",r"$\mathbf{10^{-1}}$"]) ax[0].set_ylim([1e-5, 1e-1]) #ax[0].set_ylabel(r'\mathbf{$\log_{10} \Phi\ [\mathrm{Mpc}^{-3}\: \mathrm{dex}^{-1}]}$', ax[0].set_ylabel(r'$\mathbf{log_{10} \: \Phi\ [Mpc^{-3}\: dex^{-1}]}$', fontsize = PlotScripts.global_labelsize - delta_fontsize) Obs.Get_Data_SMF() PlotScripts.Plot_SMF_z6(ax[0], errorwidth=ewidth, capsize=caps) PlotScripts.Plot_SMF_z7(ax[1], errorwidth=ewidth, capsize=caps) PlotScripts.Plot_SMF_z8(ax[2], errorwidth=ewidth, capsize=caps) #### ax[0].text(0.7, 0.9, r"$\mathbf{z = 6}$", transform = ax[0].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) ax[1].text(0.7, 0.9, r"$\mathbf{z = 7}$", transform = ax[1].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) ax[2].text(0.7, 0.9, r"$\mathbf{z = 8}$", transform = ax[2].transAxes, fontsize = PlotScripts.global_fontsize - delta_fontsize) #leg = ax[0,0].legend(loc=2, bbox_to_anchor = (0.2, -0.5), numpoints=1, labelspacing=0.1) leg = ax[0].legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize - 2) plt.tight_layout() outputFile = "{0}_paper{1}".format(output_tag, output_format) plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() ## def plot_fesc_galaxy(SnapList, PlotSnapList, simulation_norm, mean_galaxy_fesc, std_galaxy_fesc, N_galaxy_fesc, mean_halo_fesc, std_halo_fesc, N_halo_fesc, ResolutionLimit_mean, model_tags, paper_plots, mass_global, fesc_global, Ngamma_global, output_tag): """ Plots the escape fraction as a function of stellar/halo mass. Parallel compatible. Accepts 3D arrays of the escape fraction binned into Stellar Mass bins to plot the escape fraction for multiple models. Mass units are log(Msun) Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_galaxy_fesc, std_galaxy_fesc, N_galaxy_fesc : Nested 3-dimensional array, mean_galaxy_fesc[model_number0][snapshot0] = [bin0_meanfesc, ..., binN_meanfesc], with length equal to the number of models. Mean/Standard deviation for fesc in each stellar mass bin, for each [model_number] and [snapshot_number]. N_galaxy_fesc is the number of galaxies placed into each mass bin. mean_halo_fesc, std_halo_fesc, N_halo_fesc Nested 3-dimensional array, mean_halo_fesc[model_number0][snapshot0] = [bin0_meanfesc, ..., binN_meanfesc], with length equal to the number of models. Identical to previous except using the halo virial mass for the binning rather than stellar mass. ResolutionLimit_mean : array of floats with the same shape as mean_galaxy_fesc. This is the mean stellar mass for a halo with len (number of N-body simulation particles) between 'stellar_mass_halolen_lower' and 'stellar_mass_halolen_upper'. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. paper_plots: Integer. Flag to denote whether we should plot a full, 4 panel plot for the RSAGE paper. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are log(Msun). """ print("Plotting fesc as a function of stellar mass.") ## Array initialization ## master_mean_fesc_stellar, master_std_fesc_stellar, master_N_fesc_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_fesc, std_galaxy_fesc, N_galaxy_fesc, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: fig, ax = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) fig2, ax2 = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 count_x = 0 for count, model_number in enumerate(range(0, len(SnapList))): if count == 2: count_x += 1 print("There were a total of {0} galaxies over the entire redshift range.".format(sum(N_halo_fesc[model_number]))) ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## w = np.where((master_N_fesc_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_fesc_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: print(master_mean_fesc_stellar[model_number][snapshot_idx]) ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_fesc_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_fesc_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[0], rasterized = True, label = label, lw = PlotScripts.global_linewidth) #w = np.random.randint(0, # len(mass_global[model_number][snapshot_idx][0]), # size=500) #sc = ax2[count_x, count%2].scatter(mass_global[model_number][snapshot_idx][0][w], # fesc_global[model_number][snapshot_idx][0][w], # c=np.log10(Ngamma_global[model_number][snapshot_idx][0][w]*1.0e50), # alpha = 0.5,cmap='plasma') #plt.colorbar(sc) #ax2[count_x, count%2].hexbin(mass_global[model_number][snapshot_idx], # fesc_global[model_number][snapshot_idx], # C=Ngamma_global[model_number][snapshot_idx]) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: adjust_paper_plots(ax, model_tags) leg = ax[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) #leg = ax2[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) #leg.draw_frame(False) # Don't want a box frame #for t in leg.get_texts(): # Reduce the size of the text # t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) if paper_plots == 1: outputFile = './%s_scatter%s' %(output_tag, output_format) fig2.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig2) ## def plot_reionmod_galaxy(SnapList, PlotSnapList, simulation_norm, mean_galaxy_reionmod, std_galaxy_reionmod, N_galaxy_reionmod, mean_galaxy_reionmod_gnedin, std_galaxy_reionmod_gnedin, model_tags, paper_plots, output_tag): """ """ print("Reionization Modifier as a function of stellar mass.") ## Array initialization ## master_mean_reionmod_stellar, master_std_reionmod_stellar, master_N_reionmod_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_reionmod, std_galaxy_reionmod, N_galaxy_reionmod, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) master_mean_reionmod_gnedin_stellar, master_std_reionmod_gnedin_stellar, master_N_reionmod_gnedin_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_reionmod_gnedin, std_galaxy_reionmod_gnedin, N_galaxy_reionmod, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: fig, ax = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) fig2, ax2 = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 count_x = 0 for count, model_number in enumerate(range(0, len(SnapList))): if count == 2: count_x += 1 plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## w = np.where((master_N_reionmod_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_reionmod_stellar[model_number][snapshot_idx][w] = np.nan master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[0], rasterized = True, label = label, lw = PlotScripts.global_linewidth) ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[1], rasterized = True, label = label, lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break z_labels = [] for model_number in range(0, len(SnapList)): count_x = 0 plot_count = 0 for count, snapshot_idx in enumerate(range(len(SnapList[model_number]))): if count == 2: count_x += 1 if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): label = model_tags[model_number] if (model_number == 0): z_label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" z_labels.append(z_label) ## Plots as a function of stellar mass ## w = np.where((master_N_reionmod_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_reionmod_stellar[model_number][snapshot_idx][w] = np.nan master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx][w] = np.nan if (model_number == 0): print(master_mean_reionmod_stellar[model_number][snapshot_idx]) ax2[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_stellar[model_number][snapshot_idx], color = PlotScripts.colors[model_number], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) if (model_number == 0): ax2[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_reionmod_gnedin_stellar[model_number][snapshot_idx], color = 'k', ls = '--', rasterized = True, label = "Gnedin", lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: adjust_paper_plots(ax, model_tags) print(z_labels) adjust_redshift_panels(ax2, z_labels) leg = ax[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') leg = ax2[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) #leg = ax2[0,0].legend(loc="upper right", numpoints=1, labelspacing=0.1) #leg.draw_frame(False) # Don't want a box frame #for t in leg.get_texts(): # Reduce the size of the text # t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) ## Output ## outputFile = "{0}{1}".format(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) outputFile2 = "{0}_redshiftpanels{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ## def plot_nion_galaxy(SnapList, PlotSnapList, simulation_norm, mean_Ngamma_galaxy, std_Ngamma_galaxy, N_Ngamma_galaxy, model_tags, paper_plots, output_tag): """ Plots the number of ionizing photons emitted (not necessarily escaped) as a function of galaxy stellar mass. Parallel compatible. Accepts 3D arrays of the escape fraction binned into Stellar Mass bins to plot the escape fraction for multiple models. Mass units are log(Msun) Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_galaxy_Ngamma, std_galaxy_Ngamma, N_galaxy_Ngamma : Nested 3-dimensional array, mean_galaxy_Ngamma[model_number0][snapshot0] = [bin0_meanNgamma, ..., binN_meanNgamma], with length equal to the number of models. Mean/Standard deviation for Ngamma in each stellar mass bin, for each [model_number] and [snapshot_number]. N_galaxy_Ngamma is the number of galaxies placed into each mass bin. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. paper_plots: Integer. Flag to denote whether we should plot a full, 4 panel plot for the RSAGE paper. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are log(Msun). Ngamma units are 1.0e50 photons/s. """ print("Plotting Ngamma*fesc as a function of stellar mass.") ## Array initialization ## master_mean_Ngamma_stellar, master_std_Ngamma_stellar, master_N_Ngamma_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_Ngamma_galaxy, std_Ngamma_galaxy, N_Ngamma_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: fig, ax = plt.subplots(nrows=2, ncols=2, sharex='col', sharey='row', figsize=(16, 6)) delta_fontsize = 0 caps = 5 ewidth = 1.5 z_tags = np.zeros_like(model_tags, dtype=np.float32) for model_number in range(0, len(SnapList)): count_x = 0 ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for count, snapshot_idx in enumerate(range(0, len(SnapList[model_number]))): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if count == 2: count_x += 1 label = model_tags[model_number] z_tags[count] = float(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) ## Plots as a function of stellar mass ## w = np.where((master_N_Ngamma_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_Ngamma_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], np.log10(master_mean_Ngamma_stellar[model_number][snapshot_idx]*1.0e50), color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: ax[count_x, count%2].plot(master_bin_middle_stellar[model_number][snapshot_idx], np.log10(master_mean_Ngamma_stellar[model_number][snapshot_idx]*1.0e50), color = PlotScripts.colors[model_number], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: adjust_paper_plots(ax, z_tags) leg = ax[0,0].legend(loc="upper left", numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') plt.tight_layout() plt.subplots_adjust(wspace = 0.0, hspace = 0.0) ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) ## def plot_photo_galaxy(SnapList, PlotSnapList, simulation_norm, mean_photo_galaxy, std_photo_galaxy, N_photo_galaxy, model_tags, paper_plots, output_tag): """ Plots the photoionization rate as a function of galaxy stellar mass. Parallel compatible. Accepts 3D arrays of the escape fraction binned into Stellar Mass bins to plot the escape fraction for multiple models. Mass units are log(Msun) Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_photo_galaxy, std_photo_galaxy, N_photo_galaxy : Nested 3-dimensional array, mean_photo_galaxy[model_number0][snapshot0] = [bin0_meanphoto, ..., binN_meanphoto], with length equal to the number of models. Mean/Standard deviation for Photionization Rate in each stellar mass bin, for each [model_number] and [snapshot_number]. N_photo_galaxy is the number of galaxies placed into each mass bin. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. paper_plots: Integer. Flag to denote whether we should plot a full, 4 panel plot for the RSAGE paper. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are log(Msun). Ngamma units are 1.0e50 photons/s. """ print("Plotting photoionization rate as a function of stellar mass.") ## Array initialization ## master_mean_photo_stellar, master_std_photo_stellar, master_N_photo_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_photo_galaxy, std_photo_galaxy, N_photo_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: if paper_plots == 0: fig = plt.figure() ax1 = fig.add_subplot(111) else: pass for model_number in range(0, len(SnapList)): count_x = 0 ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for count, snapshot_idx in enumerate(range(0, len(SnapList[model_number]))): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## w = np.where((master_N_photo_stellar[model_number][snapshot_idx] < 4))[0] # If there are no galaxies in the bin we don't want to plot. master_mean_photo_stellar[model_number][snapshot_idx][w] = np.nan if paper_plots == 0: ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], np.log10(master_mean_photo_stellar[model_number][snapshot_idx]), color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) else: pass plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break for model_number in range(0, len(SnapList)): ax1.plot(np.nan, np.nan, color = 'k', label = model_tags[model_number], lw = PlotScripts.global_linewidth, ls = PlotScripts.linestyles[model_number]) ## Stellar Mass plots ## if paper_plots == 0: adjust_stellarmass_plot(ax1) else: pass ## Output ## outputFile = './%s%s' %(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) ## ## def plot_sfr_galaxy(SnapList, PlotSnapList, simulation_norm, mean_galaxy_sfr, std_galaxy_sfr, mean_galaxy_ssfr, std_galaxy_ssfr, N_galaxy, model_tags, output_tag): """ Plots the specific star formation rate (sSFR) as a function of stellar mass. Parallel compatible. Accepts 3D arrays of the sSFR binned into Stellar Mass bins. Mass units log(Msun). Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali mean_galaxy_ssfr, std_galaxy_ssfr, N_galaxy_ssfr : Nested 3-dimensional array, mean_galaxy_sfr[model_number0][snapshot0] = [bin0_meanssfr, ..., binN_meanssfr], with length equal to the number of models. Mean/Standard deviation for sSFR in each stellar mass bin, for each [model_number] and [snapshot_number]. N_galaxy_fesc is the number of galaxies placed into each mass bin. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Mass units are 1e10 Msun (no h). """ print("Plotting sSFR as a function of stellar mass.") ## Array initialization ## master_mean_sfr_stellar, master_std_sfr_stellar, master_N_sfr_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_sfr, std_galaxy_sfr, N_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) master_mean_ssfr_stellar, master_std_ssfr_stellar, master_N_ssfr_stellar, master_bin_middle_stellar = \ collect_across_tasks(mean_galaxy_ssfr, std_galaxy_ssfr, N_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) if rank == 0: fig = plt.figure() ax1 = fig.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(0, len(SnapList)): ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() plot_count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if (SnapList[model_number][snapshot_idx] == PlotSnapList[model_number][plot_count]): if (model_number == 0): label = r"$\mathbf{z = " + \ str(int(round(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]))) +\ "}$" else: label = "" ## Plots as a function of stellar mass ## ax1.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_sfr_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) ax2.plot(master_bin_middle_stellar[model_number][snapshot_idx], master_mean_ssfr_stellar[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], ls = PlotScripts.linestyles[model_number], rasterized = True, label = label, lw = PlotScripts.global_linewidth) plot_count += 1 if (plot_count == len(PlotSnapList[model_number])): break #for model_number in range(0, len(SnapList)): # Just plot some garbage to get the legend labels correct. #ax1.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) #ax3.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ## Stellar Mass plots ## adjust_sfr_plot(ax1) adjust_ssfr_plot(ax2) ## Output ## outputFile = "./{0}SFR{1}".format(output_tag, output_format) fig.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) outputFile = "./{0}sSFR{1}".format(output_tag, output_format) fig2.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig) ## ## def plot_ejectedfraction(SnapList, PlotSnapList, simulation_norm, mean_mvir_ejected, std_mvir_ejected, N_ejected, mean_ejected_z, std_ejected_z, N_z, model_tags, output_tag): ''' Plots the ejected fraction as a function of the halo mass. Parallel compatible. Accepts a 3D array of the ejected fraction so we can plot for multiple models and redshifts. Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. mean_mvir_ejected, std_mvir_ejected, N_ejected : Nested 3-dimensional array, mean_mvir_ejected[model_number0][snapshot0] = [bin0_meanejected, ..., binN_meanejected], with length equal to the number of models. Mean/Standard deviation for the escape fraction binned into Halo Mass bins. N_ejected is the number of data points in each bin. Bounds are given by 'm_low' and 'm_high' in bins given by 'bin_width'. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Halo Mass is in units of log10(Msun). ''' print("Plotting the Ejected Fraction as a function of halo mass.") master_mean_ejected_halo, master_std_ejected_halo, master_N_ejected_halo, master_bin_middle_halo = \ collect_across_tasks(mean_mvir_ejected, std_mvir_ejected, N_ejected, SnapList, PlotSnapList, True, m_low, m_high) master_mean_ejected_z, master_std_ejected_z, master_N_ejected_z, _ = \ collect_across_tasks(mean_ejected_z, std_ejected_z, N_z, SnapList) if rank == 0: fig1 = plt.figure() ax1 = fig1.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(0, len(SnapList)): if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() for snapshot_idx in range(0, len(PlotSnapList[model_number])): label = AllVars.SnapZ[PlotSnapList[model_number][snapshot_idx]] ax1.plot(master_bin_middle_halo[model_number][snapshot_idx], master_mean_ejected_halo[model_number][snapshot_idx], color = PlotScripts.colors[snapshot_idx], linestyle = PlotScripts.linestyles[model_number], label = label, lw = PlotScripts.global_linewidth) ax2.plot((AllVars.t_BigBang - AllVars.Lookback_Time[SnapList[model_number]]) * 1.0e3, master_mean_ejected_z[model_number], color = PlotScripts.colors[model_number], label = model_tags[model_number], ls = PlotScripts.linestyles[model_number], lw = PlotScripts.global_linewidth) for model_number in range(0, len(SnapList)): # Just plot some garbage to get the legend labels correct. ax1.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ax1.set_xlabel(r'$\log_{10}\ M_{\mathrm{vir}}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$\mathrm{Ejected \: Fraction}$', size = PlotScripts.global_fontsize) ax1.set_xlim([8.0, 12]) ax1.set_ylim([-0.05, 1.0]) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(0.1)) ax1.yaxis.set_minor_locator(mtick.MultipleLocator(0.025)) leg = ax1.legend(loc=1, numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') outputFile = "./{0}{1}".format(output_tag, output_format) fig1.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close(fig1) ax2.set_xlabel(r"$\mathbf{Time \: since \: Big \: Bang \: [Myr]}$", fontsize = PlotScripts.global_labelsize) tick_locs = np.arange(200.0, 1000.0, 100.0) tick_labels = [r"$\mathbf{%d}$" % x for x in tick_locs] ax2.xaxis.set_major_locator(mtick.MultipleLocator(100)) ax2.set_xticklabels(tick_labels, fontsize = PlotScripts.global_fontsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_ylabel(r'$\mathbf{Mean f_{ej}}$', fontsize = PlotScripts.global_labelsize) ax3 = ax2.twiny() t_plot = (AllVars.t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$\mathbf{%d}$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax3.set_xlabel(r"$\mathbf{z}$", fontsize = PlotScripts.global_labelsize) ax3.set_xlim(PlotScripts.time_xlim) ax3.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax3.set_xticklabels(z_labels, fontsize = PlotScripts.global_fontsize) # But label them as redshifts. leg = ax2.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile2 = "./{0}_z{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ## ## def plot_mvir_Ngamma(SnapList, mean_mvir_Ngamma, std_mvir_Ngamma, N_Ngamma, model_tags, output_tag,fesc_prescription=None, fesc_normalization=None, fitpath=None): ''' Plots the number of ionizing photons (pure ngamma times fesc) as a function of halo mass. Parallel compatible. The input data has been binned as a function of halo virial mass (Mvir), with the bins defined at the top of the file (m_low, m_high, bin_width). Accepts 3D arrays to plot ngamma for multiple models. Parameters ---------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model. mean_mvir_Ngamma, std_mvir_Ngamma, N_Ngamma : Nested 2-dimensional array, mean_mvir_Ngamma[model_number0][snapshot0] = [bin0_meanNgamma, ..., binN_meanNgamma], with length equal to the number of bins. Mean/Standard deviation/number of data points in each halo mass (Mvir) bin. The number of photons is in units of 1.0e50 s^-1. model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. fesc_prescription : int (optional) If this parameter is defined, we will save the Mvir-Ngamma results in a text file (not needed if not saving). Number that controls what escape fraction prescription was used to generate the escape fractions. 0 : Constant, fesc = Constant. 1 : Scaling with Halo Mass, fesc = A*Mh^B. 2 : Scaling with ejected fraction, fesc = fej*A + B. fesc_normalization : float (if fesc_prescription == 0) or `numpy.darray' with length 2 (if fesc_prescription == 1 or == 2) (optional). If this parameter is defined, we will save the Mvir-Ngamma results in a text file (not needed if not saving). Parameter not needed if you're not saving the Mvir-Ngamma results. If fesc_prescription == 0, gives the constant value for the escape fraction. If fesc_prescription == 1 or == 2, gives A and B with the form [A, B]. fitpath : string (optional) If this parameter is defined, we will save the Mvir-Ngamma results in a text file (not needed if not saving). Defines the base path for where we are saving the results. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- Ngamma is in units of 1.0e50 s^-1. ''' print("Plotting ngamma*fesc against the halo mass") ## Array initialization. ## title = [] redshift_labels = [] mean_ngammafesc_array = [] std_ngammafesc_array = [] mean_halomass_array = [] std_halomass_array = [] bin_middle_array = [] for model_number in range(0, len(SnapList)): redshift_labels.append([]) mean_ngammafesc_array.append([]) std_ngammafesc_array.append([]) mean_halomass_array.append([]) std_halomass_array.append([]) bin_middle_array.append([]) for model_number in range(0, len(SnapList)): for snapshot_idx in range(0, len(SnapList[model_number])): print("Doing Snapshot {0}".format(SnapList[model_number][snapshot_idx])) tmp = 'z = %.2f' %(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) redshift_labels[model_number].append(tmp) N = N_Ngamma[model_number][snapshot_idx] mean_ngammafesc_array[model_number], std_ngammafesc_array[model_number] = calculate_pooled_stats(mean_ngammafesc_array[model_number], std_ngammafesc_array[model_number], mean_mvir_Ngamma[model_number][snapshot_idx], std_mvir_Ngamma[model_number][snapshot_idx], N) # Collate the values from all processors. bin_middle_array[model_number].append(np.arange(m_low, m_high+bin_width, bin_width)[:-1] + bin_width * 0.5) if rank == 0: f = plt.figure() ax1 = plt.subplot(111) for model_number in range(0, len(SnapList)): count = 0 for snapshot_idx in range(0, len(SnapList[model_number])): if model_number == 0: title = redshift_labels[model_number][snapshot_idx] else: title = '' mean = np.zeros((len(mean_ngammafesc_array[model_number][snapshot_idx])), dtype = np.float32) std = np.zeros((len(mean_ngammafesc_array[model_number][snapshot_idx])), dtype=np.float32) for i in range(0, len(mean)): if(mean_ngammafesc_array[model_number][snapshot_idx][i] < 1e-10): mean[i] = np.nan std[i] = np.nan else: mean[i] = np.log10(mean_ngammafesc_array[model_number][snapshot_idx][i] * 1.0e50) # Remember that the input data is in units of 1.0e50 s^-1. std[i] = 0.434 * std_ngammafesc_array[model_number][snapshot_idx][i] / mean_ngammafesc_array[model_number][snapshot_idx][i] # We're plotting in log space so the standard deviation is 0.434*log10(std)/log10(mean). bin_middle = bin_middle_array[model_number][snapshot_idx] if (count < 4): # Only plot at most 5 lines. ax1.plot(bin_middle, mean, color = PlotScripts.colors[snapshot_idx], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = title, linewidth = PlotScripts.global_linewidth) count += 1 ## In this block we save the Mvir-Ngamma results to a file. ## if (fesc_prescription == None or fesc_normalization == None or fitpath == None): raise ValueError("You've specified you want to save the Mvir-Ngamma results but haven't provided an escape fraction prescription, normalization and base path name") # Note: All the checks that escape fraction normalization was written correctly were performed in 'calculate_fesc()', hence it will be correct by this point and we don't need to double check. if (fesc_prescription[model_number] == 0): # Slightly different naming scheme for the constant case (it only has a float for fesc_normalization). fname = "%s/fesc%d_%.3f_z%.3f.txt" %(fitpath, fesc_prescription[model_number], fesc_normalization[model_number], AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) elif (fesc_prescription[model_number] == 1 or fesc_prescription[model_number] == 2): fname = "%s/fesc%d_A%.3eB%.3f_z%.3f.txt" %(fitpath, fesc_prescription[model_number], fesc_normalization[model_number][0], fesc_normalization[model_number][1], AllVars.SnapZ[SnapList[model_number][snapshot_idx]]) f = open(fname, "w+") if not os.access(fname, os.W_OK): print("The filename is {0}".format(fname)) raise ValueError("Can't write to this file.") for i in range(0, len(bin_middle)): f.write("%.4f %.4f %.4f %d\n" %(bin_middle[i], mean[i], std[i], N_Ngamma[model_number][snapshot_idx][i])) f.close() print("Wrote successfully to file {0}".format(fname)) ## for model_number in range(0, len(SnapList)): # Just plot some garbage to get the legend labels correct. ax1.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) ax1.set_xlabel(r'$\log_{10}\ M_{\mathrm{vir}}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$\log_{10}\ \dot{N}_\gamma \: f_\mathrm{esc} \: [\mathrm{s}^{-1}]$', size = PlotScripts.global_fontsize) ax1.set_xlim([8.5, 12]) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(0.1)) leg = ax1.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize('medium') outputFile = './' + output_tag + output_format plt.savefig(outputFile, bbox_inches='tight') # Save the figure print('Saved file to'.format(outputFile)) plt.close() def plot_photoncount(SnapList, sum_nion, simulation_norm, FirstFile, LastFile, NumFiles, model_tags, output_tag): ''' Plots the ionizing emissivity as a function of redshift. We normalize the emissivity to Mpc^-3 and this function allows the read-in of only a subset of the volume. Parallel compatible. Parameters --------- SnapList : Nested array, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots for each model, defines the x-axis we plot against. sum_nion : Nested 1-dimensional array, sum_nion[z0, z1, ..., zn], with length equal to the number of redshifts. Number of escape ionizing photons (i.e., photon rate times the local escape fraction) at each redshift. In units of 1.0e50 s^-1. simulation_norm : array of ints with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation FirstFile, LastFile, NumFile : array of integers with length equal to the number of models. The file numbers for each model that were read in (defined by the range between [FirstFile, LastFile] inclusive) and the TOTAL number of files for this model (we may only be plotting a subset of the volume). model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- sum_nion is in units of 1.0e50 s^-1. ''' print("Plotting the ionizing emissivity.") sum_array = [] for model_number in range(0, len(SnapList)): if(simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") sum_array.append([]) for snapshot_idx in range(0, len(SnapList[model_number])): nion_sum_snapshot = comm.reduce(sum_nion[model_number][snapshot_idx], op = MPI.SUM, root = 0) if rank == 0: sum_array[model_number].append(nion_sum_snapshot * 1.0e50 / (pow(AllVars.BoxSize / AllVars.Hubble_h,3) * (float(LastFile[model_number] - FirstFile[model_number] + 1) / float(NumFiles[model_number])))) if (rank == 0): ax1 = plt.subplot(111) for model_number in range(0, len(SnapList)): if(simulation_norm[model_number] == 0): cosmo = AllVars.Set_Params_Mysim() if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") t = np.empty(len(SnapList[model_number])) for snapshot_idx in range(0, len(SnapList[model_number])): t[snapshot_idx] = (AllVars.t_BigBang - cosmo.lookback_time(AllVars.SnapZ[SnapList[model_number][snapshot_idx]]).value) * 1.0e3 t = [t for t, N in zip(t, sum_array[model_number]) if N > 1.0] sum_array[model_number] = [x for x in sum_array[model_number] if x > 1.0] print("The total number of ionizing photons for model {0} is {1} s^1 Mpc^-3".format(model_number, sum(sum_array[model_number]))) print(np.log10(sum_array[model_number])) ax1.plot(t, np.log10(sum_array[model_number]), color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[model_number], label = model_tags[model_number], linewidth = PlotScripts.global_linewidth) #ax1.fill_between(t, np.subtract(mean,std), np.add(mean,std), color = colors[model_number], alpha = 0.25) ax1.xaxis.set_minor_locator(mtick.MultipleLocator(PlotScripts.time_tickinterval)) #ax1.yaxis.set_minor_locator(mtick.MultipleLocator(0.025)) ax1.set_xlim(PlotScripts.time_xlim) ax1.set_ylim([48.5, 51.5]) ax2 = ax1.twiny() t_plot = (AllVars.t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$%d$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax2.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax2.set_xticklabels(z_labels) # But label them as redshifts. ax1.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_fontsize) ax1.set_ylabel(r'$\sum f_\mathrm{esc}\dot{N}_\gamma \: [\mathrm{s}^{-1}\mathrm{Mpc}^{-3}]$', fontsize = PlotScripts.global_fontsize) plot_time = 1 bouwens_z = np.arange(6,16) # Redshift range for the observations. bouwens_t = (AllVars.t_BigBang - cosmo.lookback_time(bouwens_z).value) * 1.0e3 # Corresponding values for what we will plot on the x-axis. bouwens_1sigma_lower = [50.81, 50.73, 50.60, 50.41, 50.21, 50.00, 49.80, 49.60, 49.39, 49.18] # 68% Confidence Intervals for the ionizing emissitivity from Bouwens 2015. bouwens_1sigma_upper = [51.04, 50.85, 50.71, 50.62, 50.56, 50.49, 50.43, 50.36, 50.29, 50.23] bouwens_2sigma_lower = [50.72, 50.69, 50.52, 50.27, 50.01, 49.75, 49.51, 49.24, 48.99, 48.74] # 95% CI. bouwens_2sigma_upper = [51.11, 50.90, 50.74, 50.69, 50.66, 50.64, 50.61, 50.59, 50.57, 50.55] if plot_time == 1: ax1.fill_between(bouwens_t, bouwens_1sigma_lower, bouwens_1sigma_upper, color = 'k', alpha = 0.2) ax1.fill_between(bouwens_t, bouwens_2sigma_lower, bouwens_2sigma_upper, color = 'k', alpha = 0.4, label = r"$\mathrm{Bouwens \: et \: al. \: (2015)}$") else: ax1.fill_between(bouwens_z, bouwens_1sigma_lower, bouwens_1sigma_upper, color = 'k', alpha = 0.2) ax1.fill_between(bouwens_z, bouwens_2sigma_lower, bouwens_2sigma_upper, color = 'k', alpha = 0.4, label = r"$\mathrm{Bouwens \: et \: al. \: (2015)}$") # ax1.text(0.075, 0.965, '(a)', horizontalalignment='center', verticalalignment='center', transform = ax.transAxes) ax1.text(350, 50.0, r"$68\%$", horizontalalignment='center', verticalalignment = 'center', fontsize = PlotScripts.global_labelsize) ax1.text(350, 50.8, r"$95\%$", horizontalalignment='center', verticalalignment = 'center', fontsize = PlotScripts.global_labelsize) leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) plt.tight_layout() outputFile = './{0}{1}'.format(output_tag, output_format) plt.savefig(outputFile) # Save the figure print('Saved file to {0}'.format(outputFile)) plt.close() ## ## def plot_quasars_count(SnapList, PlotList, N_quasars_z, N_quasars_boost_z, N_gal_z, mean_quasar_activity, std_quasar_activity, N_halo, N_merger_halo, N_gal, N_merger_galaxy, fesc_prescription, simulation_norm, FirstFile, LastFile, NumFile, model_tags, output_tag): ''' Parameters --------- SnapList : Nested 'array-like` of ints, SnapList[model_number0] = [snapshot0_model0, ..., snapshotN_model0], with length equal to the number of models. Snapshots that we plot the quasar density at for each model. PlotList : Nested array of ints, PlotList[model_number0]= [plotsnapshot0_model0, ..., plotsnapshotN_model0], with length equal to the number of models. Snapshots that will be plotted for the quasar activity as a function of halo mass. N_quasars_z : Nested array of floats, N_quasars_z[model_number0] = [N_quasars_z0, N_quasars_z1, ..., N_quasars_zN]. Outer array has length equal to the number of models, inner array has length equal to length of the model's SnapList. Number of quasars, THAT WENT OFF, during the given redshift. N_quasars_boost_z : Nested array of floats, N_quasars_boost_z[model_number0] = [N_quasars_boost_z0, N_quasars_boost_z1, ..., N_quasars_boost_zN]. Outer array has length equal to the number of models, inner array has length equal to length of the model's SnapList. Number of galaxies that had their escape fraction boosted by quasar activity. N_gal_z : Nested array of floats, N_gal_z[model_number0] = [N_gal_z0, N_gal_z1, ..., N_gal_zN]. Outer array has length equal to the number of models, inner array has length equal to length of the model's SnapList. Number of galaxies at each redshift. mean_quasar_activity, std_quasar_activity : Nested 2-dimensional array of floats, mean_quasar_activity[model_number0][snapshot0] = [bin0quasar_activity, ..., binNquasar_activity]. Outer array has length equal to the number of models, inner array has length equal to the length of the model's snaplist and most inner array has length equal to the number of halo bins (NB). Mean/std fraction of galaxies that had quasar go off during each snapshot as a function of halo mass. NOTE : This is for quasars going off, not for galaxies that have their escape fraction being boosted. fesc_prescription : Array with length equal to the number of models. Denotes what escape fraction prescription each model used. Quasars are only tracked when fesc_prescription == 3. simulation_norm : array with length equal to the number of models. Denotes which simulation each model uses. 0 : MySim 1 : Mini-Millennium 2 : Tiamat (down to z = 5) 3 : Extended Tiamat (down to z = 1.6ish). 4 : Britton's Simulation 5 : Kali FirstFile, LastFile, NumFile : array of integers with length equal to the number of models. The file numbers for each model that were read in (defined by the range between [FirstFile, LastFile] inclusive) and the TOTAL number of files for this model (we may only be plotting a subset of the volume). model_tags : array of strings with length equal to the number of models. Strings that contain the tag for each model. Will be placed on the plot. output_tag : string Name of the file that will be generated. File will be saved in the current directory with the output format defined by the 'output_format' variable at the beggining of the file. Returns ------- No returns. Generates and saves the plot (named via output_tag). Units ----- No relevant units. ''' print("Plotting quasar count/density") if rank == 0: fig = plt.figure() ax1 = fig.add_subplot(111) ax6 = ax1.twinx() fig2 = plt.figure() ax3 = fig2.add_subplot(111) ax5 = ax3.twinx() fig3 = plt.figure() ax7 = fig3.add_subplot(111) fig4 = plt.figure() ax50 = fig4.add_subplot(111) fig5 = plt.figure() ax55 = fig5.add_subplot(111) fig6 = plt.figure() ax56 = fig6.add_subplot(111) mean_quasar_activity_array = [] std_quasar_activity_array = [] N_quasar_activity_array = [] N_gal_halo_array = [] N_gal_array = [] merger_counts_halo_array = [] merger_counts_galaxy_array = [] bin_middle_halo_array = [] bin_middle_galaxy_array = [] for model_number in range(0, len(SnapList)): # Does this for each of the models. if (fesc_prescription[model_number] != 3): # Want to skip the models that didn't count quasars. continue ## Normalization for each model. ## if (simulation_norm[model_number] == 0): AllVars.Set_Params_Mysim() elif (simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif (simulation_norm[model_number] == 2): AllVars.Set_Params_Tiamat() elif (simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif (simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif (simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() mean_quasar_activity_array.append([]) std_quasar_activity_array.append([]) N_quasar_activity_array.append([]) N_gal_halo_array.append([]) N_gal_array.append([]) merger_counts_halo_array.append([]) merger_counts_galaxy_array.append([]) bin_middle_halo_array.append([]) bin_middle_galaxy_array.append([]) box_factor = (LastFile[model_number] - FirstFile[model_number] + 1.0)/(NumFile[model_number]) # This factor allows us to take a sub-volume of the box and scale the results to represent the entire box. print("We are plotting the quasar density using {0:.4f} of the box's volume.".format(box_factor)) norm = pow(AllVars.BoxSize,3) / pow(AllVars.Hubble_h, 3) * box_factor #### ## We perform the plotting on Rank 0 so only this rank requires the final counts array. ## if rank == 0: quasars_total = np.zeros_like((N_quasars_z[model_number])) boost_total = np.zeros_like(N_quasars_boost_z[model_number]) gal_count_total = np.zeros_like(N_gal_z[model_number]) else: quasars_total = None boost_total = None gal_count_total = None N_quasars_tmp = np.array((N_quasars_z[model_number])) # So we can use MPI.Reduce() comm.Reduce([N_quasars_tmp, MPI.DOUBLE], [quasars_total, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the number of quasars and passes back to rank 0. N_quasars_boost_tmp = np.array(N_quasars_boost_z[model_number]) # So we can use MPI.Reduce() comm.Reduce([N_quasars_boost_tmp, MPI.DOUBLE], [boost_total, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the number of galaxies that had their fesc boosted. N_gal_tmp = np.array(N_gal_z[model_number]) # So we can use MPI.Reduce() comm.Reduce([N_gal_tmp, MPI.DOUBLE], [gal_count_total, MPI.DOUBLE], op = MPI.SUM, root = 0) # Sum the number of total galaxies. for snapshot_idx in range(len(SnapList[model_number])): mean_quasar_activity_array[model_number], std_quasar_activity_array[model_number], N_quasar_activity_array[model_number] = calculate_pooled_stats(mean_quasar_activity_array[model_number], std_quasar_activity_array[model_number], N_quasar_activity_array[model_number], mean_quasar_activity[model_number][snapshot_idx], std_quasar_activity[model_number][snapshot_idx], N_halo[model_number][snapshot_idx]) if rank == 0: merger_count_halo_total = np.zeros_like((N_merger_halo[model_number][snapshot_idx])) N_gal_halo_total = np.zeros_like((N_halo[model_number][snapshot_idx])) merger_count_galaxy_total = np.zeros_like((N_merger_galaxy[model_number][snapshot_idx])) N_gal_total = np.zeros_like((N_gal[model_number][snapshot_idx])) else: merger_count_halo_total = None N_gal_halo_total = None merger_count_galaxy_total = None N_gal_total = None comm.Reduce([N_merger_halo[model_number][snapshot_idx], MPI.FLOAT], [merger_count_halo_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. comm.Reduce([N_halo[model_number][snapshot_idx], MPI.FLOAT], [N_gal_halo_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. comm.Reduce([N_merger_galaxy[model_number][snapshot_idx], MPI.FLOAT], [merger_count_galaxy_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. comm.Reduce([N_gal[model_number][snapshot_idx], MPI.FLOAT], [N_gal_total, MPI.FLOAT], op = MPI.SUM, root = 0) # Sum all the stellar mass and pass to Rank 0. if rank == 0: merger_counts_halo_array[model_number].append(merger_count_halo_total) N_gal_halo_array[model_number].append(N_gal_halo_total) merger_counts_galaxy_array[model_number].append(merger_count_galaxy_total) N_gal_array[model_number].append(N_gal_total) bin_middle_halo_array[model_number].append(np.arange(m_low, m_high+bin_width, bin_width)[:-1] + bin_width * 0.5) bin_middle_galaxy_array[model_number].append(np.arange(m_gal_low, m_gal_high+bin_width, bin_width)[:-1] + bin_width * 0.5) if rank == 0: plot_count = 0 stop_plot = 0 title = model_tags[model_number] t = np.empty(len(SnapList[model_number])) ZZ = np.empty(len(SnapList[model_number])) for snapshot_idx in range(0, len(SnapList[model_number])): t[snapshot_idx] = (AllVars.t_BigBang - AllVars.Lookback_Time[SnapList[model_number][snapshot_idx]]) * 1.0e3 ZZ[snapshot_idx] = AllVars.SnapZ[SnapList[model_number][snapshot_idx]] if (stop_plot == 0): # print("Snapshot {0} PlotSnapshot " #"{1}".format(SnapList[model_number][snapshot_idx], PlotList[model_number][plot_count])) if (SnapList[model_number][snapshot_idx] == PlotList[model_number][plot_count]): label = "z = {0:.2f}".format(AllVars.SnapZ[PlotList[model_number][plot_count]]) ax7.plot(bin_middle_halo_array[model_number][snapshot_idx], mean_quasar_activity_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) #ax50.plot(bin_middle_halo_array[model_number][snapshot_idx], merger_counts_array[model_number][snapshot_idx] / gal_count_total[snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) ax50.plot(bin_middle_halo_array[model_number][snapshot_idx], merger_counts_halo_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) #ax50.plot(bin_middle_halo_array[model_number][snapshot_idx], merger_counts_array[model_number][snapshot_idx] / N_gal_halo_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) #ax55.plot(bin_middle_galaxy_array[model_number][snapshot_idx], merger_counts_galaxy_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) ax55.plot(bin_middle_galaxy_array[model_number][snapshot_idx], merger_counts_galaxy_array[model_number][snapshot_idx] / N_gal_array[model_number][snapshot_idx], color = PlotScripts.colors[plot_count], linestyle = PlotScripts.linestyles[model_number], rasterized = True, label = label, linewidth = PlotScripts.global_linewidth) print("plot_count = {0} len(PlotList) = {1}".format(plot_count, len(PlotList[model_number]))) plot_count += 1 print("plot_count = {0} len(PlotList) = {1}".format(plot_count, len(PlotList[model_number]))) if (plot_count == len(PlotList[model_number])): stop_plot = 1 print("For Snapshot {0} at t {3} there were {1} total mergers compared to {2} total galaxies.".format(snapshot_idx, np.sum(merger_counts_galaxy_array[model_number][snapshot_idx]), np.sum(gal_count_total[snapshot_idx]), t[snapshot_idx])) if (np.sum(gal_count_total[snapshot_idx]) > 0.0 and np.sum(merger_counts_galaxy_array[model_number][snapshot_idx]) > 0.0): ax56.scatter(t[snapshot_idx], np.sum(merger_counts_galaxy_array[model_number][snapshot_idx]) / np.sum(gal_count_total[snapshot_idx]), color = 'r', rasterized = True) #ax56.scatter(t[snapshot_idx], quasars_total[snapshot_idx] / np.sum(gal_count_total[snapshot_idx]), color = 'r', rasterized = True) ax1.plot(t, quasars_total / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[0], rasterized = True, linewidth = PlotScripts.global_linewidth) p = np.where((ZZ < 15))[0] #ax1.plot(ZZ[p], quasars_total[p] / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[0], rasterized = True, linewidth = PlotScripts.global_linewidth) ax3.plot(t, boost_total, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[0], rasterized = True, label = title, linewidth = PlotScripts.global_linewidth) w = np.where((gal_count_total > 0.0))[0] # Since we're doing a division, need to only plot those redshifts that actually have galaxies. ax5.plot(t[w], np.divide(boost_total[w], gal_count_total[w]), color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[1], rasterized = True, linewidth = PlotScripts.global_linewidth) ax6.plot(t[w], gal_count_total[w] / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[1], rasterized = True, linewidth = PlotScripts.global_linewidth) #ax6.plot(ZZ[p], gal_count_total[p] / norm, color = PlotScripts.colors[model_number], linestyle = PlotScripts.linestyles[1], rasterized = True, linewidth = PlotScripts.global_linewidth) ax1.plot(np.nan, np.nan, color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[0], label = "Quasar Ejection Density") ax1.plot(np.nan, np.nan, color = PlotScripts.colors[0], linestyle = PlotScripts.linestyles[1], label = "Galaxy Density") ax3.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[0], label = "Count") ax3.plot(np.nan, np.nan, color = 'k', linestyle = PlotScripts.linestyles[1], label = "Fraction of Galaxies") ax7.set_xlabel(r'$\log_{10}\ M_\mathrm{vir}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax7.set_ylabel(r'$\mathrm{Mean \: Quasar \: Activity}$', size = PlotScripts.global_fontsize) ax50.set_xlabel(r'$\log_{10}\ M_\mathrm{vir}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) #ax50.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) ax50.set_ylabel(r'$\mathrm{Number \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) ax55.set_xlabel(r'$\log_{10}\ M_\mathrm{*}\ [M_{\odot}]$', size = PlotScripts.global_fontsize) ax55.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) #ax55.set_ylabel(r'$\mathrm{Number \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) ax56.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_labelsize) ax56.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Undergoing \: Merger}$', size = PlotScripts.global_fontsize) #ax56.set_ylabel(r'$\mathrm{Fraction \: Galaxies \: Quasar \: Activity}$', size = PlotScripts.global_fontsize) ax56.set_yscale('log', nonposy='clip') ax50.axvline(np.log10(32.0*AllVars.PartMass / AllVars.Hubble_h), color = 'k', linewidth = PlotScripts.global_linewidth, linestyle = '-.') ax1.xaxis.set_minor_locator(mtick.MultipleLocator(PlotScripts.time_tickinterval)) ax1.set_xlim(PlotScripts.time_xlim) ax1.set_yscale('log', nonposy='clip') ax3.xaxis.set_minor_locator(mtick.MultipleLocator(PlotScripts.time_tickinterval)) ax3.set_xlim(PlotScripts.time_xlim) ax3.set_yscale('log', nonposy='clip') ## Create a second axis at the top that contains the corresponding redshifts. ## ## The redshift defined in the variable 'z_plot' will be displayed. ## ax2 = ax1.twiny() ax4 = ax3.twiny() ax57 = ax56.twiny() t_plot = (AllVars.t_BigBang - AllVars.cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding time values on the bottom. z_labels = ["$%d$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax2.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax2.set_xlim(PlotScripts.time_xlim) ax2.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax2.set_xticklabels(z_labels) # But label them as redshifts. ax4.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax4.set_xlim(PlotScripts.time_xlim) ax4.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax4.set_xticklabels(z_labels) # But label them as redshifts. ax57.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax57.set_xlim(PlotScripts.time_xlim) ax57.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax57.set_xticklabels(z_labels) # But label them as redshifts. ax1.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_labelsize) #ax1.set_xlabel(r"$z$", size = PlotScripts.global_labelsize) ax1.set_ylabel(r'$N_\mathrm{Quasars} \: [\mathrm{Mpc}^{-3}]$', fontsize = PlotScripts.global_fontsize) ax6.set_ylabel(r'$N_\mathrm{Gal} \: [\mathrm{Mpc}^{-3}]$', fontsize = PlotScripts.global_fontsize) ax3.set_xlabel(r"$\mathrm{Time \: Since \: Big \: Bang \: [Myr]}$", size = PlotScripts.global_labelsize) ax3.set_ylabel(r'$N_\mathrm{Boosted}$', fontsize = PlotScripts.global_fontsize) ax5.set_ylabel(r'$\mathrm{Fraction \: Boosted}$', fontsize = PlotScripts.global_fontsize) leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax3.legend(loc='lower left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax7.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax50.legend(loc='upper right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) leg = ax55.legend(loc='upper right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) fig.tight_layout() fig2.tight_layout() fig3.tight_layout() fig5.tight_layout() fig6.tight_layout() outputFile1 = './{0}_quasardensity{1}'.format(output_tag, output_format) outputFile2 = './{0}_boostedcount{1}'.format(output_tag, output_format) outputFile3 = './{0}_quasar_activity_halo{1}'.format(output_tag, output_format) outputFile4 = './{0}_mergercount_global{1}'.format(output_tag, output_format) outputFile5 = './{0}_mergercount_global_stellarmass{1}'.format(output_tag, output_format) outputFile6 = './{0}_mergercount_total{1}'.format(output_tag, output_format) fig.savefig(outputFile1) # Save the figure fig2.savefig(outputFile2) # Save the figure fig3.savefig(outputFile3) # Save the figure fig4.savefig(outputFile4) # Save the figure fig5.savefig(outputFile5) # Save the figure fig6.savefig(outputFile6) # Save the figure print("Saved to {0}".format(outputFile1)) print("Saved to {0}".format(outputFile2)) print("Saved to {0}".format(outputFile3)) print("Saved to {0}".format(outputFile4)) print("Saved to {0}".format(outputFile5)) print("Saved to {0}".format(outputFile6)) plt.close(fig) plt.close(fig2) plt.close(fig3) ## ### ### ### ### def plot_reionmod(PlotSnapList, SnapList, simulation_norm, mean_reionmod_halo, std_reionmod_halo, N_halo, mean_reionmod_z, std_reionmod_z, N_reionmod, plot_z, model_tags, output_tag): """ Plot the reionization modifier as a function of halo mass and redshift. Parameters ---------- PlotSnapList, SnapList: 2D Nested arrays of integers. Outer length is equal to the number of models and inner length is number of snapshots we're plotting/calculated for. PlotSnapList contains the snapshots for each model we will plot for the halo mass figure. SnapList contains the snapshots for each model that we have performed calculations for. These aren't equal because we don't want to plot halo curves for ALL redshifts. simulation_norm: Array of integers. Length is equal to the number of models. Contains the simulation identifier for each model. Used to set the parameters of each model. mean_reionmod_halo, std_reionmod_halo: 3D Nested arrays of floats. Most outer length is equal to the number of models, next length is number of snapshots for each model, then inner-most length is the number of halo mass- bins (given by NB). Contains the mean/standard deviation values for the reionization modifier as a function of halo mass. NOTE: These are unique for each task. N_halo: 3D Nested arrays of floats. Lengths are identical to mean_reionmod_halo. Contains the number of halos in each halo mass bin. NOTE: These are unique for each task. mean_reionmod_z, std_reionmod_z: 2D Nested arrays of floats. Outer length is equal to the number of models, inner length is the number of snapshots for each model. NOTE: This inner length can be different to the length of PlotSnapList as we don't necessarily need to plot for every snapshot we calculate. Contains the mean/standard deviation values for the rieonization modifier as a function of redshift. NOTE: These are unique for each task. N_reionmod: 2D Nested arrays of floats. Lengths are identical to mean_reionmod_z. Contains the number of galaxies at each redshift that have non-negative reionization modifier. A negative reionization modifier is a galaxy who didn't have infall/stripping during the snapshot. NOTE: These are unique for each task. plot_z: Boolean. Denotes whether we want to plot the reionization modifier as a function of redshift. Useful because we often only calculate statistics for a subset of the snapshots to decrease computation time. For these runs, we don't want to plot for something that requires ALL snapshots. model_tags: Array of strings. Length is equal to the number of models. Contains the legend labels for each model. output_tag: String. The prefix for the output file. Returns ---------- None. Plot is saved in current directory as "./<output_tag>.<output_format>" """ master_mean_reionmod_halo, master_std_reionmod_halo, master_N_reionmod_halo, master_bin_middle = collect_across_tasks(mean_reionmod_halo, std_reionmod_halo, N_halo, SnapList, PlotSnapList, True, m_low, m_high) if plot_z: master_mean_reionmod_z, master_std_reionmod_z, master_N_reionmod_z, _ = collect_across_tasks(mean_reionmod_z, std_reionmod_z, N_reionmod) if rank == 0: fig1 = plt.figure() ax1 = fig1.add_subplot(111) if plot_z: fig2 = plt.figure() ax10 = fig2.add_subplot(111) for model_number in range(len(PlotSnapList)): if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() for snapshot_idx in range(len((PlotSnapList[model_number]))): if snapshot_idx == 0: label = model_tags[model_number] else: label = "" nonzero_bins = np.where(master_N_reionmod_halo[model_number][snapshot_idx] > 0.0)[0] ax1.plot(master_bin_middle[model_number][snapshot_idx][nonzero_bins], master_mean_reionmod_halo[model_number][snapshot_idx][nonzero_bins], label = label, ls = PlotScripts.linestyles[model_number], color = PlotScripts.colors[snapshot_idx]) if plot_z: ax10.plot((AllVars.t_BigBang - AllVars.Lookback_Time[SnapList[model_number]])*1.0e3, master_mean_reionmod_z[model_number], color = PlotScripts.colors[model_number], label = model_tags[model_number], ls = PlotScripts.linestyles[model_number], lw = 3) for count, snapshot_idx in enumerate(PlotSnapList[model_number]): #label = r"$\mathbf{z = " + str(int(round(AllVars.SnapZ[snapshot_idx]))) + "}$" label = r"$\mathbf{z = " + str(AllVars.SnapZ[snapshot_idx]) + "}$" ax1.plot(np.nan, np.nan, ls = PlotScripts.linestyles[0], color = PlotScripts.colors[count], label = label) ax1.set_xlim([8.5, 11.5]) ax1.set_ylim([0.0, 1.05]) ax1.set_xlabel(r'$\mathbf{log_{10} \: M_{vir} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax1.set_ylabel(r'$\mathbf{Mean ReionMod}$', fontsize = PlotScripts.global_labelsize) leg = ax1.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile1 = "./{0}_halo{1}".format(output_tag, output_format) fig1.savefig(outputFile1, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile1)) plt.close(fig1) if plot_z: ax10.set_xlabel(r"$\mathbf{Time \: since \: Big \: Bang \: [Myr]}$", fontsize = PlotScripts.global_labelsize) tick_locs = np.arange(200.0, 1000.0, 100.0) tick_labels = [r"$\mathbf{%d}$" % x for x in tick_locs] ax10.xaxis.set_major_locator(mtick.MultipleLocator(100)) ax10.set_xticklabels(tick_labels, fontsize = PlotScripts.global_fontsize) ax10.set_xlim(PlotScripts.time_xlim) ax10.set_ylabel(r'$\mathbf{Mean ReionMod}$', fontsize = PlotScripts.global_labelsize) ax11 = ax10.twiny() t_plot = (AllVars.t_BigBang - cosmo.lookback_time(PlotScripts.z_plot).value) * 1.0e3 # Corresponding Time values on the bottom. z_labels = ["$\mathbf{%d}$" % x for x in PlotScripts.z_plot] # Properly Latex-ize the labels. ax11.set_xlabel(r"$\mathbf{z}$", fontsize = PlotScripts.global_labelsize) ax11.set_xlim(PlotScripts.time_xlim) ax11.set_xticks(t_plot) # Set the ticks according to the time values on the bottom, ax11.set_xticklabels(z_labels, fontsize = PlotScripts.global_fontsize) # But label them as redshifts. leg = ax10.legend(loc='lower right', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile2 = "./{0}_z{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ## def plot_dust(PlotSnapList, SnapList, simulation_norm, mean_dust_galaxy, std_dust_galaxy, N_galaxy, mean_dust_halo, std_dust_halo, N_halo, plot_z, model_tags, output_tag): """ """ master_mean_dust_galaxy, master_std_dust_galaxy, master_N_dust_galaxy, master_bin_middle_galaxy = \ collect_across_tasks(mean_dust_galaxy, std_dust_galaxy, N_galaxy, SnapList, PlotSnapList, True, m_gal_low, m_gal_high) master_mean_dust_halo, master_std_dust_halo, master_N_dust_halo, master_bin_middle_halo = \ collect_across_tasks(mean_dust_halo, std_dust_halo, N_halo, SnapList, PlotSnapList, True, m_low, m_high) if rank == 0: fig1 = plt.figure() ax1 = fig1.add_subplot(111) fig2 = plt.figure() ax2 = fig2.add_subplot(111) for model_number in range(len(PlotSnapList)): if(simulation_norm[model_number] == 1): cosmo = AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): cosmo = AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): cosmo = AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): cosmo = AllVars.Set_Params_Kali() for snapshot_idx in range(len((PlotSnapList[model_number]))): if snapshot_idx == 0: label = model_tags[model_number] else: label = "" nonzero_bins = np.where(master_N_dust_galaxy[model_number][snapshot_idx] > 0.0)[0] ax1.plot(master_bin_middle_galaxy[model_number][snapshot_idx][nonzero_bins], master_mean_dust_galaxy[model_number][snapshot_idx][nonzero_bins], label = label, ls = PlotScripts.linestyles[model_number], color = PlotScripts.colors[snapshot_idx]) nonzero_bins = np.where(master_N_dust_halo[model_number][snapshot_idx] > 0.0)[0] ax2.plot(master_bin_middle_halo[model_number][snapshot_idx][nonzero_bins], master_mean_dust_halo[model_number][snapshot_idx][nonzero_bins], label = label, ls = PlotScripts.linestyles[model_number], color = PlotScripts.colors[snapshot_idx]) print(master_mean_dust_halo[model_number][snapshot_idx]) for count, snapshot_idx in enumerate(PlotSnapList[model_number]): #label = r"$\mathbf{z = " + str(int(round(AllVars.SnapZ[snapshot_idx]))) + "}$" label = r"$\mathbf{z = " + str(AllVars.SnapZ[snapshot_idx]) + "}$" ax1.plot(np.nan, np.nan, ls = PlotScripts.linestyles[0], color = PlotScripts.colors[count], label = label) ax2.plot(np.nan, np.nan, ls = PlotScripts.linestyles[0], color = PlotScripts.colors[count], label = label) ax1.set_xlim([2.0, 10.5]) #ax1.set_ylim([1.0, 6.0]) ax1.set_xlabel(r'$\mathbf{log_{10} \: M_{*} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax1.set_ylabel(r'$\mathbf{log_{10} \: \langle M_{Dust}\rangle_{M*}}$', fontsize = PlotScripts.global_labelsize) leg = ax1.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile1 = "./{0}_galaxy{1}".format(output_tag, output_format) fig1.savefig(outputFile1, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile1)) plt.close(fig1) ax2.set_xlim([6.8, 11.5]) #ax2.set_ylim([1.0, 6.0]) ax2.set_xlabel(r'$\mathbf{log_{10} \: M_{vir} \:[M_{\odot}]}$', fontsize = PlotScripts.global_labelsize) ax2.set_ylabel(r'$\mathbf{log_{10} \: \langle M_{Dust}\rangle_{Mvir}}$', fontsize = PlotScripts.global_labelsize) leg = ax2.legend(loc='upper left', numpoints=1, labelspacing=0.1) leg.draw_frame(False) # Don't want a box frame for t in leg.get_texts(): # Reduce the size of the text t.set_fontsize(PlotScripts.global_legendsize) outputFile2 = "./{0}_halo{1}".format(output_tag, output_format) fig2.savefig(outputFile2, bbox_inches='tight') # Save the figure print('Saved file to {0}'.format(outputFile2)) plt.close(fig2) ### Here ends the plotting functions. ### ### Here begins the functions that calculate various properties for the galaxies (fesc, Magnitude etc). ### def Calculate_HaloPartStellarMass(halo_part, stellar_mass, bound_low, bound_high): ''' Calculates the stellar mass for galaxies whose host halos contain a specified number of particles. Parameters ---------- halo_part : array Array containing the number of particles inside each halo. stellar_mass : array Array containing the Stellar Mass for each galaxy (entries align with HaloPart). Units of log10(Msun). bound_low, bound_high : int We calculate the Stellar Mass of galaxies whose host halo has, bound_low <= halo_part <= bound_high. Return ----- mass, mass_std : float Mean and standard deviation stellar mass of galaxies whose host halo has number of particles between the specified bounds. Units of log10(Msun) Units ----- Input Stellar Mass is in units of log10(Msun). Output mean/std Stellar Mass is in units of log10(Msun). ''' w = np.where((halo_part >= bound_low) & (halo_part <= bound_high))[0] # Find the halos with particle number between the bounds. mass = np.mean(10**(stellar_mass[w])) mass_std = np.std(10**(stellar_mass[w])) return np.log10(mass), np.log10(mass_std) ## def calculate_UV_extinction(z, L, M): ''' Calculates the observed UV magnitude after dust extinction is accounted for. Parameters ---------- z : float Redshift we are calculating the extinction at. L, M : array, length equal to the number of galaxies at this snapshot. Array containing the UV luminosities and magnitudes. Returns ------- M_UV_obs : array, length equal to the number of galaxies at this snapshot. Array containing the observed UV magnitudes. Units ----- Luminosities are in units of log10(erg s^-1 A^-1). Magnitudes are in the AB system. ''' M_UV_bins = np.arange(-24, -16, 0.1) A_mean = np.zeros((len(MUV_bins))) # A_mean is the average UV extinction for a given UV bin. for j in range(0, len(M_UV_bins)): beta = calculate_beta(M_UV_bins[j], AllVars.SnapZ[current_snap]) # Fits the beta parameter for the current redshift/UV bin. dist = np.random.normal(beta, 0.34, 10000) # Generates a normal distribution with mean beta and standard deviation of 0.34. A = 4.43 + 1.99*dist A[A < 0] = 0 # Negative extinctions don't make sense. A_Mean[j] = np.mean(A) indices = np.digitize(M, M_UV_bins) # Bins the simulation magnitude into the MUV bins. Note that digitize defines an index i if bin[i-1] <= x < bin[i] whereas I prefer bin[i] <= x < bin[i+1] dust = A_Mean[indices] flux = AllVars.Luminosity_to_Flux(L, 10.0) # Calculate the flux from a distance of 10 parsec, units of log10(erg s^-1 A^-1 cm^-2). flux_observed = flux - 0.4*dust f_nu = ALlVars.spectralflux_wavelength_to_frequency(10**flux_observed, 1600) # Spectral flux desnity in Janksy. M_UV_obs(-2.5 * np.log10(f_nu) + 8.90) # AB Magnitude from http://www.astro.ljmu.ac.uk/~ikb/convert-units/node2.html return M_UV_obs ## def update_cumulative_stats(mean_pool, std_pool, N_pool, mean_local, std_local, N_local): ''' Update the cumulative statistics (such as Stellar Mass Function, Mvir-Ngamma, fesc-z) that are saved across files. Pooled mean formulae taken : from https://www.ncbi.nlm.nih.gov/books/NBK56512/ Pooled variance formulae taken from : https://en.wikipedia.org/wiki/Pooled_variance Parameters ---------- mean_pool, std_pool, N_pool : array of floats with length equal to the number of bins (e.g. the mass bins for the Stellar Mass Function). The current mean, standard deviation and number of data points within in each bin. This is the array that will be updated in this function. mean_local, std_local, N_local : array of floats with length equal to the number of bins. The mean, standard deviation and number of data points within in each bin that will be added to the pool. Returns ------- mean_pool, std_pool, N_pool : (See above) The updated arrays with the local values added and accounted for within the pools. Units ----- All units are kept the same as the input units. Values are in real-space (not log-space). ''' N_times_mean_local = np.multiply(N_local, mean_local) N_times_var_local = np.multiply(N_local - 1, np.multiply(std_local, std_local)) # Actually N - 1 because of Bessel's Correction # https://en.wikipedia.org/wiki/Bessel%27s_correction). # N_times_mean_pool = np.add(N_times_mean_local, np.multiply(N_pool, mean_pool)) N_times_var_pool = np.add(N_times_var_local, np.multiply(N_pool - 1, np.multiply(std_pool, std_pool))) N_pool = np.add(N_local, N_pool) ''' print(mean_local) print(type(mean_local)) print((type(mean_local).__module__ == np.__name__)) print(isinstance(mean_local, list)) print(isinstance(mean_local,float64)) print(isinstance(mean_local,float32)) ''' if (((type(mean_local).__module__ == np.__name__) == True or (isinstance(mean_local, list) == True)) and isinstance(mean_local, float) == False and isinstance(mean_local, int) == False and isinstance(mean_local,float32) == False and isinstance(mean_local, float64) == False): # Checks to see if we are dealing with arrays. for i in range(0, len(N_pool)): if(N_pool[i] == 0): # This case is when we have no data points in the bin. mean_pool[i] = 0.0 else: mean_pool[i] = N_times_mean_pool[i]/N_pool[i] if(N_pool[i] < 3): # In this instance we don't have enough data points to properly calculate the standard deviation. std_pool[i] = 0.0 else: std_pool[i] = np.sqrt(N_times_var_pool[i]/ (N_pool[i] - 2)) # We have -2 because there is two instances of N_pool contains two 'N - 1' terms. else: mean_pool = N_times_mean_pool / N_pool if(N_pool < 3): std_pool = 0.0 else: std_pool = np.sqrt(N_times_var_pool / (N_pool - 2)) return mean_pool, std_pool ### Here ends the functions that deal with galaxy data manipulation. ### ################################# if __name__ == '__main__': np.seterr(divide='ignore') number_models = 4 galaxies_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/galaxies/const_0.3_z5.782" merged_galaxies_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/galaxies/const_0.3_MergedGalaxies" photo_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/grids/cifog/const_0.3_photHI" zreion_model1="/fred/oz004/jseiler/kali/self_consistent_output/rsage_constant/grids/cifog/const_0.3_reionization_redshift" galaxies_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/galaxies/fej_alpha0.40_beta0.05_z5.782" merged_galaxies_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/galaxies/fej_alpha0.40_beta0.05_MergedGalaxies" photo_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/grids/cifog/fej_alpha0.40_beta0.05_photHI" zreion_model2="/fred/oz004/jseiler/kali/self_consistent_output/rsage_fej/grids/cifog/fej_alpha0.40_beta0.05_reionization_redshift" galaxies_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/galaxies/MHneg_1e8_1e12_0.99_0.05_z5.782" merged_galaxies_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/galaxies/MHneg_1e8_1e12_0.99_0.05_MergedGalaxies" photo_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/grids/cifog/MHneg_1e8_1e12_0.99_0.05_photHI" zreion_model3="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHneg/grids/cifog/MHneg_1e8_1e12_0.99_0.05_reionization_redshift" galaxies_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/galaxies/MHpos_1e8_1e12_0.01_0.50_z5.782" merged_galaxies_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/galaxies/MHpos_1e8_1e12_0.01_0.50_MergedGalaxies" photo_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/grids/cifog/MHpos_1e8_1e12_0.01_0.50_photHI" zreion_model4="/fred/oz004/jseiler/kali/self_consistent_output/rsage_MHpos/grids/cifog/MHpos_1e8_1e12_0.01_0.50_reionization_redshift" galaxies_filepath_array = [galaxies_model1, galaxies_model2, galaxies_model3, galaxies_model4] photo_array = [photo_model1, photo_model2, photo_model3, photo_model4] zreion_array = [zreion_model1, zreion_model2, zreion_model3, zreion_model4] GridSize_array = [256, 256, 256, 256] precision_array = [2, 2, 2, 2] merged_galaxies_filepath_array = [merged_galaxies_model1, merged_galaxies_model2, merged_galaxies_model3, merged_galaxies_model4] number_substeps = [10, 10, 10, 10] # How many substeps does each model have (specified by STEPS variable within SAGE). number_snapshots = [99, 99, 99, 99] # Number of snapshots in the simulation (we don't have to do calculations for ALL snapshots). # Tiamat extended has 164 snapshots. FirstFile = [0, 0, 0, 0] # The first file number THAT WE ARE PLOTTING. #LastFile = [63, 63, 63, 63] # The last file number THAT WE ARE PLOTTING. LastFile = [0, 0, 0, 0] # The last file number THAT WE ARE PLOTTING. NumFile = [64, 64, 64, 64] # The number of files for this simulation (plotting a subset of these files is allowed). same_files = [0, 0, 0, 0] # In the case that model 1 and model 2 (index 0 and 1) have the same files, we don't want to read them in a second time. # This array will tell us if we should keep the files for the next model or otherwise throw them away. # The files will be kept until same_files[current_model_number] = 0. # For example if we had 5 models we were plotting and model 1, 2, 3 shared the same files and models 4, 5 shared different files, # Then same_files = [1, 1, 0, 1, 0] would be the correct values. done_model = np.zeros((number_models)) # We use this to keep track of if we have done a model already. model_tags = [r"$\mathbf{f_\mathrm{esc} \: Constant}$", r"$\mathbf{f_\mathrm{esc} \: \propto \: f_\mathrm{ej}}$", r"$\mathbf{f_\mathrm{esc} \: \propto \: M_\mathrm{H}^{-1}}$", r"$\mathbf{f_\mathrm{esc} \: \propto \: M_\mathrm{H}}$"] ## Constants used for each model. ## # Need to add an entry for EACH model. # halo_cut = [32, 32, 32, 32] # Only calculate properties for galaxies whose host halos have at least this many particles. # For Tiamat, z = [6, 7, 8] are snapshots [78, 64, 51] # For Kali, z = [6, 7, 8] are snapshots [93, 76, 64] #SnapList = [np.arange(0,99), np.arange(0,99)] # These are the snapshots over which the properties are calculated. NOTE: If the escape fraction is selected (fesc_prescription == 3) then this should be ALL the snapshots in the simulation as this prescriptions is temporally important. #SnapList = [np.arange(20,99), np.arange(20, 99), np.arange(20, 99)] SnapList = [[33, 50, 76, 93], [33, 50, 76, 93], [33, 50, 76, 93], [33, 50, 76, 93]] #SnapList = [[64], # [64], # [64], # [64]] #SnapList = [[33, 50, 64, 76, 93]] #SnapList = [[64], [64]] #SnapList = [np.arange(20,99)] #PlotSnapList = [[30, 50, 64, 76, 93]] #PlotSnapList = [[93, 76, 64], [93, 76, 64]] #SnapList = [[93, 76, 64], [93, 76, 64]] PlotSnapList = SnapList simulation_norm = [5, 5, 5, 5] # Changes the constants (cosmology, snapshot -> redshift mapping etc) for each simulation. # 0 for MySim (Manodeep's old one). # 1 for Mini-Millennium. # 2 for Tiamat (up to z =5). # 3 for extended Tiamat (down to z = 1.6ish). # 4 for Britton's Sim Pip # 5 for Manodeep's new simulation Kali. stellar_mass_halolen_lower = [32, 95, 95, 95] # These limits are for the number of particles in a halo. stellar_mass_halolen_upper = [50, 105, 105, 105] # We calculate the average stellar mass for galaxies whose host halos have particle count between these limits. calculate_observed_LF = [0, 0, 0, 0] # Determines whether we want to account for dust extinction when calculating the luminosity function of each model. paper_plots = 1 ############################################################################################################## ## Do a few checks to ensure all the arrays were specified properly. ## for model_number in range(0,number_models): assert(LastFile[model_number] - FirstFile[model_number] + 1 >= size) if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") if (number_snapshots[model_number] != len(AllVars.SnapZ)): # Here we do a check to ensure that the simulation we've defined correctly matches the number of snapshots we have also defined. print("The number_snapshots array is {0}".format(number_snapshots)) print("The simulation_norm array is {0}".format(simulation_norm)) print("The number of snapshots for model_number {0} has {1} but you've said there is only {2}".format(model_number, len(AllVars.SnapZ), number_snapshots[model_number])) raise ValueError("Check either that the number of snapshots has been defined properly and that the normalization option is correct.") ###################################################################### ##################### SETTING UP ARRAYS ############################## ###################################################################### ### The arrays are set up in a 3 part process. ### ### This is because our arrays are 3D nested to account for the model number and snapshots. ### # First set up the outer most array. # ## Arrays for functions of stellar mass. ## SMF = [] # Stellar Mass Function. mean_fesc_galaxy_array = [] # Mean escape fraction as a function of stellar mass. std_fesc_galaxy_array = [] # Same as above but standard devation. N_galaxy_array = [] # Number of galaxies as a function of stellar mass. mean_BHmass_galaxy_array = [] # Black hole mass as a function of stellar mass. std_BHmass_galaxy_array = [] # Same as above but standard deviation. mergers_galaxy_array = [] # Number of mergers as a function of halo mass. mean_dust_galaxy_array = [] # Mean dust mass as a function of stellar mass. std_dust_galaxy_array = [] # Same as above but standard deviation. mean_sfr_galaxy_array = [] # Mean star formation rate as a # function of stellar mass std_sfr_galaxy_array = [] # Same as above but standard deviation. mean_ssfr_galaxy_array = [] # Mean specific star formation rate as a # function of stellar mass std_ssfr_galaxy_array = [] # Same as above but standard deviation. mean_Ngamma_galaxy_array = [] # Mean number of ionizing photons emitted as # a function of stellar mass. std_Ngamma_galaxy_array = [] # Same as above but standard deviation. mean_photo_galaxy_array = [] # Mean photoionization rate. std_photo_galaxy_array = [] # Std photoionization rate. mean_reionmod_galaxy_array = [] # Mean reionization modifier using RSAGE. std_reionmod_galaxy_array = [] # Std. mean_gnedin_reionmod_galaxy_array = [] # Mean reionization modifier using Gnedin analytic prescription. std_gnedin_reionmod_galaxy_array = [] # Std. ## Arrays for functions of halo mass. ## mean_ejected_halo_array = [] # Mean ejected fractions as a function of halo mass. std_ejected_halo_array = [] # Same as above but standard deviation. mean_fesc_halo_array = [] # Mean escape fraction as a function of halo mass. std_fesc_halo_array = [] # Same as above but standard deviation. mean_Ngamma_halo_array = [] # Mean number of ionizing photons THAT ESCAPE as a function of halo mass. std_Ngamma_halo_array = [] # Same as above but standard deviation. N_halo_array = [] # Number of galaxies as a function of halo mass. mergers_halo_array = [] # Number of mergers as a function of halo mass. mean_quasar_activity_array = [] # Mean fraction of galaxies that have quasar actvitity as a function of halo mas. std_quasar_activity_array = [] # Same as above but standard deviation. mean_reionmod_halo_array = [] # Mean reionization modifier as a function of halo mass. std_reionmod_halo_array = [] # Same as above but for standard deviation. mean_dust_halo_array = [] # Mean dust mass as a function of halo mass. std_dust_halo_array = [] # Same as above but standard deviation. ## Arrays for functions of redshift. ## sum_Ngamma_z_array = [] # Total number of ionizing photons THAT ESCAPE as a functio of redshift. mean_fesc_z_array = [] # Mean number of ionizing photons THAT ESCAPE as a function of redshift. std_fesc_z_array = [] # Same as above but standard deviation. N_z = [] # Number of galaxies as a function of redshift. galaxy_halo_mass_mean = [] # Mean galaxy mass as a function of redshift. N_quasars_z = [] # This tracks how many quasars went off during a specified snapshot. N_quasars_boost_z = [] # This tracks how many galaxies are having their escape fraction boosted by quasar activity. dynamicaltime_quasars_mean_z = [] # Mean dynamical time of galaxies that have a quasar event as a function of redshift. dynamicaltime_quasars_std_z = [] # Same as above but standard deviation. dynamicaltime_all_mean_z = [] # Mean dynamical time of all galaxies. dynamicaltime_all_std_z = [] # Same as above but standard deviation. mean_reionmod_z = [] # Mean reionization modifier as a function of redshift. std_reionmod_z = [] # Same as above but for standard deviation. N_reionmod_z = [] # Number of galaxies with a non-negative reionization modifier. mean_ejected_z = [] # Mean ejected fraction as a function of redshift. std_ejected_z = [] # Same as above but for standard deviation. ## Arrays that aren't functions of other variables. ## Ngamma_global = [] mass_global = [] fesc_global = [] ## Arrays as a function of fej ## mean_Ngamma_fej = [] std_Ngamma_fej = [] N_fej = [] ## Now the outer arrays have been defined, set up the next nest level for the number of models. ## for model_number in range(0,number_models): ## Galaxy Arrays ## SMF.append([]) mean_fesc_galaxy_array.append([]) std_fesc_galaxy_array.append([]) N_galaxy_array.append([]) mean_BHmass_galaxy_array.append([]) std_BHmass_galaxy_array.append([]) mergers_galaxy_array.append([]) mean_dust_galaxy_array.append([]) std_dust_galaxy_array.append([]) mean_sfr_galaxy_array.append([]) std_sfr_galaxy_array.append([]) mean_ssfr_galaxy_array.append([]) std_ssfr_galaxy_array.append([]) mean_Ngamma_galaxy_array.append([]) std_Ngamma_galaxy_array.append([]) mean_photo_galaxy_array.append([]) std_photo_galaxy_array.append([]) mean_reionmod_galaxy_array.append([]) std_reionmod_galaxy_array.append([]) mean_gnedin_reionmod_galaxy_array.append([]) std_gnedin_reionmod_galaxy_array.append([]) ## Halo arrays. ## mean_ejected_halo_array.append([]) std_ejected_halo_array.append([]) mean_fesc_halo_array.append([]) std_fesc_halo_array.append([]) mean_Ngamma_halo_array.append([]) std_Ngamma_halo_array.append([]) N_halo_array.append([]) mergers_halo_array.append([]) mean_quasar_activity_array.append([]) std_quasar_activity_array.append([]) mean_reionmod_halo_array.append([]) std_reionmod_halo_array.append([]) mean_dust_halo_array.append([]) std_dust_halo_array.append([]) ## Redshift arrays. ## sum_Ngamma_z_array.append([]) mean_fesc_z_array.append([]) std_fesc_z_array.append([]) N_z.append([]) galaxy_halo_mass_mean.append([]) N_quasars_z.append([]) N_quasars_boost_z.append([]) dynamicaltime_quasars_mean_z.append([]) dynamicaltime_quasars_std_z.append([]) dynamicaltime_all_mean_z.append([]) dynamicaltime_all_std_z.append([]) mean_reionmod_z.append([]) std_reionmod_z.append([]) N_reionmod_z.append([]) mean_ejected_z.append([]) std_ejected_z.append([]) ## Arrays that aren't functions ## Ngamma_global.append([]) mass_global.append([]) fesc_global.append([]) ## Arrays as a function of fej ## mean_Ngamma_fej.append([]) std_Ngamma_fej.append([]) N_fej.append([]) ## And then finally set up the inner most arrays ## ## NOTE: We do the counts as float so we can keep consistency when we're calling MPI operations (just use MPI.FLOAT rather than deciding if we need to use MPI.INT) for snapshot_idx in range(len(SnapList[model_number])): ## For the arrays that are functions of stellar/halo mass, the inner most level will be an array with the statistic binned across mass ## ## E.g. SMF[model_number][snapshot_idx] will return an array whereas N_z[model_number][snapshot_idx] will return a float. ## ## Functions of stellar mass arrays. ## SMF[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_fesc_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_fesc_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) N_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_BHmass_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_BHmass_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mergers_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_dust_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_dust_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_sfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_sfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_ssfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_ssfr_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_Ngamma_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_Ngamma_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_photo_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_photo_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) mean_gnedin_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) std_gnedin_reionmod_galaxy_array[model_number].append(np.zeros((NB_gal), dtype = np.float32)) ## Function of halo mass arrays. ## mean_ejected_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_ejected_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_fesc_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_fesc_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_Ngamma_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_Ngamma_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) N_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mergers_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_quasar_activity_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_quasar_activity_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_reionmod_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_reionmod_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) mean_dust_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) std_dust_halo_array[model_number].append(np.zeros((NB), dtype = np.float32)) ## Function of Redshift arrays. ## sum_Ngamma_z_array[model_number].append(0.0) mean_fesc_z_array[model_number].append(0.0) std_fesc_z_array[model_number].append(0.0) N_z[model_number].append(0.0) galaxy_halo_mass_mean[model_number].append(0.0) N_quasars_z[model_number].append(0.0) N_quasars_boost_z[model_number].append(0.0) dynamicaltime_quasars_mean_z[model_number].append(0.0) dynamicaltime_quasars_std_z[model_number].append(0.0) dynamicaltime_all_mean_z[model_number].append(0.0) dynamicaltime_all_std_z[model_number].append(0.0) mean_reionmod_z[model_number].append(0.0) std_reionmod_z[model_number].append(0.0) N_reionmod_z[model_number].append(0.0) mean_ejected_z[model_number].append(0.0) std_ejected_z[model_number].append(0.0) Ngamma_global[model_number].append([]) mass_global[model_number].append([]) fesc_global[model_number].append([]) ## Arrays as a function of fej. ## mean_Ngamma_fej[model_number].append(np.zeros((NB_fej), dtype = np.float32)) std_Ngamma_fej[model_number].append(np.zeros((NB_fej), dtype = np.float32)) N_fej[model_number].append(np.zeros((NB_fej), dtype = np.float32)) ###################################################################### #################### ALL ARRAYS SETUP ################################ ###################################################################### ## Now it's (finally) time to read in all the data and do the actual work. ## for model_number in range(number_models): if(simulation_norm[model_number] == 1): AllVars.Set_Params_MiniMill() elif(simulation_norm[model_number] == 3): AllVars.Set_Params_Tiamat_extended() elif(simulation_norm[model_number] == 4): AllVars.Set_Params_Britton() elif(simulation_norm[model_number] == 5): AllVars.Set_Params_Kali() else: print("Simulation norm was set to {0}.".format(simulation_norm[model_number])) raise ValueError("This option has been implemented yet. Get your head in the game Jacob!") if (done_model[model_number] == 1): # If we have already done this model (i.e., we kept the files and skipped this loop), move along. assert(FirstFile[model_number] == FirstFile[model_number - 1]) assert(LastFile[model_number] == LastFile[model_number - 1]) continue for fnr in range(FirstFile[model_number] + rank, LastFile[model_number]+1, size): # Divide up the input files across the processors. GG, Gal_Desc = ReadScripts.ReadGals_SAGE(galaxies_filepath_array[model_number], fnr, number_snapshots[model_number], comm) # Read galaxies G_Merged, _ = ReadScripts.ReadGals_SAGE(merged_galaxies_filepath_array[model_number], fnr, number_snapshots[model_number], comm) # Also need the merged galaxies. G = ReadScripts.Join_Arrays(GG, G_Merged, Gal_Desc) # Then join them together for all galaxies. keep_files = 1 # Flips to 0 when we are done with this file. current_model_number = model_number # Used to differentiate between outer model_number and the inner model_number because we can keep files across model_numbers. while(keep_files == 1): ## Just a few definitions to cut down the clutter a smidge. ## current_halo_cut = halo_cut[current_model_number] NumSubsteps = number_substeps[current_model_number] do_observed_LF = calculate_observed_LF[current_model_number] for snapshot_idx in range(0, len(SnapList[current_model_number])): # Now let's calculate stats for each required redshift. current_snap = SnapList[current_model_number][snapshot_idx] # Get rid of some clutter. w_gal = np.where((G.GridHistory[:, current_snap] != -1) & (G.GridStellarMass[:, current_snap] > 0.0) & (G.LenHistory[:, current_snap] > current_halo_cut) & (G.GridSFR[:, current_snap] >= 0.0) & (G.GridFoFMass[:, current_snap] >= 0.0))[0] # Only include those galaxies that existed at the current snapshot, had positive (but not infinite) stellar/Halo mass and Star formation rate. Ensure the galaxies also resides in a halo that is sufficiently resolved. w_merged_gal = np.where((G_Merged.GridHistory[:, current_snap] != -1) & (G_Merged.GridStellarMass[:, current_snap] > 0.0) & (G_Merged.LenHistory[:, current_snap] > current_halo_cut) & (G_Merged.GridSFR[:, current_snap] >= 0.0) & (G_Merged.GridFoFMass[:, current_snap] >= 0.0) & (G_Merged.LenMergerGal[:,current_snap] > current_halo_cut))[0] print("There were {0} galaxies for snapshot {1} (Redshift {2:.3f}) model {3}.".format(len(w_gal), current_snap, AllVars.SnapZ[current_snap], current_model_number)) if (len(w_gal) == 0): continue mass_gal = np.log10(G.GridStellarMass[w_gal, current_snap] * 1.0e10 / AllVars.Hubble_h) # Msun. Log Units. w_SFR = w_gal[np.where((G.GridSFR[w_gal, current_snap] > 0.0))[0]] mass_SFR_gal = np.log10(G.GridStellarMass[w_SFR, current_snap] * \ 1.0e10 / AllVars.Hubble_h) SFR_gal = np.log10(G.GridSFR[w_SFR,current_snap]) sSFR_gal = SFR_gal - mass_SFR_gal halo_part_count = G.LenHistory[w_gal, current_snap] metallicity_gal = G.GridZ[w_gal, current_snap] metallicity_tremonti_gal = np.log10(G.GridZ[w_gal, current_snap] / 0.02) + 9.0 # Using the Tremonti relationship for metallicity. mass_central = np.log10(G.GridFoFMass[w_gal, current_snap] * 1.0e10 / AllVars.Hubble_h) # Msun. Log Units. ejected_fraction = G.EjectedFraction[w_gal, current_snap] w_dust = np.where(((G.GridDustColdGas[w_gal, current_snap] +G.GridDustHotGas[w_gal, current_snap] +G.GridDustEjectedMass[w_gal, current_snap]) > 0.0) & (G.GridType[w_gal, current_snap] == 0))[0] total_dust_gal = np.log10((G.GridDustColdGas[w_gal[w_dust], current_snap] +G.GridDustHotGas[w_gal[w_dust], current_snap] +G.GridDustEjectedMass[w_gal[w_dust], current_snap]) * 1.0e10 / AllVars.Hubble_h) mass_gal_dust = np.log10(G.GridStellarMass[w_gal[w_dust], current_snap] * 1.0e10 / AllVars.Hubble_h) mass_centralgal_dust = np.log10(G.GridFoFMass[w_gal[w_dust], current_snap] * 1.0e10 / AllVars.Hubble_h) fesc = G.Gridfesc[w_gal, current_snap] fesc[fesc < 0.0] = 0.0 Ngamma_gal = G.GridNgamma_HI[w_gal, current_snap] # 1.0e50 # photons/s. if model_number < 3: Ngamma_gal += 50.0 # Old versions of SAGE incorrectly # subtracted 50. Ngamma_gal *= fesc reionmod = G.GridReionMod[w_gal, current_snap] mass_reionmod_central = mass_central[reionmod > -1] mass_reionmod_gal = mass_gal[reionmod > -1] reionmod = reionmod[reionmod > -1] # Some satellite galaxies that don't have HotGas and hence won't be stripped. As a result reionmod = -1 for these. Ignore them. mass_BH = G.GridBHMass[w_gal, current_snap] * 1.0e10 / AllVars.Hubble_h # Msun. Not log units. L_UV = SFR_gal + 39.927 # Using relationship from STARBURST99, units of erg s^-1 A^-1. Log Units. M_UV = AllVars.Luminosity_to_ABMag(L_UV, 1600) if (do_observed_LF == 1): # Calculate the UV extinction if requested. M_UV_obs = calculate_UV_extinction(AllVars.SnapZ[current_snap], L_UV, M_UV[snap_idx]) galaxy_halo_mass_mean_local, galaxy_halo_mass_std_local = Calculate_HaloPartStellarMass(halo_part_count, mass_gal, stellar_mass_halolen_lower[current_model_number], stellar_mass_halolen_upper[current_model_number]) # This is the average stellar mass for galaxies whose halos have the specified number of particles. galaxy_halo_mass_mean[current_model_number][snapshot_idx] += pow(10, galaxy_halo_mass_mean_local) / (LastFile[current_model_number] + 1) # Adds to the average of the mean. photofield_path = "{0}_{1:03d}".format(photo_array[current_model_number], current_snap) #photo_gal = photo.calc_gal_photoion(G.GridHistory[w_gal, current_snap], # photofield_path, # GridSize_array[current_model_number], # precision_array[current_model_number]) #zreion_path = "{0}".format(zreion_array[current_model_number]) #zreion_gal = photo.calc_gal_zreion(G.GridHistory[w_gal, current_snap], # zreion_path, # GridSize_array[current_model_number], # precision_array[current_model_number]) z_0 = 8.0 z_r = 7.0 gnedin_mfilt = ga.get_filter_mass(np.array(AllVars.SnapZ[current_snap]), z_0, z_r) gnedin_reionmod_gal = 1.0 / pow(1.0 + 0.26*pow(10, gnedin_mfilt - mass_central), 3.0) ########################################### ######## BASE PROPERTIES CALCULATED ####### ########################################### # Time to calculate relevant statistics. ### Functions of Galaxies/Stellar Mass ### ## Stellar Mass Function ## (counts_local, bin_edges, bin_middle) = AllVars.Calculate_Histogram(mass_gal, bin_width, 0, m_gal_low, m_gal_high) # Bin the Stellar Mass SMF[current_model_number][snapshot_idx] += counts_local ## Escape Fraction ## (mean_fesc_galaxy_local, std_fesc_galaxy_local, N_local, sum_fesc_galaxy, bin_middle) = AllVars.Calculate_2D_Mean(mass_gal, fesc, bin_width, m_gal_low, m_gal_high) (mean_fesc_galaxy_array[current_model_number][snapshot_idx], std_fesc_galaxy_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_fesc_galaxy_array[current_model_number][snapshot_idx], std_fesc_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_fesc_galaxy_local, std_fesc_galaxy_local, N_local) ## Black Hole Mass ## (mean_BHmass_galaxy_local, std_BHmass_galaxy_local, N_local, sum_BHmass_galaxy, bin_middle) = AllVars.Calculate_2D_Mean(mass_gal, mass_BH, bin_width, m_gal_low, m_gal_high) (mean_BHmass_galaxy_array[current_model_number][snapshot_idx], std_BHmass_galaxy_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_BHmass_galaxy_array[current_model_number][snapshot_idx], std_BHmass_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_BHmass_galaxy_local, std_BHmass_galaxy_local, N_local) ## Total Dust Mass ## (mean_dust_galaxy_local, std_dust_galaxy_local, N_local, sum_dust_galaxy, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal_dust, total_dust_gal, bin_width, m_gal_low, m_gal_high) (mean_dust_galaxy_array[current_model_number][snapshot_idx], std_dust_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_dust_galaxy_array[current_model_number][snapshot_idx], std_dust_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_dust_galaxy_local, std_dust_galaxy_local, N_local) ## Star Formation Rate ## (mean_sfr_galaxy_local, std_sfr_galaxy_local, N_local, sum_sfr_galaxy, bin_middle) = AllVars.Calculate_2D_Mean( mass_SFR_gal, SFR_gal, bin_width, m_gal_low, m_gal_high) (mean_sfr_galaxy_array[current_model_number][snapshot_idx], std_sfr_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_sfr_galaxy_array[current_model_number][snapshot_idx], std_sfr_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_sfr_galaxy_local, std_sfr_galaxy_local, N_local) ## Specific Star Formation Rate ## (mean_ssfr_galaxy_local, std_ssfr_galaxy_local, N_local, sum_ssfr_galaxy, bin_middle) = AllVars.Calculate_2D_Mean( mass_SFR_gal, sSFR_gal, bin_width, m_gal_low, m_gal_high) (mean_ssfr_galaxy_array[current_model_number][snapshot_idx], std_ssfr_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_ssfr_galaxy_array[current_model_number][snapshot_idx], std_ssfr_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_ssfr_galaxy_local, std_ssfr_galaxy_local, N_local) ## Number of Ionizing Photons ## (mean_Ngamma_galaxy_local, std_Ngamma_galaxy_local, N_local, sum_Ngamma_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal, Ngamma_gal, bin_width, m_gal_low, m_gal_high) (mean_Ngamma_galaxy_array[current_model_number][snapshot_idx], std_Ngamma_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_Ngamma_galaxy_array[current_model_number][snapshot_idx], std_Ngamma_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_Ngamma_galaxy_local, std_Ngamma_galaxy_local, N_local) ## Photoionization rate ## ''' (mean_photo_galaxy_local, std_photo_galaxy_local, N_local, sum_photo_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal, photo_gal, bin_width, m_gal_low, m_gal_high) (mean_photo_galaxy_array[current_model_number][snapshot_idx], std_photo_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_photo_galaxy_array[current_model_number][snapshot_idx], std_photo_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_photo_galaxy_local, std_photo_galaxy_local, N_local) ''' ## RSAGE Reionization Modifier ## (mean_reionmod_galaxy_local, std_reionmod_galaxy_local, N_local, sum_reionmod_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_reionmod_gal, reionmod, bin_width, m_gal_low, m_gal_high) (mean_reionmod_galaxy_array[current_model_number][snapshot_idx], std_reionmod_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_reionmod_galaxy_array[current_model_number][snapshot_idx], std_reionmod_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_reionmod_galaxy_local, std_reionmod_galaxy_local, N_local) ## Gnedin Reionization Modifier ## (mean_gnedin_reionmod_galaxy_local, std_gnedin_reionmod_galaxy_local, N_local, sum_gnedin_reionmod_galaxy_local, bin_middle) = AllVars.Calculate_2D_Mean( mass_gal, gnedin_reionmod_gal, bin_width, m_gal_low, m_gal_high) (mean_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx], std_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx], std_gnedin_reionmod_galaxy_array[current_model_number][snapshot_idx], N_galaxy_array[current_model_number][snapshot_idx], mean_gnedin_reionmod_galaxy_local, std_gnedin_reionmod_galaxy_local, N_local) N_galaxy_array[current_model_number][snapshot_idx] += N_local ### Functions of Halos/Halo Mass ### ## Ejected Fraction ## (mean_ejected_halo_local, std_ejected_halo_local, N_local, sum_ejected_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_central, ejected_fraction, bin_width, m_low, m_high) (mean_ejected_halo_array[current_model_number][snapshot_idx], std_ejected_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_ejected_halo_array[current_model_number][snapshot_idx], std_ejected_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_ejected_halo_local, std_ejected_halo_local, N_local) # Then update the running total. ## Quasar Fraction ## (mean_quasar_activity_local, std_quasar_activity_local,N_local, sum_quasar_activity_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_central, G.QuasarActivity[w_gal, current_snap], bin_width, m_low, m_high) (mean_quasar_activity_array[current_model_number][snapshot_idx], std_quasar_activity_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_quasar_activity_array[current_model_number][snapshot_idx], std_quasar_activity_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_quasar_activity_local, std_quasar_activity_local, N_local) # Then update the running total. ## fesc Value ## (mean_fesc_halo_local, std_fesc_halo_local, N_local, sum_fesc_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_central, fesc, bin_width, m_low, m_high) (mean_fesc_halo_array[current_model_number][snapshot_idx], std_fesc_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_fesc_halo_array[current_model_number][snapshot_idx], std_fesc_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_fesc_halo_local, std_fesc_halo_local, N_local) # Then update the running total. ## Ngamma ## #(mean_Ngamma_halo_local, std_Ngamma_halo_local, N_local, sum_Ngamma_halo, bin_middle) \ #= AllVars.Calculate_2D_Mean(mass_central, ionizing_photons, bin_width, m_low, m_high) #mean_Ngamma_halo_local = np.divide(mean_Ngamma_halo_local, 1.0e50) ## Divide out a constant to keep the numbers manageable. #std_Ngamma_halo_local = np.divide(std_Ngamma_halo_local, 1.0e50) #(mean_Ngamma_halo_array[current_model_number][snapshot_idx], std_Ngamma_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_Ngamma_halo_array[current_model_number][snapshot_idx], std_Ngamma_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_Ngamma_halo_local, std_Ngamma_halo_local, N_local) # Then update the running total. ## Reionization Modifier ## (mean_reionmod_halo_local, std_reionmod_halo_local, N_local, sum_reionmod_halo, bin_middle) = AllVars.Calculate_2D_Mean(mass_reionmod_central, reionmod, bin_width, m_low, m_high) (mean_reionmod_halo_array[current_model_number][snapshot_idx], std_reionmod_halo_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_reionmod_halo_array[current_model_number][snapshot_idx], std_reionmod_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_reionmod_halo_local, std_reionmod_halo_local, N_local) # Then update the running total. ## Total Dust Mass ## (mean_dust_halo_local, std_dust_halo_local, N_local, sum_dust_halo, bin_middle) = AllVars.Calculate_2D_Mean( mass_centralgal_dust, total_dust_gal, bin_width, m_low, m_high) (mean_dust_halo_array[current_model_number][snapshot_idx], std_dust_halo_array[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_dust_halo_array[current_model_number][snapshot_idx], std_dust_halo_array[current_model_number][snapshot_idx], N_halo_array[current_model_number][snapshot_idx], mean_dust_halo_local, std_dust_halo_local, N_local) N_halo_array[current_model_number][snapshot_idx] += N_local ### Functions of redshift ### ## Ngamma ## #sum_Ngamma_z_array[current_model_number][snapshot_idx] += np.sum(np.divide(ionizing_photons, 1.0e50)) # Remember that we're dividing out a constant! ## fesc Value ## (mean_fesc_z_array[current_model_number][snapshot_idx], std_fesc_z_array[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_fesc_z_array[current_model_number][snapshot_idx], std_fesc_z_array[current_model_number][snapshot_idx], N_z[current_model_number][snapshot_idx], np.mean(fesc), np.std(fesc), len(w_gal)) # Updates the mean escape fraction for this redshift. ## Reionization Modifier ## (mean_reionmod_z[current_model_number][snapshot_idx], std_reionmod_z[current_model_number][snapshot_idx]) = update_cumulative_stats(mean_reionmod_z[current_model_number][snapshot_idx], std_reionmod_z[current_model_number][snapshot_idx], N_reionmod_z[current_model_number][snapshot_idx], np.mean(reionmod), np.std(reionmod), len(reionmod)) N_reionmod_z[current_model_number][snapshot_idx] += len(reionmod) ## Ejected Fraction ## (mean_ejected_z[current_model_number][snapshot_idx],std_ejected_z[current_model_number][snapshot_idx]) \ = update_cumulative_stats(mean_ejected_z[current_model_number][snapshot_idx], std_ejected_z[current_model_number][snapshot_idx], N_z[current_model_number][snapshot_idx], np.mean(ejected_fraction), np.std(ejected_fraction), len(w_gal)) N_z[current_model_number][snapshot_idx] += len(w_gal) #### Arrays that are just kept across snapshots ## Ngamma_global[current_model_number][snapshot_idx].append(Ngamma_gal) mass_global[current_model_number][snapshot_idx].append(mass_gal) fesc_global[current_model_number][snapshot_idx].append(fesc) #### Arrays that are function of fej ## (mean_Ngamma_fej_local, std_Ngamma_fej_local, N_local, sum_Ngamma_fej_local, bin_middle) = AllVars.Calculate_2D_Mean( ejected_fraction, Ngamma_gal, fej_bin_width, fej_low, fej_high) (mean_Ngamma_fej[current_model_number][snapshot_idx], std_Ngamma_fej[current_model_number][snapshot_idx]) = \ update_cumulative_stats(mean_Ngamma_fej[current_model_number][snapshot_idx], std_Ngamma_fej[current_model_number][snapshot_idx], N_fej[current_model_number][snapshot_idx], mean_Ngamma_fej_local, std_Ngamma_fej_local, N_local) N_fej[current_model_number][snapshot_idx] += N_local done_model[current_model_number] = 1 if (current_model_number < number_models): keep_files = same_files[current_model_number] # Decide if we want to keep the files loaded or throw them out. current_model_number += 1 # Update the inner loop model number. #StellarMassFunction(PlotSnapList, SMF, simulation_norm, FirstFile, # LastFile, NumFile, galaxy_halo_mass_mean, model_tags, # 1, paper_plots, "wtf") #plot_reionmod(PlotSnapList, SnapList, simulation_norm, mean_reionmod_halo_array, #std_reionmod_halo_array, N_halo_array, mean_reionmod_z, #std_reionmod_z, N_reionmod_z, False, model_tags, #"reionmod_selfcon") #plot_dust_scatter(SnapList, mass_gal_dust, mass_centralgal_dust, total_dust_gal, # "dust_scatter") #plot_dust(PlotSnapList, SnapList, simulation_norm, mean_dust_galaxy_array, # std_dust_galaxy_array, N_galaxy_array, mean_dust_halo_array, # std_dust_halo_array, N_halo_array, False, model_tags, # "dustmass_total") #plot_stellarmass_blackhole(PlotSnapList, simulation_norm, mean_BHmass_galaxy_array, # std_BHmass_galaxy_array, N_galaxy_array, # FirstFile, LastFile, NumFile, # model_tags, "StellarMass_BHMass") #plot_ejectedfraction(SnapList, PlotSnapList, simulation_norm, # mean_ejected_halo_array, std_ejected_halo_array, # N_halo_array, mean_ejected_z, std_ejected_z, N_z, # model_tags, "ejectedfraction") #plot_quasars_count(SnapList, PlotSnapList, N_quasars_z, N_quasars_boost_z, N_z, mean_quasar_activity_array, std_quasar_activity_array, N_halo_array, mergers_halo_array, SMF, mergers_galaxy_array, fesc_prescription, simulation_norm, FirstFile, LastFile, NumFile, model_tags, "SN_Prescription") plot_fesc_galaxy(SnapList, PlotSnapList, simulation_norm, mean_fesc_galaxy_array, std_fesc_galaxy_array, N_galaxy_array, mean_fesc_halo_array, std_fesc_halo_array, N_halo_array, galaxy_halo_mass_mean, model_tags, paper_plots, mass_global, fesc_global, Ngamma_global, "fesc_paper") plot_reionmod_galaxy(SnapList, PlotSnapList, simulation_norm, mean_reionmod_galaxy_array, std_reionmod_galaxy_array, N_galaxy_array, mean_gnedin_reionmod_galaxy_array, std_gnedin_reionmod_galaxy_array, model_tags, paper_plots, "reionmod") exit() #plot_nion_galaxy(SnapList, PlotSnapList, simulation_norm, # mean_Ngamma_galaxy_array, std_Ngamma_galaxy_array, # N_galaxy_array, model_tags, # paper_plots, "Ngamma") ''' plot_photo_galaxy(SnapList, PlotSnapList, simulation_norm, mean_photo_galaxy_array, std_photo_galaxy_array, N_galaxy_array, model_tags, paper_plots, "photo") ''' plot_sfr_galaxy(SnapList, PlotSnapList, simulation_norm, mean_sfr_galaxy_array, std_sfr_galaxy_array, mean_ssfr_galaxy_array, std_ssfr_galaxy_array, N_galaxy_array, model_tags, "sSFR") #plot_fej_Ngamma(SnapList, PlotSnapList, simulation_norm, # mean_Ngamma_fej, std_Ngamma_fej, # N_fej, model_tags, "Ngamma_fej") #plot_photoncount(SnapList, sum_Ngamma_z_array, simulation_norm, FirstFile, LastFile, NumFile, model_tags, "Ngamma_test") ## PARALELL COMPATIBLE #plot_mvir_Ngamma(SnapList, mean_Ngamma_halo_array, std_Ngamma_halo_array, N_halo_array, model_tags, "Mvir_Ngamma_test", fesc_prescription, fesc_normalization, "/lustre/projects/p004_swin/jseiler/tiamat/halo_ngamma/") ## PARALELL COMPATIBLE
[ 2, 48443, 14629, 14, 8800, 14, 24330, 21015, 198, 6738, 11593, 37443, 834, 1330, 3601, 62, 8818, 198, 11748, 2603, 29487, 8019, 198, 6759, 29487, 8019, 13, 1904, 10786, 46384, 11537, 198, 198, 11748, 28686, 198, 11748, 24575, 80, 198, 1...
2.082925
86,633