{"token_to_id": {"": 0, "": 1, "": 2, "": 3, "": 4, "def": 5, "class": 6, "return": 7, "if": 8, "else": 9, "elif": 10, "for": 11, "while": 12, "try": 13, "except": 14, "finally": 15, "with": 16, "as": 17, "import": 18, "from": 19, "none": 20, "true": 21, "false": 22, "and": 23, "or": 24, "not": 25, "in": 26, "is": 27, "lambda": 28, "yield": 29, "raise": 30, "pass": 31, "break": 32, "continue": 33, "self": 34, "cls": 35, "async": 36, "await": 37, "function": 38, "method": 39, "find": 40, "get": 41, "set": 42, "create": 43, "delete": 44, "add": 45, "remove": 46, "update": 47, "check": 48, "validate": 49, "parse": 50, "format": 51, "convert": 52, "calculate": 53, "compute": 54, "process": 55, "handle": 56, "that": 57, "which": 58, "the": 59, "a": 60, "an": 61, "to": 62, "of": 63, "numbers": 64, "string": 65, "list": 66, "dict": 67, "file": 68, "data": 69, "user": 70, "input": 71, "output": 72, "result": 73, "value": 74, "item": 75, "element": 76, "authentication": 77, "login": 78, "password": 79, "email": 80, "url": 81, "path": 82, "sort": 83, "search": 84, "filter": 85, "map": 86, "reduce": 87, "sum": 88, "average": 89, "maximum": 90, "minimum": 91, "count": 92, "length": 93, "size": 94, "onconnect": 95, "(": 96, ",": 97, "response": 98, ")": 99, ":": 100, "print": 101, "\"server connected: {0}\"": 102, ".": 103, "peer": 104, "taking": 105, "onmessage": 106, "payload": 107, "isbinary": 108, "\"binary message received: {0} bytes\"": 109, "len": 110, "\"text message received: {0}\"": 111, "decode": 112, "'utf8'": 113, "register": 114, "=": 115, "*": 116, "tags": 117, "kwargs": 118, "\"\"": 119, "\"\n can be used as a function or a decorator. register given function\n `f` labeled with given `tags`. the function should receive **kwargs\n and return list of errors and warnings.\n\n example::\n\n registry = checkregistry()\n @registry.register('mytag', 'anothertag')\n def my_check(apps, **kwargs):\n # ... perform checks and collect `errors` ...\n return errors\n # or\n registry.register(my_check, 'mytag', 'anothertag')\n \"": 120, "\"": 121, "can": 122, "be": 123, "used": 124, "decorator": 125, "given": 126, "`": 127, "f": 128, "labeled": 129, "should": 130, "receive": 131, "errors": 132, "warnings": 133, "example": 134, "registry": 135, "checkregistry": 136, "@": 137, "'mytag'": 138, "'anothertag'": 139, "my_check": 140, "apps": 141, "run_checks": 142, "app_configs": 143, "include_deployment_checks": 144, "\"\n run all registered checks and return list of errors and warnings.\n \"": 145, "[": 146, "]": 147, "checks": 148, "get_checks": 149, "isdisjoint": 150, "#": 151, "by": 152, "default": 153, "'database'": 154, "-": 155, "tagged": 156, "are": 157, "run": 158, "they": 159, "do": 160, "more": 161, "than": 162, "mere": 163, "static": 164, "code": 165, "analysis": 166, "database": 167, "chec": 168, "all": 169, "registered": 170, "inner": 171, "deployment_checks": 172, "'deploy'": 173, "registered_checks": 174, "safe_mask": 175, "X": 176, "mask": 177, "\"return a mask which is safe to use on x.\n\n parameters\n ----------\n x : {array-like, sparse matrix}\n data on which to apply mask.\n\n mask: array\n mask to be used on x.\n\n returns\n -------\n mask\n \"": 178, "np": 179, "asanyarray": 180, "issubdtype": 181, "dtype": 182, "int": 183, "hasattr": 184, "\"toarray\"": 185, "ind": 186, "arange": 187, "shape": 188, "0": 189, "safe": 190, "use": 191, "on": 192, "parameters": 193, "{": 194, "array": 195, "like": 196, "sparse": 197, "matrix": 198, "}": 199, "apply": 200, "returns": 201, "resample": 202, "arrays": 203, "options": 204, "matrices": 205, "consistent": 206, "way": 207, "strategy": 208, "implements": 209, "one": 210, "step": 211, "bootstrapping": 212, "procedure": 213, "sequence": 214, "scipy": 215, "same": 216, "replace": 217, "boolean": 218, "resampling": 219, "replacement": 220, "this": 221, "will": 222, "implement": 223, "sliced": 224, "random": 225, "permutations": 226, "n_samples": 227, "number": 228, "samples": 229, "generate": 230, "left": 231, "automatically": 232, "first": 233, "dimension": 234, "replac": 235, "shuffle": 236, "convenience": 237, "alias": 238, "collections": 239, "random_state": 240, "randomstate": 241, "instance": 242, "control": 243, "shuffling": 244, "reproducible": 245, "behavior": 246, "randomstat": 247, "safe_sqr": 248, "copy": 249, "\"element wise squaring of array-likes and sparse matrices.\n\n parameters\n ----------\n x : array like, matrix, sparse matrix\n\n returns\n -------\n x ** 2 : element wise square\n \"": 250, "safe_asarray": 251, "issparse": 252, "2": 253, "wise": 254, "squaring": 255, "likes": 256, "square": 257, "sqr": 258, "gen_even_slices": 259, "n": 260, "n_packs": 261, "generator": 262, "slices": 263, "going": 264, "up": 265, "examples": 266, ">": 267, "sklearn": 268, "utils": 269, "10": 270, "1": 271, "slice": 272, "doctest": 273, "+": 274, "ellipsis": 275, "9": 276, "5": 277, "4": 278, "8": 279, "3": 280, "7": 281, "gen": 282, "even": 283, "tosequence": 284, "x": 285, "\"cast iterable x to a sequence, avoiding a copy if possible.\"": 286, "isinstance": 287, "ndarray": 288, "asarray": 289, "cast": 290, "iterable": 291, "avoiding": 292, "possible": 293, "__init__": 294, "extra": 295, "''": 296, "\"\n parameters\n ----------\n extra: string\n to be added to the deprecation messages\n\n \"": 297, "added": 298, "deprecation": 299, "messages": 300, "initialize": 301, "init": 302, "constructor": 303, "__call__": 304, "obj": 305, "type": 306, "_decorate_class": 307, "_decorate_fun": 308, "call": 309, "wrapped": 310, "args": 311, "warn": 312, "msg": 313, "category": 314, "deprecationwarning": 315, "fun": 316, "'strict'": 317, "codecs": 318, "charmap_decode": 319, "decoding_table": 320, "encode": 321, "final": 322, "charmap_encode": 323, "encoding_table": 324, "remove_module": 325, "lines": 326, "readlines": 327, "pattern": 328, "get_pattern": 329, "checkmode": 330, "open": 331, "os": 332, "devnull": 333, "'w'": 334, "filename": 335, "line": 336, "re": 337, "match": 338, "strip": 339, "write": 340, "close": 341, "module": 342, "add_module": 343, "'a'": 344, "'blacklist %s\\n'": 345, "%": 346, "addition": 347, "values": 348, "key": 349, "\"looks up key in the cache and returns the value for it.\n if the key does not exist `none` is returned instead.\n\n :param key: the key to be looked up.\n \"": 350, "looks": 351, "cache": 352, "it": 353, "does": 354, "exist": 355, "returned": 356, "instead": 357, "param": 358, "looked": 359, "retrieve": 360, "\"deletes `key` from the cache. if it does not exist in the cache\n nothing happens.\n\n :param key: the key to delete.\n \"": 361, "deletes": 362, "nothing": 363, "happens": 364, "get_many": 365, "keys": 366, "\"returns a list of values for the given keys.\n for each key a item in the list is created. example::\n\n foo, bar = cache.get_many(\"": 367, "foo": 368, "\", \"": 369, "bar": 370, "\")\n\n if a key can't be looked up `none` is returned for that key\n instead.\n\n :param keys: the function accepts multiple keys as positional\n arguments.\n \"": 371, "each": 372, "created": 373, "\"foo\"": 374, "\"bar\"": 375, "'": 376, "t": 377, "accepts": 378, "multiple": 379, "positional": 380, "arguments": 381, "many": 382, "get_dict": 383, "\"works like :meth:`get_many` but returns a dict::\n\n d = cache.get_dict(\"": 384, "\")\n foo = d[\"": 385, "\"]\n bar = d[\"": 386, "\"]\n\n :param keys: the function accepts multiple keys as positional\n arguments.\n \"": 387, "zip": 388, "works": 389, "meth": 390, "but": 391, "d": 392, "timeout": 393, "\"adds a new key/value to the cache (overwrites value, if key already\n exists in the cache).\n\n :param key: the key to set\n :param value: the value for the key\n :param timeout: the cache timeout for the key (if not specified,\n it uses the default timeout).\n \"": 394, "adds": 395, "new": 396, "/": 397, "overwrites": 398, "already": 399, "exists": 400, "specified": 401, "uses": 402, "assign": 403, "\"works like :meth:`set` but does not overwrite the values of already\n existing keys.\n\n :param key: the key to set\n :param value: the value for the key\n :param timeout: the cache timeout for the key or the default\n timeout if not specified.\n \"": 404, "overwrite": 405, "existing": 406, "set_many": 407, "mapping": 408, "\"sets multiple keys and values from a mapping.\n\n :param mapping: a mapping with the keys/values to set.\n :param timeout: the cache timeout for the key (if not specified,\n it uses the default timeout).\n \"": 409, "_items": 410, "sets": 411, "delete_many": 412, "\"deletes multiple keys at once.\n\n :param keys: the function accepts multiple keys as positional\n arguments.\n \"": 413, "at": 414, "once": 415, "clear": 416, "\"clears the cache. keep in mind that not all caches support\n completely clearing the cache.\n \"": 417, "clears": 418, "keep": 419, "mind": 420, "caches": 421, "support": 422, "completely": 423, "clearing": 424, "inc": 425, "delta": 426, "\"increments the value of a key by `delta`. if the key does\n not yet exist it is initialized with `delta`.\n\n for supporting caches this is an atomic operation.\n\n :param key: the key to increment.\n :param delta: the delta to add.\n \"": 427, "increments": 428, "yet": 429, "initialized": 430, "supporting": 431, "atomic": 432, "operation": 433, "increment": 434, "dec": 435, "\"decrements the value of a key by `delta`. if the key does\n not yet exist it is initialized with `-delta`.\n\n for supporting caches this is an atomic operation.\n\n :param key: the key to increment.\n :param delta: the delta to subtract.\n \"": 436, "decrements": 437, "subtract": 438, "import_preferred_memcache_lib": 439, "servers": 440, "\"returns an initialized memcache client. used by the constructor.\"": 441, "pylibmc": 442, "importerror": 443, "client": 444, "google": 445, "appengine": 446, "api": 447, "memcache": 448, "preferred": 449, "lib": 450, "dump_object": 451, "\"dumps an object into a string for redis. by default it serializes\n integers as regular string and pickle dumps everything else.\n \"": 452, "integer_types": 453, "str": 454, "'ascii'": 455, "b": 456, "'!'": 457, "pickle": 458, "dumps": 459, "object": 460, "into": 461, "redis": 462, "serializes": 463, "integers": 464, "regular": 465, "everything": 466, "dump": 467, "load_object": 468, "\"the reversal of :meth:`dump_object`. this might be callde with\n none.\n \"": 469, "startswith": 470, "loads": 471, "valueerror": 472, "before": 473, "0.8": 474, "we": 475, "did": 476, "have": 477, "serialization": 478, "still": 479, "reversal": 480, "might": 481, "callde": 482, "load": 483, "read": 484, "sync": 485, "account_id": 486, "app": 487, "app_context": 488, "account": 489, "zendeskaccount": 490, "query": 491, "session": 492, "organization": 493, "create_session": 494, "m": 495, "class_registry": 496, "items": 497, "isclass": 498, "issubclass": 499, "base": 500, "todo": 501, "make": 502, "these": 503, "parallel": 504, "speed": 505, "execution": 506, "conn": 507, "connection": 508, "statments": 509, "cursor": 510, "has_more": 511, "cursors": 512, "__tablename__": 513, "flag_modified": 514, "'cursors'": 515, "statement": 516, "sta": 517, "__repr__": 518, "'<%s: %s>'": 519, "__class__": 520, "__name__": 521, "label": 522, "repr": 523, "entry": 524, "\"\n factory that creates an app config from an entry in installed_apps.\n \"": 525, "import_module": 526, "succeeds": 527, "may": 528, "specify": 529, "config": 530, "default_app_config": 531, "otherwise": 532, "error": 533, "track": 534, "importing": 535, "failed": 536, "fails": 537, "too": 538, "ll": 539, "trigger": 540, "again": 541, "factory": 542, "creates": 543, "installed_apps": 544, "get_model": 545, "model_name": 546, "require_ready": 547, "\"\n return the model with the given case-insensitive model_name.\n\n raise lookuperror if no model exists with this name.\n \"": 548, "check_models_ready": 549, "check_apps_ready": 550, "models": 551, "lower": 552, "keyerror": 553, "lookuperror": 554, "\"app '%s' doesn't have a '%s' model.\"": 555, "model": 556, "case": 557, "insensitive": 558, "no": 559, "name": 560, "get_models": 561, "include_auto_created": 562, "include_swapped": 563, "\"\n return an iterable of models.\n\n by default, the following models aren't included:\n\n - auto-created models for many-to-many relations without\n an explicit intermediate table,\n - models that have been swapped out.\n\n set the corresponding keyword argument to true to include such models.\n keyword arguments aren't documented; they're a private api.\n \"": 564, "mode": 565, "following": 566, "aren": 567, "included": 568, "auto": 569, "relations": 570, "without": 571, "explicit": 572, "intermediate": 573, "table": 574, "been": 575, "swapped": 576, "out": 577, "corresponding": 578, "keyword": 579, "argument": 580, "include": 581, "such": 582, "import_models": 583, "dictionary": 584, "primarily": 585, "maintained": 586, "'all_models'": 587, "attribute": 588, "appconfig": 589, "attached": 590, "all_models": 591, "module_has_submodule": 592, "models_module_name": 593, "'%s.%s'": 594, "models_module": 595, "ready": 596, "\"\n override this method in subclasses to run code when django starts.\n \"": 597, "override": 598, "subclasses": 599, "when": 600, "django": 601, "starts": 602, "load_fixture": 603, "join": 604, "fixture_path": 605, "fixture_data": 606, "json": 607, "exception": 608, "fixture": 609, "test_module_parameters": 610, "state": 611, "'present'": 612, "contact": 613, "'foo'": 614, "description": 615, "'bar'": 616, "location": 617, "'baz'": 618, "'datacenter'": 619, "p": 620, "moduleparameters": 621, "params": 622, "assert": 623, "testing": 624, "test": 625, "test_create_disabled_datacenter": 626, "set_module_args": 627, "'disabled'": 628, "'admin'": 629, "server": 630, "'localhost'": 631, "ansiblemodule": 632, "argument_spec": 633, "spec": 634, "supports_check_mode": 635, "mm": 636, "modulemanager": 637, "methods": 638, "force": 639, "specific": 640, "logic": 641, "happen": 642, "mock": 643, "side_effect": 644, "create_on_device": 645, "ret": 646, "disabled": 647, "datacenter": 648, "test_create_enabled_datacenter": 649, "'enabled'": 650, "retur": 651, "enabled": 652, "test_idempotent_disable_datacenter": 653, "current": 654, "apiparameters": 655, "'load_gtm_datacenter_disabled.json'": 656, "ha": 657, "idempotent": 658, "disable": 659, "test_getlasterror": 660, "dll": 661, "windll": 662, "\"kernel32\"": 663, "use_last_error": 664, "getmodulehandle": 665, "getmodulehandlea": 666, "argtypes": 667, "c_wchar_p": 668, "assertequal": 669, "get_last_error": 670, "126": 671, "set_last_error": 672, "32": 673, "_worker": 674, "getmodulehandlew": 675, "getlasterror": 676, "get_test_db_clone_settings": 677, "orig_settings_dict": 678, "settings_dict": 679, "source_database_name": 680, "'name'": 681, "is_in_memory_db": 682, "new_settings_dict": 683, "root": 684, "ext": 685, "splitext": 686, "'{}_{}.{}'": 687, "db": 688, "clone": 689, "settings": 690, "test_db_signature": 691, "\"\n returns a tuple that uniquely identifies a test database.\n\n this takes into account the special cases of \"": 692, "memory": 693, "\" and \"": 694, "\" for\n sqlite since the databases will be distinct despite having the same\n test name. see http://www.sqlite.org/inmemorydb.html\n \"": 695, "test_database_name": 696, "_get_test_db_name": 697, "sig": 698, "append": 699, "tuple": 700, "uniquely": 701, "identifies": 702, "takes": 703, "special": 704, "cases": 705, "\":memory:\"": 706, "sqlite": 707, "since": 708, "databases": 709, "distinct": 710, "despite": 711, "having": 712, "see": 713, "http": 714, "www": 715, "org": 716, "inmemorydb": 717, "html": 718, "signature": 719, "clean_params": 720, "cleaned_data": 721, "'params'": 722, "validationerror": 723, "'is not valid json'": 724, "clean": 725, "clean_alias": 726, "'alias'": 727, "connections": 728, "\"database alias '%s' not found\"": 729, "reformat_sql": 730, "'sql'": 731, "reformat": 732, "sql": 733, "formatting": 734, "make_hash": 735, "secret_key": 736, "endings": 737, "spaces": 738, "preserve": 739, "hash": 740, "browser": 741, "normalizes": 742, "\\": 743, "r": 744, "inputs": 745, "' '": 746, "force_text": 747, "splitlines": 748, "hashlib": 749, "sha1": 750, "'utf-8'": 751, "hexdigest": 752, "hashing": 753, "__len__": 754, "set_units": 755, "airspeed_units": 756, "altitude_units": 757, "units": 758, "update_lla": 759, "lla": 760, "modify": 761, "update_ned_history": 762, "ned": 763, "seconds": 764, "time": 765, "ned_last_time": 766, "ned_history": 767, "pop": 768, "history": 769, "update_features": 770, "feature_list": 771, "features": 772, "update_vel": 773, "vn": 774, "ve": 775, "vd": 776, "vel": 777, "update_ap": 778, "flight_mode": 779, "ap_roll": 780, "ap_pitch": 781, "ap_hdg": 782, "ap_speed": 783, "ap_altitude_ft": 784, "ap": 785, "update_pilot": 786, "aileron": 787, "elevator": 788, "throttle": 789, "rudder": 790, "pilot_ail": 791, "pilot_ele": 792, "pilot_thr": 793, "pilot_rud": 794, "pilot": 795, "update_act": 796, "act_ail": 797, "act_ele": 798, "act_thr": 799, "act_rud": 800, "act": 801, "compute_sun_moon_ned": 802, "lon_deg": 803, "lat_deg": 804, "alt_m": 805, "timestamp": 806, "datetime": 807, "utcfromtimestamp": 808, "utcnow": 809, "ed": 810, "ephem": 811, "date": 812, "'ephem time utc:'": 813, "'localtime:'": 814, "localtime": 815, "ownship": 816, "observer": 817, "lon": 818, "'%.8f'": 819, "lat": 820, "elevation": 821, "sun": 822, "moon": 823, "sun_ned": 824, "math": 825, "cos": 826, "az": 827, "alt": 828, "draw_pitch_ladder": 829, "beta_rad": 830, "0.0": 831, "a1": 832, "2.0": 833, "a2": 834, "8.0": 835, "slide_rad": 836, "psi_rad": 837, "q0": 838, "transformations": 839, "quaternion_about_axis": 840, "1.0": 841, "a0": 842, "range": 843, "35": 844, "above": 845, "horizon": 846, "right": 847, "horizontal": 848, "uv1": 849, "ladder_helper": 850, "uv2": 851, "!": 852, "cv2": 853, "frame": 854, "color": 855, "line_width": 856, "line_aa": 857, "draw": 858, "pitch": 859, "ladder": 860, "rotate_pt": 861, "center": 862, "sin": 863, "y": 864, "rotate": 865, "pt": 866, "draw_heading_bug": 867, "medium_orchid": 868, "atan2": 869, "d2r": 870, "pts": 871, "1.5": 872, "ladde": 873, "heading": 874, "bug": 875, "draw_course": 876, "yellow": 877, "filter_vn": 878, "tf_vel": 879, "filter_ve": 880, "uv3": 881, "tmp2": 882, "course": 883, "draw_label": 884, "uv": 885, "font_scale": 886, "thickness": 887, "horiz": 888, "'center'": 889, "vert": 890, "gettextsize": 891, "font": 892, "u": 893, "'above'": 894, "v": 895, "'below'": 896, "puttext": 897, "draw_lla_point": 898, "pt_ned": 899, "navpy": 900, "lla2ned": 901, "ref": 902, "rel_ned": 903, "hdist": 904, "sqrt": 905, "dist": 906, "m2sm": 907, "0.000621371": 908, "hdist_sm": 909, "<": 910, "10.0": 911, "point": 912, "draw_sticks": 913, "'auto'": 914, "h": 915, "w": 916, "lx": 917, "0.1": 918, "ly": 919, "rx": 920, "ry": 921, "r1": 922, "round": 923, "0.09": 924, "r2": 925, "sticks": 926, "draw_track": 927, "uv_list": 928, "dist_list": 929, "dn": 930, "de": 931, "dd": 932, "project_point": 933, "i": 934, "draw_features": 935, "render_w": 936, "0.25": 937, "1.25": 938, "render_h": 939, "circle": 940, "white": 941, "is_string": 942, "txt": 943, "nameerror": 944, "representation": 945, "description_bit": 946, "'content'": 947, "contents": 948, "content": 949, "'content_'": 950, "content_": 951, "'value'": 952, "'expecting a string or something with content, content_ or value attribute'": 953, "bit": 954, "paragraph": 955, "then": 956, "some": 957, "breaks": 958, "schur": 959, "'real'": 960, "lwork": 961, "overwrite_a": 962, "decomposition": 963, "A": 964, "Z": 965, "T": 966, "^": 967, "H": 968, "where": 969, "unitary": 970, "either": 971, "upper": 972, "triangular": 973, "real": 974, "quasi": 975, "form": 976, "x2": 977, "blocks": 978, "describing": 979, "complex": 980, "valued": 981, "eigenvalue": 982, "pairs": 983, "extrude": 984, "diagonal": 985, "M": 986, "decompose": 987, "'complex'": 988, "construct": 989, "ma": 990, "rsf2csf": 991, "original": 992, "transformation": 993, "also": 994, "map_options": 995, "\"builds the map options hash for the openlayers template.\"": 996, "javascript": 997, "construction": 998, "utilities": 999, "bounds": 1000, "projection": 1001, "ol_bounds": 1002, "extent": 1003, "'new openlayers.bounds(%s)'": 1004, "ol_projection": 1005, "srid": 1006, "'new openlayers.projection(\"epsg:%s\")'": 1007, "parameter": 1008, "their": 1009, "openlayers": 1010, "counterpart": 1011, "variable": 1012, "map_types": 1013, "'srid'": 1014, "'projection'": 1015, "'display_srid'": 1016, "'displayprojection'": 1017, "builds": 1018, "template": 1019, "ol": 1020, "stripescapes": 1021, "\"\n remove all ansi color escapes from the given string.\n \"": 1022, "show": 1023, "L": 1024, "_sets": 1025, "'\\x1b'": 1026, "resu": 1027, "ansi": 1028, "escapes": 1029, "parsestring": 1030, "\"\n turn a string input into a list of l{colortext} elements.\n \"": 1031, "prepend": 1032, "parts": 1033, "split": 1034, "writestring": 1035, "formattext": 1036, "s": 1037, "digits": 1038, "turn": 1039, "colortext": 1040, "elements": 1041, "parsing": 1042, "parsecolor": 1043, "\"\n handle a single ansi color sequence\n \"": 1044, "drop": 1045, "trailing": 1046, "'m'": 1047, "'0'": 1048, "';'": 1049, "log": 1050, "'invalid ansi color sequence (%d): %s'": 1051, "currentfg": 1052, "currentbg": 1053, "defaultfg": 1054, "defaultbg": 1055, "single": 1056, "pickcolor": 1057, "bold": 1058, "bold_colors": 1059, "colors": 1060, "test_is_all_white": 1061, "asserttrue": 1062, "fp": 1063, "is_all_white": 1064, "'\\t\\n\\r\\f\\v'": 1065, "assertfalse": 1066, "test_comment": 1067, "test_get_indent": 1068, "equal": 1069, "get_indent": 1070, "trailingws_comment": 1071, "leadingws_comment": 1072, "' '": 1073, "leadingws_nocomment": 1074, "indent": 1075, "undo_block_start": 1076, "undo": 1077, "block": 1078, "start": 1079, "has_prev_data": 1080, "_prev": 1081, "has": 1082, "prev": 1083, "has_more_data": 1084, "_more": 1085, "action_open_window": 1086, "cr": 1087, "uid": 1088, "ids": 1089, "context": 1090, "\"\n this function open account move line window for reconcile on given account id\n @param cr: the current row, from the database cursor,\n @param uid: the current user\u2019s id for security checks,\n @param ids: account move line reconcile select\u2019s id or list of ids\n @return: dictionary of open account move line window for reconcile on given account id\n\n \"": 1091, "'domain'": 1092, "'account_id'": 1093, "'='": 1094, "reco": 1095, "move": 1096, "window": 1097, "reconcile": 1098, "id": 1099, "row": 1100, "\u2019": 1101, "security": 1102, "select": 1103, "li": 1104, "action": 1105, "import_as": 1106, "\"\n imports the specified module (from our local directory) as the\n specified name, returning the loaded module object.\n \"": 1107, "dir": 1108, "__file__": 1109, "imp": 1110, "load_module": 1111, "find_module": 1112, "imports": 1113, "our": 1114, "local": 1115, "directory": 1116, "returning": 1117, "loaded": 1118, "rename_module": 1119, "old": 1120, "\"\n attempts to import the old module and load it under the new name.\n used for purely cosmetic name changes in python 3.x.\n \"": 1121, "sys": 1122, "modules": 1123, "attempts": 1124, "under": 1125, "purely": 1126, "cosmetic": 1127, "changes": 1128, "python": 1129, "3.": 1130, "rename": 1131, "shader": 1132, "\"defaultsurface\"": 1133, "\"surface\"": 1134, "typename": 1135, "\"compounddata\"": 1136, "\"marble\"": 1137, "ss": 1138, "testhash": 1139, "\"somethingelse\"": 1140, "assertnotequal": 1141, "\"a\"": 1142, "stringdata": 1143, "autoname": 1144, "suffix": 1145, "\" - \"": 1146, "frappe": 1147, "get_value": 1148, "\"company\"": 1149, "company": 1150, "\"abbr\"": 1151, "warehouse_name": 1152, "endswith": 1153, "update_parent_account": 1154, "getattr": 1155, "\"__islocal\"": 1156, "create_account_under": 1157, "\"warehouse\"": 1158, "\"create_account_under\"": 1159, "validate_parent_account": 1160, "warehouse_account": 1161, "\"account\"": 1162, "\"account_type\"": 1163, "\"master_name\"": 1164, "\"name\"": 1165, "\"parent_account\"": 1166, "acc_doc": 1167, "get_doc": 1168, "parent_account": 1169, "parent": 1170, "counting": 1171, "on_trash": 1172, "bin": 1173, "bins": 1174, "\"select * from `tabbin` where warehouse = %s\"": 1175, "as_dict": 1176, "'actual_qty'": 1177, "'reserved_qty'": 1178, "'ordered_qty'": 1179, "'indented_qty'": 1180, "'projected_qty'": 1181, "'planned_qty'": 1182, "throw": 1183, "_": 1184, "\"warehouse {0} can not be deleted as quantity exists for item {1}\"": 1185, "'item_code'": 1186, "\"delete from `tabbin` where name = %s\"": 1187, "wareho": 1188, "trash": 1189, "recalculate_bin_qty": 1190, "newdn": 1191, "erpnext": 1192, "repost_stock": 1193, "auto_commit_on_many_writes": 1194, "set_default": 1195, "\"allow_negative_stock\"": 1196, "\"select distinct item_code from (\n\t\t\tselect name as item_code from `tabitem` where ifnull(is_stock_item, 'yes')='yes'\n\t\t\tunion\n\t\t\tselect distinct item_code from tabbin) a\"": 1197, "\"stock settings\"": 1198, "recalculate": 1199, "qty": 1200, "setup": 1201, "_threads": 1202, "threading_setup": 1203, "test_acquire_contended": 1204, "lock": 1205, "locktype": 1206, "acquire": 1207, "N": 1208, "release": 1209, "bunch": 1210, "wait_for_started": 1211, "_wait": 1212, "finished": 1213, "wait_for_finished": 1214, "contended": 1215, "test_reacquire": 1216, "reacquire": 1217, "test_different_thread": 1218, "cannot": 1219, "different": 1220, "thread": 1221, "assertraises": 1222, "runtimeerror": 1223, "do_finish": 1224, "difference": 1225, "test_acquire": 1226, "cond": 1227, "condtype": 1228, "rlock": 1229, "condition": 1230, "acquired": 1231, "times": 1232, "threading": 1233, "test_unacquired_wait": 1234, "wait": 1235, "unacquired": 1236, "test_unacquired_notify": 1237, "notify": 1238, "test_timeout": 1239, "results": 1240, "t1": 1241, "0.2": 1242, "t2": 1243, "dt": 1244, "test_constructor": 1245, "semtype": 1246, "maxint": 1247, "test_acquire_destroy": 1248, "sem": 1249, "del": 1250, "destroy": 1251, "test_default_value": 1252, "initial": 1253, "1.": 1254, "test_with": 1255, "_with": 1256, "err": 1257, "typeerror": 1258, "test_release_unacquired": 1259, "unbounded": 1260, "releases": 1261, "allowed": 1262, "semaphore": 1263, "task": 1264, "tid": 1265, "get_ident": 1266, "started": 1267, "_can_exit": 1268, "test_timesince01": 1269, "engine": 1270, "render_to_string": 1271, "'timesince01'": 1272, "now": 1273, "timedelta": 1274, "minutes": 1275, "'1\\xa0minute'": 1276, "timesince01": 1277, "multiply": 1278, "product": 1279, "test_timesince02": 1280, "'timesince02'": 1281, "days": 1282, "'1\\xa0day'": 1283, "timesince02": 1284, "test_timesince09": 1285, "'timesince09'": 1286, "'later'": 1287, "'0\\xa0minutes'": 1288, "timesince09": 1289, "test_timesince11": 1290, "'timesince11'": 1291, "timesince11": 1292, "test_timesince12": 1293, "'timesince12'": 1294, "now_tz": 1295, "timesince12": 1296, "test_explicit_date": 1297, "timesince_filter": 1298, "2005": 1299, "12": 1300, "29": 1301, "30": 1302, "extract_from_dict": 1303, "\"\n navigate `data`, a multidimensional array (list or dictionary), and returns\n the object at `path`.\n \"": 1304, "indexerror": 1305, "navigate": 1306, "multidimensional": 1307, "extract": 1308, "get_default_scope": 1309, "scope": 1310, "userinfo_profile": 1311, "extract_uid": 1312, "\"orcid-identifier\"": 1313, "\"path\"": 1314, "extract_common_fields": 1315, "common_fields": 1316, "\"person\"": 1317, "\"emails\"": 1318, "\"email\"": 1319, "last_name": 1320, "\"family-name\"": 1321, "\"value\"": 1322, "first_name": 1323, "\"given-names\"": 1324, "common": 1325, "fields": 1326, "is_package": 1327, "'.'": 1328, "_fullname": 1329, "package": 1330, "apply_gradients": 1331, "grads_and_vars": 1332, "global_step": 1333, "tf": 1334, "_switch": 1335, "_optimizer1": 1336, "_optimizer2": 1337, "gradients": 1338, "get_slot": 1339, "var": 1340, "slot1": 1341, "slot2": 1342, "'slot named %s for variable %s populated for both '": 1343, "'optimizers'": 1344, "slot": 1345, "test_searchpath": 1346, "'\n the searchpath is based on the cachedir option and the saltenv parameter\n '": 1347, "tmp": 1348, "tempfile": 1349, "gettempdir": 1350, "opts": 1351, "deepcopy": 1352, "'cachedir'": 1353, "loader": 1354, "saltcacheloader": 1355, "saltenv": 1356, "'test'": 1357, "searchpath": 1358, "'files'": 1359, "based": 1360, "cachedir": 1361, "option": 1362, "test_mockclient": 1363, "'\n a mockfileclient is used that records all file requests normally sent\n to the master.\n '": 1364, "fc": 1365, "mockfileclient": 1366, "res": 1367, "get_source": 1368, "'hello_simple'": 1369, "windows": 1370, "unicode": 1371, "linesep": 1372, "so": 1373, "cross": 1374, "'world'": 1375, "tmpl_dir": 1376, "templates_dir": 1377, "ass": 1378, "records": 1379, "requests": 1380, "normally": 1381, "sent": 1382, "master": 1383, "mockclient": 1384, "get_test_saltenv": 1385, "'\n setup a simple jinja test environment\n '": 1386, "jinja": 1387, "environment": 1388, "simple": 1389, "test_import": 1390, "'\n you can import and use macros from other files\n '": 1391, "get_template": 1392, "'hello_import'": 1393, "render": 1394, "'hey world !a b !'": 1395, "'path'": 1396, "'salt://hello_import'": 1397, "'salt://macro'": 1398, "you": 1399, "macros": 1400, "other": 1401, "files": 1402, "test_include": 1403, "'\n you can also include a template that imports and uses macros\n '": 1404, "'hello_include'": 1405, "'salt://hello_include'": 1406, "test_include_context": 1407, "'\n context variables are passes to the included template by default.\n '": 1408, "'hi'": 1409, "'salt'": 1410, "'hey world !hi salt !'": 1411, "variables": 1412, "passes": 1413, "test_fallback": 1414, "'\n a template with a filesystem loader is returned as fallback\n if the file is not contained in the searchpath\n '": 1415, "fn_": 1416, "salt": 1417, "fopen": 1418, "fp_": 1419, "render_jinja_tmpl": 1420, "local_opts": 1421, "'world\\n'": 1422, "filesystem": 1423, "fallback": 1424, "contained": 1425, "test_fallback_noloader": 1426, "'hey world !a b !\\n'": 1427, "noloader": 1428, "test_saltenv": 1429, "'\n if the template is within the searchpath it can\n import, include and extend other templates.\n the initial template is expected to be already cached\n get_template does not request it from the master again.\n '": 1430, "monkey": 1431, "patch": 1432, "_fc": 1433, "file_client": 1434, "filen": 1435, "within": 1436, "extend": 1437, "templates": 1438, "expected": 1439, "cached": 1440, "request": 1441, "test_macro_additional_log_for_generalexc": 1442, "'\n if we failed in a macro because of e.g. a typeerror, get\n more output from trace.\n '": 1443, "'jinja error:.*division.*\n.*/macrogeneral\\(2\\):\n---\n\\{% macro mymacro\\(\\) -%\\}\n\\{\\{ 1/0 \\}\\} <======================\n\\{%- endmacro %\\}\n---.*'": 1444, "'hello_import_generalerror'": 1445, "macro": 1446, "because": 1447, "e": 1448, "g": 1449, "trace": 1450, "additional": 1451, "generalexc": 1452, "test_macro_additional_log_for_undefined": 1453, "'\n if we failed in a macro because of undefined variables, get\n more output from trace.\n '": 1454, "'jinja variable '": 1455, "' is undefined\n.*/macroundefined\\(2\\):\n---\n\\{% macro mymacro\\(\\) -%\\}\n\\{\\{b.greetee\\}\\} <-- error is here <======================\n\\{%- endmacro %\\}\n---'": 1456, "'hello_import_undefined'": 1457, "file_": 1458, "undefined": 1459, "test_macro_additional_log_syntaxerror": 1460, "'\n if we failed in a macro, get more output from trace.\n '": 1461, "'jinja syntax error: expected token .*end.*got '": 1462, "'.*\n.*/macroerror\\(2\\):\n---\n# macro\n\\{% macro mymacro\\(greeting, greetee='": 1463, "world": 1464, "'\\) -\\} <-- error is here <======================\n\\{\\{ greeting ~ '": 1465, "' ~ greetee \\}\\} !\n\\{%- endmacro %\\}\n---.*'": 1466, "'hello_import_error'": 1467, "syntaxerror": 1468, "test_non_ascii": 1469, "fn": 1470, "'non_ascii'": 1471, "'data'": 1472, "'assun\u00e7\u00e3o\\n'": 1473, "non": 1474, "ascii": 1475, "test_get_context_at_top_of_file": 1476, "'1\\n2\\n3\\n4\\n5\\n6\\n7\\n8\\n9\\na\\nb\\nc\\nd\\ne\\nf'": 1477, "get_context": 1478, "'---\\n1\\n2\\n3\\n4\\n5\\n6\\n[...]\\n---'": 1479, "top": 1480, "test_get_context_at_bottom_of_file": 1481, "15": 1482, "'---\\n[...]\\na\\nb\\nc\\nd\\ne\\nf\\n---'": 1483, "bottom": 1484, "test_get_context_with_marker": 1485, "num_lines": 1486, "marker": 1487, "' <---'": 1488, "'---\\n[...]\\n6\\n7\\n8 <---\\n9\\na\\n[...]\\n---'": 1489, "test_render_with_undefined_variable_unicode": 1490, "\"hello\\ud55c\\n\\n{{ foo }}\\n\\nfoo\"": 1491, "'jinja variable \\'": 1492, "' is undefined'": 1493, "assertraisesregexp": 1494, "saltrendererror": 1495, "test_serialize_json": 1496, "dataset": 1497, "42": 1498, "\"baz\"": 1499, "\"qux\"": 1500, "env": 1501, "extensions": 1502, "serializerextension": 1503, "rendered": 1504, "from_string": 1505, "'{{ dataset|json }}'": 1506, "serialize": 1507, "test_serialize_python": 1508, "'{{ dataset|python }}'": 1509, "pprint": 1510, "pformat": 1511, "test_load_tag": 1512, "source": 1513, "'{{ bar }}, '": 1514, "'{% load_yaml as docu %}{foo: it works, {{ bar }}: baz}{% endload %}'": 1515, "'{{ docu.foo }}'": 1516, "\"barred\"": 1517, "\"barred, it works\"": 1518, "'{{ bar }}, {% load_json as docu %}{\"foo\": \"it works\", \"{{ bar }}\": \"baz\"}{% endload %}'": 1519, "tag": 1520, "test_sequence": 1521, "filters": 1522, "'sequence'": 1523, "ensure_sequence_filter": 1524, "'{{ data | sequence | length }}'": 1525, "'1'": 1526, "'2'": 1527, "test_normlookup": 1528, "'\n sanity-check the normal dictionary-lookup syntax for our stub function\n '": 1529, "tmpl_str": 1530, "'hello, {{ salt['": 1531, "mocktest": 1532, "ping": 1533, "']() }}.'": 1534, "'hello, true.'": 1535, "sanity": 1536, "normal": 1537, "lookup": 1538, "syntax": 1539, "stub": 1540, "normlookup": 1541, "test_dotlookup": 1542, "'\n check calling a stub function using awesome dot-notation\n '": 1543, "'hello, {{ salt.mocktest.ping() }}.'": 1544, "calling": 1545, "using": 1546, "awesome": 1547, "dot": 1548, "notation": 1549, "dotlookup": 1550, "test_shadowed_dict_method": 1551, "'\n check calling a stub function with a name that shadows a ``dict``\n method name\n '": 1552, "'hello, {{ salt.mockgrains.get('": 1553, "') }}.'": 1554, "'hello, jerry.'": 1555, "shadows": 1556, "shadowed": 1557, "create_webdriver_exception_strict": 1558, "status_code": 1559, "message": 1560, "\"create the appropriate webdriverexception given the status_code.\"": 1561, "_exceptions_strict": 1562, "unknownstatuscodeexception": 1563, "\"[%s] %s\"": 1564, "appropriate": 1565, "webdriverexception": 1566, "webdriver": 1567, "strict": 1568, "create_webdriver_exception_compatibility": 1569, "_exceptions_compatibility": 1570, "compatibility": 1571, "get_object": 1572, "queryset": 1573, "\"\n returns the object the view is displaying.\n\n by default this requires `self.queryset` and a `pk` or `slug` argument\n in the urlconf, but subclasses can override this to return any object.\n \"": 1574, "custom": 1575, "provided": 1576, ";": 1577, "required": 1578, "datedetailview": 1579, "get_queryset": 1580, "next": 1581, "looking": 1582, "primary": 1583, "pk": 1584, "pk_url_kwarg": 1585, "slug": 1586, "slug_url_kwarg": 1587, "view": 1588, "displaying": 1589, "requires": 1590, "urlconf": 1591, "any": 1592, "\"\n return the `queryset` that will be used to look up the object.\n\n note that this method is called by the default implementation of\n `get_object` and may not be called if `get_object` is overridden.\n \"": 1593, "_default_manager": 1594, "improperlyconfigured": 1595, "\"%(cls)s is missing a queryset. define \"": 1596, "\"%(cls)s.model, %(cls)s.queryset, or override \"": 1597, "look": 1598, "note": 1599, "called": 1600, "implementation": 1601, "overridden": 1602, "get_slug_field": 1603, "\"\n get the name of a slug field to be used to look up by slug.\n \"": 1604, "slug_field": 1605, "field": 1606, "get_context_object_name": 1607, "\"\n get the name to use for the object.\n \"": 1608, "context_object_name": 1609, "_deferred": 1610, "_meta": 1611, "proxy_for_model": 1612, "get_context_data": 1613, "\"\n insert the single object into the context dict.\n \"": 1614, "'object'": 1615, "super": 1616, "singleobjectmixin": 1617, "insert": 1618, "get_template_names": 1619, "\"\n return a list of template names to be used for the request. may not be\n called if render_to_response is overridden. returns the following list:\n\n * the value of ``template_name`` on the view (if provided)\n * the contents of the ``template_name_field`` field on the\n object instance that the view is operating upon (if available)\n * ``/.html``\n \"": 1620, "names": 1621, "singleobjecttemplateresponsemixin": 1622, "render_to_response": 1623, "template_name": 1624, "template_name_field": 1625, "ope": 1626, "register_chooser": 1627, "chooser": 1628, "\"adds a model chooser definition to the registry.\"": 1629, "register_simple_chooser": 1630, "choosers": 1631, "definition": 1632, "\"\n generates a model chooser definition from a model, and adds it to the\n registry.\n \"": 1633, "'{}chooser'": 1634, "object_name": 1635, "attrs": 1636, "'model'": 1637, "generates": 1638, "flash_nat_pause": 1639, "my": 1640, "_get_unique_filename": 1641, "flash": 1642, "nat": 1643, "pause": 1644, "flash_final_wave": 1645, "wave": 1646, "sign": 1647, "K": 1648, "piece": 1649, "dsa": 1650, "longer": 1651, "sub": 1652, "group": 1653, "order": 1654, "q": 1655, "byte": 1656, "long": 1657, "secret": 1658, "chosen": 1659, "randomly": 1660, "closed": 1661, "recommended": 1662, "attention": 1663, "selection": 1664, "crucial": 1665, "generating": 1666, "larger": 1667, "modulus": 1668, "secure": 1669, "smaller": 1670, "wil": 1671, "verify": 1672, "\"verify the validity of a dsa signature.\n\n :parameter m: the expected message.\n :type m: byte string or long\n\n :parameter signature: the dsa signature to verify.\n :type signature: a tuple with 2 longs as return by `sign`\n\n :return: true if the signature is correct, false otherwise.\n \"": 1673, "pubkey": 1674, "validity": 1675, "longs": 1676, "correct": 1677, "publickey": 1678, "__getstate__": 1679, "k": 1680, "keydata": 1681, "attributeerror": 1682, "getstate": 1683, "bits": 1684, "randfunc": 1685, "progress_func": 1686, "fresh": 1687, "must": 1688, "64": 1689, "interval": 1690, "512": 1691, "1024": 1692, "callable": 1693, "generation": 1694, "accept": 1695, "integer": 1696, "tup": 1697, "valid": 1698, "components": 1699, "prime": 1700, "equations": 1701, "mod": 1702, "public": 1703, "2.": 1704, "finite": 1705, "4.": 1706, "S": 1707, "instrumentationsetup": 1708, "host_driven_test_root": 1709, "official_build": 1710, "instrumentation_options": 1711, "\"creates a list of host-driven instrumentation tests and a runner factory.\n\n args:\n host_driven_test_root: directory where the host-driven tests are.\n official_build: true if this is an official build.\n instrumentation_options: an instrumentationoptions object.\n\n returns:\n a tuple of (testrunnerfactory, tests).\n \"": 1712, "test_collection": 1713, "test_info_collection": 1714, "testinfocollection": 1715, "all_tests": 1716, "_getalltests": 1717, "instru": 1718, "host": 1719, "driven": 1720, "instrumentation": 1721, "tests": 1722, "runner": 1723, "official": 1724, "build": 1725, "instrumentationoptions": 1726, "testrunnerfac": 1727, "testrunnerfactory": 1728, "device": 1729, "shard_index": 1730, "test_runner": 1731, "hostdriventestrunner": 1732, "tool": 1733, "connect": 1734, "'utility function to connect to database.\n\n use this method to connect to the pipeline database.\n additional databases can be attached here as well.\n\n returns an sqlite3 database handle.\n '": 1735, "dbh": 1736, "sqlite3": 1737, "\"database_name\"": 1738, "\"gtf_database\"": 1739, "\"can't find database '%s'\"": 1740, "'attach database '": 1741, "' as annotations'": 1742, "cc": 1743, "execute": 1744, "utility": 1745, "pipeline": 1746, "here": 1747, "well": 1748, "countreads": 1749, "infile": 1750, "outfile": 1751, "'count number of reads in input files.'": 1752, "'printf \"nreads \\\\t\" >> %(outfile)s'": 1753, "P": 1754, "'samtools view %(infile)s | wc -l | xargs printf >> %(outfile)s'": 1755, "reads": 1756, "strandspecificity": 1757, "'this function will determine the strand specificity of your library\n from the bam file'": 1758, "iterations": 1759, "\"1000000\"": 1760, "pipelinebamstats": 1761, "getstrandspecificity": 1762, "determine": 1763, "strand": 1764, "specificity": 1765, "your": 1766, "library": 1767, "bam": 1768, "intbam": 1769, "'make an intermediate bam file if there is no sequence infomation.\n if there is no sequence quality then make a softlink. picard tools\n has an issue when quality score infomation is missing'": 1770, "\"bam_sequence_stripped\"": 1771, "addpseudosequencequality": 1772, "copybamfile": 1773, "there": 1774, "infomation": 1775, "quality": 1776, "softlink": 1777, "picard": 1778, "tools": 1779, "issue": 1780, "score": 1781, "missing": 1782, "buildpicardstats": 1783, "infiles": 1784, "' build picard alignment stats '": 1785, "reffile": 1786, "against": 1787, "transcriptome": 1788, "switch": 1789, "genomic": 1790, "reference": 1791, "transcriptomic": 1792, "sequences": 1793, "\"transcriptome.dir\"": 1794, "\"refcoding.fa\"": 1795, "buildpicardalignmentstats": 1796, "alignment": 1797, "stats": 1798, "buildpicardduplicationstats": 1799, "'get duplicate stats from picard markduplicates '": 1800, "duplicate": 1801, "markduplicates": 1802, "buildbamstats": 1803, "'count number of reads mapped, duplicates, etc.\n\n excludes regions overlapping repetitive rna sequences\n\n parameters\n ----------\n infiles : list\n infiles[0] : str\n input filename in :term:`bam` format\n infiles[1] : str\n input filename with number of reads per sample\n\n outfile : str\n output filename with read stats\n\n annotations_interface_rna_gtf : str\n :term:`parms`. :term:`gtf` format file with repetitive rna\n '": 1804, "rna_file": 1805, "\"annotations_interface_rna_gff\"": 1806, "job_memory": 1807, "\"32g\"": 1808, "mapped": 1809, "duplicates": 1810, "etc": 1811, "excludes": 1812, "regions": 1813, "overlapping": 1814, "repetitive": 1815, "rna": 1816, "term": 1817, "per": 1818, "sample": 1819, "wi": 1820, "processgenomiccontext": 1821, "'\n this module process genomic context file.\n it assigns each and every features of context\n file to a specific catagory. it helps us to\n understand heiarchical classification\n of features.\n '": 1822, "definebedfeatures": 1823, "assigns": 1824, "every": 1825, "catagory": 1826, "helps": 1827, "us": 1828, "understand": 1829, "heiarchical": 1830, "classification": 1831, "buildcontextstats": 1832, "' build mapping context stats '": 1833, "summarizetagswithincontext": 1834, "buildidxstats": 1835, "'gets idxstats for bam file so number of reads per chromosome can\n be plotted later'": 1836, "'samtools idxstats %(infile)s > %(outfile)s'": 1837, "gets": 1838, "idxstats": 1839, "chromosome": 1840, "plotted": 1841, "later": 1842, "buildexonvalidation": 1843, "'compare the alignments to the exon models to quantify exon\n overrun/underrun\n\n expectation is that reads should not extend beyond known exons.\n\n parameters\n ----------\n infiles : list\n infiles[0] : str\n input filename in :term:`bam` format\n infiles[1] : str\n input filename in :term:`gtf` format\n\n outfile : str\n output filename in :term:`gtf` format with exon validation stats\n '": 1844, "exons": 1845, "cat": 1846, "|": 1847, "cgat": 1848, "bam_vs_gtf": 1849, "compare": 1850, "alignments": 1851, "exon": 1852, "quantify": 1853, "overrun": 1854, "underrun": 1855, "expectation": 1856, "beyond": 1857, "known": 1858, "gtf": 1859, "outfil": 1860, "buildtranscriptlevelreadcounts": 1861, "gene": 1862, "overlap": 1863, "positions": 1864, "protein": 1865, "coding": 1866, "transcripts": 1867, "tsv": 1868, "paired": 1869, "end": 1870, "mate": 1871, "counted": 1872, "thus": 1873, "actual": 1874, "counts": 1875, "buildintronlevelreadcounts": 1876, "introns": 1877, "geneset": 1878, "approximat": 1879, "buildtranscriptprofiles": 1880, "'build gene coverage profiles\n\n polya-rna-seq is expected to show a bias towards the 3'": 1881, "meta": 1882, "profile": 1883, "depth": 1884, "across": 1885, "defined": 1886, "plots": 1887, "saved": 1888, "full": 1889, "focus": 1890, "views": 1891, "coverage": 1892, "profiles": 1893, "polya": 1894, "seq": 1895, "bias": 1896, "towards": 1897, "buildpicardrnaseqmetrics": 1898, "'get duplicate stats from picard rnaseqmetrics '": 1899, "strandness": 1900, "tophat": 1901, "style": 1902, "\"strandness\"": 1903, "\"rf\"": 1904, "\"r\"": 1905, "\"second_read_transcription_strand\"": 1906, "\"fr\"": 1907, "\"f\"": 1908, "\"first_read_transcription_strand\"": 1909, "\"none\"": 1910, "rnaseqmetrics": 1911, "loadpicardstats": 1912, "'merge alignment stats into single tables.'": 1913, "loadpicardalignmentstats": 1914, "merge": 1915, "tables": 1916, "loadpicardduplicationstats": 1917, "outfiles": 1918, "loadbamstats": 1919, "' load bam statistics into bam_stats table '": 1920, "statistics": 1921, "bam_stats": 1922, "loadcontextstats": 1923, "' load context mapping statistics into context_stats table '": 1924, "loadsummarizedcontextstats": 1925, "context_stats": 1926, "loadidxstats": 1927, "'merge idxstats files into single dataframe and load\n to database\n\n loads tables into the database\n * mapped_reads_per_chromosome\n\n arguments\n ---------\n infiles : list\n list where each element is a string of the filename containing samtools\n idxstats output. filename format is expected to be '": 1928, "'\n outfile : string\n logfile. the table name will be derived from `outfile`.'": 1929, "dataframe": 1930, "mapped_reads_per_chromosome": 1931, "containing": 1932, "samtools": 1933, "sam": 1934, "loadexonvalidation": 1935, "' load individual and merged exon validation stats\n\n for each sample, the exon validation stats are loaded into a table\n named by sample and mapper\n [sample]_[mapper]_overrun\n\n the merge alignment stats for all samples are merged and loaded\n into single table called exon_validation\n\n parameters\n ----------\n infiles : list\n input filenames with exon validation stats\n outfile : str\n output filename\n '": 1936, "\".exon.validation.tsv.gz\"": 1937, "mergeandload": 1938, "infi": 1939, "individual": 1940, "merged": 1941, "validation": 1942, "named": 1943, "mapper": 1944, "_overrun": 1945, "exon_validation": 1946, "loadpicardrnaseqmetrics": 1947, "loadcountreads": 1948, "' load read counts count_reads table '": 1949, "count_reads": 1950, "loadtranscriptprofile": 1951, "' merge transcript profiles into a single table'": 1952, "transcript": 1953, "loadstrandspecificity": 1954, "' merge strand specificity data into a single table'": 1955, "'a dummy task to run all tasks in the pipeline'": 1956, "dummy": 1957, "tasks": 1958, "renderrreport": 1959, "'build r markdown report '": 1960, "report_path": 1961, "abspath": 1962, "dirname": 1963, "'pipeline_docs'": 1964, "'pipeline_bamstats'": 1965, "'r_report'": 1966, "'cp %(report_path)s/* r_report.dir ; cd r_report.dir ; r -e \"rmarkdown::render_site()\"'": 1967, "R": 1968, "markdown": 1969, "report": 1970, "renderjupyterreport": 1971, "'build jupyter notebook report'": 1972, "'jupyter_report'": 1973, "cp": 1974, "jupyter_report": 1975, "cd": 1976, "jupyter": 1977, "nbconvert": 1978, "executepreprocessor": 1979, "ipynb": 1980, "allow": 1981, "mkdir": 1982, "_site": 1983, "mv": 1984, "notebook": 1985, "rendermultiqc": 1986, "'build mulitqc report'": 1987, "'lang=en_gb.utf-8 multiqc . -f;\n mv multiqc_report.html multiqc_report.dir/'": 1988, "mulitqc": 1989, "build_report": 1990, "'report dummy task to build reports'": 1991, "reports": 1992, "__getitem__": 1993, "nd": 1994, "getitem": 1995, "text": 1996, "dispatch": 1997, "transforms": 1998, "set_attr": 1999, "idx": 2000, "_idx": 2001, "attr": 2002, "__add__": 2003, "__iter__": 2004, "child": 2005, "children": 2006, "iter": 2007, "iterate": 2008, "iteration": 2009, "setup_platform": 2010, "hass": 2011, "add_devices": 2012, "discovery_info": 2013, "\"set up the myq component.\"": 2014, "pymyq": 2015, "myqapi": 2016, "username": 2017, "conf_username": 2018, "conf_password": 2019, "brand": 2020, "conf_type": 2021, "persistent_notification": 2022, "get_component": 2023, "'persistent_notification'": 2024, "myq": 2025, "is_supported_brand": 2026, "\"unsupported type. see documentation\"": 2027, "is_login_valid": 2028, "incorrect": 2029, "component": 2030, "platform": 2031, "should_poll": 2032, "\"poll for state.\"": 2033, "poll": 2034, "\"return the name of the garage door if any.\"": 2035, "_name": 2036, "default_name": 2037, "garage": 2038, "door": 2039, "is_closed": 2040, "\"return true if cover is closed, else false.\"": 2041, "_status": 2042, "state_closed": 2043, "cover": 2044, "close_cover": 2045, "\"issue close command to cover.\"": 2046, "close_device": 2047, "device_id": 2048, "command": 2049, "open_cover": 2050, "\"issue open command to cover.\"": 2051, "open_device": 2052, "\"update status of cover.\"": 2053, "get_status": 2054, "status": 2055, "makecasignedcert": 2056, "common_name": 2057, "private_key": 2058, "ca_cert": 2059, "ca_private_key": 2060, "serial_number": 2061, "\"make a cert and sign it with the ca's private key.\"": 2062, "public_key": 2063, "getpublickey": 2064, "builder": 2065, "x509": 2066, "certificatebuilder": 2067, "issuer_name": 2068, "getissuer": 2069, "subject": 2070, "nameattribute": 2071, "oid": 2072, "nameoid": 2073, "subject_name": 2074, "valid_from": 2075, "rdfvalue": 2076, "rdfdatetime": 2077, "duration": 2078, "\"1d\"": 2079, "valid_until": 2080, "rdfdat": 2081, "cert": 2082, "ca": 2083, "private": 2084, "makecacert": 2085, "\"grr\"": 2086, "issuer_cn": 2087, "\"grr_test\"": 2088, "issuer_c": 2089, "\"us\"": 2090, "\"generate a ca certificate.\n\n args:\n private_key: the private key to use.\n common_name: name for cert.\n issuer_cn: name for issuer.\n issuer_c: country for issuer.\n\n returns:\n the certificate.\n \"": 2091, "issuer": 2092, "country_name": 2093, "x50": 2094, "certificate": 2095, "country": 2096, "task_vars": 2097, "_supports_async": 2098, "actionmodule": 2099, "effect": 2100, "off": 2101, "nanny": 2102, "'warn'": 2103, "_task": 2104, "C": 2105, "command_warnings": 2106, "wrap_async": 2107, "async_val": 2108, "_connection": 2109, "has_native_async": 2110, "merge_hash": 2111, "_execute_module": 2112, "serve": 2113, "document_root": 2114, "show_indexes": 2115, "below": 2116, "structure": 2117, "put": 2118, "'^(?p.*)$'": 2119, "'document_root'": 2120, "'/path/to/my/files/'": 2121, "provide": 2122, "'d like to serve a basic index\n of the directory. this index view will use the template hardcoded below,\n but if you'": 2123, "was_modified_since": 2124, "header": 2125, "mtime": 2126, "\"\n was something modified since the user last downloaded it?\n\n header\n this is the value of the if-modified-since header. if this is none,\n i'll just return true.\n\n mtime\n this is the modification time of the item we're talking about.\n\n size\n this is the size of the item we're talking about.\n \"": 2127, "matches": 2128, "\"^([^;]+)(; length=([0-9]+))?$\"": 2129, "ignorecase": 2130, "header_mtime": 2131, "parse_h": 2132, "was": 2133, "something": 2134, "modified": 2135, "last": 2136, "downloaded": 2137, "?": 2138, "I": 2139, "'ll just return true.\n\nmtime\n this is the modification time of the item we'": 2140, "talking": 2141, "about": 2142, "display": 2143, "panel": 2144, "hide": 2145, "update_panels": 2146, "curses": 2147, "doupdate": 2148, "parse_options": 2149, "parser": 2150, "optionparser": 2151, "usage": 2152, "version": 2153, "add_option": 2154, "\"-b\"": 2155, "\"--bind\"": 2156, "\"store\"": 2157, "\"string\"": 2158, "\"0.0.0.0:8000\"": 2159, "dest": 2160, "\"bind\"": 2161, "help": 2162, "\"bind to address:[port]\"": 2163, "\"-d\"": 2164, "\"--debug\"": 2165, "\"store_true\"": 2166, "\"debug\"": 2167, "\"enable debug mode\"": 2168, "parse_args": 2169, "main": 2170, "configure": 2171, "\"run\"": 2172, "system": 2173, "nodeserver": 2174, "\"initialize our ``chatserver`` component.\n\n this uses the convenience ``init`` method which is called after the\n component is proeprly constructed and initialized and passed the\n same args and kwargs that were passed during construction.\n \"": 2175, "clients": 2176, "debug": 2177, "debugger": 2178, "\":\"": 2179, "bind": 2180, "address": 2181, "port": 2182, "chatserver": 2183, "after": 2184, "proeprly": 2185, "constructed": 2186, "passed": 2187, "were": 2188, "during": 2189, "disconnect": 2190, "sock": 2191, "\"disconnect event -- triggered for disconnecting clients\"": 2192, "event": 2193, "triggered": 2194, "disconnecting": 2195, "onerror": 2196, "excinfo": 2197, "chmod": 2198, "stat": 2199, "s_iwusr": 2200, "read_data_from_mistral": 2201, "'/work/bb1029/b324045/work1/work/member/post/'": 2202, "'member_t_2m_ts_seasmean.nc'": 2203, "'t_2m'": 2204, "mistral": 2205, "work": 2206, "\"\n\n :rtype: object\n \"": 2207, "cmd": 2208, "'scp $mistral:'": 2209, "' ./'": 2210, "'wget users.met.fu-berlin.de/~bijanfallah/'": 2211, "nc": 2212, "netcdffile": 2213, "name2": 2214, "attrname": 2215, "ncattrs": 2216, "'-----------------'": 2217, "rtype": 2218, "get_delivered_items_cost": 2219, "dn_items": 2220, "\"select dn.project_name, sum(dn_item.base_net_amount) as amount\n\t\tfrom `tabdelivery note` dn, `tabdelivery note item` dn_item\n\t\twhere dn.name = dn_item.parent and dn.docstatus = 1 and ifnull(dn.project_name, '') != ''\n\t\tgroup by dn.project_name\"": 2221, "si_items": 2222, "si": 2223, "project_name": 2224, "si_item": 2225, "base_net_amount": 2226, "amount": 2227, "tabsales": 2228, "invoice": 2229, "docstatus": 2230, "ifnull": 2231, "update_stock": 2232, "is_pos": 2233, "delivered": 2234, "cost": 2235, "to_dict": 2236, "book": 2237, "available": 2238, "metadata": 2239, "discovered": 2240, "authors": 2241, "embeddability": 2242, "identifiers": 2243, "publishers": 2244, "rating": 2245, "review": 2246, "subjects": 2247, "title": 2248, "viewability": 2249, "notes": 2250, "plural": 2251, "lists": 2252, "singular": 2253, "strings": 2254, "usually": 2255, "being": 2256, "joins": 2257, "subtitle": 2258, "space": 2259, "getthumbnaillink": 2260, "\"returns the atom.link object representing the thumbnail uri.\"": 2261, "link": 2262, "rel": 2263, "thumbnail_rel": 2264, "atom": 2265, "representing": 2266, "thumbnail": 2267, "uri": 2268, "getinfolink": 2269, "\"\n returns the atom.link object representing the human-readable info uri.\n \"": 2270, "info_rel": 2271, "human": 2272, "readable": 2273, "info": 2274, "getpreviewlink": 2275, "\"returns the atom.link object representing the preview uri.\"": 2276, "preview_rel": 2277, "preview": 2278, "getannotationlink": 2279, "\"\n returns the atom.link object representing the annotation uri.\n note that the use of www.books in the href of this link seems to make\n this information useless. using books.service.annotation_feed and \n book_server to construct your uri seems to work better.\n \"": 2280, "annotation_rel": 2281, "annotation": 2282, "books": 2283, "href": 2284, "seems": 2285, "information": 2286, "useless": 2287, "service": 2288, "annotation_feed": 2289, "book_server": 2290, "better": 2291, "set_rating": 2292, "\"set user's rating. must be an integral string between 1 nad 5\"": 2293, "'3'": 2294, "'4'": 2295, "'5'": 2296, "integral": 2297, "between": 2298, "nad": 2299, "set_review": 2300, "'text'": 2301, "lang": 2302, "'en'": 2303, "\"set user's review text\"": 2304, "get_label": 2305, "\"get users label for the item as a string\"": 2306, "scheme": 2307, "label_scheme": 2308, "users": 2309, "set_label": 2310, "\"clear pre-existing label for the item and set term as the label.\"": 2311, "remove_label": 2312, "pre": 2313, "\"clear the user's label for the item\"": 2314, "ln": 2315, "j": 2316, "enumerate": 2317, "clean_annotations": 2318, "\"clear all annotations from an item. useful for taking an item from\n another user's library/annotation feed and adding it to the \n authenticated user's library without adopting annotations.\"": 2319, "annotations": 2320, "useful": 2321, "another": 2322, "'s library/annotation feed and adding it to the \nauthenticated user'": 2323, "adopting": 2324, "get_google_id": 2325, "\"get google's id of the item.\"": 2326, "'/'": 2327, "func": 2328, "core": 2329, "check_setting_app_dirs_loaders": 2330, "test_app_dirs_and_loaders": 2331, "\"\n error if template loaders are specified and app_dirs is true.\n \"": 2332, "e001": 2333, "loaders": 2334, "app_dirs": 2335, "dirs": 2336, "test_app_dirs_removed": 2337, "templates_app_dirs_and_loaders": 2338, "'app_dirs'": 2339, "removed": 2340, "index": 2341, "docname": 2342, "\"render documentation page.\"": 2343, "document": 2344, "blueprint": 2345, "get_document": 2346, "cfg": 2347, "\"documentation_index\"": 2348, "documentnotfounderror": 2349, "abort": 2350, "404": 2351, "additional_breadcrumbs": 2352, "'title'": 2353, "'url'": 2354, "url_for": 2355, "'.index'": 2356, "render_template": 2357, "'documentation/index.html'": 2358, "breadcrumbs": 2359, "current_breadcrumbs": 2360, "documentation": 2361, "page": 2362, "documentation_package": 2363, "\"return documentation package.\"": 2364, "invenio_docs": 2365, "import_string": 2366, "'documentation_package'": 2367, "docs": 2368, "\"return an instance of sphinx ``websupport``.\"": 2369, "builddir": 2370, "current_app": 2371, "instance_path": 2372, "'docs'": 2373, "websupport": 2374, "srcdir": 2375, "__path__": 2376, "staticroot": 2377, "url_prefix": 2378, "'static'": 2379, "docroot": 2380, "sphinx": 2381, "send_static_file": 2382, "\"return static file.\"": 2383, "cache_timeout": 2384, "get_send_file_max_age": 2385, "send_from_directory": 2386, "\"docs\"": 2387, "\"static\"": 2388, "send": 2389, "correct_html": 2390, "div_math": 2391, "insert_idx": 2392, "\"separates out
from the parent tag

. anything\n in between is put into its own parent tag of

\"": 2394, "current_idx": 2395, "el": 2396, "util": 2397, "etree": 2398, "'p'": 2399, "ensure": 2400, "empty": 2401, "inserted": 2402, "isspace": 2403, "separates": 2404, "div": 2405, "\"math\"": 2406, "anything": 2407, "its": 2408, "own": 2409, "copyextractedfiles": 2410, "filenames": 2411, "\".png\"": 2412, "\".jpg\"": 2413, "rootpath": 2414, "subdirs": 2415, "walk": 2416, "\"./tmp\"": 2417, "'\\\\'": 2418, "shutil": 2419, "unregister": 2420, "callback": 2421, "resource_type": 2422, "\"unregister callback from the registry.\n\n :param callback: the callback.\n :param resource_type: must be a valid resource type.\n \"": 2423, "\"unregistering callback for %s\"": 2424, "_validate_resource_type": 2425, "_delete_callback": 2426, "resource": 2427, "get_callback": 2428, "_callbacks": 2429, "rpc_exc": 2430, "callbacknotfound": 2431, "get_callbacks": 2432, "\"return the callback if found, none otherwise.\n\n :param resource_type: must be a valid resource type.\n \"": 2433, "callbacks": 2434, "found": 2435, "add_arguments": 2436, "add_argument": 2437, "'--%s'": 2438, "usermodel": 2439, "username_field": 2440, "'specifies the login for the superuser.'": 2441, "'--noinput'": 2442, "'store_false'": 2443, "'interactive'": 2444, "'tells django to not prompt the user for input of any kind. '": 2445, "'you must use --%s with --noinput, along with an option for '": 2446, "'any other required field. superusers created with --noinput will '": 2447, "able": 2448, "get_input_data": 2449, "\"\n override this method if you want to customize data inputs or\n validation exceptions.\n \"": 2450, "raw_value": 2451, "val": 2452, "exceptions": 2453, "stderr": 2454, "\"error: %s\"": 2455, "'; '": 2456, "want": 2457, "customize": 2458, "discard": 2459, "incomplete": 2460, "memsample": 2461, "used_values": 2462, "__str__": 2463, "\"\\t\"": 2464, "cpu_sample": 2465, "writer": 2466, "pid": 2467, "ppid": 2468, "start_time": 2469, "last_cpu_ns": 2470, "last_blkio_delay_ns": 2471, "last_swapin_delay_ns": 2472, "set_parent": 2473, "processmap": 2474, "1000": 2475, "2000": 2476, "kernel": 2477, "threads": 2478, "\"missing config_proc_events: no parent for pid '%i' ('%s') with ppid '%i'\"": 2479, "ensure_ofc_tenant": 2480, "tenant_id": 2481, "exists_ofc_tenant": 2482, "create_ofc_tenant": 2483, "ofc": 2484, "tenant": 2485, "desc": 2486, "\"id=%s at openstack.\"": 2487, "ofc_tenant_id": 2488, "driver": 2489, "create_tenant": 2490, "_add_ofc_item": 2491, "\"ofc_tenant\"": 2492, "delete_ofc_tenant": 2493, "_get_ofc_id": 2494, "convert_ofc_tenant_id": 2495, "delete_tenant": 2496, "_del_ofc_item": 2497, "create_ofc_port": 2498, "port_id": 2499, "ofc_net_id": 2500, "\"ofc_network\"": 2501, "'network_id'": 2502, "convert_ofc_network_id": 2503, "'tenant_id'": 2504, "portinfo": 2505, "ndb": 2506, "get_portinfo": 2507, "nexc": 2508, "portinfonotfound": 2509, "ofc_port_id": 2510, "create_port": 2511, "\"ofc_port\"": 2512, "create_ofc_packet_filter": 2513, "filter_id": 2514, "filter_dict": 2515, "in_port_id": 2516, "'in_port'": 2517, "ofc_pf_id": 2518, "packet": 2519, "filtering": 2520, "delete_ofc_router": 2521, "router_id": 2522, "router": 2523, "ofc_router_id": 2524, "\"ofc_router\"": 2525, "delete_router": 2526, "add_ofc_router_interface": 2527, "network_id": 2528, "cidr": 2529, "ip_address": 2530, "mac_address": 2531, "'%s/%s'": 2532, "'ip_address'": 2533, "netaddr": 2534, "ipnetwork": 2535, "'cidr'": 2536, "prefixlen": 2537, "'mac_address'": 2538, "ofc_inf_id": 2539, "add_router_interface": 2540, "interface": 2541, "delete_ofc_router_interface": 2542, "maintain": 2543, "delete_router_interface": 2544, "__delitem__": 2545, "nobj": 2546, "delitem": 2547, "iteritems": 2548, "cur": 2549, "cur2": 2550, "me": 2551, "clear_prefix": 2552, "prefix": 2553, "\" remove from `self` all the items with the given `prefix`. \"": 2554, "test_split_name_parts": 2555, "\"bibauthority - test get_type_from_authority_id\"": 2556, "invenio": 2557, "legacy": 2558, "bibauthority": 2559, "cfg_bibauthority_prefix_sep": 2560, "get_type_from_control_no": 2561, "\"journal\"": 2562, "control_no": 2563, "\"(cern)abcd1234\"": 2564, "'('": 2565, "get_type_from_authority_id": 2566, "on_message": 2567, "\"handle new messages which are received from the fan.\"": 2568, "prevent": 2569, "refreshing": 2570, "needed": 2571, "_old_value": 2572, "_logger": 2573, "\"message received for %s device: %s\"": 2574, "schedule_update_ha_state": 2575, "received": 2576, "fan": 2577, "unit_of_measurement": 2578, "\"return the unit the value is expressed in.\"": 2579, "sensor_units": 2580, "_sensor_type": 2581, "unit": 2582, "expressed": 2583, "measurement": 2584, "icon": 2585, "\"return the icon for this sensor.\"": 2586, "sensor_icons": 2587, "sensor": 2588, "unique_id": 2589, "\"return the sensor's unique id.\"": 2590, "\"{self._device.serial}-{self._sensor_type}\"": 2591, "unique": 2592, "\"return filter life in hours.\"": 2593, "_device": 2594, "filter_life": 2595, "life": 2596, "hours": 2597, "teardown": 2598, "tb": 2599, "clean_stage_name": 2600, "stages": 2601, "get_stages": 2602, "'*'": 2603, "20": 2604, "\"error\"": 2605, "'error: unknown stage \"{}\". stage must be one of the '": 2606, "'following (remember to include surrounding quotation marks):'": 2607, "stage_name": 2608, "'{:>3} - \"{}\"'": 2609, "exit": 2610, "doclean": 2611, "stage": 2612, "parse_arguments": 2613, "argparse": 2614, "argumentparser": 2615, "\"genome-wide reconstruction of complex structural variants\"": 2616, "\"config\"": 2617, "\"path to configuration.json file\"": 2618, "\"--restart\"": 2619, "metavar": 2620, "\"from-stage\"": 2621, "\"restart from this stage\"": 2622, "\"--local\"": 2623, "\"run locally in single processor mode\"": 2624, "\"--multiprocessing\"": 2625, "\"run locally using multiprocessing\"": 2626, "\"run in debug mode\"": 2627, "get_schema_facts": 2628, "schema": 2629, "facts": 2630, "\"\n select schema_name, schema_owner, create_time\n from schemata\n where not is_system_schema and schema_name not in ('public')\n and (? = '' or schema_name ilike ?)\n \"": 2631, "rows": 2632, "fetchmany": 2633, "100": 2634, "schema_name": 2635, "'owner'": 2636, "schema_owner": 2637, "'create_time'": 2638, "create_time": 2639, "__html__": 2640, "__radd__": 2641, "'__html__'": 2642, "string_types": 2643, "escape": 2644, "notimplemented": 2645, "radd": 2646, "__mod__": 2647, "arg": 2648, "_markupescapehelper": 2649, "text_type": 2650, "unescape": 2651, "\"unescape markup again into an text_type string. this also resolves\n known html4 and xhtml entities:\n\n >>> markup(\"": 2652, "&": 2653, "raquo": 2654, "em": 2655, "\").unescape()\n u'main \\xbb about'\n \"": 2656, "markupsafe": 2657, "_constants": 2658, "html_entities": 2659, "handle_match": 2660, "unichr": 2661, "'#x'": 2662, "16": 2663, "markup": 2664, "resolves": 2665, "html4": 2666, "xhtml": 2667, "entities": 2668, "\"main » about\"": 2669, "'main \\xbb about'": 2670, "striptags": 2671, "\"unescape markup into an text_type string and strip all tags. this\n also resolves known html4 and xhtml entities. whitespace is\n normalized to one:\n\n >>> markup(\"": 2672, "\").striptags()\n u'main \\xbb about'\n \"": 2673, "stripped": 2674, "_striptags_re": 2675, "whitespace": 2676, "normalized": 2677, "\"main » about\"": 2678, "'main \\xbb about'": 2679, "\"escape the string. works like :func:`escape` with the difference\n that for subclasses of :class:`markup` this function would return the\n correct subclass.\n \"": 2680, "rv": 2681, "would": 2682, "subclass": 2683, "rpartition": 2684, "sep": 2685, "formatter": 2686, "escapeformatter": 2687, "_magicformatmapping": 2688, "vformat": 2689, "test_invalid_login": 2690, "browse_to_url": 2691, "reverse": 2692, "'auth_login'": 2693, "login_url": 2694, "current_url": 2695, "process_login_form": 2696, "'passwordinvalid'": 2697, "least": 2698, "wait_until_present": 2699, "'.errorlist li'": 2700, "invalid": 2701, "authenticate": 2702, "test_can_retrieve_username": 2703, "'forgot_username'": 2704, "email_elem": 2705, "find_element_by_name": 2706, "'email'": 2707, "send_keys": 2708, "click": 2709, "'form input[type=\"submit\"]'": 2710, "wait_until_text_present": 2711, "'email sent'": 2712, "mail": 2713, "outbox": 2714, "peek": 2715, "from_idx": 2716, "to_idx": 2717, "_stk": 2718, "\"( x -- )\n take something and return it\"": 2719, "pmlr": 2720, "debug_write": 2721, "\"popping more items than exist on stack!\\n\"": 2722, "level": 2723, "\"warn\"": 2724, "stackoverflow": 2725, "com": 2726, "34633242": 2727, "4532996": 2728, "slower": 2729, "doesn": 2730, "seem": 2731, "optimised": 2732, "avoid": 2733, "take": 2734, "push": 2735, "\"( -- x ... )\n put somethings at idx\"": 2736, "somethings": 2737, "pick": 2738, "\"( x -- x )\n pick somethings from a range of indicies\"": 2739, "indicies": 2740, "\"( x -- )\n drop items without returning (cheaper pop)\"": 2741, "cheaper": 2742, "dup": 2743, "\"( y -- y y )\n duplicate something and push\"": 2744, "errtype": 2745, "\"range\"": 2746, "dupn": 2747, "\"( x y -- x y x y )\n dup count items from an idx\"": 2748, "swap": 2749, "\"( x y -- y x )\n swap two things at an index\"": 2750, "reversed": 2751, "two": 2752, "things": 2753, "rot": 2754, "\"( w x y z -- x y z w )\n rotate things left, at an index\"": 2755, "l": 2756, "z": 2757, "urot": 2758, "\"( w x y z -- z w x y )\n rotate things right, at an index\"": 2759, "framelevel": 2760, "err_data": 2761, "\"type\"": 2762, "\"fatal\"": 2763, "\"\\n\"": 2764, "\"is_fatal\"": 2765, "debug_fmt": 2766, "\" \"": 2767, "getmedian": 2768, "\"median\"": 2769, "abs": 2770, "float": 2771, "test_00_message_compose_wizard": 2772, "\" tests designed for the mail.compose.message wizard updated by email_template. \"": 2773, "mail_compose": 2774, "'mail.compose.message'": 2775, "res_users": 2776, "'signature'": 2777, "'a@a.a'": 2778, "user_admin": 2779, "browse": 2780, "p_a_id": 2781, "partner_id": 2782, "group_pigs": 2783, "mail_group": 2784, "group_pigs_id": 2785, "group_bird": 2786, "group_bird_id": 2787, "_s": 2788, "designed": 2789, "compose": 2790, "wizard": 2791, "updated": 2792, "email_template": 2793, "00": 2794, "test_10_email_templating": 2795, "group_model_id": 2796, "'ir.model'": 2797, "'mail.group'": 2798, "'email.template'": 2799, "email_template_id": 2800, "'model_id'": 2801, "'pigs template'": 2802, "'email_from'": 2803, "'raoul grosbedon '": 2804, "templating": 2805, "get_lastrowid": 2806, "create_cursor": 2807, "\"select last_insert_id()\"": 2808, "lastrowid": 2809, "fetchone": 2810, "event_register": 2811, "post": 2812, "'event'": 2813, "'main_object'": 2814, "'range'": 2815, "\"website_event.event_description_full\"": 2816, "registration_confirm": 2817, "attendees": 2818, "'event.registration'": 2819, "registrations": 2820, "_process_registration_details": 2821, "registration": 2822, "'event_id'": 2823, "sudo": 2824, "_prepare_attendee_values": 2825, "\"website_event.registration_complete\"": 2826, "'attendees'": 2827, "confirm": 2828, "sdn": 2829, "to_string": 2830, "hour": 2831, "23": 2832, "minute": 2833, "59": 2834, "second": 2835, "sd": 2836, "test_make_scorer": 2837, "make_scorer": 2838, "assert_raises": 2839, "needs_threshold": 2840, "needs_proba": 2841, "scorer": 2842, "test_regression_scorers": 2843, "regression": 2844, "scorers": 2845, "diabetes": 2846, "load_diabetes": 2847, "target": 2848, "x_train": 2849, "x_test": 2850, "y_train": 2851, "y_test": 2852, "train_test_split": 2853, "clf": 2854, "ridge": 2855, "fit": 2856, "score1": 2857, "get_scorer": 2858, "'r2'": 2859, "score2": 2860, "r2_score": 2861, "predict": 2862, "assert_almost_equal": 2863, "test_scoring_is_not_metric": 2864, "assert_raises_regexp": 2865, "'make_scorer'": 2866, "check_scoring": 2867, "logisticregression": 2868, "f1_score": 2869, "roc_auc_score": 2870, "kmeans": 2871, "adjusted_rand_score": 2872, "scoring": 2873, "metric": 2874, "handle_help": 2875, "stdout": 2876, "textwrap": 2877, "dedent": 2878, "'\\\n to migrate a template from .lang to fluent, use the subcommands like so\n\n ./manage.py fluent recipe bedrock/app/templates/app/some.html\n\n # edit ids in lib/fluent_migrations/app/some.py\n\n ./manage.py fluent template bedrock/app/templates/app/some.html\n ./manage.py fluent ftl bedrock/app/templates/app/some.html\n\n more documentation on https://bedrock.readthedocs.io/en/latest/fluent-conversion.html.\n '": 2879, "create_ftl": 2880, "recipe_or_template": 2881, "locales": 2882, "_fluent_ftl": 2883, "ftlcreator": 2884, "ftl_creator": 2885, "locale": 2886, "ftl": 2887, "test_main": 2888, "run_unittest": 2889, "testcgitb": 2890, "get_albums_by_name": 2891, "\"\n given a list of names this method will\n return album objects for each matching name.\n\n if a name is not found then it is silently ignored.\n\n this method returns a dictionary mapping name\n to album object.\n \"": 2892, "albums": 2893, "album_names": 2894, "album": 2895, "graph": 2896, "objects": 2897, "matching": 2898, "silently": 2899, "ignored": 2900, "get_albums_by_id": 2901, "\"\n given a list of ids this method will\n return album objects for each matching id.\n\n if an id is not found then it is silently ignored.\n\n this method returns a dictionary mapping id to\n album object\n \"": 2902, "album_id": 2903, "album_ids": 2904, "get_all_albums_by_id": 2905, "\"\n this method returns a dictionary of all\n albums with album ids as the keys\n \"": 2906, "album_data": 2907, "get_all_albums_by_name": 2908, "\"\n this method returns a dictionary of all\n albums with album names as the keys\n \"": 2909, "get_photos": 2910, "\"\n get a list of photo objects\n \"": 2911, "photos_data": 2912, "get_connections": 2913, "'photos'": 2914, "photos": 2915, "photo_data": 2916, "photomodel": 2917, "'me ='": 2918, "'id'": 2919, "fetch": 2920, "orig": 2921, "'images'": 2922, "'source'": 2923, "photo": 2924, "test_upconvert": 2925, "day": 2926, "14": 2927, "week": 2928, "72": 2929, "'d'": 2930, "240": 2931, "'h'": 2932, "360": 2933, "6": 2934, "86400": 2935, "millisecond": 2936, "5000": 2937, "'s'": 2938, "microsecond": 2939, "5000000": 2940, "nanosecond": 2941, "5000000000": 2942, "offset": 2943, "to_unit": 2944, "equals": 2945, "upconvert": 2946, "transform": 2947, "test_cannot_upconvert": 2948, "ibiserror": 2949, "test_downconvert_day": 2950, "24": 2951, "1440": 2952, "'ms'": 2953, "86400000": 2954, "'us'": 2955, "86400000000": 2956, "'ns'": 2957, "8640000000": 2958, "downconvert": 2959, "test_combine_with_different_kinds": 2960, "1441": 2961, "1010": 2962, "3910": 2963, "_check_cases": 2964, "combine": 2965, "kinds": 2966, "test_timedelta_generic_api": 2967, "weeks": 2968, "milliseconds": 2969, "microseconds": 2970, "nanoseconds": 2971, "generic": 2972, "test_offset_timestamp_expr": 2973, "c": 2974, "expr": 2975, "ir": 2976, "timestamparray": 2977, "op": 2978, "ops": 2979, "timestampdelta": 2980, "load_plugin": 2981, "\" returns plugin available in this module\n \"": 2982, "hosttestplugincopymethod_smart": 2983, "plugin": 2984, "\" configure plugin, this function should be called before plugin execute() method is used.\n \"": 2985, "capability": 2986, "\" executes capability by name.\n each capability may directly just call some command line\n program or execute building pythonic function\n \"": 2987, "check_parameters": 2988, "image_path": 2989, "'image_path'": 2990, "destination_disk": 2991, "'destination_disk'": 2992, "target_mcu": 2993, "'target_mcu'": 2994, "mount": 2995, "check_mount_point_ready": 2996, "blocking": 2997, "executes": 2998, "directly": 2999, "just": 3000, "program": 3001, "building": 3002, "pythonic": 3003, "filter_dictionary": 3004, "\" lms implementation, adds filtering by user partition, course id and user \"": 3005, "get_group_for_user_partition": 3006, "user_partition": 3007, "course_key": 3008, "\" returns the specified user's group for user partition \"": 3009, "scheme_supports_assignment": 3010, "get_group_for_user": 3011, "retu": 3012, "lms": 3013, "partition": 3014, "field_dictionary": 3015, "\" add course if provided otherwise add courses in which the user is enrolled in \"": 3016, "lmssearchfiltergenerator": 3017, "'user'": 3018, "'course'": 3019, "'course_id'": 3020, "user_enrollments": 3021, "_enrollments_for_user": 3022, "enrollment": 3023, "course_id": 3024, "only": 3025, "courses": 3026, "enrolled": 3027, "exclude_dictionary": 3028, "\" if we are not on a microsite, then exclude any microsites that are defined \"": 3029, "course_org_filter": 3030, "microsite": 3031, "'course_org_filter'": 3032, "ensuring": 3033, "those": 3034, "org_filter_out_set": 3035, "get_all_orgs": 3036, "'org'": 3037, "exclude": 3038, "microsites": 3039, "user_parti": 3040, "get_group_ids_for_user": 3041, "\" collect user partition group ids for user for this course \"": 3042, "partition_groups": 3043, "user_partitions": 3044, "include_schemes": 3045, "partition_group_ids": 3046, "partition_group": 3047, "collect": 3048, "get_git_revision": 3049, "\"\n returns the git revision (e.g. aab4afc016b742c6d4b157427e192942d0e131fe),\n or unknown is getting the git revision fails\n\n relies on gitpython (see http://gitorious.org/git-python)\n \"": 3050, "git": 3051, "unknown": 3052, "gitrepo": 3053, "rev_list": 3054, "\"head\"": 3055, "gitcommanderror": 3056, "revision": 3057, "aab4afc016b742c6d4b157427e192942d0e131fe": 3058, "getting": 3059, "relies": 3060, "gitpython": 3061, "gitorious": 3062, "this_is_easybuild": 3063, "\"standard starting message\"": 3064, "top_version": 3065, "max": 3066, "framework_version": 3067, "easyblocks_version": 3068, "bootstrap_eb": 3069, "py": 3070, "script": 3071, "hard": 3072, "adjust": 3073, "sufficient": 3074, "care": 3075, "\"this is easybuild %s (framework: %s, easyblocks: %s) on host %s.\"": 3076, "gethostname": 3077, "standard": 3078, "starting": 3079, "easybuild": 3080, "test_api_parameters": 3081, "'load_net_service_policy_1.json'": 3082, "'my description'": 3083, "timer_policy": 3084, "'/common/foo'": 3085, "port_misuse_policy": 3086, "'/common/bar'": 3087, "test_create_selfip": 3088, "'timer1'": 3089, "'misuse1'": 3090, "'common'": 3091, "provider": 3092, "'password'": 3093, "mo": 3094, "selfip": 3095, "handle_data": 3096, "short_ema": 3097, "short_ema_trans": 3098, "long_ema": 3099, "long_ema_trans": 3100, "buy": 3101, "sell": 3102, "invested": 3103, "'aapl'": 3104, "fal": 3105, "vals": 3106, "bypass": 3107, "obsolete": 3108, "resequencing": 3109, "'line_ids'": 3110, "'ebanking_import'": 3111, "osv": 3112, "account_bank_statement": 3113, "save": 3114, "interpret": 3115, "execution_context": 3116, "\"\n interpret a marker and return a result depending on environment.\n\n :param marker: the marker to interpret.\n :type marker: str\n :param execution_context: the context used for name lookup.\n :type execution_context: mapping\n \"": 3117, "evaluator": 3118, "evaluate": 3119, "depending": 3120, "get_fragment": 3121, "\"\n get the part of the source which is causing a problem.\n \"": 3122, "fragment_len": 3123, "'%r'": 3124, "'...'": 3125, "part": 3126, "causing": 3127, "problem": 3128, "fragment": 3129, "get_handler": 3130, "node_type": 3131, "\"\n get a handler for the specified ast node type.\n \"": 3132, "'do_%s'": 3133, "handler": 3134, "ast": 3135, "node": 3136, "\"\n evaluate a source string or node, using ``filename`` when\n displaying errors.\n \"": 3137, "'mode'": 3138, "'eval'": 3139, "'filename'": 3140, "'syntax error %s'": 3141, "low": 3142, "do_attribute": 3143, "get_attr_key": 3144, "allowed_values": 3145, "'invalid expression: %s'": 3146, "do_compare": 3147, "sanity_check": 3148, "lhsnode": 3149, "rhsnode": 3150, "klhs": 3151, "krhs": 3152, "col_offset": 3153, "comparison": 3154, "poll_message": 3155, "receiver": 3156, "recv": 3157, "binary": 3158, "ints": 3159, "byte_list": 3160, "ord": 3161, "poll_short": 3162, "type_str": 3163, "verbose": 3164, "raw_data": 3165, "npa_s": 3166, "npa_f": 3167, "astype": 3168, "10000.0": 3169, "fmt": 3170, "\"<%di\"": 3171, "struct": 3172, "unpack": 3173, "short": 3174, "static_slotname": 3175, "\"\n returns a string to be used as the slot\n for the static placeholder field.\n \"": 3176, "placeholder": 3177, "slotname": 3178, "has_publish_permission": 3179, "is_superuser": 3180, "has_perm": 3181, "app_label": 3182, "get_permission_codename": 3183, "'change'": 3184, "'publish_page'": 3185, "publish": 3186, "permission": 3187, "get_xframe_options_value": 3188, "\"\n gets the value to set for the x_frame_options header.\n\n by default this uses the value from the x_frame_options django\n settings. if not found in settings, defaults to 'sameorigin'.\n\n this method can be overridden if needed, allowing it to vary based on\n the request or response.\n \"": 3189, "'x_frame_options'": 3190, "'sameorigin'": 3191, "x_frame_options": 3192, "defaults": 3193, "allowing": 3194, "vary": 3195, "xframe": 3196, "merge_objects": 3197, "base_obj": 3198, "client_obj": 3199, "\"\n :param base_obj: has the base attributes which are overwritten if they exist\n in the client_obj and remain if they do not exist in the client_obj\n :param client_obj: the object that the client wants to create.\n :return: the merged objects\n \"": 3200, "client_obj_cp": 3201, "base_key": 3202, "base_val": 3203, "attributes": 3204, "overwritten": 3205, "remain": 3206, "wants": 3207, "extend_object_field": 3208, "field_name": 3209, "\"\n :param base_obj: an object which has a property `field_name` that is a list\n :param client_obj: an object which has a property `field_name` that is a list.\n a copy of this object is returned with `field_name` modified\n :param field_name: the name of the list field\n :type field_name: str\n :return: the client_obj with the property `field_name` being the two properties appended\n \"": 3210, "base_obj_field": 3211, "client_obj_field": 3212, "property": 3213, "gen_pod": 3214, "k8s": 3215, "v1pod": 3216, "\"generates pod\"": 3217, "ud_pod": 3218, "pod": 3219, "containers": 3220, "container": 3221, "make_unique_pod_id": 3222, "extract_xcom": 3223, "add_sidecar": 3224, "\"adds sidecar\"": 3225, "pod_cp": 3226, "volumes": 3227, "poddefaults": 3228, "volume": 3229, "volume_mounts": 3230, "volume_mount": 3231, "sidecar_container": 3232, "sidecar": 3233, "from_obj": 3234, "optional": 3235, "\"converts to pod from obj\"": 3236, "podgenerator": 3237, "'cannot convert a non-dictionary or non-podgenerator '": 3238, "'object into a kubernetesexecutorconfig'": 3239, "constant": 3240, "executorloader": 3241, "rather": 3242, "executor": 3243, "mechanism": 3244, "causes": 3245, "cyclic": 3246, "converts": 3247, "reconcile_pods": 3248, "base_pod": 3249, "client_pod": 3250, "\"\n :param base_pod: has the base attributes which are overwritten if they exist\n in the client pod and remain if they do not exist in the client_pod\n :type base_pod: k8s.v1pod\n :param client_pod: the pod that the client wants to create.\n :type client_pod: k8s.v1pod\n :return: the merged pods\n\n this can't be done recursively as certain fields some overwritten, and some concatenated.\n \"": 3251, "pods": 3252, "reconcile_specs": 3253, "base_spec": 3254, "v1podspec": 3255, "client_spec": 3256, "\"\n :param base_spec: has the base attributes which are overwritten if they exist\n in the client_spec and remain if they do not exist in the client_spec\n :type base_spec: k8s.v1podspec\n :param client_spec: the spec that the client wants to create.\n :type client_spec: k8s.v1podspec\n :return: the merged specs\n \"": 3257, "sp": 3258, "specs": 3259, "reconcile_containers": 3260, "base_containers": 3261, "v1container": 3262, "client_containers": 3263, "clien": 3264, "construct_pod": 3265, "dag_id": 3266, "task_id": 3267, "pod_id": 3268, "try_number": 3269, "kube_executor_config": 3270, "worker_config": 3271, "namespace": 3272, "worker_uuid": 3273, "\"\n construct a pod by gathering and consolidating the configuration from 3 places:\n - airflow.cfg\n - executor_config\n - dynamic arguments\n \"": 3274, "dynamic_pod": 3275, "labels": 3276, "gathering": 3277, "consolidating": 3278, "configuration": 3279, "places": 3280, "airflow": 3281, "executor_config": 3282, "dynamic": 3283, "deserialize_model_file": 3284, "\"\n :param path: path to the file\n :return: a kubernetes.client.models.v1pod\n\n unfortunately we need access to the private method\n ``_apiclient__deserialize_model`` from the kubernetes client.\n this issue is tracked here; https://github.com/kubernetes-client/python/issues/977.\n \"": 3285, "api_client": 3286, "apiclient": 3287, "stream": 3288, "yaml": 3289, "safe_load": 3290, "kubernetes": 3291, "unfortunately": 3292, "need": 3293, "access": 3294, "_apiclient__deserialize_model": 3295, "tracked": 3296, "https": 3297, "github": 3298, "issues": 3299, "977.": 3300, "deserialize": 3301, "\"\n kubernetes pod names must be <= 253 chars and must pass the following regex for\n validation\n ``^[a-z0-9]([-a-z0-9]*[a-z0-9])?(\\\\.[a-z0-9]([-a-z0-9]*[a-z0-9])?)*$``\n\n :param dag_id: a dag_id with only alphanumeric characters\n :return: ``str`` valid pod name of appropriate length\n \"": 3302, "safe_uuid": 3303, "uuid": 3304, "uuid4": 3305, "hex": 3306, "safe_pod_id": 3307, "max_pod_id_len": 3308, "\"-\"": 3309, "253": 3310, "chars": 3311, "regex": 3312, "z0": 3313, "$": 3314, "alphanumeric": 3315, "characters": 3316, "validate_pod_generator_args": 3317, "given_args": 3318, "\"\n :param given_args: the arguments passed to the podgenerator constructor.\n :type given_args: dict\n :return: none\n\n validate that if `pod` or `pod_template_file` are set that the user is not attempting\n to configure the pod with the other arguments.\n \"": 3319, "pod_args": 3320, "inspect": 3321, "predicate": 3322, "pod_template_file": 3323, "attempting": 3324, "\"\n :param k: an arg to podgenerator\n :type k: string\n :param v: the parameter of the given arg\n :type v: inspect.parameter\n :return: bool\n\n returns true if the podgenerator argument has no default arguments\n or the default argument is none, and it is not one of the listed field\n in `non_empty_fields`.\n \"": 3325, "non_empty_fields": 3326, "'pod'": 3327, "'pod_template_file'": 3328, "'extract_xcom'": 3329, "'service_account_name'": 3330, "'image_pull_policy'": 3331, "bool": 3332, "listed": 3333, "heapsorted_by_value": 3334, "maxheap": 3335, "\"\n takes an arbitrary mapping and, treating the values as priority keys, sorts\n its items by priority via heapsort using a pqdict.\n\n returns:\n a list of the dictionary items sorted by value\n\n \"": 3336, "pq": 3337, "pqdict": 3338, "maxpq": 3339, "arbitrary": 3340, "treating": 3341, "priority": 3342, "sorts": 3343, "via": 3344, "heapsort": 3345, "sorted": 3346, "sorting": 3347, "heapsorted": 3348, "__lt__": 3349, "lt": 3350, "minpq": 3351, "create_entry": 3352, "minpqdentry": 3353, "min": 3354, "fromfunction": 3355, "pkeygen": 3356, "fromkeys": 3357, "\"\n provide a key function that determines priorities by which to heapify\n the elements of an iterable into a pqd.\n\n \"": 3358, "dkey": 3359, "determines": 3360, "priorities": 3361, "heapify": 3362, "pqd": 3363, "__contains__": 3364, "\"\n return true if dkey is in the pqd else return false.\n\n \"": 3365, "nodefinder": 3366, "contains": 3367, "__setitem__": 3368, "pkey": 3369, "\"\n assign priority to dictionary key.\n\n \"": 3370, "heap": 3371, "finder": 3372, "pos": 3373, "_swim": 3374, "parent_pos": 3375, "child_pos": 3376, "setitem": 3377, "__copy__": 3378, "\"\n return a new pqd with the same dkeys associated with the same priority\n keys.\n\n \"": 3379, "pqds": 3380, "behave": 3381, "schedules": 3382, "dkeys": 3383, "copies": 3384, "entries": 3385, "pqdentry": 3386, "mutable": 3387, "shared": 3388, "pqdicts": 3389, "maps": 3390, "associated": 3391, "popitem": 3392, "\"\n extract top priority item. raises keyerror if pqd is empty.\n\n \"": 3393, "_sink": 3394, "raises": 3395, "additem": 3396, "\"\n add a new item. raises keyerror if item is already in the pqd.\n\n \"": 3397, "updateitem": 3398, "new_pkey": 3399, "\"\n update the priority key of an existing item. raises keyerror if item is\n not in the pqd.\n\n \"": 3400, "iterkeys": 3401, "\"\n destructive heapsort iterator over dictionary keys, ordered by priority\n key.\n\n \"": 3402, "destructive": 3403, "iterator": 3404, "over": 3405, "ordered": 3406, "itervalues": 3407, "\"\n destructive heapsort iterator over priority keys.\n\n \"": 3408, "get_history": 3409, "logger": 3410, "warning": 3411, "'do not support account of type %s'": 3412, "__getattr__": 3413, "manager": 3414, "__dict__": 3415, "'manager'": 3416, "kill": 3417, "\"destroy the service object.\"": 3418, "stop": 3419, "timers": 3420, "\"exception occurs when timer stops\"": 3421, "periodic_tasks": 3422, "raise_on_error": 3423, "\"tasks to be run at a periodic interval.\"": 3424, "ctxt": 3425, "get_admin_context": 3426, "periodic": 3427, "report_state": 3428, "\"update the state of this service.\"": 3429, "gongysh": 3430, "neutron": 3431, "testusedefaultgraph": 3432, "as_default": 3433, "'/cpu:0'": 3434, "constant_op": 3435, "6.0": 3436, "7.0": 3437, "math_ops": 3438, "matmul": 3439, "'matmul'": 3440, "eval": 3441, "assertallequal": 3442, "42.0": 3443, "testmanycpus": 3444, "keveman": 3445, "listdevices": 3446, "devices": 3447, "config_pb2": 3448, "configproto": 3449, "device_count": 3450, "'cpu'": 3451, "inp": 3452, "'w1'": 3453, "testfetchtuple": 3454, "sess": 3455, "control_flow_ops": 3456, "no_op": 3457, "tensor": 3458, "44.0": 3459, "tuple_runner": 3460, "make_callable": 3461, "testfetchtensorobject": 3462, "results_with_list": 3463, "4.0": 3464, "results_with_single": 3465, "results_with_get": 3466, "a_val": 3467, "b_val": 3468, "fetches": 3469, "testfetchoperationobject": 3470, "'testfetchoperationobject_v'": 3471, "initializer": 3472, "v_val": 3473, "testfeedsparseplaceholderpartialshape": 3474, "indices": 3475, "int64": 3476, "float32": 3477, "array_ops": 3478, "sparse_placeholder": 3479, "'placeholder1'": 3480, "sp_indices": 3481, "identity": 3482, "sp_values": 3483, "sp_shape": 3484, "dense_shape": 3485, "sp2": 3486, "sparse_tensor": 3487, "sparsetensor": 3488, "feed": 3489, "testfeedindexedslices": 3490, "indexedslices": 3491, "ind_values": 3492, "testfetchindexedsliceswithoutdenseshape": 3493, "ind_out": 3494, "values_out": 3495, "indices_out": 3496, "dense_shape_out": 3497, "dens": 3498, "testextendwithstatefuloperations": 3499, "'testextendwithstatefuloperations_v'": 3500, "3.0": 3501, "assign_e_to_v": 3502, "state_ops": 3503, "e_val": 3504, "6.": 3505, "testoperationrunmethod": 3506, "'b'": 3507, "assign_a_to_v": 3508, "assign_b_to_v": 3509, "feed_dict": 3510, "'b:0'": 3511, "asse": 3512, "testdefaultgraph": 3513, "get_default_graph": 3514, "'testdefaultgraph_v'": 3515, "matmu": 3516, "testuseaftercloseconcurrent": 3517, "5.0": 3518, "update_thread": 3519, "assertraiseswithpredicatematch": 3520, "'attempted to use a closed session.'": 3521, "sleep": 3522, "testnotentered": 3523, "pylint": 3524, "protected": 3525, "_default_session_stack": 3526, "get_default": 3527, "enable": 3528, "c_1": 3529, "c_2": 3530, "'no default session is registered.'": 3531, "testinteractiveplaceprunedgraph": 3532, "interactivesession": 3533, "bad": 3534, "currently": 3535, "gpu": 3536, "kernels": 3537, "why": 3538, "placing": 3539, "combo": 3540, "'/device:gpu:0'": 3541, "assertraise": 3542, "testdefaultsessionplaceprunedgraph": 3543, "invalidargumenterror": 3544, "though": 3545, "don": 3546, "testsharedgraph": 3547, "sess1": 3548, "sess2": 3549, "testmakecallablewithfeedlistandrunoptions": 3550, "ph": 3551, "dtypes": 3552, "tensor_runner": 3553, "feed_list": 3554, "accept_options": 3555, "run_options": 3556, "runoptions": 3557, "trace_level": 3558, "full_trace": 3559, "run_metadata": 3560, "runmetadata": 3561, "step_stats": 3562, "dev_stats": 3563, "assertallclose": 3564, "41.0": 3565, "testinvalidtargetfails": 3566, "notfounderror": 3567, "'no session factory registered for the given session options'": 3568, "'invalid_target'": 3569, "testfeedshapecompatibility": 3570, "some_tensor": 3571, "new_shape": 3572, "reshaped_tensor": 3573, "reshape": 3574, "'cannot feed value of shape'": 3575, "'may not be fed'": 3576, "testinfershapestrue": 3577, "graph_options": 3578, "graphoptions": 3579, "infer_shapes": 3580, "'_output_shapes'": 3581, "graph_def": 3582, "lint": 3583, "regarding": 3584, "'unused'": 3585, "testbuildcostmodel": 3586, "allow_soft_placement": 3587, "build_cost_model": 3588, "xrange": 3589, "120": 3590, "testasdefault": 3591, "37": 3592, "remains": 3593, "captured": 3594, "testinvalidargument": 3595, "'target must be a string'": 3596, "'config must be a tf.configproto'": 3597, "'graph must be a tf.graph'": 3598, "testtimeoutwithshortoperations": 3599, "num_epochs": 3600, "data_flow_ops": 3601, "fifoqueue": 3602, "capacity": 3603, "50": 3604, "int32": 3605, "shapes": 3606, "enqueue_op": 3607, "enqueue_many": 3608, "operation_timeout_in_ms": 3609, "10000": 3610, "testbuildgrapherrordirect": 3611, "runtestbuildgrapherror": 3612, "testbuildgrapherrordist": 3613, "server_lib": 3614, "create_local_server": 3615, "testdeviceattributescanonicalization": 3616, "_deviceattributes": 3617, "'/job:worker/replica:0/task:3/cpu:1'": 3618, "'type'": 3619, "1337": 3620, "memory_limit_bytes": 3621, "'/job:worker/replica:0/task:3/device:cpu:1'": 3622, "device_type": 3623, "str_repr": 3624, "'%s'": 3625, "'_deviceattributes'": 3626, "runtestaddfunctiontosession": 3627, "\"add a function to a session after the graph has already been run.\"": 3628, "defun": 3629, "extract_links": 3630, "wrapper": 3631, "_extract_links": 3632, "body": 3633, "encoding": 3634, "links": 3635, "\"this extractor matches with any url, since\n it doesn't contain any patterns\"": 3636, "extractor": 3637, "contain": 3638, "patterns": 3639, "registeruserfunction": 3640, "io_environ": 3641, "'registeruserfunction'": 3642, "reset_delivery_states": 3643, "' remove the state set when resources are unpacked. '": 3644, "remove_state": 3645, "'kubernetes-e2e.installed'": 3646, "resources": 3647, "unpacked": 3648, "reset": 3649, "delivery": 3650, "states": 3651, "messaging": 3652, "' probe our relations to determine the propper messaging to the\n end user '": 3653, "missing_services": 3654, "is_state": 3655, "'kubernetes-master.available'": 3656, "'kubernetes-master'": 3657, "'certificates.available'": 3658, "'certificates'": 3659, "'relations'": 3660, "'relation'": 3661, "services": 3662, "','": 3663, "'missing {0}: {1}'": 3664, "hooken": 3665, "probe": 3666, "propper": 3667, "install_kubernetes_e2e": 3668, "' deliver the e2e and kubectl components from the binary resource stream\n packages declared in the charm '": 3669, "charm_dir": 3670, "getenv": 3671, "'charm_dir'": 3672, "arch": 3673, "determine_arch": 3674, "resource_get": 3675, "'e2e_{}'": 3676, "archive": 3677, "hookenv": 3678, "'error fetching the {} resource.'": 3679, "status_set": 3680, "'blocked'": 3681, "'missing {} resource.'": 3682, "deliver": 3683, "e2e": 3684, "kubectl": 3685, "packages": 3686, "declared": 3687, "charm": 3688, "install": 3689, "prepare_kubeconfig_certificates": 3690, "' prepare the data to feed to create the kubeconfig file. '": 3691, "layer_options": 3692, "layer": 3693, "'tls-client'": 3694, "paths": 3695, "tls": 3696, "kubeconfig": 3697, "'ca_certificate_path'": 3698, "'client_key_path'": 3699, "'client_certificate_path'": 3700, "get_kube_api_servers": 3701, "pedantry": 3702, "kubeconfig_path": 3703, "'/home/ubuntu/.kube/config'": 3704, "ubuntu": 3705, "create_kubeconfig": 3706, "prepare": 3707, "certificates": 3708, "set_app_version": 3709, "' declare the application version to juju '": 3710, "'kubectl'": 3711, "'version'": 3712, "'--client'": 3713, "subprocess": 3714, "calledprocesserror": 3715, "check_output": 3716, "\"missing kubeconfig causes errors. skipping version set.\"": 3717, "git_version": 3718, "'gitversion:\"v'": 3719, "version_from": 3720, "'\",'": 3721, "application_version_set": 3722, "rstrip": 3723, "declare": 3724, "application": 3725, "juju": 3726, "'ubuntu'": 3727, "'juju-context'": 3728, "cluster": 3729, "'juju-cluster'": 3730, "'create a configuration for kubernetes based on path using the supplied\n arguments for values of the kubernetes server, ca, key, certificate, user\n context and cluster.'": 3731, "'kubectl config --kubeconfig={0} set-cluster {1} '": 3732, "'--server={2} --certificate-authority={3} --embed-certs=true'": 3733, "check_call": 3734, "supplied": 3735, "'return the kubernetes api server address and port for this\n relationship.'": 3736, "hosts": 3737, "relation": 3738, "'hosts'": 3739, "'https://{0}:{1}'": 3740, "'hostname'": 3741, "'port'": 3742, "relationship": 3743, "kube": 3744, "' dpkg wrapper to surface the architecture we are tied to'": 3745, "'dpkg'": 3746, "'--print-architecture'": 3747, "dpkg": 3748, "surface": 3749, "architecture": 3750, "tied": 3751, "msgdlg": 3752, "caption": 3753, "'ofaimage'": 3754, "wx": 3755, "yes_no": 3756, "cancel": 3757, "\"common messagedialog.\"": 3758, "dlg": 3759, "messagedialog": 3760, "showmodal": 3761, "onfileexit": 3762, "\"exits the program.\"": 3763, "skip": 3764, "exits": 3765, "onsize": 3766, "\"window 'size' event.\"": 3767, "refresh": 3768, "'size'": 3769, "onerasebackground": 3770, "\"window 'erase background' event.\"": 3771, "'erase background'": 3772, "populatetree": 3773, "ds": 3774, "\" populate the tree in the left window with the [desired]\n dataset values\"": 3775, "dcmdsroot": 3776, "dstreeview": 3777, "addroot": 3778, "\"dicom objects\"": 3779, "deletechildren": 3780, "recurse_tree": 3781, "expandall": 3782, "populate": 3783, "tree": 3784, "desired": 3785, "\" order the dicom tags \"": 3786, "data_element": 3787, "ip": 3788, "appenditem": 3789, "vr": 3790, "\"sq\"": 3791, "sq_item_description": 3792, "\" sequence\"": 3793, "item_text": 3794, "\"%s %d\"": 3795, "sq_item_desc": 3796, "dicom": 3797, "recurse": 3798, "onfileopen": 3799, "\"opens a selected file.\"": 3800, "filedialog": 3801, "'choose a file to add.'": 3802, "'*.*'": 3803, "id_ok": 3804, "fullpath": 3805, "getpath": 3806, "imagefile": 3807, "getfilename": 3808, "checkdicmheader": 3809, "show_file": 3810, "opens": 3811, "selected": 3812, "onpaint": 3813, "\"window 'paint' event.\"": 3814, "dc": 3815, "paintdc": 3816, "imview": 3817, "buffereddc": 3818, "paint": 3819, "background": 3820, "isn": 3821, "boring": 3822, "setbackground": 3823, "brush": 3824, "\"white\"": 3825, "setbrush": 3826, "\"grey\"": 3827, "crossdiag_hatch": 3828, "windowsize": 3829, "getsizetuple": 3830, "drawrectangle": 3831, "bmpx0": 3832, "bmpy0": 3833, "bitmap": 3834, "loadcentered": 3835, "width": 3836, "'paint'": 3837, "convertwxtopil": 3838, "bmp": 3839, "\" convert wx.image into pil image. \"": 3840, "getwidth": 3841, "height": 3842, "getheight": 3843, "im": 3844, "emptyimage": 3845, "fromarray": 3846, "\"rgba\"": 3847, "getdata": 3848, "img": 3849, "image": 3850, "pil": 3851, "convertpiltowx": 3852, "alpha": 3853, "\" convert pil image into wx.image. \"": 3854, "setdata": 3855, "\"rgb\"": 3856, "tostring": 3857, "setalphadata": 3858, "new_image": 3859, "'rgb'": 3860, "get_lut_value": 3861, "\"apply the rgb look-up table for the given data and window/level value.\"": 3862, "have_numpy": 3863, "\"numpy is not available. see http://numpy.scipy.org/ to download and install\"": 3864, "piecewise": 3865, "0.5": 3866, "rgb": 3867, "lut": 3868, "\" load the dicom file, make sure it contains at least one\n image, and set it up for display by onpaint(). ** be\n careful not to pass a unicode string to read_file or it will\n give you 'fp object does not have a defer_size attribute,\n or some such.\"": 3869, "read_file": 3870, "change": 3871, "'pixeldata'": 3872, "dimage": 3873, "loadpil_lut": 3874, "dimag": 3875, "sure": 3876, "careful": 3877, "give": 3878, "defer_size": 3879, "oninit": 3880, "\"create the image application.\"": 3881, "imframe": 3882, "'wximage example'": 3883, "testbadval": 3884, "c1": 3885, "valueconstrainterror": 3886, "'constraint check fails'": 3887, "testgoodval": 3888, "testcmp3": 3889, "constraint": 3890, "constraintsunion": 3891, "constraintsintersection": 3892, "singlevalueconstraint": 3893, "valuerangeconstraint": 3894, "'__cmp__() fails'": 3895, "init_tof_0": 3896, "xshut": 3897, "\"xshut port low resets the device.\"": 3898, "rpi_gpio": 3899, "setup_output": 3900, "write_output": 3901, "resets": 3902, "tof": 3903, "init_tof_1": 3904, "\"xshut port high enables the device.\"": 3905, "high": 3906, "enables": 3907, "setup_method": 3908, "\"prob\"": 3909, "global_seed": 3910, "314159": 3911, "mech": 3912, "exponentialcategorical": 3913, "test_inf_epsilon": 3914, "utility_list": 3915, "\"b\"": 3916, "\"c\"": 3917, "epsilon": 3918, "\"inf\"": 3919, "_mech": 3920, "randomise": 3921, "inf": 3922, "test_nonzero_delta": 3923, "nonzero": 3924, "test_non_string_hierarchy": 3925, "hierarchy": 3926, "test_wrong_utilities": 3927, "\"2\"": 3928, "wrong": 3929, "test_outside_domain": 3930, "\"d\"": 3931, "outside": 3932, "domain": 3933, "test_get_utility_list": 3934, "_utility_list": 3935, "test_bias": 3936, "notimplementederror": 3937, "test_variance": 3938, "variance": 3939, "train": 3940, "ts_column": 3941, "x_columns": 3942, "y_max_lag": 3943, "x_max_lag": 3944, "no_intercept": 3945, "arx": 3946, "training": 3947, "autoregressive": 3948, "exogenous": 3949, "column": 3950, "series": 3951, "regressors": 3952, "lag": 3953, "dependent": 3954, "na": 3955, "tc": 3956, "tkcontext": 3957, "implicit": 3958, "\"load arxmodel from given path\"": 3959, "arxmodel": 3960, "get_scala_obj": 3961, "\"gets reference to the arxmodel scala object\"": 3962, "sc": 3963, "_jvm": 3964, "trustedanalytics": 3965, "sparktk": 3966, "timeseries": 3967, "scala": 3968, "\"\n the maximum lag order for the dependent (time series) values.\n \"": 3969, "_scala": 3970, "ymaxlag": 3971, "\"\n the maximum lag order for exogenous variables.\n \"": 3972, "xmaxlag": 3973, "\"\n an intercept term (zero if none desired), from the trained model.\n \"": 3974, "intercept": 3975, "zero": 3976, "trained": 3977, "coefficients": 3978, "\"\n coefficient values from the trained model.\n \"": 3979, "_tc": 3980, "jutils": 3981, "from_scala_seq": 3982, "coefficient": 3983, "\"\n a boolean flag indicating if the intercept should be dropped.\n \"": 3984, "nointercept": 3985, "flag": 3986, "indicating": 3987, "dropped": 3988, "predicted": 3989, "columns": 3990, "predicted_y": 3991, "predicting": 3992, "ts": 3993, "\"\n save the trained model to the specified path.\n\n parameters\n ----------\n :param path: (str) path to save\n \"": 3994, "_scala_sc": 3995, "export_to_mar": 3996, "\"\n exports the trained model as a model archive (.mar) to the specified path.\n\n parameters\n ----------\n\n :param path: (str) path to save the trained model\n :returns (str) full path to the saved .mar file\n\n \"": 3997, "basestring": 3998, "\"path parameter must be a str, but received %s\"": 3999, "exporttomar": 4000, "exports": 4001, "mar": 4002, "export": 4003, "sequence_classifier": 4004, "decoding": 4005, "sampling_decoding": 4006, "\"returns predictions and loss for sequence of predictions.\n\n args:\n decoding: list of tensors with predictions.\n labels: list of tensors with labels.\n sampling_decoding: optional, list of tensor with predictions to be used\n in sampling. e.g. they shouldn't have dependncy on outputs.\n if not provided, decoding is used.\n name: operation name.\n\n returns:\n predictions and losses tensors.\n \"": 4007, "name_scope": 4008, "\"sequence_classifier\"": 4009, "predictions": 4010, "xent_list": 4011, "loss": 4012, "tensors": 4013, "sampling": 4014, "E": 4015, "shouldn": 4016, "dependncy": 4017, "outputs": 4018, "classifier": 4019, "seq2seq_inputs": 4020, "input_length": 4021, "output_length": 4022, "sentinel": 4023, "processes": 4024, "batch_size": 4025, "embed_dim": 4026, "decoder": 4027, "zeros": 4028, "due": 4029, "fact": 4030, "inferred": 4031, "o": 4032, "seq2seq": 4033, "rnn_decoder": 4034, "decoder_inputs": 4035, "initial_state": 4036, "cell": 4037, "\"rnn decoder that creates training and sampling sub-graphs.\n\n args:\n decoder_inputs: inputs for decoder, list of tensors.\n this is used only in training sub-graph.\n initial_state: initial state for the decoder.\n cell: rnn cell to use for decoder.\n scope: scope to use, if none new will be produced.\n\n returns:\n list of tensors for outputs and states for training and sampling sub-graphs.\n \"": 4038, "vs": 4039, "variable_scope": 4040, "\"dnn_decoder\"": 4041, "sampling_states": 4042, "rnn": 4043, "graphs": 4044, "produc": 4045, "rnn_seq2seq": 4046, "encoder_inputs": 4047, "encoder_cell": 4048, "decoder_cell": 4049, "encoder": 4050, "produced": 4051, "enc": 4052, "namedany": 4053, "fully": 4054, "qualified": 4055, "global": 4056, "describes": 4057, "imported": 4058, "subsequent": 4059, "treated": 4060, "came": 4061, "'twisted.python.reflect.namedany'": 4062, "invalidname": 4063, "whi": 4064, "get_charset_name": 4065, "\"big5\"": 4066, "charset": 4067, "test_user": 4068, "\"test\"": 4069, "deleteuser": 4070, "\"user1\"": 4071, "\"password\"": 4072, "\"11111\"": 4073, "\"surname\"": 4074, "\"lastname\"": 4075, "\"organistion\"": 4076, "\"organisation\"": 4077, "\"user1@aaa.ccc\"": 4078, "adduser": 4079, "activate": 4080, "generatepw": 4081, "groups": 4082, "currentuser": 4083, "assert_": 4084, "secureupdate": 4085, "updategroups": 4086, "\"group:author\"": 4087, "getgroups": 4088, "au": 4089, "test_login": 4090, "identityfield": 4091, "\"adminxxxxx\"": 4092, "\"admin\"": 4093, "\"admin@aaa.cccxxxxx\"": 4094, "admi": 4095, "checksubpackage": 4096, "'__map__'": 4097, "__map__": 4098, "test_virtual_module_identity": 4099, "path1": 4100, "path2": 4101, "local1": 4102, "local2": 4103, "virtual": 4104, "test_margulis_gabber_galil_graph": 4105, "conversion": 4106, "adjacency": 4107, "computing": 4108, "eigenvalues": 4109, "could": 4110, "done": 4111, "numpy": 4112, "linalg": 4113, "has_scipy": 4114, "margulis_gabber_galil_graph": 4115, "assert_equal": 4116, "number_of_nodes": 4117, "degree": 4118, "margulis": 4119, "gabber": 4120, "galil": 4121, "test_chordal_cycle_graph": 4122, "\"test for the :func:`networkx.chordal_cycle_graph` function.\"": 4123, "is_scipy_available": 4124, "skiptest": 4125, "'scipy is not available'": 4126, "primes": 4127, "11": 4128, "G": 4129, "chordal_cycle_graph": 4130, "networkx": 4131, "chordal": 4132, "cycle": 4133, "addbuttonstotoolbar": 4134, "button": 4135, "toggling": 4136, "executions": 4137, "eye_open_icon": 4138, "qtgui": 4139, "qicon": 4140, "vistrails_root_directory": 4141, "'gui/resources/images/eye.png'": 4142, "portvisibilityaction": 4143, "qaction": 4144, "\"show/hide port visibility toggle buttons\"": 4145, "showportvisibility": 4146, "setcheckable": 4147, "setchecke": 4148, "build_widget": 4149, "name_label": 4150, "qlabel": 4151, "\"name:\"": 4152, "name_edit": 4153, "qlineedit": 4154, "qtcore": 4155, "signal": 4156, "'editingfinished()'": 4157, "name_editing_finished": 4158, "setminimumsize": 4159, "22": 4160, "type_label": 4161, "\"type:\"": 4162, "type_edit": 4163, "package_label": 4164, "\"package:\"": 4165, "package_edit": 4166, "namespace_label": 4167, "\"namespace:\"": 4168, "namespace_edit": 4169, "\"id:\"": 4170, "widget": 4171, "setreadonly": 4172, "read_only": 4173, "ports_lists": 4174, "vistrails": 4175, "gui": 4176, "vistrails_window": 4177, "_app": 4178, "configure_module": 4179, "show_documentation": 4180, "show_annotations": 4181, "tab_widget": 4182, "setcurrentwidget": 4183, "editnextavailablecell": 4184, "reqfunc": 4185, "counter": 4186, "buildmatch": 4187, "ui": 4188, "repo": 4189, "'return tuple of (match function, list enabled).'": 4190, "has_section": 4191, "'acl: %s not enabled\\n'": 4192, "pats": 4193, "pat": 4194, "configitems": 4195, "_usermatch": 4196, "'acl: %s enabled, %d entries for user %s\\n'": 4197, "branch": 4198, "acl": 4199, "'s an asterisk (meaning \"any branch\"), always return true;\n # otherwise, test if b is in pats\n if '": 4200, "fromrna": 4201, "meshobject": 4202, "globalsettings": 4203, "bpy": 4204, "scene": 4205, "ogre_mesh_exporter": 4206, "meshsettings": 4207, "meshexportsettings": 4208, "fixupaxistoy": 4209, "requirematerials": 4210, "requirematerials_override": 4211, "applymodifiers": 4212, "applymodifiers_override": 4213, "skeletonnamefollowmesh": 4214, "skeletonnamefollowmesh_override": 4215, "globa": 4216, "norm": 4217, "uvs": 4218, "mnormal": 4219, "muvs": 4220, "mcolors": 4221, "color1": 4222, "color2": 4223, "serializeboneassignments": 4224, "'%s\\t\\n'": 4225, "vertexwithnoboneassignements": 4226, "vertex": 4227, "mvertexdata": 4228, "mboneweights": 4229, "boneweight": 4230, "'%s\\t\\t\\n'": 4231, "mboneindex": 4232, "mboneweight": 4233, "logmanager": 4234, "logmessage": 4235, "\"there are %d vertices with no bone assignements!\"": 4236, "vertexwithn": 4237, "vertexcount": 4238, "mvertexbuffer": 4239, "materialattribute": 4240, "mmaterial": 4241, "' material=\"%s\"'": 4242, "'\\t\\t\\n'": 4243, "'true'": 4244, "msharevertexbuffer": 4245, "'false'": 4246, "65536": 4247, "face": 4248, "'\\t\\t\\t\\n'": 4249, "mfacedata": 4250, "'\\t\\t\\t\\t\\n'": 4251, "'\\t\\t\\t\\n'": 4252, "test_update_rc_files": 4253, "pmock": 4254, "\"\n test happy path on updating env files. assuming everything\n exists and is in place.\n \"": 4255, "return_value": 4256, "'line1'": 4257, "'line2'": 4258, "update_rc_files": 4259, "'test1'": 4260, "'test2'": 4261, "write_lines": 4262, "assert_called_with": 4263, "happy": 4264, "updating": 4265, "assuming": 4266, "place": 4267, "rc": 4268, "test_update_rc_files_with_nonexistant_path": 4269, "\"\n test an unhappy path if the bashrc/users do not exist.\n \"": 4270, "pytest": 4271, "oserror": 4272, "exinfo": 4273, "unhappy": 4274, "bashrc": 4275, "nonexistant": 4276, "test_package_installation": 4277, "hemock": 4278, "ftmock": 4279, "\"\n verify we are calling the known essentials to build and syndicate\n kubes.\n \"": 4280, "pkgs": 4281, "'build-essential'": 4282, "'git'": 4283, "'make'": 4284, "'nginx'": 4285, "'python-pip'": 4286, "install_packages": 4287, "'installing debian packages'": 4288, "filter_installed_packages": 4289, "essentials": 4290, "syndicate": 4291, "kubes": 4292, "installation": 4293, "test_go_download": 4294, "aumock": 4295, "\"\n test that we are actually handing off to charm-helpers to\n download a specific archive of go. this is non-configurable so\n its reasonably safe to assume we're going to always do this,\n and when it changes we shall curse the brittleness of this test.\n \"": 4296, "ins_mock": 4297, "download_go": 4298, "'https://storage.googleapis.com/golang/go1.4.2.linux-amd64.tar.gz'": 4299, "'5020af94b52b65cc9b6f11d50a67e4bae07b0aff'": 4300, "assert_cal": 4301, "actually": 4302, "handing": 4303, "helpers": 4304, "download": 4305, "go": 4306, "configurable": 4307, "reasonably": 4308, "assume": 4309, "always": 4310, "shall": 4311, "curse": 4312, "brittleness": 4313, "test_clone_repository": 4314, "spmock": 4315, "\"\n we're not using a unit-tested git library - so ensure our subprocess\n call is consistent. if we change this, we want to know we've broken it.\n \"": 4316, "clone_repository": 4317, "'https://github.com/googlecloudplatform/kubernetes.git'": 4318, "direct": 4319, "'/opt/kubernetes'": 4320, "'clone'": 4321, "'re not using a unit-tested git library - so ensure our subprocess\ncall is consistent. if we change this, we want to know we'": 4322, "broken": 4323, "repository": 4324, "tested": 4325, "test_install_main": 4326, "urmock": 4327, "crmock": 4328, "dgmock": 4329, "ipmock": 4330, "\"\n ensure the driver/main method is calling all the supporting methods.\n \"": 4331, "'export goroot=/usr/local/go\\n'": 4332, "'export path=$path:$goroot/bin\\n'": 4333, "'export kube_master_ip=0.0.0.0\\n'": 4334, "'export kubernetes_master=http://$kube_master_ip\\n'": 4335, "assert_called_once": 4336, "open_port": 4337, "asser": 4338, "test_can_specify_attributes_gridded": 4339, "\"%s,%s:%s\"": 4340, "valid_echamham_variable_1": 4341, "valid_echamham_variable_2": 4342, "escape_colons": 4343, "valid_echamham_filename": 4344, "'%s+%s'": 4345, "'kg m^-3'": 4346, "'-o'": 4347, "'var_out:'": 4348, "output_filename": 4349, "'-a'": 4350, "'att1=val1,att2=val2'": 4351, "evaluate_cmd": 4352, "netcdf4": 4353, "assert_that": 4354, "'var_out'": 4355, "att1": 4356, "is_": 4357, "'val1'": 4358, "var_": 4359, "gridded": 4360, "test_can_specify_units_gridded": 4361, "test_can_specify_units_gridded_no_output_var": 4362, "\"%s:%s\"": 4363, "valid_hadgem_variable": 4364, "valid_hadgem_filename": 4365, "\"od550aer\"": 4366, "\"ppm\"": 4367, "\"-o\"": 4368, "\"-a\"": 4369, "\"att1=val1\"": 4370, "'calculated_variable'": 4371, "'ppm'": 4372, "get_template_sources": 4373, "template_dirs": 4374, "\"\r\n returns the absolute paths to \"": 4375, "\", when appended to each\r\n directory in \"": 4376, "\". any paths that don't lie inside one of the\r\n template dirs are excluded from the result set, for security reasons.\r\n \"": 4377, "get_app_template_dirs": 4378, "'templates'": 4379, "template_dir": 4380, "safe_join": 4381, "suspiciousfileoperation": 4382, "absolute": 4383, "\"template_name\"": 4384, "appended": 4385, "\"template_dirs\"": 4386, "lie": 4387, "inside": 4388, "excluded": 4389, "reasons": 4390, "sources": 4391, "load_template_source": 4392, "filepath": 4393, "io": 4394, "file_charset": 4395, "ioerror": 4396, "templatedoesnotexist": 4397, "test_zero_as_autoval": 4398, "\"\n zero as id for autofield should raise exception in mysql, because mysql\n does not allow zero for automatic primary key.\n \"": 4399, "valid_country": 4400, "'germany'": 4401, "iso_two_letter": 4402, "'de'": 4403, "invalid_country": 4404, "'poland'": 4405, "'pl'": 4406, "bulk_create": 4407, "autofield": 4408, "mysql": 4409, "automatic": 4410, "autoval": 4411, "test_batch_same_vals": 4412, "had": 4413, "collapsed": 4414, "restaurant": 4415, "batch": 4416, "test_large_batch": 4417, "override_settings": 4418, "queries": 4419, "twofields": 4420, "f1": 4421, "f2": 4422, "1001": 4423, "f1__gte": 4424, "450": 4425, "f1__lte": 4426, "550": 4427, "101": 4428, "f2__gte": 4429, "901": 4430, "large": 4431, "test_large_batch_mixed": 4432, "\"\n test inserting a large batch with objects having primary key set\n mixed together with objects without pk set.\n \"": 4433, "100000": 4434, "101000": 4435, "'t assume much about the id'": 4436, "inserting": 4437, "mixed": 4438, "together": 4439, "test_large_batch_mixed_efficiency": 4440, "efficiency": 4441, "test_explicit_batch_size_efficiency": 4442, "objs": 4443, "assertnumqueries": 4444, "dbapi": 4445, "warn_deprecated": 4446, "\"google cloud sql now recommends creating connections via the \"": 4447, "\"mysqldb dialect directly, using the url format \"": 4448, "\"mysql+mysqldb://root@/?unix_socket=/cloudsql/\"": 4449, "\":\"": 4450, "googleappengine": 4451, "trunk": 4452, "storage": 4453, "speckle": 4454, "backend": 4455, "118": 4456, "ticket": 4457, "2649": 4458, "14224679": 4459, "34549": 4460, "googl": 4461, "create_connect_args": 4462, "translate_connect_args": 4463, "_is_dev_environment": 4464, "'dsn'": 4465, "'instance'": 4466, "skipping": 4467, "traditional": 4468, "rdbms": 4469, "\"\n yield unique values in iterable, preserving order.\n \"": 4470, "seen": 4471, "preserving": 4472, "handle_requires": 4473, "pkg_info": 4474, "\"\n place the runtime requirements from pkg_info into metadata.\n \"": 4475, "may_requires": 4476, "ordereddefaultdict": 4477, "get_all": 4478, "extra_match": 4479, "extra_re": 4480, "groupdict": 4481, "'condition'": 4482, "'extra'": 4483, "'package'": 4484, "' and '": 4485, "runtime": 4486, "requirements": 4487, "pkginfo_to_dict": 4488, "distribution": 4489, "\"\n convert pkg-info to a prototype metadata 2.0 (pep 426) dict.\n\n the description is included under the key ['description'] rather than\n being written to a separate file.\n\n path: path to pkg-info file\n distribution: optional distutils distribution()\n \"": 4490, "ordereddict": 4491, "\"generator\"": 4492, "\"bdist_wheel (\"": 4493, "wheel_version": 4494, "\")\"": 4495, "read_pkg_info": 4496, "pkg": 4497, "prototype": 4498, "pep": 4499, "426": 4500, "'description'": 4501, "written": 4502, "separate": 4503, "distutils": 4504, "pkginfo": 4505, "requires_to_requires_dist": 4506, "requirement": 4507, "\"compose the version predicates for requirement in pep 345 fashion.\"": 4508, "requires_dist": 4509, "ver": 4510, "\" (%s)\"": 4511, "predicates": 4512, "345": 4513, "fashion": 4514, "convert_requirements": 4515, "\"yield requires-dist: strings for parsed requirements strings.\"": 4516, "req": 4517, "parsed_requirement": 4518, "pkg_resources": 4519, "extras": 4520, "\",\"": 4521, "\"[%s]\"": 4522, "parsed": 4523, "generate_requirements": 4524, "extras_require": 4525, "\"\n convert requirements from a setup()-style dictionary to ('requires-dist', 'requirement')\n and ('provides-extra', 'extra') tuples.\n\n extras_require is a dictionary of {extra: [requirements]} as passed to setup(),\n using the empty extra {'': [requirements]} to hold install_requires.\n \"": 4526, "depends": 4527, "':'": 4528, "setuptools": 4529, "safe_extra": 4530, "'requires-dist'": 4531, "'requirement'": 4532, "'provides-extra'": 4533, "tuples": 4534, "hold": 4535, "install_requires": 4536, "pkginfo_to_metadata": 4537, "egg_info_path": 4538, "pkginfo_path": 4539, "\"\n convert .egg-info directory with pkg-info to the metadata 1.3 aka\n old-draft metadata 2.0 format.\n \"": 4540, "replace_header": 4541, "'metadata-version'": 4542, "'2.0'": 4543, "requires_path": 4544, "'requires.txt'": 4545, "requires_file": 4546, "reqs": 4547, "split_sections": 4548, "egg": 4549, "1.3": 4550, "aka": 4551, "draft": 4552, "pkginfo_unicode": 4553, "\"hack to coax unicode out of an email message() - python 3.3+\"": 4554, "'raw_items'": 4555, "3.2": 4556, "raw_items": 4557, "'surrogateescape'": 4558, "hack": 4559, "coax": 4560, "3.3": 4561, "dedent_description": 4562, "\"\n dedent and convert pkg_info['description'] to unicode.\n \"": 4563, "handling": 4564, "sorta": 4565, "surrogates": 4566, "description_lines": 4567, "description_dedent": 4568, "'\\n'": 4569, "long_description": 4570, "blank": 4571, "indented": 4572, "lstrip": 4573, "sort_key": 4574, "both": 4575, "'t be compared\n # against strings in python 3.\n key, value = item\n if key.condition is none:\n return '": 4576, "'dds'": 4577, "from_db_value": 4578, "expression": 4579, "geometry": 4580, "get_tenant_by_name": 4581, "tenants": 4582, "tenants_client": 4583, "list_tenants": 4584, "'tenants'": 4585, "projects_client": 4586, "list_projects": 4587, "'projects'": 4588, "create_test_user": 4589, "data_utils": 4590, "rand_password": 4591, "rand_name": 4592, "'test_user'": 4593, "'@testmail.tm'": 4594, "users_client": 4595, "create_user": 4596, "addcleanup": 4597, "test_utils": 4598, "call_and_ignore_notfound_exc": 4599, "delete_user": 4600, "setup_test_role": 4601, "domain_id": 4602, "\"set up a test role.\"": 4603, "'test_role'": 4604, "'domain_id'": 4605, "role": 4606, "roles_client": 4607, "create_role": 4608, "'role'": 4609, "delete_role": 4610, "skip_checks": 4611, "baseidentityv2admintest": 4612, "conf": 4613, "identity_feature_enabled": 4614, "api_v2_admin": 4615, "skipexception": 4616, "'identity v2 admin not available'": 4617, "setup_clients": 4618, "os_admin": 4619, "identity_client": 4620, "non_admin_client": 4621, "os_primary": 4622, "token_client": 4623, "non_admin_tenants_client": 4624, "non_admin_roles_client": 4625, "non_admin_users_client": 4626, "os_primar": 4627, "setup_test_user": 4628, "\"set up a test user.\"": 4629, "setup_test_tenant": 4630, "tenantid": 4631, "\"set up a test tenant.\"": 4632, "'test_tenant'": 4633, "'desc'": 4634, "'tenant'": 4635, "create_domain": 4636, "\"create a domain.\"": 4637, "'test_domain'": 4638, "domains_client": 4639, "delete_domain": 4640, "mpavlase": 4641, "necessary": 4642, "deleting": 4643, "forbidden": 4644, "update_domain": 4645, "setup_test_project": 4646, "\"set up a test project.\"": 4647, "'test_project'": 4648, "'test_description'": 4649, "project": 4650, "create_project": 4651, "'project'": 4652, "delete_project": 4653, "setup_test_domain": 4654, "\"set up a test domain.\"": 4655, "entity_id": 4656, "\"return entity id.\"": 4657, "\"{0}.{1}\"": 4658, "_id": 4659, "entity": 4660, "device_state_attributes": 4661, "\"return device specific state attributes.\"": 4662, "p_id": 4663, "_api": 4664, "store": 4665, "\"store group/person data and ids.\"": 4666, "_store": 4667, "person": 4668, "app_backend": 4669, "_app_backend": 4670, "platform_backend": 4671, "closebrowser": 4672, "\"close the browser instance.\"": 4673, "'method'": 4674, "'browser.close'": 4675, "_browser_websocket": 4676, "syncrequest": 4677, "60": 4678, "isalive": 4679, "\"whether the devtools server is available and connectable.\"": 4680, "_devtools_http": 4681, "devtools_http": 4682, "devtoolsclientconnectionerror": 4683, "whether": 4684, "devtools": 4685, "connectable": 4686, "getversion": 4687, "\"return the version dict as provided by the devtools agent.\"": 4688, "requestjson": 4689, "agent": 4690, "requestnewtab": 4691, "in_new_window": 4692, "\"creates a new tab, either in new window or current window.\n\n returns:\n a dict of a parsed json object as returned by devtools. example:\n if an error is present, the dict will contain an 'error' key.\n if no error is present, the result is present in the 'result' key:\n {\n \"": 4693, "\": {\n \"": 4694, "targetid": 4695, "\": \"": 4696, "\" # this is the id for the tab.\n }\n }\n \"": 4697, "'target.createtarget'": 4698, "bla": 4699, "tab": 4700, "present": 4701, "'error'": 4702, "'result'": 4703, "\"result\"": 4704, "targ": 4705, "closetab": 4706, "tab_id": 4707, "\"closes the tab with the given id.\n\n raises:\n devtools_http.devtoolsclientconnectionerror\n tabnotfounderror\n \"": 4708, "'close/%s'": 4709, "devtoolsclienturlerror": 4710, "tabnotfounderror": 4711, "'unable to close tab, tab id not found: %s'": 4712, "six": 4713, "reraise": 4714, "exc_info": 4715, "closes": 4716, "activatetab": 4717, "\"activates the tab with the given id.\n\n raises:\n devtools_http.devtoolsclientconnectionerror\n tabnotfounderror\n \"": 4718, "'activate/%s'": 4719, "'unable to activate tab, tab id not found: %s'": 4720, "activates": 4721, "geturl": 4722, "\"returns the url of the tab with |tab_id|, as reported by devtools.\n\n raises:\n devtools_http.devtoolsclientconnectionerror\n \"": 4723, "_listinspectablecontexts": 4724, "reported": 4725, "isinspectable": 4726, "\"whether the tab with |tab_id| is inspectable, as reported by devtools.\n\n raises:\n devtools_http.devtoolsclientconnectionerror\n \"": 4727, "contexts": 4728, "inspectable": 4729, "getupdatedinspectablecontexts": 4730, "\"returns an updated instance of _devtoolscontextmapbackend.\"": 4731, "_devtools_context_map_backend": 4732, "_update": 4733, "_devtoolscontextmapbackend": 4734, "startchrometracing": 4735, "trace_config": 4736, "transfer_mode": 4737, "\"\n args:\n trace_config: an tracing_config.tracingconfig instance.\n transfer_mode: defaults to using 'returnasstream' transfer mode\n for chrome tracing. can be set to 'reportevents'.\n timeout: time waited for websocket to receive a response.\n \"": 4738, "_tracing_backend": 4739, "enable_chrome_trace": 4740, "starttracing": 4741, "chrome_trace_config": 4742, "tracing_config": 4743, "tracingconfig": 4744, "'returnasstream'": 4745, "transfer": 4746, "chrome": 4747, "tracing": 4748, "'reportevents'": 4749, "waited": 4750, "websocket": 4751, "firsttabbackend": 4752, "\"obtain the inspector backend for the firstly created tab.\"": 4753, "_iterinspectorbackends": 4754, "'page'": 4755, "obtain": 4756, "inspector": 4757, "firstly": 4758, "dumpmemory": 4759, "detail_level": 4760, "responses": 4761, "detail": 4762, "'detailed'": 4763, "'light'": 4764, "'background'": 4765, "guid": 4766, "generated": 4767, "successful": 4768, "tracingtimeoutexception": 4769, "tracingunrecoverableexception": 4770, "tracingunexpectedresponseexception": 4771, "setmemorypressurenotificationssuppressed": 4772, "suppressed": 4773, "\"enable/disable suppressing memory pressure notifications.\n\n args:\n suppressed: if true, memory pressure notifications will be suppressed.\n timeout: the timeout in seconds.\n\n raises:\n memorytimeoutexception: if more than |timeout| seconds has passed\n since the last time any data is received.\n memoryunrecoverableexception: if there is a websocket error.\n memoryunexpectedresponseexception: if the response contains an error\n or does not contain the expected result.\n \"": 4774, "suppressing": 4775, "pressure": 4776, "notifications": 4777, "memorytimeoutexception": 4778, "memo": 4779, "simulatememorypressurenotification": 4780, "pressure_level": 4781, "simulate": 4782, "notification": 4783, "'moderate'": 4784, "'critical'": 4785, "memoryunrecoverableexception": 4786, "memoryunexpectedresponseexception": 4787, "window_manager_backend": 4788, "\"return the window manager backend.\n\n this should be called by a cros backend only.\n \"": 4789, "_createwindowmanagerbackendifneeded": 4790, "_wm_backend": 4791, "cros": 4792, "getwindowfortarget": 4793, "target_id": 4794, "'browser.getwindowfortarget'": 4795, "'targetid'": 4796, "\"the most up to date contexts data.\n\n returned in the order returned by devtools agent.\"": 4797, "_contexts": 4798, "most": 4799, "getcontextinfo": 4800, "context_id": 4801, "'cannot find a context with id=%s'": 4802, "getinspectorbackend": 4803, "\"gets an inspectorbackend instance for the given context_id.\n\n this lazily creates inspectorbackend for the context_id if it does\n not exist yet. otherwise, it will return the cached instance.\"": 4804, "_inspector_backends_dict": 4805, "new_backend": 4806, "inspector_backend": 4807, "inspectorbackend": 4808, "_devtools_client": 4809, "lazily": 4810, "topology": 4811, "\"for geos unary topology functions.\"": 4812, "geom_ptr": 4813, "restype": 4814, "errcheck": 4815, "check_geom": 4816, "geos": 4817, "unary": 4818, "functions": 4819, "mock_scan_fixture": 4820, "\"mock pyatv.scan.\"": 4821, "\"homeassistant.components.apple_tv.config_flow.scan\"": 4822, "mock_scan": 4823, "_scan": 4824, "loop": 4825, "identifier": 4826, "protocol": 4827, "pyatv": 4828, "scan": 4829, "dmap_pin_fixture": 4830, "\"homeassistant.components.apple_tv.config_flow.randrange\"": 4831, "mock_pin": 4832, "1111": 4833, "dmap": 4834, "pin": 4835, "pairing": 4836, "\"homeassistant.components.apple_tv.config_flow.pair\"": 4837, "mock_pair": 4838, "_pair": 4839, "mockpairinghandler": 4840, "net": 4841, "get_service": 4842, "always_fail": 4843, "pairing_mock": 4844, "_begin": 4845, "_close": 4846, "begin": 4847, "full_device": 4848, "dmap_pin": 4849, "create_conf": 4850, "\"127.0.0.1\"": 4851, "\"mrp device\"": 4852, "mrpservice": 4853, "\"mrpid\"": 4854, "5555": 4855, "dmapservice": 4856, "\"dmapid\"": 4857, "6666": 4858, "airplayservice": 4859, "\"airplayid\"": 4860, "7777": 4861, "mrp_device": 4862, "mrp": 4863, "dmap_device": 4864, "\"dmap device\"": 4865, "dmap_device_with_credentials": 4866, "\"dummy_creds\"": 4867, "credentials": 4868, "airplay_device": 4869, "\"airplay device\"": 4870, "airplay": 4871, "assertdriverstartsuccessful": 4872, "expected_logs": 4873, "expected_display": 4874, "pixel_tests": 4875, "outputcapture": 4876, "assert_outputs": 4877, "_server_process": 4878, "\"display\"": 4879, "__realize_cb": 4880, "set_type_hint": 4881, "gdk": 4882, "windowtypehint": 4883, "dialog": 4884, "get_window": 4885, "set_accept_focus": 4886, "parent_window_xid": 4887, "gdkx11": 4888, "x11window": 4889, "foreign_new_for_display": 4890, "set_transient_for": 4891, "modal": 4892, "hot": 4893, "sl": 4894, "4601": 4895, "shell": 4896, "push_modal": 4897, "realize": 4898, "cb": 4899, "__size_changed_cb": 4900, "_calculate_max_columns": 4901, "changed": 4902, "grab_focus": 4903, "grab": 4904, "_main_view": 4905, "get_child": 4906, "__key_press_event_cb": 4907, "keyval": 4908, "key_return": 4909, "_selected": 4910, "show_section_view": 4911, "key_escape": 4912, "_toolbar": 4913, "_main_toolbar": 4914, "__stop_clicked_cb": 4915, "__cancel_clicked_cb": 4916, "clicked": 4917, "fix": 4918, "3188": 4919, "is_active": 4920, "pres": 4921, "press": 4922, "__clear_icon_pressed_cb": 4923, "icon_pos": 4924, "pressed": 4925, "_set_toolbar": 4926, "_section_toolbar": 4927, "get_icon": 4928, "set_from_icon_name": 4929, "_options": 4930, "'icon'": 4931, "gtk": 4932, "iconsize": 4933, "large_toolbar": 4934, "props": 4935, "xo_color": 4936, "'color'": 4937, "get_title": 4938, "set_text": 4939, "_current_option": 4940, "__import__": 4941, "'cpsection'": 4942, "'view'": 4943, "glob": 4944, "section": 4945, "set_section_view_auto_close": 4946, "\"automatically close the control panel if there is \"": 4947, "\"\n \"": 4948, "_section_view": 4949, "auto_close": 4950, "\"nothing to do\"": 4951, "'alerts'": 4952, "accept_button": 4953, "set_sensitive": 4954, "_show_main_view": 4955, "__quit_timeout_cb": 4956, "unbusy": 4957, "alert": 4958, "timeoutalert": 4959, "'an activity is not responding.'": 4960, "'you may lose unsaved work if you continue.'": 4961, "'response'": 4962, "__quit_accept_cb": 4963, "add_alert": 4964, "quit": 4965, "__select_option_cb": 4966, "__close_request_cb": 4967, "do_get_property": 4968, "pspec": 4969, "'icon-name'": 4970, "_icon_name": 4971, "'pixel-size'": 4972, "_pixel_size": 4973, "'xo-color'": 4974, "_xo_color": 4975, "_title": 4976, "ce": 4977, "line_num": 4978, "config_error": 4979, "test_attribute_mc_required_env": 4980, "capsys": 4981, "errorline": 4982, "configexception": 4983, "mc_config": 4984, "ef1_prod_pp": 4985, "load_now": 4986, "itemwithaa": 4987, "next_line_num": 4988, "setattr": 4989, "'aa'": 4990, "prod": 4991, "mc_required": 4992, "pp": 4993, "\"hello\"": 4994, "_sout": 4995, "serr": 4996, "readouterr": 4997, "lines_in": 4998, "start_file_line": 4999, "config_error_never_received_value_expected": 5000, "prod1": 5001, "'^configerror: '": 5002, "mc": 5003, "test_attribute_mc_required_mc_force_env": 5004, "mc_force": 5005, "config_error_mc_required_expected": 5006, "pp1": 5007, "replace_ids": 5008, "_mc": 5009, "test_attribute_mc_required_default": 5010, "configerro": 5011, "test_attribute_mc_required_default_resolved_with_default_value_in_mc_init": 5012, "itemwithaamcinitresolve": 5013, "mc_init": 5014, "aa": 5015, "'hello'": 5016, "resolved": 5017, "test_attribute_mc_required_args_partial_set_in_init_overridden_in_with": 5018, "configitem": 5019, "partial": 5020, "assignment": 5021, "rq": 5022, "test_attribute_mc_required_args_set_in_init_overridden_in_with": 5023, "test_attribute_mc_required_init_args_missing_previous_item": 5024, "mcrequiredininitl1": 5025, "mcrequiredininitl3": 5026, "'^file \"{}/invalid_values_classes.py\", line 8'": 5027, "_utils": 5028, "mc_required_expected": 5029, "previous": 5030, "sense": 5031, "\"hi\"": 5032, "config2": 5033, "config0": 5034, "nextrange": 5035, "test01_getslice": 5036, "'slice retrieval'": 5037, "pl": 5038, "ul": 5039, "lists_of_len": 5040, "limits_plus": 5041, "'slice [%d:]'": 5042, "'slice [:%d]'": 5043, "'slice [%d:%d]'": 5044, "step_range": 5045, "'slice [%d:%d:%d]'": 5046, "retrieval": 5047, "test01": 5048, "getslice": 5049, "test02_setslice": 5050, "'slice assignment'": 5051, "setfcn": 5052, "slen": 5053, "limit": 5054, "ssl": 5055, "'set slice [:]'": 5056, "'set slice [%d:]'": 5057, "test02": 5058, "setslice": 5059, "test03_delslice": 5060, "'delete slice'": 5061, "'del slice [:]'": 5062, "'del slice [%d:]'": 5063, "test03": 5064, "delslice": 5065, "test04_get_set_del_single": 5066, "'get/set/delete single item'": 5067, "'get single item [%d]'": 5068, "'set single item [%d]'": 5069, "singl": 5070, "test04": 5071, "test05_out_of_range_exceptions": 5072, "'out of range exceptions'": 5073, "getfcn": 5074, "delfcn": 5075, "'set index %d'": 5076, "'get index %d'": 5077, "'del index %d'": 5078, "test05": 5079, "test06_list_methods": 5080, "'list methods'": 5081, "40": 5082, "'append'": 5083, "55": 5084, "'extend'": 5085, "'reverse'": 5086, "'insert at %d'": 5087, "test06": 5088, "test07_allowed_types": 5089, "'type-restricted list'": 5090, "_allowed": 5091, "70": 5092, "80": 5093, "'goodbye'": 5094, "restricted": 5095, "test07": 5096, "types": 5097, "test08_min_length": 5098, "'length limits'": 5099, "_minlength": 5100, "_maxlength": 5101, "limits": 5102, "test08": 5103, "test09_iterable_check": 5104, "'error on assigning non-iterable to slice'": 5105, "assigning": 5106, "test09": 5107, "test10_checkindex": 5108, "'index check'": 5109, "_checkindex": 5110, "'_checkindex(neg index)'": 5111, "'_checkindex(pos index)'": 5112, "test10": 5113, "checkindex": 5114, "test_12_arithmetic": 5115, "'arithmetic'": 5116, "al": 5117, "'add'": 5118, "'type of add result'": 5119, "'radd'": 5120, "'type of radd result'": 5121, "objid": 5122, "'in-place add'": 5123, "'in-place add id'": 5124, "arithmetic": 5125, "popfcn": 5126, "indexfcn": 5127, "test_reraise": 5128, "exc1": 5129, "exc2": 5130, "fail": 5131, "\"no exception raised\"": 5132, "test_finally_reraise": 5133, "\"caught\"": 5134, "test_nested_reraise": 5135, "nested_reraise": 5136, "nested": 5137, "test_with_reraise1": 5138, "reraise1": 5139, "test_yield_reraise": 5140, "stopiteration": 5141, "test_new_returns_invalid_instance": 5142, "11627.": 5143, "myexception": 5144, "__new__": 5145, "test_c_exception_context": 5146, "assertisinstance": 5147, "__context__": 5148, "zerodivisionerror": 5149, "test_3118": 5150, "caused": 5151, "cleared": 5152, "3118": 5153, "test_tuples": 5154, "\" return a dict or list of dict objects. return none to discard \"": 5155, "'new_key'": 5156, "'new_value'": 5157, "x_dump_line": 5158, "alist": 5159, "stride": 5160, "dpos": 5161, "\"%5d%s\"": 5162, "\" =\"": 5163, "fprintf": 5164, "logfile": 5165, "\"did=%d name=%r etype=%d dids(left=%d right=%d root=%d parent=%d kids=%r) first_sid=%d tot_size=%d\\n\"": 5166, "etype": 5167, "left_did": 5168, "right_did": 5169, "root_did": 5170, "first_sid": 5171, "tot_size": 5172, "cre_lo": 5173, "cre_hi": 5174, "mod_lo": 5175, "mod_hi": 5176, "tsinfo": 5177, "\"timestamp info\"": 5178, "get_named_stream": 5179, "qname": 5180, "_dir_search": 5181, "\"/\"": 5182, "min_size_std_stream": 5183, "_get_stream": 5184, "mem": 5185, "sat": 5186, "sec_size": 5187, "seen_id": 5188, "sscs": 5189, "ssat": 5190, "short_sec_size": 5191, "\" (from sscs)\"": 5192, "test_2": 5193, "appendpath": 5194, "'prefix/sub'": 5195, "'prefix'": 5196, "'sub'": 5197, "'sup/name'": 5198, "'sup'": 5199, "'/prefix/sub'": 5200, "'/prefix/name'": 5201, "ajoin": 5202, "test_1": 5203, "minrelpath": 5204, "'aa/bb'": 5205, "'..'": 5206, "'aa/..'": 5207, "'aa/../bb'": 5208, "'bb'": 5209, "'aa/bb/..'": 5210, "'aa/bb/../..'": 5211, "'aa/bb/../cc/../dd'": 5212, "'aa/dd'": 5213, "'.././..'": 5214, "'../..'": 5215, "'aa/bb/.././../dd'": 5216, "'dd'": 5217, "test_gpaths": 5218, "local_path": 5219, "ls": 5220, "gpaths": 5221, "'command/*.py'": 5222, "'command'": 5223, "'build_src.py'": 5224, "'system_info.py'": 5225, "test_get_shared_lib_extension": 5226, "get_shared_lib_extension": 5227, "is_python_ext": 5228, "'linux'": 5229, "'.so'": 5230, "'gnukfreebsd'": 5231, "'darwin'": 5232, "'.dylib'": 5233, "'win'": 5234, "'.dll'": 5235, "crash": 5236, "extension": 5237, "symmetric": 5238, "reconstruction": 5239, "tensor_like": 5240, "theano": 5241, "symbolic": 5242, "minibatch": 5243, "assumed": 5244, "indexing": 5245, "dimensions": 5246, "seconde": 5247, "applied": 5248, "optionally": 5249, "represents": 5250, "intractable": 5251, "optimized": 5252, "get_gradients": 5253, "pylearn2": 5254, "get_data_specs": 5255, "test_not_handled": 5256, "urllib2": 5257, "httperror": 5258, "cm": 5259, "resp": 5260, "\"/not_existing\"": 5261, "assertequals": 5262, "handled": 5263, "test_get_kumoscale": 5264, "mock_get_info": 5265, "success_prov_response": 5266, "_get_kumoscale": 5267, "'https://1.2.3.4:8090'": 5268, "'token'": 5269, "'cert'": 5270, "mgmt_ips": 5271, "'1.2.3.4'": 5272, "'8090'": 5273, "token": 5274, "kumoscale": 5275, "test_delete_volume_success": 5276, "mock_delete_volume": 5277, "testvol": 5278, "_stub_volume": 5279, "delete_volume": 5280, "assert_any_call": 5281, "assertisnone": 5282, "success": 5283, "test_delete_volume_exception": 5284, "volumebackendapiexception": 5285, "test_initialize_connection_host_probe_failure": 5286, "mock_host_probe": 5287, "mock_publish": 5288, "mock_get_volumes_by_uuid": 5289, "mock_get_targets": 5290, "mock_get_backend_by_id": 5291, "testconn": 5292, "_stub_connector": 5293, "prov_target": 5294, "targetentity": 5295, "'target.nqn'": 5296, "prov_backend1": 5297, "prov_portal": 5298, "portalentity": 5299, "4420": 5300, "'tcp'": 5301, "backendentity": 5302, "failure": 5303, "test_initialize_connection_volumes_failure": 5304, "test_initialize_connection_volumes_exception": 5305, "pro": 5306, "test_initialize_connection_backend_failure": 5307, "prov_targets_response": 5308, "provisionerresponse": 5309, "return_v": 5310, "test_initialize_connection_no_backend": 5311, "success_prov_respo": 5312, "test_get_volume_stats_tenants_failure": 5313, "mock_get_tenants": 5314, "fail_prov_response": 5315, "expected_stats": 5316, "'total_capacity_gb'": 5317, "'unknown'": 5318, "'free_capacity_gb'": 5319, "assertdictequal": 5320, "get_volume_stats": 5321, "test_create_snapshot_success": 5322, "mock_create_snapshot": 5323, "testsnap": 5324, "_stub_snapshot": 5325, "create_snapshot": 5326, "call_args": 5327, "mock_call": 5328, "volumeid": 5329, "'volume_id'": 5330, "snapshotid": 5331, "snapshot": 5332, "test_delete_snapshot_success": 5333, "mock_delete_snapshot": 5334, "delete_snapshot": 5335, "test_delete_snapshot_exception": 5336, "test_create_cloned_volume_exception": 5337, "mock_clone_volume": 5338, "create_cloned_volume": 5339, "cloned": 5340, "test_convert_host_name": 5341, "'ks-node3-000c2960a794-000c2960a797'": 5342, "_convert_host_name": 5343, "md5": 5344, "usedforsecurity": 5345, "test_ensure_export": 5346, "ensure_export": 5347, "get_bader_volumes": 5348, "'return volumes in ang**3'": 5349, "bohr": 5350, "ang": 5351, "bader": 5352, "write_atom_volume": 5353, "atomlist": 5354, "'write bader atom volumes to cube files.\n atomlist = [0,2] #for example\n\n -p sel_atom write the selected atomic volumes, read from the\n subsequent list of volumes.\n '": 5355, "'bader -p sel_atom %s %s'": 5356, "densityfile": 5357, "cube": 5358, "sel_atom": 5359, "write_bader_volume": 5360, "\"write bader atom volumes to cube files.\n\n ::\n \n atomlist = [0,2] # for example\n \n -p sel_bader write the selected bader volumes, read from the\n subsequent list of volumes.\n \"": 5361, "'bader -p sel_bader %s %s'": 5362, "sel_bader": 5363, "write_atom_index": 5364, "' -p atom_index write the atomic volume index to a charge\n density file.\n '": 5365, "'bader -p atom_index %s'": 5366, "atom_index": 5367, "charge": 5368, "density": 5369, "write_bader_index": 5370, "'\n -p bader_index write the bader volume index to a charge\n density file.\n '": 5371, "'bader -p bader_index %s'": 5372, "bader_index": 5373, "write_all_atom": 5374, "'\n -p all_atom combine all volumes associated with an atom and\n write to file. this is done for all atoms and written to files\n named bvatxxxx.dat. the volumes associated with atoms are\n those for which the maximum in charge density within the\n volume is closest to the atom.\n '": 5375, "'bader -p all_atom %s'": 5376, "all_atom": 5377, "atoms": 5378, "bvatxxxx": 5379, "dat": 5380, "closest": 5381, "write_all_bader": 5382, "'\n -p all_bader write all bader volumes (containing charge above\n threshold of 0.0001) to a file. the charge distribution in\n each volume is written to a separate file, named\n bvolxxxx.dat. it will either be of a chgcar format or a cube\n file format, depending on the format of the initial charge\n density file. these files can be quite large, so this option\n should be used with caution.\n '": 5383, "'bader -p all_bader %s'": 5384, "all_bader": 5385, "threshold": 5386, "0.0001": 5387, "bvolxxxx": 5388, "chgcar": 5389, "fi": 5390, "staff_grading_service": 5391, "\"\n return a staff grading service instance--if settings.mock_staff_grading is true,\n returns a mock one, otherwise a real one.\n\n caches the result, so changing the setting after the first call to this\n function will have no effect.\n \"": 5392, "_service": 5393, "mock_staff_grading": 5394, "mockstaffgradingservice": 5395, "staffgradingservice": 5396, "open_ended_grading_interface": 5397, "staff": 5398, "grading": 5399, "changing": 5400, "setting": 5401, "get_next": 5402, "thing": 5403, "grade": 5404, "'success'": 5405, "'submission_id'": 5406, "submission": 5407, "back": 5408, "'submission'": 5409, "'rubric'": 5410, "rubric": 5411, "'message'": 5412, "went": 5413, "rendere": 5414, "get_problem_list": 5415, "problems": 5416, "problem_list": 5417, "dicts": 5418, "problem_name": 5419, "num_graded": 5420, "graded": 5421, "num_pending": 5422, "sitting": 5423, "queue": 5424, "min_for_ml": 5425, "num": 5426, "save_grade": 5427, "\"\n save the grade and feedback for a submission, and, if all goes well, return\n the next thing to grade.\n\n expects the following post parameters:\n 'score': int\n 'feedback': string\n 'submission_id': int\n\n returns the same thing as get_next, except that additional error messages\n are possible if something goes wrong with saving the grade.\n \"": 5428, "slashseparatedcoursekey": 5429, "from_deprecated_string": 5430, "_check_access": 5431, "'post'": 5432, "http404": 5433, "feedback": 5434, "goes": 5435, "expects": 5436, "'score'": 5437, "'feedback'": 5438, "get_notifications": 5439, "to_deprecated_string": 5440, "get_notifications_url": 5441, "'course_id:{}'": 5442, "'staff_needs_to_grade:{}'": 5443, "'staff_needs_to_grade'": 5444, "_record_result": 5445, "'get_notifications'": 5446, "schedule_create_share": 5447, "request_spec": 5448, "filter_properties": 5449, "\"picks a host that is up at random.\"": 5450, "topic": 5451, "share_topic": 5452, "_schedule": 5453, "share_id": 5454, "'share_id'": 5455, "snapshot_id": 5456, "'snapshot_id'": 5457, "updated_share": 5458, "share_update_db": 5459, "share_rpcapi": 5460, "create_share": 5461, "picks": 5462, "schedule": 5463, "share": 5464, "set_markup": 5465, "_layout": 5466, "set_font_description": 5467, "font_desc": 5468, "set_wrap": 5469, "wrap_mode": 5470, "wrap": 5471, "get_cursor_pos": 5472, "get_extents": 5473, "extents": 5474, "set_allocation": 5475, "allocation": 5476, "reset_attrs": 5477, "~": 5478, "set_attributes": 5479, "pango": 5480, "attrlist": 5481, "get_text": 5482, "_default_attrs": 5483, "is_max": 5484, "get_position": 5485, "this_pos": 5486, "other_pos": 5487, "set_position": 5488, "same_line": 5489, "get_current_line": 5490, "keep_going": 5491, "get_iter": 5492, "get_line": 5493, "start_index": 5494, "le": 5495, "at_last_line": 5496, "next_line": 5497, "get_current_word": 5498, "layout": 5499, "get_index": 5500, "word_terminators": 5501, "next_char": 5502, "word": 5503, "append_paragraph": 5504, "vspacing": 5505, "_new_layout": 5506, "select_all": 5507, "_select_all": 5508, "queue_draw": 5509, "queue_resize": 5510, "\"updates the status of a machine that failed with initialization.\"": 5511, "instance_id": 5512, "getrequiredparameter": 5513, "'instance_id'": 5514, "getoptionalparameter": 5515, "'log'": 5516, "old_instance": 5517, "gqlquery": 5518, "'select * from clientmachine where client_id = :1'": 5519, "logging": 5520, "'the given instance id \"%s\" does not match any machines.'": 5521, "500": 5522, "enum": 5523, "machine_status": 5524, "initializing": 5525, "machine": 5526, "ins": 5527, "updates": 5528, "initialization": 5529, "getorset": 5530, "65533": 5531, "temp": 5532, "'t adjust the first value since that'": 5533, "readatn": 5534, "readint": 5535, "grammartype": 5536, "atntype": 5537, "fromordinal": 5538, "maxtokentype": 5539, "atn": 5540, "readsets": 5541, "readunicode": 5542, "iset": 5543, "intervalset": 5544, "containseof": 5545, "addone": 5546, "i1": 5547, "i2": 5548, "addrange": 5549, "exclusive": 5550, "readedges": 5551, "nedges": 5552, "src": 5553, "trg": 5554, "ttype": 5555, "arg1": 5556, "arg2": 5557, "arg3": 5558, "trans": 5559, "edgefactory": 5560, "srcstate": 5561, "addtransition": 5562, "edges": 5563, "rule": 5564, "derived": 5565, "serialized": 5566, "readdecisions": 5567, "ndecisions": 5568, "decstate": 5569, "decisiontostate": 5570, "decision": 5571, "generaterulebypasstransition": 5572, "bypassstart": 5573, "basicblockstartstate": 5574, "ruleindex": 5575, "addstate": 5576, "bypassstop": 5577, "blockendstate": 5578, "endstate": 5579, "definedecisionstate": 5580, "startstate": 5581, "excludetransition": 5582, "ruletostartstate": 5583, "isprecedencerule": 5584, "beginning": 5585, "starloopentrystate": 5586, "endst": 5587, "readint32": 5588, "readuuid": 5589, "readlong": 5590, "allbits": 5591, "xffffffffffffffff": 5592, "\"new selfcite tables\"": 5593, "estimate": 5594, "'-'": 5595, "year_val": 5596, "month_val": 5597, "day_val": 5598, "year": 5599, "month": 5600, "month_choices": 5601, "months": 5602, "select_html": 5603, "choices": 5604, "month_field": 5605, "day_choices": 5606, "__unicode__": 5607, "\"%s - %s\"": 5608, "paneles": 5609, "modulo": 5610, "get_results": 5611, "\"\n returns a tuple of the value scored, possible, and a list of messages\n in the result set.\n \"": 5612, "out_of": 5613, "scored": 5614, "results_list": 5615, "msgs": 5616, "new_nc_file": 5617, "\"\n make a new temporary netcdf file for the scope of the test\n \"": 5618, "nc_file_path": 5619, "\"example.nc\"": 5620, "\"file exists: %s\"": 5621, "\"w\"": 5622, "temporary": 5623, "netcdf": 5624, "test_coord_data_vars": 5625, "\"check that coordinate data variables are properly handled\"": 5626, "mocktimeseries": 5627, "createdimension": 5628, "\"siglev\"": 5629, "createvariable": 5630, "\"temp\"": 5631, "float64": 5632, "\"time\"": 5633, "fill_value": 5634, "99999999999999999999.0": 5635, "coordinates": 5636, "\"sigma noexist\"": 5637, "\"sigma\"": 5638, "cf": 5639, "nug": 5640, "coordinate": 5641, "sigma": 5642, "referred": 5643, "properly": 5644, "coord": 5645, "vars": 5646, "load_dataset": 5647, "nc_dataset": 5648, "\"\n return a loaded nc dataset for the given path\n \"": 5649, "\"nc_dataset should be a string\"": 5650, "test_check_data_types": 5651, "\"\n invoke check_data_types() and loop through all variables to check data\n types. pertains to 2.2. the netcdf data types char, byte, short, int,\n float or real, and double are all acceptable. netcdf4 allows string as\n data type, which is also acceptable.\n \"": 5652, "static_files": 5653, "\"rutgers\"": 5654, "check_data_types": 5655, "invoke": 5656, "through": 5657, "pertains": 5658, "2.2": 5659, "char": 5660, "double": 5661, "acceptable": 5662, "allows": 5663, "test_check_child_attr_data_types": 5664, "\"\n tests check_child_attr_data_types() to ensure the attributes specified in section 2.5.1\n have a matching data type to their parent variables.\"": 5665, "mockdataset": 5666, "gives": 5667, "good": 5668, "check_child_attr_data_ty": 5669, "check_child_attr_data_types": 5670, "2.5": 5671, "test_naming_conventions": 5672, "\"\n section 2.3 naming conventions\n\n variable, dimension and attr names should begin with a letter and be composed of letters, digits, and underscores.\n \"": 5673, "compliant": 5674, "check_naming_conventions": 5675, "\"bad\"": 5676, "check_naming_convent": 5677, "2.3": 5678, "naming": 5679, "conventions": 5680, "letter": 5681, "composed": 5682, "letters": 5683, "underscores": 5684, "test_check_names_unique": 5685, "\"\n 2.3 names should not be distinguished purely by case, i.e., if case is disregarded, no two names should be the same.\n \"": 5686, "check_names_unique": 5687, "num_var": 5688, "\"chap2\"": 5689, "distinguished": 5690, "disregarded": 5691, "test_check_dimension_names": 5692, "\"\n 2.4 a variable may have any number of dimensions, including zero, and the dimensions must all have different names.\n \"": 5693, "\"bad_data_type\"": 5694, "check_dimension_names": 5695, "\"no_reason has two or more dimensions named time\"": 5696, "2.4": 5697, "including": 5698, "test_check_dimension_order": 5699, "\"\n 2.4 if any or all of the dimensions of a variable have the interpretations of \"": 5700, "\" (t), \"": 5701, "\" (z),\n \"": 5702, "latitude": 5703, "\" (y), or \"": 5704, "longitude": 5705, "\" (x) then we recommend, those dimensions to appear in the relative order t, then z, then y,\n then x in the cdl definition corresponding to the file. all other dimensions should, whenever possible, be placed to the\n left of the spatiotemporal dimensions.\n \"": 5706, "check_d": 5707, "interpretations": 5708, "\"date or time\"": 5709, "\"height or depth\"": 5710, "\"latitude\"": 5711, "Y": 5712, "\"longitude\"": 5713, "recommend": 5714, "appear": 5715, "relative": 5716, "cdl": 5717, "test_check_fill_value_outside_valid_range": 5718, "\"\n 2.5.1 the _fillvalue should be outside the range specified by valid_range (if used) for a variable.\n \"": 5719, "check_fill_value_outside_valid_range": 5720, "\"salinity:_fillvalue (1.0) should be outside the \"": 5721, "\"range specified by valid_min/valid_max (-10, 10)\"": 5722, "check_fill_valu": 5723, "_fillvalue": 5724, "valid_range": 5725, "fill": 5726, "test_check_conventions_are_cf_16": 5727, "\"\n \u00a72.6.1 the nug defined global attribute conventions to the string value\n \"": 5728, "1.6": 5729, "\"cf-1.6\"": 5730, "check_conventions_version": 5731, "\"cf-1.6 ,acdd\"": 5732, "\"conv_multi\"": 5733, "noconv": 5734, "\u00a7": 5735, "2.6": 5736, "test_check_convention_globals": 5737, "\"\n load up a dataset and ensure title and history global attrs are checked\n properly (\u00a72.6.2).\n \"": 5738, "check_convention_globals": 5739, "checked": 5740, "convention": 5741, "globals": 5742, "test_check_convention_possibly_var_attrs": 5743, "\"\n \u00a72.6.2 the units attribute is required for all variables that represent dimensional quantities\n (except for boundary variables defined in section 7.1, \"": 5744, "boundaries": 5745, "\" and climatology variables\n defined in section 7.4, \"": 5746, "climatological": 5747, "dimensionless": 5748, "quantities": 5749, "however": 5750, "specifying": 5751, "represent": 5752, "dimensional": 5753, "boundary": 5754, "7.1": 5755, "\"cell boundaries\"": 5756, "climatology": 5757, "7.4": 5758, "\"climatological statistics\"": 5759, "possibly": 5760, "test_check_standard_name": 5761, "\"\n 3.3 a standard name is associated with a variable via the attribute standard_name which takes a\n string value comprised of a standard name optionally followed by one or more blanks and a\n standard name modifier\n \"": 5762, "\"2dim\"": 5763, "check_standard_name": 5764, "failing": 5765, "standard_name": 5766, "comprised": 5767, "followed": 5768, "blanks": 5769, "modifier": 5770, "test_cell_measures": 5771, "\"cell_measure\"": 5772, "check_cell_measures": 5773, "\"bad_cell_measure1\"": 5774, "\"the cell_measures attribute for variable ps is formatted incorrectly. \"": 5775, "measures": 5776, "test_climatology_cell_methods": 5777, "\"\n checks that climatology cell_methods strings are properly validated\n \"": 5778, "\"climatology\"": 5779, "check_climatological_statistics": 5780, "\"time: mean within days time: mean over days\"": 5781, "temp_var": 5782, "\"temperature\"": 5783, "mockvariable": 5784, "te": 5785, "cell_methods": 5786, "validated": 5787, "test_check_ancillary_variables": 5788, "\"\n test to ensure that ancillary variables are properly checked\n \"": 5789, "check_ancillary_variables": 5790, "result_dict": 5791, "\"\u00a73.4 ancillary data\"": 5792, "\"bad_reference\"": 5793, "resul": 5794, "ancillary": 5795, "test_download_standard_name_table": 5796, "\"\n test that a user can download a specific standard name table\n \"": 5797, "\"35\"": 5798, "data_directory": 5799, "create_cached_data_dir": 5800, "\"cf-standard-name-table-test-{0}.xml\"": 5801, "download_cf_standard_name_table": 5802, "isfile": 5803, "std_names": 5804, "standardnametable": 5805, "test_bad_standard_name_table": 5806, "\"\n test that failure in case a bad standard name table is passed.\n \"": 5807, "ever": 5808, "reached": 5809, "\"dummy_non_existent_file.ext\"": 5810, "nc_obj": 5811, "standard_name_table": 5812, "_find_cf_standard_name_table": 5813, "test_check_flags": 5814, "\"test that the check for flags works as expected.\"": 5815, "check_flags": 5816, "perfect": 5817, "scores": 5818, "imperfect": 5819, "flags": 5820, "test_check_flag_masks": 5821, "\"ghrsst\"": 5822, "masks": 5823, "test_check_bad_units": 5824, "\"load a dataset with units that are expected to fail (bad_units.nc).\n there are 6 variables in this dataset, three of which should give\n an error:\n - time, with units \"": 5825, "\" (should be since )\n - lat, with units \"": 5826, "degrees_e": 5827, "\" (should be degrees)\n - lev, with units \"": 5828, "\" (deprecated)\"": 5829, "check_units": 5830, "assert_result_is_good": 5831, "bad_units": 5832, "three": 5833, "\"s\"": 5834, "epoch": 5835, "\"degrees_e\"": 5836, "degrees": 5837, "lev": 5838, "\"level\"": 5839, "deprecated": 5840, "test_latitude": 5841, "\"\n section 4.1 latitude coordinate\n \"": 5842, "compliance": 5843, "\"example-grid\"": 5844, "check_latitude": 5845, "4.1": 5846, "test_longitude": 5847, "\"\n section 4.2 longitude coordinate\n \"": 5848, "check_longitude": 5849, "4.2": 5850, "test_is_vertical_coordinate": 5851, "\"\n section 4.3 qualifiers for vertical coordinate\n\n note: the standard doesn't explicitly say that vertical coordinates must be a\n coordinate type.\n \"": 5852, "attach": 5853, "mock_variable": 5854, "proper": 5855, "known_name": 5856, "\"depth\"": 5857, "is_vertical_coordinate": 5858, "\"not_known\"": 5859, "axis": 5860, "axis_set": 5861, "\"z\"": 5862, "4.3": 5863, "qualifiers": 5864, "vertical": 5865, "explicitly": 5866, "say": 5867, "test_vertical_dimension": 5868, "\"\n section 4.3.1 dimensional vertical coordinate\n \"": 5869, "check_dimensional_vertical_coordinate": 5870, "\"\u00a74.3 vertical coordinate\"": 5871, "\"illegal-vertical\"": 5872, "test_appendix_d": 5873, "appendix": 5874, "D": 5875, "definitions": 5876, "ones": 5877, "formulas": 5878, "gridpoint": 5879, "terms": 5880, "formula_t": 5881, "test_dimensionless_vertical": 5882, "\"\n section 4.3.2\n \"": 5883, "affirmative": 5884, "\"dimensionless\"": 5885, "check_dimensionless_vertical_coordinates": 5886, "negative": 5887, "load_datase": 5888, "test_is_time_variable": 5889, "var1": 5890, "is_time_variable": 5891, "\"not_time\"": 5892, "var2": 5893, "var3": 5894, "\"t\"": 5895, "\"maybe_time\"": 5896, "var4": 5897, "\"seconds since 1900-01-01\"": 5898, "test_dimensionless_standard_names": 5899, "\"check that dimensionless standard names are properly detected\"": 5900, "std_names_xml_root": 5901, "_std_names": 5902, "_root": 5903, "canonical_units": 5904, "cfutil": 5905, "is_dimensionless_standard_name": 5906, "\"sea_water_temperature\"": 5907, "\"sea_water_practical_salinity\"": 5908, "detected": 5909, "test_check_calendar": 5910, "\"load a dataset with an invalid calendar attribute (non-comp/bad.nc).\n this dataset has a variable, \"": 5911, "\" with calendar attribute \"": 5912, "nope": 5913, "\".\"": 5914, "check_calendar": 5915, "4.4": 5916, "calendar": 5917, "comp": 5918, "\"nope\"": 5919, "test_check_two_dimensional": 5920, "check_grid_coordinates": 5921, "\"bad2dim\"": 5922, "5.6": 5923, "horizo": 5924, "test_is_geophysical": 5925, "cf_role": 5926, "processed": 5927, "file_format": 5928, "\"netcdf4\"": 5929, "\"netcdf file of wrong format (not netcdf4) was created for checking\"": 5930, "is_geophysical": 5931, "\"j\"": 5932, "\"test probably fails because var.dtype.kind or var.dtype.char \"": 5933, "geophysical": 5934, "test_check_attr_type": 5935, "\"\n check that the check_attr_type method checks\n grid_mapping attribute types correctly.\n \"": 5936, "att_name": 5937, "\"test_att\"": 5938, "att": 5939, "45": 5940, "att_type": 5941, "\"n\"": 5942, "numeric": 5943, "_check_attr_type": 5944, "check_attr_type": 5945, "grid_mapping": 5946, "correctly": 5947, "test_check_grid_mapping_attr_condition": 5948, "\"\n ensure the check_grid_mapping_attr_condition() method works as expected.\n \"": 5949, "attr_name": 5950, "\"latitude_of_projection_origin\"": 5951, "_check_grid_mapping_attr_condition": 5952, "\"longitude_of_projection_origin\"": 5953, "\"longitude_of_prime_meridian\"": 5954, "check_grid_mapping_attr_condition": 5955, "grid": 5956, "test_compress_packed": 5957, "\"tests compressed indexed coordinates\"": 5958, "\"reduced_horizontal_grid\"": 5959, "check_compression_gathering": 5960, "compressed": 5961, "indexed": 5962, "compress": 5963, "packed": 5964, "test_featuretype_is_case_insensitive": 5965, "\"\n tests that the featuretype attribute is case insensitive\n \"": 5966, "featuretype": 5967, "\"timeseriesprofile\"": 5968, "check_feature_type": 5969, "\"trajectoryprofile\"": 5970, "test_check_units": 5971, "\"\n ensure that container variables are not checked for units but geophysical variables are\n \"": 5972, "\"units_check\"": 5973, "'t keep track of the variables names for checks that passed, so\n # we can make a strict assertion about how many checks were performed\n # and if there were errors, which there shouldn'": 5974, "fixme": 5975, "badams": 5976, "grouping": 5977, "test_check_duplicates": 5978, "\"\n test to verify that the check identifies duplicate axes. load the\n duplicate_axis.nc dataset and verify the duplicate axes are accounted\n for.\n \"": 5979, "\"duplicate_axis\"": 5980, "check_duplicate_axis": 5981, "'temp'": 5982, "lon_rho": 5983, "lo": 5984, "axes": 5985, "duplicate_axis": 5986, "accounted": 5987, "test_check_multi_dimensional_coords": 5988, "\"\n test to verify that multi dimensional coordinates are checked for\n sharing names with dimensions\n \"": 5989, "\"multi-dim-coordinates\"": 5990, "check_multi_dimensional_coords": 5991, "multi": 5992, "sharing": 5993, "coords": 5994, "test_variable_feature_check": 5995, "\"bad-trajectory\"": 5996, "check_variable_features": 5997, "\"\u00a79.1 features and feature types\"": 5998, "\"trajectory-complete\"": 5999, "feature": 6000, "test_check_cell_methods": 6001, "\"load a dataset (climatology.nc) and check the cell methods.\n this dataset has variable \"": 6002, "temperature": 6003, "\" which has valid cell_methods\n format, cell_methods attribute, and valid names within the\n cell_methods attribute.\"": 6004, "check_cell_methods": 6005, "itertools": 6006, "chain": 6007, "test_temporal_unit_conversion": 6008, "units_convertible": 6009, "\"hours\"": 6010, "\"seconds\"": 6011, "\"hours since 2000-01-01\"": 6012, "temporal": 6013, "test_check_actual_range": 6014, "\"test the check_actual_range method works as expected\"": 6015, "operating": 6016, "manually": 6017, "dims": 6018, "actual_range": 6019, "check_actual_range": 6020, "test_check_cell_boundaries": 6021, "\"check our over-ridden check_cell_boundaries emthod behaves as expected\"": 6022, "\"grid-boundaries\"": 6023, "check_cell_boundaries": 6024, "\"cf_example_cell_measures\"": 6025, "check_cell_boundarie": 6026, "ridden": 6027, "emthod": 6028, "behaves": 6029, "test_check_grid_mapping_crs_wkt": 6030, "\"mapping\"": 6031, "valid_crs_check": 6032, "\"wgs84\"": 6033, "crs_wkt": 6034, "check_grid_mapping": 6035, "assertin": 6036, "\"crs_wkt attribute must be a string\"": 6037, "ogc": 6038, "crs": 6039, "wkt": 6040, "\"epsg:3785\"": 6041, "check_": 6042, "test_check_grid_mapping_coordinates": 6043, "\"\n checks that coordinates variables referred to by a grid mapping\n are well-formed and exist.\n \"": 6044, "\"grid_mapping_coordinates\"": 6045, "valid_grid_mapping": 6046, "valid_grid_mapping_2": 6047, "malformed_sep": 6048, "formed": 6049, "test_check_grid_mapping_vert_datum_geoid_name": 6050, "\"checks that geoid_name works proerly\"": 6051, "geoid_name": 6052, "\"navd88\"": 6053, "geopotential_datum_name": 6054, "geoid_name_good": 6055, "geopotential_datum_name_bad": 6056, "proerly": 6057, "datum": 6058, "geoid": 6059, "test_check_conventions_are_cf_1_7": 6060, "\"ensure the check_conventions_are_cf_1_7() check works as expected\"": 6061, "setncattr": 6062, "\"conventions\"": 6063, "\"cf-1.9999\"": 6064, "se": 6065, "check_conventions_are_cf_1_7": 6066, "test_check_dimensionless_vertical_coordinate_1_7": 6067, "\"\n unit test for _check_dimensionless_vertical_coordinate_1_7 method.\n \"": 6068, "deprecated_units": 6069, "\"layer\"": 6070, "\"sigma_level\"": 6071, "ret_val": 6072, "\"lev\"": 6073, "\"standard_name\"": 6074, "\"atmosphere_sigma_coordinate\"": 6075, "_check_dimensionless_vertical_coordinate_1_7": 6076, "test_check_gmattr_existence_condition_ell_pmerid_hdatum": 6077, "\"reference_ellipsoid_name\"": 6078, "\"blah\"": 6079, "\"prime_meridian_name\"": 6080, "\"horizontal_datum_name\"": 6081, "_check_gmattr_existence_condition_ell_pmerid_hdatum": 6082, "gmattr": 6083, "existence": 6084, "ell": 6085, "pmerid": 6086, "hdatum": 6087, "test_update_concept_with_invalid_record": 6088, "concept": 6089, "create_concept": 6090, "mnemonic": 6091, "user1": 6092, "source1": 6093, "testfile": 6094, "'./integration_tests/fixtures/concept_without_fully_specified_name.json'": 6095, "'rb'": 6096, "stderr_stub": 6097, "teststream": 6098, "create_source": 6099, "validation_schema": 6100, "custom_validation_schema_openmrs": 6101, "importer": 6102, "conceptsimporter": 6103, "save_validation_errors": 6104, "import_concepts": 6105, "total": 6106, "openmr": 6107, "record": 6108, "test_import_concepts_into_openmrs_validated_source_with_valid_records": 6109, "test_file": 6110, "'./integration_tests/fixtures/concepts_for_openmrs_validation.json'": 6111, "openmrs_must_have_exactly_one_preferred_name": 6112, "getvalue": 6113, "concepts": 6114, "openmrs": 6115, "test_import_job_for_one_invalid_record": 6116, "stdout_stub": 6117, "invalid_json_file": 6118, "'./integration_tests/fixtures/one_invalid_mapping.json'": 6119, "mappingsimporter": 6120, "import_mappings": 6121, "'cannot map concept to itself.'": 6122, "job": 6123, "test_import_job_for_change_in_data": 6124, "map_type": 6125, "'same-as'": 6126, "from_concept": 6127, "concept3": 6128, "to_source": 6129, "to_concept_code": 6130, "'413532003'": 6131, "external_id": 6132, "'junk'": 6133, "'parent_resource'": 6134, "persist_new": 6135, "source_version": 6136, "sourceversion": 6137, "get_latest_version_of": 6138, "testreadfirstidsfromfile": 6139, "test_resource_ids": 6140, "'testdata'": 6141, "'resource_ids'": 6142, "base_dir": 6143, "src_dir": 6144, "id_dict": 6145, "misc": 6146, "_readfirstidsfromfile": 6147, "'shared_intermediate_dir'": 6148, "'out/release/obj/gen'": 6149, "'bar/file.grd'": 6150, "testelsiness": 6151, "grd": 6152, "parsegrdforunittest": 6153, "\"true\"": 6154, "\"ids_yes1\"": 6155, "\"ids_no1\"": 6156, "\"ids_yes2\"": 6157, "\"ids_no2\"": 6158, "testifbadchildrennesting": 6159, "includes": 6160, "xml": 6161, "stringio": 6162, "'\n \n \n \n \n \n \n \n \n '": 6165, "grit": 6166, "unexpectedchild": 6167, "grd_reader": 6168, "test_process_404": 6169, "\"a 404 should not trigger anything\"": 6170, "requestfactory": 6171, "\"dummy_url\"": 6172, "redirectmiddleware": 6173, "process_exception": 6174, "test_redirect_exceptions": 6175, "\"\n unit tests for handling of redirect exceptions.\n \"": 6176, "test_url": 6177, "'/test_url'": 6178, "redirect": 6179, "302": 6180, "target_url": 6181, "_headers": 6182, "'location'": 6183, "auto_commit": 6184, "raise_exception": 6185, "\" create communication for partner, if not already existing.\n \"": 6186, "comm_obj": 6187, "'partner.communication.job'": 6188, "with_context": 6189, "no_print": 6190, "'partner_communication.default_communication'": 6191, "filtered": 6192, "mail_message_id": 6193, "communication": 6194, "'email_id'": 6195, "communicatio": 6196, "partner": 6197, "exe_name": 6198, "'*/sw4lite'": 6199, "install_tree": 6200, "'tests'": 6201, "create_config": 6202, "'test_app'": 6203, "imgur_client_id": 6204, "test_api_init": 6205, "'the default interface will only initialize the app name and payload.'": 6206, "apiprovider": 6207, "test_api_issue_payload": 6208, "'\n if the payload is related to an issue (or an issue comment in an issue/pr),\n then this should'": 6209, "commonly": 6210, "related": 6211, "stuff": 6212, "'\n\n payload = {\n '": 6213, "': {\n '": 6214, "': {\n '": 6215, "': '": 6216, "foobar": 6217, "'\n },\n '": 6218, "',\n '": 6219, "': [\n { '": 6220, "' },\n { '": 6221, "' }\n ],\n '": 6222, "': 200,\n '": 6223, "updated_at": 6224, "1970": 6225, "01": 6226, "t00": 6227, "comment": 6228, "pr": 6229, "test_api_pr_payload": 6230, "'\n if the payload is related to a pr, then the commonly used pr attributes\n should'": 6231, "pull_request": 6232, "assignee": 6233, "baz": 6234, "': 50,\n '": 6235, "confi": 6236, "test_api_other_events": 6237, "'test for payload belonging to other events such as comment, label, etc.'": 6238, "hypothetical": 6239, "'sender'": 6240, "'login'": 6241, "'someone'": 6242, "'label'": 6243, "'repository'": 6244, "'comment'": 6245, "'body'": 6246, "'hello, world!'": 6247, "'issue'": 6248, "belonging": 6249, "events": 6250, "test_api_imgur_upload": 6251, "'test imgur api upload'": 6252, "post_image_to_imgur": 6253, "'some data'": 6254, "'foobar'": 6255, "test_valid_request": 6256, "headers": 6257, "'authorization'": 6258, "'client-id foobar'": 6259, "'https://api.imgur.com/3/image'": 6260, "imgur": 6261, "upload": 6262, "test_contributors_update": 6263, "'\n contributors list (cache) live only for an hour (by default). once it'": 6264, "outdated": 6265, "get_contributors": 6266, "calls": 6267, "fetch_contributors": 6268, "writes": 6269, "contributors": 6270, "testapi": 6271, "fetched": 6272, "live": 6273, "get_group_visibility_label": 6274, "\"\n returns the group visibility label shown for the thread.\n \"": 6275, "_get_element_text": 6276, "\".group-visibility-label\"": 6277, "visibility": 6278, "shown": 6279, "get_response_total_text": 6280, "\"returns the response count text, or none if not present\"": 6281, "\".response-count\"": 6282, "get_num_displayed_responses": 6283, "\"returns the number of responses actually rendered\"": 6284, "_find_within": 6285, "\".discussion-response\"": 6286, "displayed": 6287, "get_shown_responses_text": 6288, "\"returns the shown response count text, or none if not present\"": 6289, "\".response-display-count\"": 6290, "get_load_responses_button_text": 6291, "\"returns the load more responses button text, or none if not present\"": 6292, "\".load-response-button\"": 6293, "load_more_responses": 6294, "\"clicks the load more responses button and waits for responses to load\"": 6295, "emptypromise": 6296, "is_ajax_finished": 6297, "\"loading more responses\"": 6298, "fulfill": 6299, "clicks": 6300, "waits": 6301, "has_add_response_button": 6302, "\"returns true if the add response button is visible, false otherwise\"": 6303, "_is_element_visible": 6304, "\".add-response-btn\"": 6305, "visible": 6306, "click_add_response_button": 6307, "\"\n clicks the add response button and ensures that the response text\n field receives focus\n \"": 6308, "\".discussion-reply-new textarea:focus\"": 6309, "\"response field received focus\"": 6310, "ensures": 6311, "receives": 6312, "is_response_editor_visible": 6313, "response_id": 6314, "\"returns true if the response editor is present, false otherwise\"": 6315, "\".response_{} .edit-post-body\"": 6316, "editor": 6317, "is_response_visible": 6318, "comment_id": 6319, "\"returns true if the response is viewable onscreen\"": 6320, "\".response_{} .response-body\"": 6321, "viewable": 6322, "onscreen": 6323, "is_response_editable": 6324, "\"returns true if the edit response button is present, false otherwise\"": 6325, "_secondary_action_menu_open": 6326, "\".response_{} .discussion-response\"": 6327, "\".response_{} .discussion-response .action-edit\"": 6328, "edit": 6329, "editable": 6330, "start_response_edit": 6331, "\"click the edit button for the response, loading the editing view\"": 6332, "\"response edit started\"": 6333, "loading": 6334, "editing": 6335, "get_link_href": 6336, "\"extracts href attribute of the referenced link\"": 6337, "link_href": 6338, "\".post-body p a\"": 6339, "'href'": 6340, "extracts": 6341, "referenced": 6342, "is_response_endorsed": 6343, "\"endorsed\"": 6344, "\".response_{} .discussion-response .posted-details\"": 6345, "endorsed": 6346, "endorse_response": 6347, "\".response_{} .discussion-response .action-endorse\"": 6348, "wait_for_ajax": 6349, "endorse": 6350, "set_response_editor_value": 6351, "new_body": 6352, "\"replace the contents of the response editor\"": 6353, "\".response_{} .discussion-response .wmd-input\"": 6354, "submit_response_edit": 6355, "new_response_body": 6356, "\"click the submit button on the response editor\"": 6357, "\".response_{} .discussion-response .post-update\"": 6358, "get_response_body": 6359, "\"comment edit succeeded\"": 6360, "submit": 6361, "is_show_comments_visible": 6362, "\"returns true if the \"": 6363, "comments": 6364, "\" link is visible for a response\"": 6365, "\".response_{} .action-show-comments\"": 6366, "\"show comments\"": 6367, "show_comments": 6368, "\"click the \"": 6369, "\" link for a response\"": 6370, "\".response_{} .comments\"": 6371, "\"comments shown\"": 6372, "is_add_comment_visible": 6373, "\" form is visible for a response\"": 6374, "\"#wmd-input-comment-body-{}\"": 6375, "\"add comment\"": 6376, "is_comment_visible": 6377, "\"returns true if the comment is viewable onscreen\"": 6378, "\"#comment_{} .response-body\"": 6379, "is_comment_deletable": 6380, "\"returns true if the delete comment button is present, false otherwise\"": 6381, "\"#comment_{}\"": 6382, "\"#comment_{} .action-delete\"": 6383, "deletable": 6384, "is_comment_editable": 6385, "\"returns true if the edit comment button is present, false otherwise\"": 6386, "\"#comment_{} .action-edit\"": 6387, "is_comment_editor_visible": 6388, "\"returns true if the comment editor is present, false otherwise\"": 6389, "\".edit-comment-body[data-id='{}']\"": 6390, "start_comment_edit": 6391, "\"click the edit button for the comment, loading the editing view\"": 6392, "old_body": 6393, "get_comment_body": 6394, "_get_comment_editor_value": 6395, "set_comment_editor_value": 6396, "\"replace the contents of the comment editor\"": 6397, "\"#comment_{} .wmd-input\"": 6398, "submit_comment_edit": 6399, "new_comment_body": 6400, "\"click the submit button on the comment editor\"": 6401, "\"#comment_{} .post-update\"": 6402, "cancel_comment_edit": 6403, "original_body": 6404, "\"click the cancel button on the comment editor\"": 6405, "\"#comment_{} .post-cancel\"": 6406, "\"comment edit was canceled\"": 6407, "is_browser_on_page": 6408, "\"\n return true if the browser is on the right page else false.\n \"": 6409, "css": 6410, "\"body.discussion .forum-nav-sort-control\"": 6411, "get_selected_sort_preference": 6412, "\"\n return the text of option that is selected for sorting.\n \"": 6413, "\"body.discussion .forum-nav-sort-control option\"": 6414, "is_selected": 6415, "get_attribute": 6416, "preference": 6417, "change_sort_preference": 6418, "sort_by": 6419, "\"\n change the option of sorting by clicking on new option.\n \"": 6420, "\"body.discussion .forum-nav-sort-control option[value='{0}']\"": 6421, "clicking": 6422, "refresh_page": 6423, "\"\n reload the page.\n \"": 6424, "reload": 6425, "is_window_on_top": 6426, "\"\n check if window's scroll is at top\n \"": 6427, "execute_script": 6428, "\"return $('html, body').offset().top\"": 6429, "scroll": 6430, "click_and_open_thread": 6431, "thread_id": 6432, "\"\n click specific thread on the list.\n \"": 6433, "thread_selector": 6434, "\"li[data-id='{}']\"": 6435, "_thread_is_rendered_successfully": 6436, "\"thread has been rendered\"": 6437, "check_threads_rendered_successfully": 6438, "thread_count": 6439, "\"\n count the number of threads available on page.\n \"": 6440, "\".forum-nav-thread\"": 6441, "successfully": 6442, "check_window_is_on_top": 6443, "\"\n check window is on top of the page\n \"": 6444, "\"window is on top\"": 6445, "is_discussion_expanded": 6446, "\".discussion\"": 6447, "discussion": 6448, "expanded": 6449, "expand_discussion": 6450, "\"click the link to expand the discussion\"": 6451, "\".discussion-show\"": 6452, "\"discussion expanded\"": 6453, "expand": 6454, "has_thread": 6455, "\"returns true if this page is showing the thread with the specified id.\"": 6456, "'.discussion-thread#thread_{}'": 6457, "showing": 6458, "element_exists": 6459, "selector": 6460, "_discussion_selector": 6461, "\"clicks the link to expand the thread\"": 6462, "\".forum-thread-expand\"": 6463, "\"thread expanded\"": 6464, "check_if_selector_is_focused": 6465, "\"\n check if selector is focused\n \"": 6466, "\"return $('{}').is(':focus')\"": 6467, "focused": 6468, "get_current_page": 6469, "check_func": 6470, "current_page": 6471, "\"nav.discussion-paginator li.current-page\"": 6472, "promise": 6473, "'discussion-paginator current page has text'": 6474, "dismiss_alert_message": 6475, "\"\n dismiss any search alert message containing the specified text.\n \"": 6476, "_match_messages": 6477, "\".search-alert\"": 6478, "elem": 6479, "alert_id": 6480, "\"id\"": 6481, "\"{}#{} a.dismiss\"": 6482, "alert_selector": 6483, "\"waiting for dismissed alerts to disappear\"": 6484, "dismiss": 6485, "click_new_post_button": 6486, "\"\n clicks the 'new post' button.\n \"": 6487, "new_post_button": 6488, "new_post_form": 6489, "\"new post action succeeded\"": 6490, "'new post'": 6491, "\"\n returns the new post button.\n \"": 6492, "\"ol.course-tabs .new-post-btn\"": 6493, "\"\n returns the new post form.\n \"": 6494, "\".forum-new-post-form\"": 6495, "get_cols": 6496, "cols": 6497, "col": 6498, "'get_cols'": 6499, "as_sql": 6500, "qn": 6501, "prepare_node": 6502, "allow_joins": 6503, "'prepare'": 6504, "test_1_b": 6505, "_fields_": 6506, "finalized": 6507, "\"_fields_\"": 6508, "test_context": 6509, "\"\n verify learner profile page context data.\n \"": 6510, "'/url'": 6511, "learner_profile_context": 6512, "is_staff": 6513, "'default_public_account_fields'": 6514, "account_visibility_configuration": 6515, "'public_fields'": 6516, "'accounts_api_url'": 6517, "\"accounts_api\"": 6518, "'username'": 6519, "learner": 6520, "test_view": 6521, "\"\n verify learner profile page view.\n \"": 6522, "profile_path": 6523, "'learner_profile'": 6524, "context_data": 6525, "test_undefined_profile_page": 6526, "\"\n verify that a 404 is returned for a non-existent profile page.\n \"": 6527, "\"no_such_user\"": 6528, "existent": 6529, "applystealshortcut": 6530, "clearshortcut": 6531, "ismodifierlessallowed": 6532, "setclearbuttonsshown": 6533, "setshortcut": 6534, "test_set_vm_state_and_notify": 6535, "mock_save": 6536, "mock_add": 6537, "mock_get": 6538, "expected_uuid": 6539, "'fake-uuid'": 6540, "instance_properties": 6541, "'other-uuid'": 6542, "vm_state": 6543, "'fake-vm-state'": 6544, "'fake-service'": 6545, "'fake-method'": 6546, "'exc_info'": 6547, "'instance_properties'": 6548, "vm": 6549, "test_populate_filter_props_no_retry": 6550, "_test_populate_filter_props": 6551, "with_retry": 6552, "retry": 6553, "test_setup_instance_group_with_no_group": 6554, "mock_ggd": 6555, "'uuid'": 6556, "filter_props": 6557, "'group_hosts'": 6558, "'hostc'": 6559, "scheduler_utils": 6560, "setup_instance_group": 6561, "assert_called_once_with": 6562, "assertnotin": 6563, "'group_updated'": 6564, "'group_policies'": 6565, "test_setup_instance_group_with_filter_not_configured": 6566, "novalidhost": 6567, "reason": 6568, "'whatever'": 6569, "configured": 6570, "smartobject": 6571, "\" sets the current object data, schema information and initializes\n the attribute box and wizard buttons.\n \"": 6572, "schemainfo": 6573, "objectclassattributeinfo": 6574, "getservermeta": 6575, "processdata": 6576, "initattributebox": 6577, "currentpagewidget": 6578, "initializes": 6579, "box": 6580, "buttons": 6581, "\" compute all attributes which can be added according to the data of\n the object. single values which are already given are sorted out.\n \"": 6582, "possiblemust": 6583, "possiblemay": 6584, "getpossibleattributes": 6585, "objectclass": 6586, "usedattributes": 6587, "objectattributes": 6588, "'objectclass'": 6589, "getattributelist": 6590, "singleattributes": 6591, "issingle": 6592, "useda": 6593, "according": 6594, "initclasspage": 6595, "currentpage": 6596, "qwizard": 6597, "finishbutton": 6598, "setdisabled": 6599, "classbox": 6600, "mustattributebox": 6601, "attributebox": 6602, "currenttext": 6603, "mustset": 6604, "mayset": 6605, "getallobjectclassesforattr": 6606, "classlist": 6607, "union": 6608, "hasstructuralclass": 6609, "structlist": 6610, "isstructural": 6611, "assert_dump_recorded": 6612, "\"\n verify that this order does have a dump of information from the\n payment processor.\n \"": 6613, "processor_reply_dump": 6614, "payment": 6615, "processor": 6616, "recorded": 6617, "test_process_payment_raises_exception": 6618, "purchased_callback": 6619, "unused": 6620, "orderitem": 6621, "unit_cost": 6622, "line_cost": 6623, "_signed_callback_params": 6624, "process_postpay_callback": 6625, "test_process_payment_invalid_order": 6626, "\"98272\"": 6627, "expect": 6628, "\"inconsistent data\"": 6629, "'error_html'": 6630, "test_process_user_cancelled": 6631, "decimal": 6632, "\"abcd\"": 6633, "'decision'": 6634, "'cancel'": 6635, "\"you have cancelled this transaction\"": 6636, "cancelled": 6637, "test_process_missing_parameters": 6638, "missing_param": 6639, "signed": 6640, "past": 6641, "'signed_field_names'": 6642, "'reason_code,message'": 6643, "_signature": 6644, "test_get_processor_exception_html": 6645, "error_string": 6646, "\"\n tests the processor exception html message\n \"": 6647, "exception_type": 6648, "ccprocessorsignatureexception": 6649, "ccprocessorwrongamountexception": 6650, "ccprocessordataexception": 6651, "error_msg": 6652, "_get_processor_exception_html": 6653, "payment_support_email": 6654, "'sorry!'": 6655, "cleanup": 6656, "\"these users are not managed by rally, so don't touch them.\"": 6657, "managed": 6658, "rally": 6659, "touch": 6660, "them": 6661, "loglevel": 6662, "xbmc": 6663, "logdebug": 6664, "addon": 6665, "getsetting": 6666, "\"logenabled\"": 6667, "\"utf-8\"": 6668, "'%s: %s'": 6669, "addon_id": 6670, "getdimvalue": 6671, "dim": 6672, "00000000": 6673, "ff000000": 6674, "black": 6675, "\"dimlevel\"": 6676, "dim_level": 6677, "'00000000'": 6678, "combine_sents": 6679, "sent_set": 6680, "'\n parameter: sent_set ==> 2d sentences set\n ==> type: list[list[list]]\n\n return: sents1d ==> 1d sentences set\n ==> type: list[list]\n\n this function will combine 2d sentence set \n into 1d sentence set. \n e.g.\n [\n [[sent1], [sent2], [sent3], ..., [sentn]]\n ...\n [[sent1], [sent2], [sent3], ..., [sentn]]\n ]\n ==> \n [\n [sentences1],\n ...\n [sentencesn]\n ]\n '": 6681, "sents1d": 6682, "doc": 6683, "sentences": 6684, "sentence": 6685, "sent1": 6686, "sent2": 6687, "sent3": 6688, "sentn": 6689, "sents": 6690, "shuffle_index": 6691, "length_of_indices_ls": 6692, "'\n ----------\n parameter: \n ----------\n length_of_indices_ls: type = int \n\n ----------\n return: \n ----------\n a shuffled numpy array of indices \n '": 6693, "shuffled": 6694, "padding": 6695, "batch_input_list": 6696, "'\n ----------\n parameter: \n ----------\n batch_input_list: type = list(list) \n\n ----------\n return: \n ----------\n numpy.ndarray: shape == (n_batch, max_time_step) \n '": 6697, "n_batch": 6698, "max_time_step": 6699, "padding_result": 6700, "concatenate": 6701, "batc": 6702, "mask_generator": 6703, "indices_matrix": 6704, "'\n ----------\n parameter: \n ----------\n indices_matrix: type = list[list] \n\n ----------\n return: \n ----------\n mask : type = np.ndarray\n a mask matrix of a batch of varied length instances\n '": 6705, "len_ls": 6706, "max_len": 6707, "varied": 6708, "instances": 6709, "mlp_mask_generator": 6710, "wemb_size": 6711, "'\n ----------\n parameter: \n ----------\n indices_matrix: type = list[list] \n\n ----------\n return: \n ----------\n mask : type = np.ndarray\n mask.shape = (n_batch, wemb_size)\n '": 6712, "mlp": 6713, "fake_input_generator": 6714, "max_index": 6715, "batch_number": 6716, "length_range": 6717, "'\n ----------\n parameter: \n ----------\n max_index: type = int \n batch_number: type = int \n length_range: tuple(int), len(length_range) = 2 \n e.g. (50, 70)\n\n ----------\n return: \n ----------\n fake_data: type = list[list]\n format: fake_data.shape[0] = batch_number\n length_range[0] <= len(fake_data[i]) <= length_range[1]\n 0 <= fake_data[i][j] <= max_index\n '": 6718, "randint": 6719, "fake_data": 6720, "length_": 6721, "fake": 6722, "painter": 6723, "isselected": 6724, "setpen": 6725, "selection_pen": 6726, "drawrect": 6727, "_selectionrect": 6728, "mousemoveevent": 6729, "resizing": 6730, "_objref": 6731, "scenepos": 6732, "haschanged": 6733, "resizex": 6734, "xside": 6735, "resizey": 6736, "yside": 6737, "_tryandresize": 6738, "updatecache": 6739, "_layerref": 6740, "_updateposition": 6741, "_updatesize": 6742, "kpeditoritem": 6743, "elapsed": 6744, "start_ts": 6745, "remaining": 6746, "goto": 6747, "incr": 6748, "finish": 6749, "standard_tensorboard_wsgi": 6750, "logdir": 6751, "purge_orphaned_data": 6752, "reload_interval": 6753, "plugins": 6754, "\"construct a tensorboardwsgiapp with standard plugins and multiplexer.\n\n args:\n logdir: the path to the directory containing events files.\n purge_orphaned_data: whether to purge orphaned data.\n reload_interval: the interval at which the backend reloads more data in\n seconds.\n plugins: a list of plugins for tensorboard to initialize.\n\n returns:\n the new tensorboard wsgi application.\n \"": 6755, "multiplexer": 6756, "event_multiplexer": 6757, "eventmultiplexer": 6758, "size_guidance": 6759, "defaul": 6760, "tensorboardwsgiapp": 6761, "purge": 6762, "orphaned": 6763, "reloads": 6764, "tensorboard": 6765, "wsgi": 6766, "parse_event_files_spec": 6767, "parses": 6768, "comma": 6769, "separated": 6770, "specifications": 6771, "specification": 6772, "'group_name:/path/to/directory'": 6773, "'/path/to/directory'": 6774, "latter": 6775, "unnamed": 6776, "forward": 6777, "slash": 6778, "interpreted": 6779, "'/foo:bar/baz'": 6780, "globs": 6781, "supported": 6782, "reload_multiplexer": 6783, "path_to_run": 6784, "\"loads all runs into the multiplexer.\n\n args:\n multiplexer: the `eventmultiplexer` to add runs to and reload.\n path_to_run: a dict mapping from paths to run names, where `none` as the run\n name is interpreted as a run name equal to the path.\n \"": 6785, "'tensorboard reload process beginning'": 6786, "addrunsfromdirectory": 6787, "'tensorboard reload process: reload the whole multiplexer'": 6788, "runs": 6789, "start_reloading_multiplexer": 6790, "load_interval": 6791, "reloadmultiplexer": 6792, "immediately": 6793, "how": 6794, "reloading": 6795, "get_tensorboard_tag": 6796, "\"read the tensorboard tag number, and return it or an empty string.\"": 6797, "resource_loader": 6798, "load_resource": 6799, "'tensorboard/tag'": 6800, "plot": 6801, "toserial": 6802, "\"input data must be a str type\"": 6803, "serial": 6804, "'baud'": 6805, "9600": 6806, "ser": 6807, "baudrate": 6808, "isopen": 6809, "toprinter": 6810, "printer": 6811, "pri": 6812, "test_inherited_models_content_type": 6813, "\"\n test that genericrelations on inherited classes use the correct content\n type.\n \"": 6814, "\"south park\"": 6815, "\"chubby's\"": 6816, "l1": 6817, "content_object": 6818, "l2": 6819, "genericrelations": 6820, "inherited": 6821, "classes": 6822, "test_reverse_relation_pk": 6823, "\"\n test that the correct column name is used for the primary key on the\n originating model of a query. see #12664.\n \"": 6824, "'chef'": 6825, "street": 6826, "'123 anywhere place'": 6827, "city": 6828, "'conifer'": 6829, "'co'": 6830, "zipcode": 6831, "'80433'": 6832, "qs": 6833, "addresses__zipcode": 6834, "originating": 6835, "12664.": 6836, "test_q_object_or": 6837, "\"\n tests that sql query parameters for generic relations are properly\n grouped when or is used.\n\n test for bug http://code.djangoproject.com/ticket/11535\n\n in this bug the first query (below) works while the second, with the\n query parameters the same but in reverse order, does not.\n\n the issue is that the generic relation conditions do not get properly\n grouped in parentheses.\n \"": 6838, "note_contact": 6839, "org_contact": 6840, "grouped": 6841, "djangoproject": 6842, "11535": 6843, "test_generic_relation_ordering": 6844, "\"\n test that ordering over a generic relation does not include extraneous\n duplicate results, nor excludes rows not participating in the relation.\n \"": 6845, "p1": 6846, "p2": 6847, "\"the city\"": 6848, "\"chubby's intl.\"": 6849, "order_by": 6850, "'links__id'": 6851, "count_places": 6852, "ordering": 6853, "extraneous": 6854, "nor": 6855, "participating": 6856, "test_target_model_is_unsaved": 6857, "\"test related to #13085\"": 6858, "orm": 6859, "dev1": 6860, "developer": 6861, "'joe'": 6862, "'deserves promotion'": 6863, "integrityerror": 6864, "13085": 6865, "unsaved": 6866, "test_target_model_len_zero": 6867, "\"test for #13085 -- __len__() returns 0\"": 6868, "team1": 6869, "team": 6870, "'backend devs'": 6871, "'deserve a bonus'": 6872, "'impossible arguments to gfk.get_content_type!'": 6873, "\"saving model with genericforeignkey to model instance whose \"": 6874, "\"__len__ method returns 0 shouldn't fail.\"": 6875, "test_target_model_nonzero_false": 6876, "__nonzero__": 6877, "'t currently fail.\n # this test validates that\n g1 = guild.objects.create(name='": 6878, "guild": 6879, "')\n note = note(note='": 6880, "g1": 6881, "test_gfk_to_model_with_empty_pk": 6882, "saving": 6883, "genericforeignkey": 6884, "charfield": 6885, "b1": 6886, "board": 6887, "'vp'": 6888, "gfk": 6889, "test_ticket_20378": 6890, "couple": 6891, "haslinkthing": 6892, "autopk": 6893, "hs1": 6894, "hs2": 6895, "hs3": 6896, "hs4": 6897, "assertquerysetequal": 6898, "20378": 6899, "iterpcapfile": 6900, "fd": 6901, "reuse": 6902, "'\n figure out if it'": 6903, "tcpdump": 6904, "pcapng": 6905, "pcap_file_header": 6906, "vsparse": 6907, "fast": 6908, "seek": 6909, "magic": 6910, "pcapng_blocktype_section_header": 6911, "_iterpcapngfile": 6912, "_iterpcapfile": 6913, "figure": 6914, "pcb_optsize": 6915, "pad4bytes": 6916, "optsize": 6917, "vsgetfield": 6918, "'bytes'": 6919, "vssetlength": 6920, "pcb": 6921, "bytez": 6922, "'\n we need the tsresol value to adjust timestamp values, so pull it\n out here\n '": 6923, "pcapng_block_parent": 6924, "tsresol": 6925, "tsoffset": 6926, "'pcapng_interface_description_block: searching options'": 6927, "opt": 6928, "opt_if_tsresol": 6929, "bytes": 6930, "'got tsresol: 0x%x\\n'": 6931, "eli": 6932, "pull": 6933, "setpcaptimestamp": 6934, "idb": 6935, "'\n adds a libpcap compatible tvsec and tvusec fields, based on the pcapng timestamp\n '": 6936, "orange": 6937, "snaplen": 6938, "tstamp": 6939, "tstamphi": 6940, "tstamplow": 6941, "scale": 6942, "1000000": 6943, "capture": 6944, "assumes": 6945, "resolution": 6946, "x80": 6947, "power": 6948, "x7f": 6949, "libpcap": 6950, "compatible": 6951, "tvsec": 6952, "tvusec": 6953, "test_suite": 6954, "suite": 6955, "unittest": 6956, "testsuite": 6957, "addtest": 6958, "makesuite": 6959, "testelement": 6960, "test_taggedvalues": 6961, "\"test that we can update tagged values of more than one element\n \"": 6962, "e1": 6963, "e2": 6964, "settaggedvalue": 6965, "\"x\"": 6966, "gettaggedvalue": 6967, "taggedvalues": 6968, "test_mysql_list_esc": 6969, "sql_query_dict": 6970, "mysql_list_esc": 6971, "\"1,2\"": 6972, "esc": 6973, "test_mysql_list_compare_with_none": 6974, "_mysql_clause": 6975, "'x!='": 6976, "\" ((x is not null) and (x not in (1,2,3)) ) \"": 6977, "'x'": 6978, "\" ((x is null) or (x in (1,2,3)) ) \"": 6979, "setup_logging": 6980, "\"configure logging module for rotating logs and console output\"": 6981, "rotate_cfg": 6982, "\"filename\"": 6983, "\"log_file\"": 6984, "\"maxbytes\"": 6985, "\"backupcount\"": 6986, "rotate_fmt": 6987, "\"%(asctime)s %(levelname)-8s %(message)s\"": 6988, "console_fmt": 6989, "\"%(levelname)-8s %(message)s\"": 6990, "getlogger": 6991, "setlevel": 6992, "handlers": 6993, "rotatingfilehandler": 6994, "setformatter": 6995, "rotating": 6996, "logs": 6997, "console": 6998, "\"start a new reddit session\"": 6999, "\"logging in...\"": 7000, "\"login failed\"": 7001, "reddit": 7002, "get_requests": 7003, "\"fetch and return all new pms matching configured subject\"": 7004, "\"fetching new messages...\"": 7005, "pending": 7006, "get_unread": 7007, "\"failed to get new messages\"": 7008, "\"subject\"": 7009, "\"limit_read\"": 7010, "mark_as_read": 7011, "got": 7012, "pms": 7013, "process_request": 7014, "subreddit": 7015, "\"read flair request message and set if possible\"": 7016, "author": 7017, "flair": 7018, "\"blacklist\"": 7019, "\"skipping blacklisted user: %s\"": 7020, "flairs": 7021, "set_flair": 7022, "\"error setting flair to %s for %s\"": 7023, "get_wiki_page": 7024, "\"fetching wiki page...\"": 7025, "\"cache_file\"": 7026, "\"no cache file found\"": 7027, "st_mtime": 7028, "\"cache_time\"": 7029, "\"using valid cache\"": 7030, "wiki_page": 7031, "wiki": 7032, "test_rol_x86_64": 7033, "binary_path": 7034, "test_location": 7035, "\"/x86_64/test_rol.exe\"": 7036, "proj": 7037, "angr": 7038, "blank_state": 7039, "addr": 7040, "x401000": 7041, "r_rax": 7042, "bvs": 7043, "'rax'": 7044, "regs": 7045, "rax": 7046, "pg": 7047, "simgr": 7048, "immutable": 7049, "explore": 7050, "x401013": 7051, "x401010": 7052, "found_state": 7053, "nose": 7054, "x37b7ab70": 7055, "rol": 7056, "x86": 7057, "\"\n returns the target output for this task.\n in this case, a successful execution of this task will create a file on the local filesystem.\n\n :return: the target output for this task.\n :rtype: object (:py:class:`luigi.target.target`)\n \"": 7058, "luigi": 7059, "localtarget": 7060, "'/tmp/_docs-%s.ldj'": 7061, "\"\n this task's dependencies:\n\n * :py:class:`~.fakedocuments`\n\n :return: object (:py:class:`luigi.task.task`)\n \"": 7062, "fakedocuments": 7063, "dependencies": 7064, "key_index": 7065, "'accepts a parameter, :key:, and returns an integer value\n representing its index in the ordered list of keys.'": 7066, "_keys": 7067, "'accepts a :key:, :value:, and :index: parameter and inserts\n a new key, value member at the desired index.\n\n note: inserting with a negative index will have the following behavior:\n >>> l = [1, 2, 3, 4]\n >>> l.insert(-1, 5)\n >>> l\n [1, 2, 3, 5, 4]\n '": 7068, "_d": 7069, "inserts": 7070, "member": 7071, "reorder_keys": 7072, "'accepts a :keys: parameter, an iterable of keys in the\n desired new order. the :keys: parameter must contain all\n existing keys.'": 7073, "'the supplied number of keys does not match.'": 7074, "'the supplied keys do not match the current set of keys.'": 7075, "reorder": 7076, "\"return a copy of the _keys list instead of iterating over it as the mutablemapping does by default.\n \"": 7077, "iterating": 7078, "mutablemapping": 7079, "test_get_default_password_validators": 7080, "validators": 7081, "get_default_password_validators": 7082, "'commonpasswordvalidator'": 7083, "'minimumlengthvalidator'": 7084, "min_length": 7085, "test_help_text": 7086, "minimumlengthvalidator": 7087, "get_help_text": 7088, "\"your password must contain at least 8 characters.\"": 7089, "test_validate": 7090, "'testclient'": 7091, "'testclient@example.com'": 7092, "'client'": 7093, "expected_error": 7094, "\"the password is too similar to the %s.\"": 7095, "userattributesimilarityvalidator": 7096, "\"username\"": 7097, "test_ascii_validator": 7098, "valid_usernames": 7099, "'glenn'": 7100, "'jean-marc'": 7101, "invalid_usernames": 7102, "\"o'connell\"": 7103, "'\u00e9ric'": 7104, "'jean marc'": 7105, "\"\u0623\u062d\u0645\u062f\"": 7106, "asciiusernamevalidator": 7107, "subtest": 7108, "validator": 7109, "poll_and_publish": 7110, "\"polling counter and publish into pipeline.\"": 7111, "polling": 7112, "create_polling_task": 7113, "\"create an empty polling task.\"": 7114, "setup_polling_tasks": 7115, "polling_tasks": 7116, "\"setup polling tasks\"": 7117, "pollster": 7118, "pipeline_manager": 7119, "pipelines": 7120, "pollster_manager": 7121, "\"pipeline:{},pollster:{}\"": 7122, "get_counter_names": 7123, "\"counter:{}\"": 7124, "support_counter": 7125, "\"pipeline support counter:{}\"": 7126, "initialize_service_hook": 7127, "\"service:{}\"": 7128, "\"intervak:{} task:{}\"": 7129, "openstack": 7130, "threadgroup": 7131, "\u306a": 7132, "\u3069": 7133, "\u304c": 7134, "\u547c": 7135, "\u3070": 7136, "\u308c": 7137, "\u308b": 7138, "tg": 7139, "add_timer": 7140, "interval_task": 7141, "hook": 7142, "makestatefuldispatcher": 7143, "determined": 7144, "_state": 7145, "subsidiary": 7146, "'_%s_%s'": 7147, "giv": 7148, "switchtobodymode": 7149, "\"\n switch to body parsing mode - interpret any more bytes delivered as\n part of the message body and deliver them to the given decoder.\n \"": 7150, "\"already in body mode\"": 7151, "bodydecoder": 7152, "setrawmode": 7153, "linereceived": 7154, "\"\n handle one line from a response.\n \"": 7155, "lf": 7156, "'\\r'": 7157, "statusreceived": 7158, "' \\t'": 7159, "_partialheader": 7160, "rawdatareceived": 7161, "\"\n pass data from the message body to the body decoder object.\n \"": 7162, "datareceived": 7163, "isconnectioncontrolheader": 7164, "\"\n return c{true} if the given lower-cased name is the name of a\n connection control header (rather than an entity header).\n\n according to rfc 2616, section 14.10, the tokens in the connection\n header are probably relevant here. however, i am not sure what the\n practical consequences of either implementing or ignoring that are.\n so i leave it unimplemented for the time being.\n \"": 7165, "connection_control_headers": 7166, "cased": 7167, "rfc": 7168, "2616": 7169, "14.10": 7170, "tokens": 7171, "probably": 7172, "relevant": 7173, "am": 7174, "what": 7175, "practical": 7176, "consequences": 7177, "implementin": 7178, "\"\n callback invoked whenever the first line of a new message is received.\n override this.\n\n @param status: the first line of an http request or response message\n without trailing i{cr lf}.\n @type status: c{str}\n \"": 7179, "invoked": 7180, "whenever": 7181, "headerreceived": 7182, "\"\n store the given header in c{self.headers}.\n \"": 7183, "connheaders": 7184, "addrawheader": 7185, "allheadersreceived": 7186, "\"\n callback invoked after the last header is passed to c{headerreceived}.\n override this to change to the c{body} or c{done} state.\n \"": 7187, "\"\n override so that we know if any response has been received.\n \"": 7188, "_everreceiveddata": 7189, "httpparser": 7190, "know": 7191, "parseversion": 7192, "strversion": 7193, "\"\n parse version strings of the form protocol '/' major '.' minor. e.g.\n 'http/1.1'. returns (protocol, major, minor). will raise valueerror\n on bad syntax.\n \"": 7194, "proto": 7195, "strnumber": 7196, "major": 7197, "minor": 7198, "badresponseversion": 7199, "\"version may not be negative\"": 7200, "st": 7201, "'http/1.1'": 7202, "writeto": 7203, "transport": 7204, "\"\n format this l{request} as an http/1.1 request and write it to the given\n transport. if bodyproducer is not none, it will be associated with an\n l{iconsumer}.\n\n @return: a l{deferred} which fires with c{none} when the request has\n been completely written to the transport or with a l{failure} if\n there is any problem generating the request bytes.\n \"": 7205, "bodyproducer": 7206, "unknown_length": 7207, "_writetochunke": 7208, "1.1": 7209, "iconsumer": 7210, "deferred": 7211, "fires": 7212, "stopwriting": 7213, "\"\n stop writing this request to the transport. this can only be called\n after c{writeto} and before the l{deferred} returned by c{writeto}\n fires. it should cancel any asynchronous task started by c{writeto}.\n the l{deferred} returned by c{writeto} need not be fired if this method\n is called.\n \"": 7214, "fired": 7215, "_callappfunction": 7216, "stopproducing": 7217, "writing": 7218, "asynchronous": 7219, "deliverbody": 7220, "\"\n dispatch the given l{iprotocol} depending of the current state of the\n response.\n \"": 7221, "iprotocol": 7222, "registerproducer": 7223, "producer": 7224, "streaming": 7225, "\"\n register the given producer with c{self.transport}.\n \"": 7226, "unregisterproducer": 7227, "\"\n indicate that the request body is complete and finish the request.\n \"": 7228, "_allownomorewrites": 7229, "indicate": 7230, "complete": 7231, "\"\n proxy the stoppage to the underlying producer, unless this proxy has\n been stopped.\n \"": 7232, "_producer": 7233, "proxy": 7234, "stoppage": 7235, "underlying": 7236, "unless": 7237, "stopped": 7238, "resumeproducing": 7239, "\"\n proxy the resumption to the underlying producer, unless this proxy has\n been stopped.\n \"": 7240, "resumption": 7241, "pauseproducing": 7242, "\"\n proxy the pause to the underlying producer, unless this proxy has been\n stopped.\n \"": 7243, "connectionlost": 7244, "\"\n the underlying transport went away. if appropriate, notify the parser\n object.\n \"": 7245, "away": 7246, "\"\n close the connection and cause all outstanding l{request} l{deferred}s\n to fire with an error.\n \"": 7247, "\"connection_lost\"": 7248, "succeed": 7249, "loseconnection": 7250, "'aborting'": 7251, "_abortdeferreds": 7252, "cause": 7253, "outstanding": 7254, "fire": 7255, "ebproduced": 7256, "'t call the encoder'": 7257, "chunk": 7258, "cancelrequest": 7259, "ign": 7260, "'s deferred if it'": 7261, "trying": 7262, "'transmitting'": 7263, "'transmitting_after_receiving_response'": 7264, "_requestdeferred": 7265, "abortconnection": 7266, "_disconnectparser": 7267, "cancellederror": 7268, "get_form_kwargs": 7269, "testwizard": 7270, "'kwargs_test'": 7271, "test_form_kwargs": 7272, "get_request": 7273, "testform": 7274, "as_view": 7275, "'start'": 7276, "step1": 7277, "customkwargsstep1": 7278, "get_form": 7279, "test_init": 7280, "sessionwizardview": 7281, "templateresponse": 7282, "main_func": 7283, "pconfs": 7284, "pconf": 7285, "pconf_class": 7286, "family": 7287, "class_sps": 7288, "structures": 7289, "protein_conformation__protein__parent__family__slug__startswith": 7290, "sps": 7291, "execfile_": 7292, "fname": 7293, "\".pyc\"": 7294, "_get_codeobj": 7295, "compile": 7296, "'exec'": 7297, "exec_": 7298, "execfile": 7299, "bytes_to_str": 7300, "'latin1'": 7301, "wrap_error": 7302, "kw": 7303, "\"\n wrap socket.error, ioerror, oserror, select.error to raise new specialized\n exceptions of python 3.3 like interruptederror (pep 3151).\n \"": 7304, "socket": 7305, "exc": 7306, "'winerror'": 7307, "_wrap_error": 7308, "_map_errno": 7309, "winerror": 7310, "\"file not found\"": 7311, "errno": 7312, "ex": 7313, "enoent": 7314, "specialized": 7315, "interruptederror": 7316, "3151": 7317, "urlsplit": 7318, "allow_fragments": 7319, "\"parse a url into 5 components:\n :///?#\n return a 5-tuple: (scheme, netloc, path, query, fragment).\n note that we don't break the components up in smaller bits\n (e.g. netloc is a single string) and we don't expand % escapes.\"": 7320, "_parse_cache": 7321, "netloc": 7322, "'t break the components up in smaller bits\n(e.g. netloc is a single string) and we don'": 7323, "judge_is_mac_same": 7324, "mac1": 7325, "mac2": 7326, "\"judge whether two macs are the same\"": 7327, "list1": 7328, "findall": 7329, "'([0-9a-fa-f]+)'": 7330, "list2": 7331, "judge": 7332, "macs": 7333, "mac": 7334, "get_interface_type": 7335, "\"gets the type of interface, such as 10ge, eth-trunk, vlanif...\"": 7336, "iftype": 7337, "'ge'": 7338, "'10ge'": 7339, "'25ge'": 7340, "'4x10ge'": 7341, "'40ge'": 7342, "'100ge'": 7343, "ge": 7344, "eth": 7345, "vlanif": 7346, "check_config_if_same": 7347, "\"judge whether current config is the same as what we excepted\"": 7348, "'absent'": 7349, "dldp_intf_conf": 7350, "'dldpenable'": 7351, "mode_enable": 7352, "'dldpcompatibleenable'": 7353, "local_mac": 7354, "'dldplocalmac'": 7355, "fla": 7356, "excepted": 7357, "check_macaddr": 7358, "\"check mac-address whether valid\"": 7359, "valid_char": 7360, "'0123456789abcdef-'": 7361, "mac_list": 7362, "macaddr": 7363, "check_params": 7364, "\"check all input params\"": 7365, "fail_json": 7366, "'error: interface name cannot be empty.'": 7367, "intf_type": 7368, "'error: interface name of %s '": 7369, "'is error.'": 7370, "init_module": 7371, "\"init module object\"": 7372, "check_response": 7373, "xml_str": 7374, "xml_name": 7375, "\"check if response message is already succeed\"": 7376, "\"\"": 7377, "'error: %s failed.'": 7378, "get_dldp_intf_exist_config": 7379, "\"get current dldp existed config\"": 7380, "dldp_conf": 7381, "ce_nc_get_intf_dldp_config": 7382, "con_obj": 7383, "get_nc_config": 7384, "\"\"": 7385, "'disable'": 7386, "'xmlns=\"urn:ietf:params:xml:ns:netconf:base:1.0\"'": 7387, "dldp": 7388, "existed": 7389, "intf": 7390, "config_intf_dldp": 7391, "\"config global dldp\"": 7392, "same_conf": 7393, "\"present\"": 7394, "'enable'": 7395, "get_existing": 7396, "\"get existing info\"": 7397, "'interface'": 7398, "'mode_enable'": 7399, "'local_mac'": 7400, "'reset'": 7401, "get_proposed": 7402, "\"get proposed result \"": 7403, "proposed": 7404, "get_update_cmd": 7405, "\"get updatede commands\"": 7406, "updates_cmd": 7407, "\"dldp enable\"": 7408, "\"undo dldp enable\"": 7409, "updatede": 7410, "commands": 7411, "get_end_state": 7412, "\"get end state info\"": 7413, "end_state": 7414, "show_result": 7415, "\"show result\"": 7416, "'changed'": 7417, "'proposed'": 7418, "'existing'": 7419, "'end_state'": 7420, "'updates'": 7421, "exit_json": 7422, "\"excute task\"": 7423, "excute": 7424, "halton": 7425, "floor": 7426, "floortonearest": 7427, "nearest": 7428, "getmultiplier": 7429, "percent_complete": 7430, "rad": 7431, "radians": 7432, "pi": 7433, "multiplier": 7434, "multplier": 7435, "preload_check": 7436, "\"raises an exception if the environment is not correctly configured.\n\n raises:\n importerror: if the check detects that the environment is not correctly\n configured, and attempting to load the tensorflow runtime will fail.\n \"": 7437, "\"nt\"": 7438, "attempt": 7439, "dlls": 7440, "actionable": 7441, "ctypes": 7442, "build_info": 7443, "\"msvcp_dll_name\"": 7444, "ctype": 7445, "detects": 7446, "tensorflow": 7447, "preload": 7448, "test_string_form_referencing": 7449, "\"\n regression test for #1661 and #1662\n\n string form referencing of models works, both as pre and post\n reference, on all relatedfield types.\n \"": 7450, "\"foo1\"": 7451, "\"foo2\"": 7452, "w1": 7453, "whiz": 7454, "\"whiz1\"": 7455, "\"bar1\"": 7456, "fwd": 7457, "1661": 7458, "1662": 7459, "referencing": 7460, "relatedfield": 7461, "test_unicode_chars_in_queries": 7462, "\"\n regression tests for #3937\n\n make sure we can use unicode characters in queries.\n if these tests fail on mysql, it's a problem with the test setup.\n a properly configured utf-8 database can handle this.\n \"": 7463, "fx": 7464, "'bjorn'": 7465, "friend": 7466, "'fran\u00e7ois'": 7467, "friend__contains": 7468, "'\\xe7'": 7469, "3937": 7470, "utf": 7471, "test_queries_on_textfields": 7472, "\"\n regression tests for #5087\n\n make sure we can perform queries on textfields.\n \"": 7473, "article": 7474, "'the quick brown fox jumps over the lazy dog.'": 7475, "text__exact": 7476, "text__contains": 7477, "'quick brown fox'": 7478, "5087": 7479, "perform": 7480, "textfields": 7481, "test_ipaddress_on_postgresql": 7482, "\"\n regression test for #708\n\n \"": 7483, "\" queries on ip address fields require casting with host() (on postgresql).\n \"": 7484, "'ip test'": 7485, "'the body'": 7486, "submitted_from": 7487, "'192.0.2.100'": 7488, "assertsequenceequal": 7489, "submitted_from__contains": 7490, "'192.0.2'": 7491, "searches": 7492, "subnet": 7493, "'32'": 7494, "708": 7495, "\"like\"": 7496, "require": 7497, "casting": 7498, "postgresql": 7499, "ipaddress": 7500, "test_assignement": 7501, "'change_contenttype'": 7502, "userobjectpermission": 7503, "'contenttypes.change_contenttype'": 7504, "assignement": 7505, "test_assignement_and_remove": 7506, "remove_perm": 7507, "test_get_for_object": 7508, "'get_group_perms_for_object'": 7509, "perms": 7510, "groupobjectpermission": 7511, "get_for_object": 7512, "to_assign": 7513, "'delete_contenttype'": 7514, "perm": 7515, "codenames": 7516, "test_assign_validation": 7517, "doesnotexist": 7518, "'change_user'": 7519, "'test_user_assign_validation'": 7520, "contenttype": 7521, "get_for_model": 7522, "codename": 7523, "'change_group'": 7524, "create_info": 7525, "content_type": 7526, "object_pk": 7527, "grou": 7528, "test_unicode": 7529, "obj_perm": 7530, "\"change_group\"": 7531, "test_has_perm_notauthed": 7532, "anonymoususer": 7533, "\"change_user\"": 7534, "notauthed": 7535, "test_has_perm_wrong_app": 7536, "wrongapperror": 7537, "\"no_app.change_user\"": 7538, "test_error_classes": 7539, "guardianerror": 7540, "guardian_errors": 7541, "notusernorgroup": 7542, "_test_error_class": 7543, "blueprint_is_module": 7544, "bp": 7545, "\"used to figure out if something is actually a module\"": 7546, "test_running_build": 7547, "tmpdir": 7548, "caplog": 7549, "user_params": 7550, "sources_dir": 7551, "sources_dir_exists": 7552, "sources_dir_empty": 7553, "remote_dir": 7554, "remote_dir_exists": 7555, "remote_dir_empty": 7556, "export_failed": 7557, "\"\n test if proper result is returned and if plugin works\n \"": 7558, "sources_dir_path": 7559, "strpath": 7560, "mknod": 7561, "'stub.srpm'": 7562, "remote_dir_path": 7563, "remote_di": 7564, "running": 7565, "test_failed_build": 7566, "\"\n test if proper error state is returned when build inside build\n container failed\n \"": 7567, "flexmock": 7568, "should_receive": 7569, "'check_output'": 7570, "and_raise": 7571, "'cmd'": 7572, "'stub stdout'": 7573, "workflow": 7574, "mock_workflow": 7575, "mocked_tasker": 7576, "tasker": 7577, "'wait'": 7578, "and_return": 7579, "buildsteppluginsrunner": 7580, "sourcecontainerplugin": 7581, "'args'": 7582, "check_check_output": 7583, "'skopeo'": 7584, "'copy'": 7585, "'oci:%s'": 7586, "temp_image_output_dir": 7587, "'docker-archive:%s'": 7588, "temp_image_export_dir": 7589, "exported_squashed_image_name": 7590, "returncode": 7591, "\"failed\"": 7592, "args_expect": 7593, "'bsi'": 7594, "'-d'": 7595, "locate_oc_binary": 7596, "' find and return oc binary file '": 7597, "openshift": 7598, "ansible": 7599, "3410": 7600, "oc": 7601, "usr": 7602, "environ": 7603, "defpath": 7604, "pathsep": 7605, "additional_path_lookups": 7606, "oc_binary": 7607, "'oc'": 7608, "naive": 7609, "which_result": 7610, "locate": 7611, "separator": 7612, "' getter method for separator '": 7613, "_separator": 7614, "getter": 7615, "yaml_dict": 7616, "' getter method for yaml_dict '": 7617, "__yaml_dict": 7618, "parse_key": 7619, "'parse the key allowing the appropriate separator'": 7620, "common_separators": 7621, "yedit": 7622, "com_sep": 7623, "re_key": 7624, "valid_key": 7625, "'validate the incoming key'": 7626, "re_valid_key": 7627, "incoming": 7628, "remove_entry": 7629, "' remove data at location key '": 7630, "key_indexes": 7631, "arr_ind": 7632, "dict_key": 7633, "add_entry": 7634, "' get an item from a dictionary with key notation a.b.c\n d = {'": 7635, "': {'": 7636, "'}}}\n key = a#b\n return c\n '": 7637, "noqa": 7638, "e501": 7639, "'c'": 7640, "get_entry": 7641, "'}}}\n key = a.b\n return c\n '": 7642, "' read from file '": 7643, "file_exists": 7644, "yfd": 7645, "' return whether file exists '": 7646, "' check if value exists at path'": 7647, "rval": 7648, "brea": 7649, "'append value to a list'": 7650, "audit": 7651, "maybe": 7652, "makes": 7653, "' put path, value into a dict '": 7654, "didn": 7655, "ruamel": 7656, "pyyaml": 7657, "tmp_copy": 7658, "round_trip_dump": 7659, "default_flow_style": 7660, "roundtriploader": 7661, "get_curr_value": 7662, "invalue": 7663, "val_type": 7664, "'return the current value'": 7665, "curr_value": 7666, "'yaml'": 7667, "'json'": 7668, "curr": 7669, "parse_value": 7670, "inc_value": 7671, "vtype": 7672, "'determine value type passed'": 7673, "true_bools": 7674, "'y'": 7675, "'yes'": 7676, "'on'": 7677, "false_bools": 7678, "'n'": 7679, "'no'": 7680, "'off'": 7681, "'t specify value_type as string\n # we will convert to bool if it matches any of the above cases\n if isinstance(inc_value, str) and '": 7682, "process_edits": 7683, "edits": 7684, "yamlfile": 7685, "'run through a list of edits and process them one-by-one'": 7686, "'value_type'": 7687, "'action'": 7688, "'update'": 7689, "'curr_value'": 7690, "'curr_value_format'": 7691, "'key'": 7692, "run_ansible": 7693, "'perform the idempotent crud operations'": 7694, "'src'": 7695, "backup": 7696, "'backup'": 7697, "'separator'": 7698, "'state'": 7699, "'failed'": 7700, "'msg'": 7701, "'error opening file [{}]. verify that the '": 7702, "crud": 7703, "operations": 7704, "openshift_cmd": 7705, "oadm": 7706, "output_type": 7707, "input_data": 7708, "'base command for oc '": 7709, "cmds": 7710, "'adm'": 7711, "all_namespaces": 7712, "'--all-namespaces'": 7713, "'none'": 7714, "'emtpy'": 7715, "'-n'": 7716, "_run": 7717, "create_tmp_file_from_contents": 7718, "rname": 7719, "ftype": 7720, "' create a file in tmp with name and contents'": 7721, "create_tmpfile": 7722, "'roundtripdumper'": 7723, "_write": 7724, "dumper": 7725, "roundtripdumper": 7726, "safe_dump": 7727, "create_tmpfile_copy": 7728, "inc_file": 7729, "'create a temporary copy of a file'": 7730, "tmpfile": 7731, "'lib_openshift-'": 7732, "atexit": 7733, "'tmp'": 7734, "' generates and returns a temporary file name '": 7735, "namedtemporaryfile": 7736, "create_tmp_files_from_contents": 7737, "'turn an array of dict: filename, content into a files array'": 7738, "basename": 7739, "find_result": 7740, "' find the specified result by name'": 7741, "'metadata'": 7742, "get_resource_file": 7743, "sfile": 7744, "sfile_type": 7745, "' return the service file '": 7746, "sfd": 7747, "'roundtriploader'": 7748, "conten": 7749, "filter_versions": 7750, "' filter the oc version output '": 7751, "version_dict": 7752, "version_search": 7753, "'openshift'": 7754, "'kubernetes'": 7755, "horrible": 7756, "\"oc version in 3.2 does not return an \"": 7757, "\" version\n if \"": 7758, "versions": 7759, "add_custom_versions": 7760, "' create custom versions strings '": 7761, "versions_dict": 7762, "tech": 7763, "'v'": 7764, "'_numeric'": 7765, "'+'": 7766, "\"v3.3.0.33\"": 7767, "\"3.3\"": 7768, "'_short'": 7769, "openshift_installed": 7770, "' check if openshift is installed '": 7771, "rpm": 7772, "transaction_set": 7773, "transactionset": 7774, "rpmquery": 7775, "dbmatch": 7776, "\"atomic-openshift\"": 7777, "installed": 7778, "check_def_equal": 7779, "user_def": 7780, "result_def": 7781, "skip_keys": 7782, "' given a user defined definition, compare it with the results given back by our query. '": 7783, "autogenerated": 7784, "'status'": 7785, "config_options": 7786, "' return config options '": 7787, "to_option_list": 7788, "ascommalist": 7789, "'return all options as a string\n if ascommalist is set to the name of a key, and\n the value of that key is a dict, format the dict\n as a list of comma delimited key=value pairs'": 7790, "stringify": 7791, "delimited": 7792, "' return the options hash as cli params in a string\n if ascommalist is set to the name of a key, and\n the value of that key is a dict, format the dict\n as a list of comma delimited key=value pairs '": 7793, "'include'": 7794, "'{}={}'": 7795, "kk": 7796, "cli": 7797, "current_labels": 7798, "'property for the current labels'": 7799, "_curr_labels": 7800, "'labels'": 7801, "compare_labels": 7802, "host_labels": 7803, "' compare incoming labels against current labels'": 7804, "all_user_labels_exist": 7805, "' return whether all the labels already exist '": 7806, "current_host_labels": 7807, "rbool": 7808, "any_label_exists": 7809, "' return whether any single label already exists '": 7810, "get_user_keys": 7811, "' go through list of user key:values and return all keys '": 7812, "user_keys": 7813, "get_current_label_keys": 7814, "' collect all the current label keys '": 7815, "current_label_keys": 7816, "get_extra_current_labels": 7817, "' return list of labels that are currently stored, but aren'": 7818, "extra_labels": 7819, "user_label_keys": 7820, "current_key": 7821, "stored": 7822, "extra_current_labels": 7823, "' return whether there are labels currently stored that user\n hasn'": 7824, "hasn": 7825, "cmd_template": 7826, "' boilerplate oc command for modifying lables on this object '": 7827, "let": 7828, "\"label\"": 7829, "kind": 7830, "\"--selector\"": 7831, "boilerplate": 7832, "modifying": 7833, "lables": 7834, "lit": 7835, "\"\n creates a :class:`~pyspark.sql.column` of literal value.\n\n .. versionadded:: 1.3.0\n\n examples\n --------\n >>> df.select(lit(5).alias('height')).withcolumn('spark_user', lit(true)).take(1)\n [row(height=5, spark_user=true)]\n \"": 7836, "_invoke_function": 7837, "\"lit\"": 7838, "pyspark": 7839, "literal": 7840, "versionadded": 7841, "df": 7842, "'height'": 7843, "withcolumn": 7844, "'spark_user'": 7845, "spark_user": 7846, "\"\n returns a :class:`~pyspark.sql.column` based on the given column name.'\n examples\n --------\n >>> col('x')\n column<'x'>\n >>> column('x')\n column<'x'>\n \"": 7847, "\"col\"": 7848, "'\nexamples\n--------\n>>> col('": 7849, "')\ncolumn<'": 7850, "'>\n>>> column('": 7851, "asc": 7852, "\"\n returns a sort expression based on the ascending order of the given column name.\n \"": 7853, "\"asc\"": 7854, "ascending": 7855, "\"\n returns a sort expression based on the descending order of the given column name.\n \"": 7856, "\"desc\"": 7857, "descending": 7858, "\"\n computes the square root of the specified float value.\n \"": 7859, "_invoke_function_over_column": 7860, "\"sqrt\"": 7861, "computes": 7862, "\"\n computes the absolute value.\n \"": 7863, "\"abs\"": 7864, "\"\n aggregate function: returns the number of items in a group.\n \"": 7865, "\"count\"": 7866, "aggregate": 7867, "\"\n aggregate function: returns the sum of all values in the expression.\n \"": 7868, "\"sum\"": 7869, "avg": 7870, "\"\n aggregate function: returns the average of the values in a group.\n \"": 7871, "\"avg\"": 7872, "mean": 7873, "\"mean\"": 7874, "sumdistinct": 7875, "\"\n aggregate function: returns the sum of distinct values in the expression.\n\n .. deprecated:: 3.2.0\n use :func:`sum_distinct` instead.\n \"": 7876, "\"deprecated in 3.2, use sum_distinct instead.\"": 7877, "futurewarning": 7878, "sum_distinct": 7879, "\"\n aggregate function: returns the sum of distinct values in the expression.\n \"": 7880, "\"sum_distinct\"": 7881, "\"\n aggregate function: returns the product of the values in a group.\n\n .. versionadded:: 3.2.0\n\n parameters\n ----------\n col : str, :class:`column`\n column containing values to be multiplied together\n\n examples\n --------\n >>> df = spark.range(1, 10).todf('x').withcolumn('mod3', col('x') % 3)\n >>> prods = df.groupby('mod3').agg(product('x').alias('product'))\n >>> prods.orderby('mod3').show()\n +----+-------+\n |mod3|product|\n +----+-------+\n | 0| 162.0|\n | 1| 28.0|\n | 2| 80.0|\n +----+-------+\n\n \"": 7882, "multiplied": 7883, "spark": 7884, "todf": 7885, "'mod3'": 7886, "prods": 7887, "acos": 7888, "\"\n .. versionadded:: 1.4.0\n\n returns\n -------\n :class:`~pyspark.sql.column`\n inverse cosine of `col`, as if computed by `java.lang.math.acos()`\n \"": 7889, "\"acos\"": 7890, "1.4": 7891, "inverse": 7892, "cosine": 7893, "computed": 7894, "java": 7895, "acosh": 7896, "\"\n computes inverse hyperbolic cosine of the input column.\n\n .. versionadded:: 3.1.0\n\n returns\n -------\n :class:`~pyspark.sql.column`\n \"": 7897, "\"acosh\"": 7898, "hyperbolic": 7899, "3.1": 7900, "asin": 7901, "\"\n .. versionadded:: 1.3.0\n\n\n returns\n -------\n :class:`~pyspark.sql.column`\n inverse sine of `col`, as if computed by `java.lang.math.asin()`\n \"": 7902, "\"asin\"": 7903, "sine": 7904, "asinh": 7905, "\"\n computes inverse hyperbolic sine of the input column.\n\n .. versionadded:: 3.1.0\n\n returns\n -------\n :class:`~pyspark.sql.column`\n \"": 7906, "\"asinh\"": 7907, "atan": 7908, "\"\n .. versionadded:: 1.4.0\n\n returns\n -------\n :class:`~pyspark.sql.column`\n inverse tangent of `col`, as if computed by `java.lang.math.atan()`\n \"": 7909, "\"atan\"": 7910, "tangent": 7911, "atanh": 7912, "\"\n computes inverse hyperbolic tangent of the input column.\n\n .. versionadded:: 3.1.0\n\n returns\n -------\n :class:`~pyspark.sql.column`\n \"": 7913, "\"atanh\"": 7914, "cbrt": 7915, "\"\n computes the cube-root of the given value.\n \"": 7916, "\"cbrt\"": 7917, "ceil": 7918, "\"\n computes the ceiling of the given value.\n \"": 7919, "\"ceil\"": 7920, "ceiling": 7921, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n angle in radians\n\n returns\n -------\n :class:`~pyspark.sql.column`\n cosine of the angle, as if computed by `java.lang.math.cos()`.\n \"": 7922, "\"cos\"": 7923, "angle": 7924, "cosh": 7925, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n hyperbolic angle\n\n returns\n -------\n :class:`~pyspark.sql.column`\n hyperbolic cosine of the angle, as if computed by `java.lang.math.cosh()`\n \"": 7926, "\"cosh\"": 7927, "exp": 7928, "\"\n computes the exponential of the given value.\n \"": 7929, "\"exp\"": 7930, "exponential": 7931, "expm1": 7932, "\"\n computes the exponential of the given value minus one.\n \"": 7933, "\"expm1\"": 7934, "minus": 7935, "\"\n computes the floor of the given value.\n \"": 7936, "\"floor\"": 7937, "log10": 7938, "\"\n computes the logarithm of the given value in base 10.\n \"": 7939, "\"log10\"": 7940, "logarithm": 7941, "10.": 7942, "log1p": 7943, "\"\n computes the natural logarithm of the given value plus one.\n \"": 7944, "\"log1p\"": 7945, "natural": 7946, "plus": 7947, "rint": 7948, "\"\n returns the double value that is closest in value to the argument and\n is equal to a mathematical integer.\n \"": 7949, "\"rint\"": 7950, "mathematical": 7951, "signum": 7952, "\"\n computes the signum of the given value.\n \"": 7953, "\"signum\"": 7954, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n\n returns\n -------\n :class:`~pyspark.sql.column`\n sine of the angle, as if computed by `java.lang.math.sin()`\n \"": 7955, "\"sin\"": 7956, "sinh": 7957, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n hyperbolic angle\n\n returns\n -------\n :class:`~pyspark.sql.column`\n hyperbolic sine of the given value,\n as if computed by `java.lang.math.sinh()`\n \"": 7958, "\"sinh\"": 7959, "tan": 7960, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n angle in radians\n\n returns\n -------\n :class:`~pyspark.sql.column`\n tangent of the given value, as if computed by `java.lang.math.tan()`\n \"": 7961, "\"tan\"": 7962, "tanh": 7963, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n hyperbolic angle\n\n returns\n -------\n :class:`~pyspark.sql.column`\n hyperbolic tangent of the given value\n as if computed by `java.lang.math.tanh()`\n \"": 7964, "\"tanh\"": 7965, "todegrees": 7966, "\"\n .. deprecated:: 2.1.0\n use :func:`degrees` instead.\n \"": 7967, "\"deprecated in 2.1, use degrees instead.\"": 7968, "2.1": 7969, "toradians": 7970, "\"\n .. deprecated:: 2.1.0\n use :func:`radians` instead.\n \"": 7971, "\"deprecated in 2.1, use radians instead.\"": 7972, "bitwisenot": 7973, "\"\n computes bitwise not.\n\n .. deprecated:: 3.2.0\n use :func:`bitwise_not` instead.\n \"": 7974, "\"deprecated in 3.2, use bitwise_not instead.\"": 7975, "bitwise_not": 7976, "bitwise": 7977, "\"\n computes bitwise not.\n \"": 7978, "\"bitwise_not\"": 7979, "asc_nulls_first": 7980, "\"\n returns a sort expression based on the ascending order of the given\n column name, and null values return before non-null values.\n \"": 7981, "\"asc_nulls_first\"": 7982, "null": 7983, "nulls": 7984, "asc_nulls_last": 7985, "\"\n returns a sort expression based on the ascending order of the given\n column name, and null values appear after non-null values.\n \"": 7986, "\"asc_nulls_last\"": 7987, "desc_nulls_first": 7988, "\"\n returns a sort expression based on the descending order of the given\n column name, and null values appear before non-null values.\n \"": 7989, "\"desc_nulls_first\"": 7990, "desc_nulls_last": 7991, "\"\n returns a sort expression based on the descending order of the given\n column name, and null values appear after non-null values.\n \"": 7992, "\"desc_nulls_last\"": 7993, "stddev": 7994, "\"\n aggregate function: alias for stddev_samp.\n \"": 7995, "\"stddev\"": 7996, "stddev_samp": 7997, "\"\n aggregate function: returns the unbiased sample standard deviation of\n the expression in a group.\n \"": 7998, "\"stddev_samp\"": 7999, "unbiased": 8000, "deviation": 8001, "samp": 8002, "stddev_pop": 8003, "\"\n aggregate function: returns population standard deviation of\n the expression in a group.\n \"": 8004, "\"stddev_pop\"": 8005, "population": 8006, "\"\n aggregate function: alias for var_samp\n \"": 8007, "\"variance\"": 8008, "var_samp": 8009, "\"\n aggregate function: returns the unbiased sample variance of\n the values in a group.\n \"": 8010, "\"var_samp\"": 8011, "var_pop": 8012, "\"\n aggregate function: returns the population variance of the values in a group.\n \"": 8013, "\"var_pop\"": 8014, "skewness": 8015, "\"\n aggregate function: returns the skewness of the values in a group.\n \"": 8016, "\"skewness\"": 8017, "kurtosis": 8018, "\"\n aggregate function: returns the kurtosis of the values in a group.\n \"": 8019, "\"kurtosis\"": 8020, "collect_list": 8021, "\"\n aggregate function: returns a list of objects with duplicates.\n\n .. versionadded:: 1.6.0\n\n notes\n -----\n the function is non-deterministic because the order of collected results depends\n on the order of the rows which may be non-deterministic after a shuffle.\n\n examples\n --------\n >>> df2 = spark.createdataframe([(2,), (5,), (5,)], ('age',))\n >>> df2.agg(collect_list('age')).collect()\n [row(collect_list(age)=[2, 5, 5])]\n \"": 8022, "\"collect_list\"": 8023, "deterministic": 8024, "collected": 8025, "df2": 8026, "createdata": 8027, "collect_set": 8028, "\"\n aggregate function: returns a set of objects with duplicate elements eliminated.\n\n .. versionadded:: 1.6.0\n\n notes\n -----\n the function is non-deterministic because the order of collected results depends\n on the order of the rows which may be non-deterministic after a shuffle.\n\n examples\n --------\n >>> df2 = spark.createdataframe([(2,), (5,), (5,)], ('age',))\n >>> df2.agg(collect_set('age')).collect()\n [row(collect_set(age)=[5, 2])]\n \"": 8029, "\"collect_set\"": 8030, "eliminated": 8031, "\"\n converts an angle measured in radians to an approximately equivalent angle\n measured in degrees.\n\n .. versionadded:: 2.1.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n angle in radians\n\n returns\n -------\n :class:`~pyspark.sql.column`\n angle in degrees, as if computed by `java.lang.math.todegrees()`\n \"": 8032, "\"degrees\"": 8033, "measured": 8034, "approximately": 8035, "equivalent": 8036, "\"\n converts an angle measured in degrees to an approximately equivalent angle\n measured in radians.\n\n .. versionadded:: 2.1.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n angle in degrees\n\n returns\n -------\n :class:`~pyspark.sql.column`\n angle in radians, as if computed by `java.lang.math.toradians()`\n \"": 8037, "\"radians\"": 8038, "col1": 8039, "col2": 8040, "\"\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n col1 : str, :class:`~pyspark.sql.column` or float\n coordinate on y-axis\n col2 : str, :class:`~pyspark.sql.column` or float\n coordinate on x-axis\n\n returns\n -------\n :class:`~pyspark.sql.column`\n the `theta` component of the point\n (`r`, `theta`)\n in polar coordinates that corresponds to the point\n (`x`, `y`) in cartesian coordinates,\n as if computed by `java.lang.math.atan2()`\n \"": 8041, "_invoke_binary_math_function": 8042, "\"atan2\"": 8043, "co": 8044, "theta": 8045, "hypot": 8046, "\"\n computes ``sqrt(a^2 + b^2)`` without intermediate overflow or underflow.\n \"": 8047, "\"hypot\"": 8048, "overflow": 8049, "underflow": 8050, "pow": 8051, "\"\n returns the value of the first argument raised to the power of the second argument.\n \"": 8052, "\"pow\"": 8053, "raised": 8054, "row_number": 8055, "\"\n window function: returns a sequential number starting at 1 within a window partition.\n \"": 8056, "\"row_number\"": 8057, "sequential": 8058, "dense_rank": 8059, "rank": 8060, "gaps": 8061, "leaves": 8062, "ranking": 8063, "ties": 8064, "competition": 8065, "people": 8066, "tie": 8067, "third": 8068, "making": 8069, "coming": 8070, "fifth": 8071, "equi": 8072, "secon": 8073, "dense": 8074, "funct": 8075, "cume_dist": 8076, "\"\n window function: returns the cumulative distribution of values within a window partition,\n i.e. the fraction of rows that are below the current row.\n \"": 8077, "\"cume_dist\"": 8078, "cumulative": 8079, "fraction": 8080, "cume": 8081, "percent_rank": 8082, "\"\n window function: returns the relative rank (i.e. percentile) of rows within a window partition.\n \"": 8083, "\"percent_rank\"": 8084, "percentile": 8085, "percent": 8086, "approxcountdistinct": 8087, "rsd": 8088, "\"\n .. deprecated:: 2.1.0\n use :func:`approx_count_distinct` instead.\n \"": 8089, "\"deprecated in 2.1, use approx_count_distinct instead.\"": 8090, "approx_count_distinct": 8091, "\"aggregate function: returns a new :class:`~pyspark.sql.column` for approximate distinct count\n of column `col`.\n\n .. versionadded:: 2.1.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n rsd : float, optional\n maximum relative standard deviation allowed (default = 0.05).\n for rsd < 0.01, it is more efficient to use :func:`count_distinct`\n\n examples\n --------\n >>> df.agg(approx_count_distinct(df.age).alias('distinct_ages')).collect()\n [row(distinct_ages=2)]\n \"": 8092, "sparkcont": 8093, "approximate": 8094, "0.05": 8095, "approx": 8096, "broadcast": 8097, "\"marks a dataframe as small enough for use in broadcast joins.\"": 8098, "sparkcontext": 8099, "_active_spark_context": 8100, "_jdf": 8101, "sql_ctx": 8102, "marks": 8103, "small": 8104, "enough": 8105, "coalesce": 8106, "\"returns the first column that is not null.\n\n .. versionadded:: 1.4.0\n\n examples\n --------\n >>> cdf = spark.createdataframe([(none, none), (1, none), (none, 2)], (\"": 8107, "\"))\n >>> cdf.show()\n +----+----+\n | a| b|\n +----+----+\n |null|null|\n | 1|null|\n |null| 2|\n +----+----+\n\n >>> cdf.select(coalesce(cdf[\"": 8108, "\"], cdf[\"": 8109, "cdf": 8110, "createdataframe": 8111, "corr": 8112, "\"returns a new :class:`~pyspark.sql.column` for the pearson correlation coefficient for\n ``col1`` and ``col2``.\n\n .. versionadded:: 1.6.0\n\n examples\n --------\n >>> a = range(20)\n >>> b = [2 * x for x in range(20)]\n >>> df = spark.createdataframe(zip(a, b), [\"": 8113, "\"])\n >>> df.agg(corr(\"": 8114, "\").alias('c')).collect()\n [row(c=1.0)]\n \"": 8115, "_to_java_column": 8116, "pearson": 8117, "correlation": 8118, "agg": 8119, "covar_pop": 8120, "\"returns a new :class:`~pyspark.sql.column` for the population covariance of ``col1`` and\n ``col2``.\n\n .. versionadded:: 2.0.0\n\n examples\n --------\n >>> a = [1] * 10\n >>> b = [1] * 10\n >>> df = spark.createdataframe(zip(a, b), [\"": 8121, "\"])\n >>> df.agg(covar_pop(\"": 8122, "\").alias('c')).collect()\n [row(c=0.0)]\n \"": 8123, "covariance": 8124, "covar": 8125, "covar_samp": 8126, "\"returns a new :class:`~pyspark.sql.column` for the sample covariance of ``col1`` and\n ``col2``.\n\n .. versionadded:: 2.0.0\n\n examples\n --------\n >>> a = [1] * 10\n >>> b = [1] * 10\n >>> df = spark.createdataframe(zip(a, b), [\"": 8127, "\"])\n >>> df.agg(covar_samp(\"": 8128, "countdistinct": 8129, "\"returns a new :class:`~pyspark.sql.column` for distinct count of ``col`` or ``cols``.\n\n an alias of :func:`count_distinct`, and it is encouraged to use :func:`count_distinct`\n directly.\n\n .. versionadded:: 1.3.0\n \"": 8130, "count_distinct": 8131, "encouraged": 8132, "\"returns a new :class:`column` for distinct count of ``col`` or ``cols``.\n\n .. versionadded:: 3.2.0\n\n examples\n --------\n >>> df.agg(count_distinct(df.age, df.name).alias('c')).collect()\n [row(c=2)]\n\n >>> df.agg(count_distinct(\"": 8133, "age": 8134, "\").alias('c')).collect()\n [row(c=2)]\n \"": 8135, "jc": 8136, "_to_seq": 8137, "\"age\"": 8138, "ignorenulls": 8139, "\"aggregate function: returns the first value in a group.\n\n the function by default returns the first values it sees. it will return the first non-null\n value it sees when ignorenulls is set to true. if all values are null, then null is returned.\n\n .. versionadded:: 1.3.0\n\n notes\n -----\n the function is non-deterministic because its results depends on the order of the\n rows which may be non-deterministic after a shuffle.\n \"": 8140, "ignoren": 8141, "sees": 8142, "\"\n aggregate function: indicates whether a specified column in a group by list is aggregated\n or not, returns 1 for aggregated or 0 for not aggregated in the result set.\n\n .. versionadded:: 2.0.0\n\n examples\n --------\n >>> df.cube(\"": 8143, "\").agg(grouping(\"": 8144, "\"), sum(\"": 8145, "\")).orderby(\"": 8146, "\").show()\n +-----+--------------+--------+\n | name|grouping(name)|sum(age)|\n +-----+--------------+--------+\n | null| 1| 7|\n |alice| 0| 2|\n | bob| 0| 5|\n +-----+--------------+--------+\n \"": 8147, "indicates": 8148, "aggregated": 8149, "orderby": 8150, "grouping_id": 8151, "\"\n aggregate function: returns the level of grouping, equals to\n\n (grouping(c1) << (n-1)) + (grouping(c2) << (n-2)) + ... + grouping(cn)\n\n .. versionadded:: 2.0.0\n\n notes\n -----\n the list of columns should match with grouping columns exactly, or empty (means all\n the grouping columns).\n\n examples\n --------\n >>> df.cube(\"": 8152, "\").agg(grouping_id(), sum(\"": 8153, "alice": 8154, "c2": 8155, "cn": 8156, "exactly": 8157, "means": 8158, "input_file_name": 8159, "\"creates a string column for the file name of the current spark task.\n \"": 8160, "isnan": 8161, "\"an expression that returns true iff the column is nan.\n\n .. versionadded:: 1.6.0\n\n examples\n --------\n >>> df = spark.createdataframe([(1.0, float('nan')), (float('nan'), 2.0)], (\"": 8162, "\"))\n >>> df.select(isnan(\"": 8163, "\").alias(\"": 8164, "\"), isnan(df.a).alias(\"": 8165, "\")).collect()\n [row(r1=false, r2=false), row(r1=true, r2=true)]\n \"": 8166, "iff": 8167, "nan": 8168, "'nan'": 8169, "\"r1\"": 8170, "\"r2\"": 8171, "isnull": 8172, "\"an expression that returns true iff the column is null.\n\n .. versionadded:: 1.6.0\n\n examples\n --------\n >>> df = spark.createdataframe([(1, none), (none, 2)], (\"": 8173, "\"))\n >>> df.select(isnull(\"": 8174, "\"), isnull(df.a).alias(\"": 8175, "\"aggregate function: returns the last value in a group.\n\n the function by default returns the last values it sees. it will return the last non-null\n value it sees when ignorenulls is set to true. if all values are null, then null is returned.\n\n .. versionadded:: 1.3.0\n\n notes\n -----\n the function is non-deterministic because its results depends on the order of the\n rows which may be non-deterministic after a shuffle.\n \"": 8176, "monotonically_increasing_id": 8177, "monotonically": 8178, "increasing": 8179, "guaranteed": 8180, "consecutive": 8181, "puts": 8182, "31": 8183, "33": 8184, "assumption": 8185, "less": 8186, "billion": 8187, "partitions": 8188, "nanvl": 8189, "\"returns col1 if it is not nan, or col2 if col1 is nan.\n\n both inputs should be floating point columns (:class:`doubletype` or :class:`floattype`).\n\n .. versionadded:: 1.6.0\n\n examples\n --------\n >>> df = spark.createdataframe([(1.0, float('nan')), (float('nan'), 2.0)], (\"": 8190, "\"))\n >>> df.select(nanvl(\"": 8191, "\"), nanvl(df.a, df.b).alias(\"": 8192, "\")).collect()\n [row(r1=1.0, r2=1.0), row(r1=2.0, r2=2.0)]\n \"": 8193, "floating": 8194, "doubletype": 8195, "floattype": 8196, "percentile_approx": 8197, "percentage": 8198, "accuracy": 8199, "smallest": 8200, "greatest": 8201, "positive": 8202, "controls": 8203, "approximation": 8204, "higher": 8205, "yields": 8206, "relativ": 8207, "rand": 8208, "seed": 8209, "\"generates a random column with independent and identically distributed (i.i.d.) samples\n uniformly distributed in [0.0, 1.0).\n\n .. versionadded:: 1.4.0\n\n notes\n -----\n the function is non-deterministic in general case.\n\n examples\n --------\n >>> df.withcolumn('rand', rand(seed=42) * 3).collect()\n [row(age=2, name='alice', rand=2.4052597283576684),\n row(age=5, name='bob', rand=2.3913904055683974)]\n \"": 8210, "independent": 8211, "identically": 8212, "distributed": 8213, "uniformly": 8214, "general": 8215, "'rand'": 8216, "randn": 8217, "\"generates a column with independent and identically distributed (i.i.d.) samples from\n the standard normal distribution.\n\n .. versionadded:: 1.4.0\n\n notes\n -----\n the function is non-deterministic in general case.\n\n examples\n --------\n >>> df.withcolumn('randn', randn(seed=42)).collect()\n [row(age=2, name='alice', randn=1.1027054481455365),\n row(age=5, name='bob', randn=0.7400395449950132)]\n \"": 8218, "'randn'": 8219, "\"\n round the given value to `scale` decimal places using half_up rounding mode if `scale` >= 0\n or at integral part when `scale` < 0.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([(2.5,)], ['a']).select(round('a', 0).alias('r')).collect()\n [row(r=3.0)]\n \"": 8220, "half_up": 8221, "rounding": 8222, "0.": 8223, "'r'": 8224, "bround": 8225, "\"\n round the given value to `scale` decimal places using half_even rounding mode if `scale` >= 0\n or at integral part when `scale` < 0.\n\n .. versionadded:: 2.0.0\n\n examples\n --------\n >>> spark.createdataframe([(2.5,)], ['a']).select(bround('a', 0).alias('r')).collect()\n [row(r=2.0)]\n \"": 8226, "half_even": 8227, "shiftleft": 8228, "numbits": 8229, "\"shift the given value numbits left.\n\n .. versionadded:: 1.5.0\n\n .. deprecated:: 3.2.0\n use :func:`shiftleft` instead.\n \"": 8230, "\"deprecated in 3.2, use shiftleft instead.\"": 8231, "shift": 8232, "\"shift the given value numbits left.\n\n .. versionadded:: 3.2.0\n\n examples\n --------\n >>> spark.createdataframe([(21,)], ['a']).select(shiftleft('a', 1).alias('r')).collect()\n [row(r=42)]\n \"": 8233, "21": 8234, "shiftright": 8235, "\"(signed) shift the given value numbits right.\n\n .. versionadded:: 1.5.0\n\n .. deprecated:: 3.2.0\n use :func:`shiftright` instead.\n \"": 8236, "\"deprecated in 3.2, use shiftright instead.\"": 8237, "\"(signed) shift the given value numbits right.\n\n .. versionadded:: 3.2.0\n\n examples\n --------\n >>> spark.createdataframe([(42,)], ['a']).select(shiftright('a', 1).alias('r')).collect()\n [row(r=21)]\n \"": 8238, "shiftrightunsigned": 8239, "\"unsigned shift the given value numbits right.\n\n .. versionadded:: 1.5.0\n\n .. deprecated:: 3.2.0\n use :func:`shiftrightunsigned` instead.\n \"": 8240, "\"deprecated in 3.2, use shiftrightunsigned instead.\"": 8241, "unsigned": 8242, "\"unsigned shift the given value numbits right.\n\n .. versionadded:: 3.2.0\n\n examples\n --------\n >>> df = spark.createdataframe([(-42,)], ['a'])\n >>> df.select(shiftrightunsigned('a', 1).alias('r')).collect()\n [row(r=9223372036854775787)]\n \"": 8243, "9223372036854775787": 8244, "spark_partition_id": 8245, "\"a column for partition id.\n\n .. versionadded:: 1.6.0\n\n notes\n -----\n this is non deterministic because it depends on data partitioning and task scheduling.\n\n examples\n --------\n >>> df.repartition(1).select(spark_partition_id().alias(\"": 8246, "\")).collect()\n [row(pid=0), row(pid=0)]\n \"": 8247, "partitioning": 8248, "scheduling": 8249, "repartition": 8250, "\"pid\"": 8251, "\"creates a new struct column.\n\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n cols : list, set, str or :class:`~pyspark.sql.column`\n column names or :class:`~pyspark.sql.column`\\\\s to contain in the output struct.\n\n examples\n --------\n >>> df.select(struct('age', 'name').alias(\"": 8252, "\")).collect()\n [row(struct=row(age=2, name='alice')), row(struct=row(age=5, name='bob'))]\n >>> df.select(struct([df.age, df.name]).alias(\"": 8253, "\")).collect()\n [row(struct=row(age=2, name='alice')), row(struct=row(age=5, name='bob'))]\n \"": 8254, "'age'": 8255, "\"struct\"": 8256, "ro": 8257, "\"\n returns the greatest value of the list of column names, skipping null values.\n this function takes at least 2 parameters. it will return null iff all parameters are null.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([(1, 4, 3)], ['a', 'b', 'c'])\n >>> df.select(greatest(df.a, df.b, df.c).alias(\"": 8258, "\")).collect()\n [row(greatest=4)]\n \"": 8259, "\"greatest should take at least two columns\"": 8260, "functi": 8261, "\"\n returns the least value of the list of column names, skipping null values.\n this function takes at least 2 parameters. it will return null iff all parameters are null.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([(1, 4, 3)], ['a', 'b', 'c'])\n >>> df.select(least(df.a, df.b, df.c).alias(\"": 8262, "\")).collect()\n [row(least=1)]\n \"": 8263, "\"least should take at least two columns\"": 8264, "\"evaluates a list of conditions and returns one of multiple possible result expressions.\n if :func:`pyspark.sql.column.otherwise` is not invoked, none is returned for unmatched\n conditions.\n\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n condition : :class:`~pyspark.sql.column`\n a boolean :class:`~pyspark.sql.column` expression.\n value :\n a literal value, or a :class:`~pyspark.sql.column` expression.\n\n >>> df.select(when(df['age'] == 2, 3).otherwise(4).alias(\"": 8265, "evaluates": 8266, "conditions": 8267, "expressions": 8268, "unmatched": 8269, "log2": 8270, "\"returns the base-2 logarithm of the argument.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([(4,)], ['a']).select(log2('a').alias('log2')).collect()\n [row(log2=2.0)]\n \"": 8271, "'log2'": 8272, "conv": 8273, "frombase": 8274, "tobase": 8275, "\"\n convert a number in a string column from one base to another.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([(\"": 8276, "010101": 8277, "\",)], ['n'])\n >>> df.select(conv(df.n, 2, 16).alias('hex')).collect()\n [row(hex='15')]\n \"": 8278, "\"010101\"": 8279, "'hex'": 8280, "'15'": 8281, "factorial": 8282, "\"\n computes the factorial of the given value.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([(5,)], ['n'])\n >>> df.select(factorial(df.n).alias('f')).collect()\n [row(f=120)]\n \"": 8283, "'f'": 8284, "lead": 8285, "sq": 8286, "nth_value": 8287, "th": 8288, "nth": 8289, "ntile": 8290, "\"\n window function: returns the ntile group id (from 1 to `n` inclusive)\n in an ordered window partition. for example, if `n` is 4, the first\n quarter of the rows will get value 1, the second quarter will get 2,\n the third quarter will get 3, and the last quarter will get 4.\n\n this is equivalent to the ntile function in sql.\n\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n n : int\n an integer\n \"": 8291, "inclusive": 8292, "quarter": 8293, "nti": 8294, "current_date": 8295, "\"\n returns the current date at the start of query evaluation as a :class:`datetype` column.\n all calls of current_date within the same query return the same value.\n \"": 8296, "evaluation": 8297, "datetype": 8298, "current_timestamp": 8299, "\"\n returns the current timestamp at the start of query evaluation as a :class:`timestamptype`\n column. all calls of current_timestamp within the same query return the same value.\n \"": 8300, "timestamptype": 8301, "date_format": 8302, "yyyy": 8303, "'18.03.1993'": 8304, "_datetime": 8305, "apache": 8306, "latest": 8307, "ht": 8308, "\"\n extract the year of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(year('dt').alias('year')).collect()\n [row(year=2015)]\n \"": 8309, "'2015-04-08'": 8310, "'dt'": 8311, "'year'": 8312, "2015": 8313, "\"\n extract the quarter of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(quarter('dt').alias('quarter')).collect()\n [row(quarter=2)]\n \"": 8314, "'quarter'": 8315, "\"\n extract the month of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(month('dt').alias('month')).collect()\n [row(month=4)]\n \"": 8316, "'month'": 8317, "dayofweek": 8318, "\"\n extract the day of the week of a given date as integer.\n\n .. versionadded:: 2.3.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(dayofweek('dt').alias('day')).collect()\n [row(day=4)]\n \"": 8319, "'day'": 8320, "dayofmonth": 8321, "\"\n extract the day of the month of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(dayofmonth('dt').alias('day')).collect()\n [row(day=8)]\n \"": 8322, "dayofyear": 8323, "\"\n extract the day of the year of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(dayofyear('dt').alias('day')).collect()\n [row(day=98)]\n \"": 8324, "98": 8325, "\"\n extract the hours of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08 13:08:15',)], ['ts'])\n >>> df.select(hour('ts').alias('hour')).collect()\n [row(hour=13)]\n \"": 8326, "'2015-04-08 13:08:15'": 8327, "'ts'": 8328, "'hour'": 8329, "13": 8330, "\"\n extract the minutes of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08 13:08:15',)], ['ts'])\n >>> df.select(minute('ts').alias('minute')).collect()\n [row(minute=8)]\n \"": 8331, "'minute'": 8332, "\"\n extract the seconds of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08 13:08:15',)], ['ts'])\n >>> df.select(second('ts').alias('second')).collect()\n [row(second=15)]\n \"": 8333, "'second'": 8334, "weekofyear": 8335, "\"\n extract the week number of a given date as integer.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(weekofyear(df.dt).alias('week')).collect()\n [row(week=15)]\n \"": 8336, "'week'": 8337, "date_add": 8338, "\"\n returns the date that is `days` days after `start`\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(date_add(df.dt, 1).alias('next_date')).collect()\n [row(next_date=datetime.date(2015, 4, 9))]\n \"": 8339, "'next_date'": 8340, "next_date": 8341, "date_sub": 8342, "\"\n returns the date that is `days` days before `start`\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(date_sub(df.dt, 1).alias('prev_date')).collect()\n [row(prev_date=datetime.date(2015, 4, 7))]\n \"": 8343, "'prev_date'": 8344, "prev_date": 8345, "datediff": 8346, "\"\n returns the number of days from `start` to `end`.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08','2015-05-10')], ['d1', 'd2'])\n >>> df.select(datediff(df.d2, df.d1).alias('diff')).collect()\n [row(diff=32)]\n \"": 8347, "'2015-05-10'": 8348, "'d1'": 8349, "'d2'": 8350, "d2": 8351, "d1": 8352, "'diff'": 8353, "diff": 8354, "add_months": 8355, "\"\n returns the date that is `months` months after `start`\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-04-08',)], ['dt'])\n >>> df.select(add_months(df.dt, 1).alias('next_month')).collect()\n [row(next_month=datetime.date(2015, 5, 8))]\n \"": 8356, "'next_month'": 8357, "next_month": 8358, "months_between": 8359, "date1": 8360, "date2": 8361, "roundoff": 8362, "dates": 8363, "rounded": 8364, "'1997-02-28 10:30:00'": 8365, "'1996-10-30'": 8366, "'date1'": 8367, "'date2'": 8368, "to_date": 8369, "\"converts a :class:`~pyspark.sql.column` into :class:`pyspark.sql.types.datetype`\n using the optionally specified format. specify formats according to `datetime pattern`_.\n by default, it follows casting rules to :class:`pyspark.sql.types.datetype` if the format\n is omitted. equivalent to ``col.cast(\"": 8370, "'t'": 8371, "to_d": 8372, "formats": 8373, "follows": 8374, "rules": 8375, "omitted": 8376, "da": 8377, "to_timestamp": 8378, "\"converts a :class:`~pyspark.sql.column` into :class:`pyspark.sql.types.timestamptype`\n using the optionally specified format. specify formats according to `datetime pattern`_.\n by default, it follows casting rules to :class:`pyspark.sql.types.timestamptype` if the format\n is omitted. equivalent to ``col.cast(\"": 8379, "trunc": 8380, "\"\n returns date truncated to the unit specified by the format.\n\n .. versionadded:: 1.5.0\n\n parameters\n ----------\n date : :class:`~pyspark.sql.column` or str\n format : str\n 'year', 'yyyy', 'yy' or 'month', 'mon', 'mm'\n\n examples\n --------\n >>> df = spark.createdataframe([('1997-02-28',)], ['d'])\n >>> df.select(trunc(df.d, 'year').alias('year')).collect()\n [row(year=datetime.date(1997, 1, 1))]\n >>> df.select(trunc(df.d, 'mon').alias('month')).collect()\n [row(month=datetime.date(1997, 2, 1))]\n \"": 8381, "sparkconte": 8382, "truncated": 8383, "'yyyy'": 8384, "'yy'": 8385, "'mon'": 8386, "'mm'": 8387, "'1997-02-28'": 8388, "sele": 8389, "date_trunc": 8390, "'1997-02-28 05:02:11'": 8391, "1997": 8392, "date_": 8393, "next_day": 8394, "\"\n returns the first date which is later than the value of the date column.\n\n day of the week parameter is case insensitive, and accepts:\n \"": 8395, "mon": 8396, "tue": 8397, "wed": 8398, "thu": 8399, "fri": 8400, "\".\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('2015-07-27',)], ['d'])\n >>> df.select(next_day(df.d, 'sun').alias('date')).collect()\n [row(date=datetime.date(2015, 8, 2))]\n \"": 8401, "\"mon\"": 8402, "\"tue\"": 8403, "\"wed\"": 8404, "\"thu\"": 8405, "\"fri\"": 8406, "\"sat\"": 8407, "\"sun\"": 8408, "'2015-07-27'": 8409, "selec": 8410, "last_day": 8411, "\"\n returns the last day of the month which the given date belongs to.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('1997-02-10',)], ['d'])\n >>> df.select(last_day(df.d).alias('date')).collect()\n [row(date=datetime.date(1997, 2, 28))]\n \"": 8412, "belongs": 8413, "'1997-02-10'": 8414, "'date'": 8415, "28": 8416, "from_unixtime": 8417, "\"yyyy-mm-dd hh:mm:ss\"": 8418, "\"\n converts the number of seconds from unix epoch (1970-01-01 00:00:00 utc) to a string\n representing the timestamp of that moment in the current system time zone in the given\n format.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.conf.set(\"": 8419, "timezone": 8420, "america": 8421, "los_angeles": 8422, "\")\n >>> time_df = spark.createdataframe([(1428476400,)], ['unix_time'])\n >>> time_df.select(from_unixtime('unix_time').alias('ts')).collect()\n [row(ts='2015-04-08 00:00:00')]\n >>> spark.conf.unset(\"": 8423, "unix": 8424, "utc": 8425, "moment": 8426, "zone": 8427, "\"spark.sql.session.timezone\"": 8428, "\"america/los_angeles\"": 8429, "unixtime": 8430, "unix_timestamp": 8431, "'yyyy-mm-dd hh:mm:ss'": 8432, "\"\n convert time string with given pattern ('yyyy-mm-dd hh:mm:ss', by default)\n to unix time stamp (in seconds), using the default timezone and the default\n locale, return null if fail.\n\n if `timestamp` is none, then it returns current timestamp.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.conf.set(\"": 8433, "time_df": 8434, "'yyyy-mm-dd'": 8435, "unix_t": 8436, "stamp": 8437, "from_utc_timestamp": 8438, "tz": 8439, "agnostic": 8440, "interprets": 8441, "renders": 8442, "confusing": 8443, "stri": 8444, "micr": 8445, "to_utc_timestamp": 8446, "micro": 8447, "timestamp_seconds": 8448, "\"\n .. versionadded:: 3.1.0\n\n examples\n --------\n >>> from pyspark.sql.functions import timestamp_seconds\n >>> spark.conf.set(\"": 8449, "\")\n >>> time_df = spark.createdataframe([(1230219000,)], ['unix_time'])\n >>> time_df.select(timestamp_seconds(time_df.unix_time).alias('ts')).show()\n +-------------------+\n | ts|\n +-------------------+\n |2008-12-25 07:30:00|\n +-------------------+\n >>> spark.conf.unset(\"": 8450, "\")\n \"": 8451, "_acti": 8452, "1230219000": 8453, "'unix_time'": 8454, "unix_time": 8455, "timecolumn": 8456, "windowduration": 8457, "slideduration": 8458, "starttime": 8459, "bucketize": 8460, "ends": 8461, "05": 8462, "precision": 8463, "durations": 8464, "'1 second'": 8465, "'1 day 12 hours'": 8466, "'2 minutes'": 8467, "crc32": 8468, "\"\n calculates the cyclic redundancy check value (crc32) of a binary column and\n returns the value as a bigint.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('abc',)], ['a']).select(crc32('a').alias('crc32')).collect()\n [row(crc32=2743272264)]\n \"": 8469, "calculates": 8470, "redundancy": 8471, "bigint": 8472, "'abc'": 8473, "'crc32'": 8474, "2743272264": 8475, "\"calculates the md5 digest and returns the value as a 32 character hex string.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('abc',)], ['a']).select(md5('a').alias('hash')).collect()\n [row(hash='902fbdd2b1df0c4f70b4a5d23525e932')]\n \"": 8476, "digest": 8477, "character": 8478, "'hash'": 8479, "'902fbdd2b1df0c4f70b4a5d23525e932'": 8480, "\"returns the hex string result of sha-1.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('abc',)], ['a']).select(sha1('a').alias('hash')).collect()\n [row(hash='3c01bdbb26f358bab27f267924aa2c9a03fcfdb8')]\n \"": 8481, "sha": 8482, "'3c01bdbb26f358bab27f267924aa2c9a03fcfdb8'": 8483, "sha2": 8484, "224": 8485, "256": 8486, "384": 8487, "digests": 8488, "'3bc51062973c458d5a6f2d8d64a023246354ad7e064b1e4e009ec8a0699a3043'": 8489, "cd9fb1e148ccd8442e5aa74904cc73bf6fb54d1d54d333bd596aa9bb4bb4e": 8490, "dige": 8491, "\"calculates the hash code of given columns, and returns the result as an int column.\n\n .. versionadded:: 2.0.0\n\n examples\n --------\n >>> spark.createdataframe([('abc',)], ['a']).select(hash('a').alias('hash')).collect()\n [row(hash=-757602832)]\n \"": 8492, "757602832": 8493, "xxhash64": 8494, "\"calculates the hash code of given columns using the 64-bit variant of the xxhash algorithm,\n and returns the result as a long column.\n\n .. versionadded:: 3.0.0\n\n examples\n --------\n >>> spark.createdataframe([('abc',)], ['a']).select(xxhash64('a').alias('hash')).collect()\n [row(hash=4105715581806190027)]\n \"": 8495, "variant": 8496, "xxhash": 8497, "algorithm": 8498, "4105715581806190027": 8499, "assert_true": 8500, "errmsg": 8501, "throws": 8502, "erro": 8503, "createdat": 8504, "raise_error": 8505, "\"\n throws an exception with the provided error message.\n \"": 8506, "\"errmsg should be a column or a str, got {}\"": 8507, "_create_column_from_literal": 8508, "\"\n converts a string expression to upper case.\n \"": 8509, "\"upper\"": 8510, "\"\n converts a string expression to lower case.\n \"": 8511, "\"lower\"": 8512, "\"\n computes the numeric value of the first character of the string column.\n \"": 8513, "\"ascii\"": 8514, "base64": 8515, "\"\n computes the base64 encoding of a binary column and returns it as a string column.\n \"": 8516, "\"base64\"": 8517, "unbase64": 8518, "\"\n decodes a base64 encoded string column and returns it as a binary column.\n \"": 8519, "\"unbase64\"": 8520, "decodes": 8521, "encoded": 8522, "ltrim": 8523, "\"\n trim the spaces from left end for the specified string value.\n \"": 8524, "\"ltrim\"": 8525, "trim": 8526, "rtrim": 8527, "\"\n trim the spaces from right end for the specified string value.\n \"": 8528, "\"rtrim\"": 8529, "\"\n trim the spaces from both ends for the specified string column.\n \"": 8530, "\"trim\"": 8531, "concat_ws": 8532, "\"\n concatenates multiple input string columns together into a single string column,\n using the given separator.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('abcd','123')], ['s', 'd'])\n >>> df.select(concat_ws('-', df.s, df.d).alias('s')).collect()\n [row(s='abcd-123')]\n \"": 8533, "concatenates": 8534, "'abcd'": 8535, "'123'": 8536, "'abcd-123'": 8537, "concat": 8538, "ws": 8539, "format_number": 8540, "\"\n formats the number x to a format like '#,--#,--#.--', rounded to d decimal places\n with half_even round mode, and returns the result as a string.\n\n .. versionadded:: 1.5.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n the column name of the numeric value to be formatted\n d : int\n the n decimal places\n\n >>> spark.createdataframe([(5,)], ['a']).select(format_number('a', 4).alias('v')).collect()\n [row(v='5.0000')]\n \"": 8541, "functio": 8542, "'#,--#,--#.--'": 8543, "formatted": 8544, "format_string": 8545, "\"\n formats the arguments in printf-style and returns the result as a string column.\n\n .. versionadded:: 1.5.0\n\n parameters\n ----------\n format : str\n string that can contain embedded format tags and used as result column's value\n cols : :class:`~pyspark.sql.column` or str\n column names or :class:`~pyspark.sql.column`\\\\s to be used in formatting\n\n examples\n --------\n >>> df = spark.createdataframe([(5, \"": 8546, "hello": 8547, "'%d %s'": 8548, "he": 8549, "printf": 8550, "embedded": 8551, "pys": 8552, "instr": 8553, "substr": 8554, "\"\n locate the position of the first occurrence of substr column in the given string.\n returns null if either of the arguments are null.\n\n .. versionadded:: 1.5.0\n\n notes\n -----\n the position is not zero based, but 1 based index. returns 0 if substr\n could not be found in str.\n\n >>> df = spark.createdataframe([('abcd',)], ['s',])\n >>> df.select(instr(df.s, 'b').alias('s')).collect()\n [row(s=2)]\n \"": 8555, "position": 8556, "occurrence": 8557, "overlay": 8558, "\"\n overlay the specified portion of `src` with `replace`,\n starting from byte position `pos` of `src` and proceeding for `len` bytes.\n\n .. versionadded:: 3.0.0\n\n examples\n --------\n >>> df = spark.createdataframe([(\"": 8559, "spark_sql": 8560, "\")], (\"": 8561, "\"))\n >>> df.select(overlay(\"": 8562, "\", 7).alias(\"": 8563, "overlayed": 8564, "\")).show()\n +----------+\n | overlayed|\n +----------+\n |spark_core|\n +----------+\n \"": 8565, "portion": 8566, "proceeding": 8567, "\"spark_sql\"": 8568, "\"core\"": 8569, "\"y\"": 8570, "\"overlayed\"": 8571, "sho": 8572, "language": 8573, "splits": 8574, "words": 8575, "'language'": 8576, "'country'": 8577, "substring": 8578, "\"\n substring starts at `pos` and is of length `len` when str is string type or\n returns the slice of byte array that starts at `pos` in byte and is of length `len`\n when str is binary type.\n\n .. versionadded:: 1.5.0\n\n notes\n -----\n the position is not zero based, but 1 based index.\n\n examples\n --------\n >>> df = spark.createdataframe([('abcd',)], ['s',])\n >>> df.select(substring(df.s, 1, 2).alias('s')).collect()\n [row(s='ab')]\n \"": 8579, "substring_index": 8580, "delim": 8581, "occurrences": 8582, "delimiter": 8583, "performs": 8584, "sensitive": 8585, "searching": 8586, "'a.b.c.d'": 8587, "substring_i": 8588, "levenshtein": 8589, "\"computes the levenshtein distance of the two given strings.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df0 = spark.createdataframe([('kitten', 'sitting',)], ['l', 'r'])\n >>> df0.select(levenshtein('l', 'r').alias('d')).collect()\n [row(d=3)]\n \"": 8590, "distance": 8591, "df0": 8592, "'kitten'": 8593, "'sitting'": 8594, "'l'": 8595, "stringtype": 8596, "positi": 8597, "lpad": 8598, "pad": 8599, "\"\n left-pad the string column to width `len` with `pad`.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('abcd',)], ['s',])\n >>> df.select(lpad(df.s, 6, '#').alias('s')).collect()\n [row(s='##abcd')]\n \"": 8600, "'#'": 8601, "'##abcd'": 8602, "rpad": 8603, "\"\n right-pad the string column to width `len` with `pad`.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('abcd',)], ['s',])\n >>> df.select(rpad(df.s, 6, '#').alias('s')).collect()\n [row(s='abcd##')]\n \"": 8604, "'abcd##'": 8605, "repeat": 8606, "\"\n repeats a string column n times, and returns it as a new string column.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('ab',)], ['s',])\n >>> df.select(repeat(df.s, 3).alias('s')).collect()\n [row(s='ababab')]\n \"": 8607, "repeats": 8608, "'ab'": 8609, "'ababab'": 8610, "regexp_extract": 8611, "matched": 8612, "'100-200'": 8613, "'str'": 8614, "'(\\d+)-(\\d+)'": 8615, "'100'": 8616, "'(\\d+)'": 8617, "regexp_": 8618, "regexp": 8619, "regexp_replace": 8620, "\"replace all substrings of the specified string value that match regexp with rep.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('100-200',)], ['str'])\n >>> df.select(regexp_replace('str', r'(\\d+)', '--').alias('d')).collect()\n [row(d='-----')]\n \"": 8621, "substrings": 8622, "rep": 8623, "'--'": 8624, "'-----'": 8625, "initcap": 8626, "\"translate the first letter of each word to upper case in the sentence.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('ab cd',)], ['a']).select(initcap(\"": 8627, "\").alias('v')).collect()\n [row(v='ab cd')]\n \"": 8628, "translate": 8629, "'ab cd'": 8630, "soundex": 8631, "\"\n returns the soundex encoding for a string\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([(\"": 8632, "peters": 8633, "\",),(\"": 8634, "uhrbach": 8635, "\",)], ['name'])\n >>> df.select(soundex(df.name).alias(\"": 8636, "\")).collect()\n [row(soundex='p362'), row(soundex='u612')]\n \"": 8637, "\"peters\"": 8638, "\"uhrbach\"": 8639, "\"soundex\"": 8640, "'p362'": 8641, "'u612'": 8642, "\"returns the string representation of the binary value of the given column.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df.select(bin(df.age).alias('c')).collect()\n [row(c='10'), row(c='101')]\n \"": 8643, "'10'": 8644, "'101'": 8645, "\"computes hex value of the given column, which could be :class:`pyspark.sql.types.stringtype`,\n :class:`pyspark.sql.types.binarytype`, :class:`pyspark.sql.types.integertype` or\n :class:`pyspark.sql.types.longtype`.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('abc', 3)], ['a', 'b']).select(hex('a'), hex('b')).collect()\n [row(hex(a)='414243', hex(b)='3')]\n \"": 8646, "binarytype": 8647, "integertype": 8648, "longtype": 8649, "unhex": 8650, "\"inverse of hex. interprets each pair of characters as a hexadecimal number\n and converts to the byte representation of number.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('414243',)], ['a']).select(unhex('a')).collect()\n [row(unhex(a)=bytearray(b'abc'))]\n \"": 8651, "pair": 8652, "hexadecimal": 8653, "'414243'": 8654, "bytearray": 8655, "\"computes the character length of string data or number of bytes of binary data.\n the length of character data includes the trailing spaces. the length of binary data\n includes binary zeros.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('abc ',)], ['a']).select(length('a').alias('length')).collect()\n [row(length=4)]\n \"": 8656, "'abc '": 8657, "ali": 8658, "srccol": 8659, "\"a function translate any character in the `srccol` by a character in `matching`.\n the characters in `replace` is corresponding to the characters in `matching`.\n the translate will happen when any character in the string matching with the character\n in the `matching`.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> spark.createdataframe([('translate',)], ['a']).select(translate('a', \"": 8660, "rnlt": 8661, "123": 8662, "\") \\\\\n ... .alias('r')).collect()\n [row(r='1a2s3ae')]\n \"": 8663, "create_map": 8664, "\"creates a new map column.\n\n .. versionadded:: 2.0.0\n\n parameters\n ----------\n cols : :class:`~pyspark.sql.column` or str\n column names or :class:`~pyspark.sql.column`\\\\s that are\n grouped as key-value pairs, e.g. (key1, value1, key2, value2, ...).\n\n examples\n --------\n >>> df.select(create_map('name', 'age').alias(\"": 8665, "\")).collect()\n [row(map={'alice': 2}), row(map={'bob': 5})]\n >>> df.select(create_map([df.name, df.age]).alias(\"": 8666, "\")).collect()\n [row(map={'alice': 2}), row(map={'bob': 5})]\n \"": 8667, "key1": 8668, "value1": 8669, "key2": 8670, "value2": 8671, "map_from_arrays": 8672, "\"creates a new map from two arrays.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col1 : :class:`~pyspark.sql.column` or str\n name of column containing a set of keys. all elements should not be null\n col2 : :class:`~pyspark.sql.column` or str\n name of column containing a set of values\n\n examples\n --------\n >>> df = spark.createdataframe([([2, 5], ['a', 'b'])], ['k', 'v'])\n >>> df.select(map_from_arrays(df.k, df.v).alias(\"": 8673, "\"creates a new array column.\n\n .. versionadded:: 1.4.0\n\n parameters\n ----------\n cols : :class:`~pyspark.sql.column` or str\n column names or :class:`~pyspark.sql.column`\\\\s that have\n the same data type.\n\n examples\n --------\n >>> df.select(array('age', 'age').alias(\"": 8674, "arr": 8675, "\")).collect()\n [row(arr=[2, 2]), row(arr=[5, 5])]\n >>> df.select(array([df.age, df.age]).alias(\"": 8676, "\")).collect()\n [row(arr=[2, 2]), row(arr=[5, 5])]\n \"": 8677, "\"arr\"": 8678, "array_contains": 8679, "\"\n collection function: returns null if the array is null, true if the array contains the\n given value, and false otherwise.\n\n .. versionadded:: 1.5.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column containing array\n value :\n value or column to check for in array\n\n examples\n --------\n >>> df = spark.createdataframe([([\"": 8680, "\"],), ([],)], ['data'])\n >>> df.select(array_contains(df.data, \"": 8681, "collection": 8682, "arrays_overlap": 8683, "\"\n collection function: returns true if the arrays contain any common non-null element; if not,\n returns null if both the arrays are non-empty and any of them contains a null element; returns\n false otherwise.\n\n .. versionadded:: 2.4.0\n\n examples\n --------\n >>> df = spark.createdataframe([([\"": 8684, "\"], [\"": 8685, "\"]), ([\"": 8686, "\"])], ['x', 'y'])\n >>> df.select(arrays_overlap(df.x, df.y).alias(\"": 8687, "\")).collect()\n [row(overlap=true), row(overlap=false)]\n \"": 8688, "sli": 8689, "array_join": 8690, "null_replacement": 8691, "\"\n concatenates the elements of `column` using the `delimiter`. null values are replaced with\n `null_replacement` if set, otherwise they are ignored.\n\n .. versionadded:: 2.4.0\n\n examples\n --------\n >>> df = spark.createdataframe([([\"": 8692, "\"],), ([\"": 8693, "\", none],)], ['data'])\n >>> df.select(array_join(df.data, \"": 8694, "joined": 8695, "\")).collect()\n [row(joined='a,b,c'), row(joined='a')]\n >>> df.select(array_join(df.data, \"": 8696, "\")).collect()\n [row(joined='a,b,c'), row(joined='a,null')]\n \"": 8697, "replaced": 8698, "\"\n concatenates multiple input columns together into a single column.\n the function works with strings, binary and compatible array columns.\n\n .. versionadded:: 1.5.0\n\n examples\n --------\n >>> df = spark.createdataframe([('abcd','123')], ['s', 'd'])\n >>> df.select(concat(df.s, df.d).alias('s')).collect()\n [row(s='abcd123')]\n\n >>> df = spark.createdataframe([([1, 2], [3, 4], [5]), ([1, 2], none, [3])], ['a', 'b', 'c'])\n >>> df.select(concat(df.a, df.b, df.c).alias(\"": 8699, "\")).collect()\n [row(arr=[1, 2, 3, 4, 5]), row(arr=none)]\n \"": 8700, "array_position": 8701, "\"\n collection function: locates the position of the first occurrence of the given value\n in the given array. returns null if either of the arguments are null.\n\n .. versionadded:: 2.4.0\n\n notes\n -----\n the position is not zero based, but 1 based index. returns 0 if the given\n value could not be found in the array.\n\n examples\n --------\n >>> df = spark.createdataframe([([\"": 8702, "\"],), ([],)], ['data'])\n >>> df.select(array_position(df.data, \"": 8703, "locates": 8704, "element_at": 8705, "extraction": 8706, "\"\n collection function: returns element of array at given index in extraction if col is array.\n returns value for the given key in extraction if col is map.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column containing array or map\n extraction :\n index to check for in array or key to check for in map\n\n notes\n -----\n the position is not zero based, but 1 based index.\n\n examples\n --------\n >>> df = spark.createdataframe([([\"": 8707, "array_remove": 8708, "\"\n collection function: remove all elements that equal to element from the given array.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column containing array\n element :\n element to be removed from the array\n\n examples\n --------\n >>> df = spark.createdataframe([([1, 2, 3, 1, 1],), ([],)], ['data'])\n >>> df.select(array_remove(df.data, 1)).collect()\n [row(array_remove(data, 1)=[2, 3]), row(array_remove(data, 1)=[])]\n \"": 8709, "_active_sp": 8710, "crea": 8711, "array_distinct": 8712, "\"\n collection function: removes duplicate values from the array.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> df = spark.createdataframe([([1, 2, 3, 2],), ([4, 5, 5, 4],)], ['data'])\n >>> df.select(array_distinct(df.data)).collect()\n [row(array_distinct(data)=[1, 2, 3]), row(array_distinct(data)=[4, 5])]\n \"": 8713, "removes": 8714, "array_intersect": 8715, "\"\n collection function: returns an array of the elements in the intersection of col1 and col2,\n without duplicates.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col1 : :class:`~pyspark.sql.column` or str\n name of column containing array\n col2 : :class:`~pyspark.sql.column` or str\n name of column containing array\n\n examples\n --------\n >>> from pyspark.sql import row\n >>> df = spark.createdataframe([row(c1=[\"": 8716, "\"], c2=[\"": 8717, "collec": 8718, "intersection": 8719, "intersect": 8720, "array_union": 8721, "\"\n collection function: returns an array of the elements in the union of col1 and col2,\n without duplicates.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col1 : :class:`~pyspark.sql.column` or str\n name of column containing array\n col2 : :class:`~pyspark.sql.column` or str\n name of column containing array\n\n examples\n --------\n >>> from pyspark.sql import row\n >>> df = spark.createdataframe([row(c1=[\"": 8722, "ar": 8723, "contai": 8724, "array_except": 8725, "\"\n collection function: returns an array of the elements in col1 but not in col2,\n without duplicates.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col1 : :class:`~pyspark.sql.column` or str\n name of column containing array\n col2 : :class:`~pyspark.sql.column` or str\n name of column containing array\n\n examples\n --------\n >>> from pyspark.sql import row\n >>> df = spark.createdataframe([row(c1=[\"": 8726, "array_": 8727, "explode": 8728, "\"\n returns a new row for each element in the given array or map.\n uses the default column name `col` for elements in the array and\n `key` and `value` for elements in the map unless specified otherwise.\n\n .. versionadded:: 1.4.0\n\n examples\n --------\n >>> from pyspark.sql import row\n >>> edf = spark.createdataframe([row(a=1, intlist=[1,2,3], mapfield={\"": 8729, "\"})])\n >>> edf.select(explode(edf.intlist).alias(\"": 8730, "anint": 8731, "\")).collect()\n [row(anint=1), row(anint=2), row(anint=3)]\n\n >>> edf.select(explode(edf.mapfield).alias(\"": 8732, "edf": 8733, "createdataf": 8734, "posexplode": 8735, "\"\n returns a new row for each element with position in the given array or map.\n uses the default column name `pos` for position, and `col` for elements in the\n array and `key` and `value` for elements in the map unless specified otherwise.\n\n .. versionadded:: 2.1.0\n\n examples\n --------\n >>> from pyspark.sql import row\n >>> edf = spark.createdataframe([row(a=1, intlist=[1,2,3], mapfield={\"": 8736, "intlist": 8737, "explode_outer": 8738, "\"\n returns a new row for each element in the given array or map.\n unlike explode, if the array/map is null or empty then null is produced.\n uses the default column name `col` for elements in the array and\n `key` and `value` for elements in the map unless specified otherwise.\n\n .. versionadded:: 2.3.0\n\n examples\n --------\n >>> df = spark.createdataframe(\n ... [(1, [\"": 8739, "\"], {\"": 8740, "\": 1.0}), (2, [], {}), (3, none, none)],\n ... (\"": 8741, "an_array": 8742, "a_map": 8743, "\")\n ... )\n >>> df.select(\"": 8744, "\", explode_outer(\"": 8745, "unlike": 8746, "exam": 8747, "outer": 8748, "posexplode_outer": 8749, "\"\n returns a new row for each element with position in the given array or map.\n unlike posexplode, if the array/map is null or empty then the row (null, null) is produced.\n uses the default column name `pos` for position, and `col` for elements in the\n array and `key` and `value` for elements in the map unless specified otherwise.\n\n .. versionadded:: 2.3.0\n\n examples\n --------\n >>> df = spark.createdataframe(\n ... [(1, [\"": 8750, "get_json_object": 8751, "\"\n extracts json object from a json string based on json path specified, and returns json string\n of the extracted json object. it will return null if the input json string is invalid.\n\n .. versionadded:: 1.6.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n string column in json format\n path : str\n path to the json object to extract\n\n examples\n --------\n >>> data = [(\"": 8752, "\", '''{\"": 8753, "\"}'''), (\"": 8754, "value12": 8755, "\"}''')]\n >>> df = spark.createdataframe(data, (\"": 8756, "extracted": 8757, "json_tuple": 8758, "\"creates a new row for a json column according to the given field names.\n\n .. versionadded:: 1.6.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n string column in json format\n fields : str\n fields to extract\n\n examples\n --------\n >>> data = [(\"": 8759, "jstring": 8760, "'f1'": 8761, "'f2'": 8762, "c0": 8763, "'value1'": 8764, "'value2'": 8765, "\"1\"": 8766, "\"f1\"": 8767, "\"value1\"": 8768, "\"f2\"": 8769, "from_json": 8770, "maptype": 8771, "structtype": 8772, "arraytype": 8773, "unparseable": 8774, "datatype": 8775, "versionchanged": 8776, "to_json": 8777, "unsupported": 8778, "converting": 8779, "datasource": 8780, "schema_of_json": 8781, "infers": 8782, "ddl": 8783, "foldable": 8784, "versionchan": 8785, "optio": 8786, "schema_of_csv": 8787, "csv": 8788, "to_csv": 8789, "cont": 8790, "\"\n collection function: returns the length of the array or map stored in the column.\n\n .. versionadded:: 1.5.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> df = spark.createdataframe([([1, 2, 3],),([1],),([],)], ['data'])\n >>> df.select(size(df.data)).collect()\n [row(size(data)=3), row(size(data)=1), row(size(data)=0)]\n \"": 8791, "array_min": 8792, "\"\n collection function: returns the minimum value of the array.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> df = spark.createdataframe([([2, 1, 3],), ([none, 10, -1],)], ['data'])\n >>> df.select(array_min(df.data).alias('min')).collect()\n [row(min=1), row(min=-1)]\n \"": 8793, "array_m": 8794, "array_max": 8795, "\"\n collection function: returns the maximum value of the array.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> df = spark.createdataframe([([2, 1, 3],), ([none, 10, -1],)], ['data'])\n >>> df.select(array_max(df.data).alias('max')).collect()\n [row(max=3), row(max=10)]\n \"": 8796, "sort_array": 8797, "placed": 8798, "array_sort": 8799, "\"\n collection function: sorts the input array in ascending order. the elements of the input array\n must be orderable. null elements will be placed at the end of the returned array.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> df = spark.createdataframe([([2, 1, none, 3],),([1],),([],)], ['data'])\n >>> df.select(array_sort(df.data).alias('r')).collect()\n [row(r=[1, 2, 3, none]), row(r=[1]), row(r=[])]\n \"": 8800, "sparkco": 8801, "orderable": 8802, "'spark sql'": 8803, "'lqs kraps'": 8804, "flatten": 8805, "\"\n collection function: creates a single array from an array of arrays.\n if a structure of nested arrays is deeper than two levels,\n only one level of nesting is removed.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> df = spark.createdataframe([([[1, 2, 3], [4, 5], [6]],), ([none, [4, 5]],)], ['data'])\n >>> df.select(flatten(df.data).alias('r')).collect()\n [row(r=[1, 2, 3, 4, 5, 6]), row(r=none)]\n \"": 8806, "deeper": 8807, "levels": 8808, "nesting": 8809, "map_keys": 8810, "\"\n collection function: returns an unordered array containing the keys of the map.\n\n .. versionadded:: 2.3.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> from pyspark.sql.functions import map_keys\n >>> df = spark.sql(\"": 8811, "\")\n >>> df.select(map_keys(\"": 8812, "\")).show()\n +------+\n | keys|\n +------+\n |[1, 2]|\n +------+\n \"": 8813, "unordered": 8814, "map_values": 8815, "\"\n collection function: returns an unordered array containing the values of the map.\n\n .. versionadded:: 2.3.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> from pyspark.sql.functions import map_values\n >>> df = spark.sql(\"": 8816, "\")\n >>> df.select(map_values(\"": 8817, "\")).show()\n +------+\n |values|\n +------+\n |[a, b]|\n +------+\n \"": 8818, "_j": 8819, "map_entries": 8820, "\"\n collection function: returns an unordered array of all entries in the given map.\n\n .. versionadded:: 3.0.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> from pyspark.sql.functions import map_entries\n >>> df = spark.sql(\"": 8821, "\")\n >>> df.select(map_entries(\"": 8822, "\")).show()\n +----------------+\n | entries|\n +----------------+\n |[{1, a}, {2, b}]|\n +----------------+\n \"": 8823, "map_from_entries": 8824, "\"\n collection function: returns a map created from the given array of entries.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n col : :class:`~pyspark.sql.column` or str\n name of column or expression\n\n examples\n --------\n >>> from pyspark.sql.functions import map_from_entries\n >>> df = spark.sql(\"": 8825, "\")\n >>> df.select(map_from_entries(\"": 8826, "array_repeat": 8827, "\"\n collection function: creates an array containing a column repeated count times.\n\n .. versionadded:: 2.4.0\n\n examples\n --------\n >>> df = spark.createdataframe([('ab',)], ['data'])\n >>> df.select(array_repeat(df.data, 3).alias('r')).collect()\n [row(r=['ab', 'ab', 'ab'])]\n \"": 8828, "repeated": 8829, "arrays_zip": 8830, "structs": 8831, "'vals1'": 8832, "'vals2'": 8833, "vals1": 8834, "vals2": 8835, "'zipped'": 8836, "zipped": 8837, "map_concat": 8838, "\"returns the union of all the given maps.\n\n .. versionadded:: 2.4.0\n\n parameters\n ----------\n cols : :class:`~pyspark.sql.column` or str\n column names or :class:`~pyspark.sql.column`\\\\s\n\n examples\n --------\n >>> from pyspark.sql.functions import map_concat\n >>> df = spark.sql(\"": 8839, "map1": 8840, "map2": 8841, "\")\n >>> df.select(map_concat(\"": 8842, "map3": 8843, "truncate": 8844, "incrementing": 8845, "df1": 8846, "'c1'": 8847, "'c2'": 8848, "'c3'": 8849, "sequen": 8850, "from_csv": 8851, "forall": 8852, "holds": 8853, "userdefinedfunctions": 8854, "27052": 8855, "jira": 8856, "br": 8857, "clas": 8858, "forms": 8859, "initialvalue": 8860, "applies": 8861, "operator": 8862, "reduces": 8863, "converted": 8864, "applying": 8865, "__": 8866, "parame": 8867, "zip_with": 8868, "shorter": 8869, "x1": 8870, "nam": 8871, "transform_keys": 8872, "applications": 8873, "userdefinedfunction": 8874, "transform_values": 8875, "userdefinedfunc": 8876, "map_filter": 8877, "whose": 8878, "satisfy": 8879, "27": 8880, "map_zip_with": 8881, "ternary": 8882, "v1": 8883, "v2": 8884, "userdefinedf": 8885, "years": 8886, "\"\n partition transform function: a transform for timestamps and dates\n to partition data into years.\n\n .. versionadded:: 3.1.0\n\n examples\n --------\n >>> df.writeto(\"": 8887, "catalog": 8888, "\").partitionedby( # doctest: +skip\n ... years(\"": 8889, "\")\n ... ).createorreplace()\n\n notes\n -----\n this function can be used only in combination with\n :py:meth:`~pyspark.sql.readwriter.dataframewriterv2.partitionedby`\n method of the `dataframewriterv2`.\n\n \"": 8890, "_to_java_c": 8891, "timestamps": 8892, "\"catalog.db.table\"": 8893, "partitionedby": 8894, "\"ts\"": 8895, "createorreplace": 8896, "\"\n partition transform function: a transform for timestamps and dates\n to partition data into months.\n\n .. versionadded:: 3.1.0\n\n examples\n --------\n >>> df.writeto(\"": 8897, "\").partitionedby(\n ... months(\"": 8898, "\")\n ... ).createorreplace() # doctest: +skip\n\n notes\n -----\n this function can be used only in combination with\n :py:meth:`~pyspark.sql.readwriter.dataframewriterv2.partitionedby`\n method of the `dataframewriterv2`.\n\n \"": 8899, "_to_ja": 8900, "\"\n partition transform function: a transform for timestamps and dates\n to partition data into days.\n\n .. versionadded:: 3.1.0\n\n examples\n --------\n >>> df.writeto(\"": 8901, "\").partitionedby( # doctest: +skip\n ... days(\"": 8902, "_to_java_colum": 8903, "\"\n partition transform function: a transform for timestamps\n to partition data into hours.\n\n .. versionadded:: 3.1.0\n\n examples\n --------\n >>> df.writeto(\"": 8904, "\").partitionedby( # doctest: +skip\n ... hours(\"": 8905, "combinati": 8906, "bucket": 8907, "numbuckets": 8908, "\"\n partition transform function: a transform for any type that partitions\n by a hash of the input column.\n\n .. versionadded:: 3.1.0\n\n examples\n --------\n >>> df.writeto(\"": 8909, "\").partitionedby( # doctest: +skip\n ... bucket(42, \"": 8910, "udf": 8911, "returntype": 8912, "standalone": 8913, "is_on": 8914, "\" true if switch is on. \"": 8915, "_brightness": 8916, "brightness": 8917, "\" brightness of this light between 0..255. \"": 8918, "light": 8919, "255.": 8920, "turn_off": 8921, "\" turns the switch off. \"": 8922, "tellstick": 8923, "turns": 8924, "turn_on": 8925, "\" turns the switch on. \"": 8926, "attr_brightness": 8927, "255": 8928, "startclientthread": 8929, "client_port": 8930, "clientthread": 8931, "handlerequestinput": 8932, "requestinput": 8933, "'requestinput: ok'": 8934, "handle_request_input": 8935, "pydev_localhost": 8936, "'starting client with:'": 8937, "pydev_lo": 8938, "testserver": 8939, "server_port": 8940, "getfreeaddresses": 8941, "serverthread": 8942, "'starting server with:'": 8943, "get_localhost": 8944, "pydevconsole": 8945, "startserver": 8946, "grouped_iteration_values": 8947, "_iterations": 8948, "test_path": 8949, "_test_path": 8950, "run_single": 8951, "time_out_ms": 8952, "should_run_pixel_test": 8953, "run_test": 8954, "driverinput": 8955, "image_hash": 8956, "stop_when_done": 8957, "wait_until_ready": 8958, "create_connection": 8959, "'8080'": 8960, "until": 8961, "get_external_messages": 8962, "\"return a table of externalized messages.\n\n the table is lazzy instancied (loaded once when called the first time).\"": 8963, "_messages": 8964, "propfile": 8965, "resource_filename": 8966, "messages_file": 8967, "\"ru\"": 8968, "propdict": 8969, "propline": 8970, "propdef": 8971, "punctuation": 8972, "':= '": 8973, "mi": 8974, "externalized": 8975, "lazzy": 8976, "instancied": 8977, "external": 8978, "notify_download": 8979, "ep_name": 8980, "sickbeard": 8981, "plex_notify_ondownload": 8982, "_notify_pmc": 8983, "notifystrings": 8984, "update_library": 8985, "ep_obj": 8986, "\"handles updating the plex media server host via http api\n\n plex media server currently only supports updating the whole video library and not a specific path.\n\n returns:\n returns none for no issue, else a string of host with connection issues\n\n \"": 8987, "use_plex": 8988, "plex_update_library": 8989, "plex_server_host": 8990, "'plex: no plex media server host specified, check your settings'": 8991, "logg": 8992, "handles": 8993, "plex": 8994, "media": 8995, "supports": 8996, "whole": 8997, "video": 8998, "get_phrases_from_module": 8999, "\"\n gets phrases from a module.\n\n arguments:\n module -- a module reference\n\n returns:\n the list of phrases in this module.\n \"": 9000, "'words'": 9001, "phrases": 9002, "get_keyword_phrases": 9003, "\"\n gets the keyword phrases from the keywords file in the jasper data dir.\n\n returns:\n a list of keyword phrases.\n \"": 9004, "jasperpath": 9005, "'keyword_phrases'": 9006, "phrase": 9007, "keywords": 9008, "jasper": 9009, "get_all_phrases": 9010, "\"\n gets phrases from all modules.\n\n returns:\n a list of phrases in all modules plus additional phrases passed to this\n function.\n \"": 9011, "brain": 9012, "get_modules": 9013, "phrases_to_revision": 9014, "\"\n calculates a revision from phrases by using the sha1 hash function.\n\n arguments:\n phrases -- a list of phrases\n\n returns:\n a revision string for given phrases.\n \"": 9015, "sorted_phrases": 9016, "joined_phrases": 9017, "revision_file": 9018, "\"\n returns:\n the path of the the revision file as string\n \"": 9019, "'revision'": 9020, "is_compiled": 9021, "\"\n checks if the vocabulary is compiled by checking if the revision file\n is readable. this method should be overridden by subclasses to check\n for class-specific additional files, too.\n\n returns:\n true if the dictionary is compiled, else false\n \"": 9022, "r_ok": 9023, "vocabulary": 9024, "compiled": 9025, "checking": 9026, "compiled_revision": 9027, "\"\n reads the compiled revision from the revision file.\n\n returns:\n the revision of this vocabulary (i.e. the string\n inside the revision file), or none if is_compiled\n if false\n \"": 9028, "\"compiled_revision is '%s'\"": 9029, "matches_phrases": 9030, "\"\n convenience method to check if this vocabulary exactly contains the\n phrases passed to this method.\n\n arguments:\n phrases -- a list of phrases\n\n returns:\n true if phrases exactly matches the phrases inside this\n vocabulary.\n\n \"": 9031, "compiles": 9032, "compilation": 9033, "forced": 9034, "regardless": 9035, "necessity": 9036, "preliminary": 9037, "skipped": 9038, "meant": 9039, "_compile_vocabulary": 9040, "forces": 9041, "languagemodel_file": 9042, "\"\n returns:\n the path of the the pocketsphinx languagemodel file as string\n \"": 9043, "'languagemodel'": 9044, "pocketsphinx": 9045, "languagemodel": 9046, "dictionary_file": 9047, "\"\n returns:\n the path of the pocketsphinx dictionary file as string\n \"": 9048, "'dictionary'": 9049, "decoder_kwargs": 9050, "\"\n convenience property to use this vocabulary with the __init__() method\n of the pocketsphinx.decoder class.\n\n returns:\n a dict containing kwargs for the pocketsphinx.decoder.__init__()\n method.\n\n example:\n decoder = pocketsphinx.decoder(**vocab_instance.decoder_kwargs,\n hmm='/path/to/hmm')\n\n \"": 9051, "'lm'": 9052, "'dict'": 9053, "vocab_instance": 9054, "dfa_file": 9055, "\"\n returns:\n the path of the the julius dfa file as string\n \"": 9056, "'dfa'": 9057, "julius": 9058, "dfa": 9059, "dict_file": 9060, "\"\n returns:\n the path of the the julius dict file as string\n \"": 9061, "translate_word": 9062, "_dict": 9063, "checkrepotools": 9064, "\"check if the needed repo tools exist.\"": 9065, "svnadmin": 9066, "\"error: can not find svnadmin command.\"": 9067, "svnlook": 9068, "\"error: can not find svnlook command.\"": 9069, "createrepo": 9070, "\" create repsitory and layout \"": 9071, "svncreaterepo": 9072, "repopath": 9073, "categories": 9074, "createcategory": 9075, "\"create repository: \"": 9076, "repsitory": 9077, "installhooks": 9078, "\" install hooks in repository \"": 9079, "repohooklist": 9080, "'pre-commit'": 9081, "'post-commit'": 9082, "repohook": 9083, "repohookpath": 9084, "const_repohookpath": 9085, "hooks": 9086, "executable": 9087, "copyfile": 9088, "const_hookfilepath": 9089, "setexecutable": 9090, "\"install hook: \"": 9091, "\" -> \"": 9092, "hookspath": 9093, "installtemplates": 9094, "\" install templates in repository \"": 9095, "frontend": 9096, "\"tmpl\"": 9097, "'subdms template'": 9098, "defaulttype": 9099, "\"gen\"": 9100, "doctypes": 9101, "createproject": 9102, "tmpl": 9103, "tmpltypes": 9104, "walkrepo": 9105, "\" walk the repo and list the content. \"": 9106, "repolist": 9107, "svncmd": 9108, "recursivels": 9109, "walkrepoleafs": 9110, "\" walk the repo and list all files. \"": 9111, "filenamelist": 9112, "\"kind\"": 9113, "filekind": 9114, "walkreponodes": 9115, "\" walk the repo and list all paths. \"": 9116, "pathnamelist": 9117, "upgraderepo": 9118, "\" upgrade layout in repo. \"": 9119, "projpath": 9120, "repourl": 9121, "\"/p\"": 9122, "trunkpath": 9123, "\"/trunk\"": 9124, "splitpath": 9125, "rsplit": 9126, "\"///\"": 9127, "old_path": 9128, "new_path": 9129, "\"upgrade document path\"": 9130, "upgrade": 9131, "upgradefilename": 9132, "\" upgrade document file names. \"": 9133, "old_name": 9134, "docext": 9135, "new_base": 9136, "new_baselist": 9137, "new_basename": 9138, "\"p-\"": 9139, "new_name": 9140, "tounicode": 9141, "articles": 9142, "\"convert a list of articles utf-8 encoded to unicode strings.\"": 9143, "art": 9144, "'utf_8'": 9145, "todicts": 9146, "\"given a list of utf-8 encoded articles, build two dictionary (one\n utf-8 encoded and another one with unicode keys) for faster matches.\"": 9147, "uarticles": 9148, "faster": 9149, "addtrailingspace": 9150, "\"from the given list of utf-8 encoded articles, return two\n lists (one utf-8 encoded and another one in unicode) where a space\n is added at the end - if the last char is not ' or -.\"": 9151, "_sparticles": 9152, "_spunicodearticles": 9153, "\"'\"": 9154, "articlesdictsforlang": 9155, "\"return dictionaries of articles specific for the given language, or the\n default one if the language is not known.\"": 9156, "_art_cache": 9157, "artdicts": 9158, "lang_articlesget": 9159, "generic_articles": 9160, "dictionaries": 9161, "sparticlesforlang": 9162, "\"return lists of articles (plus optional spaces) specific for the\n given language, or the default one if the language is not known.\"": 9163, "_sp_art_cache": 9164, "sparticles": 9165, "add_signal_handler": 9166, "sigint": 9167, "create_task": 9168, "run_forever": 9169, "' create an options parser for bin/ansible '": 9170, "base_parser": 9171, "'%prog -u [options] []'": 9172, "connect_opts": 9173, "vault_opts": 9174, "runtask_opts": 9175, "subset_opts": 9176, "inventory_opts": 9177, "module_opts": 9178, "runas_prompt_opts": 9179, "\"pulls playbooks from a vcs repo and executes them for the local host\"": 9180, "'--purge'": 9181, "ac": 9182, "do_done": 9183, "process_partners": 9184, "partner_ids": 9185, "partner_obj": 9186, "pool": 9187, "'res.partner'": 9188, "partner_ids_to_print": 9189, "nbmanuals": 9190, "manuals": 9191, "nbmails": 9192, "nbunknownmails": 9193, "nbprints": 9194, "resulttext": 9195, "'account_followup.stat.by.partner'": 9196, "max_followup_id": 9197, "manual_action": 9198, "do_partner_manual_action": 9199, "partners": 9200, "startelement": 9201, "ec2_build_list_params": 9202, "'%sblockdevicemapping'": 9203, "_build_list_params": 9204, "ec2": 9205, "autoscale_build_list_params": 9206, "'%sblockdevicemappings.member'": 9207, "autoscale": 9208, "construct_mapping": 9209, "deep": 9210, "safeconstructor": 9211, "replicate": 9212, "overwriting": 9213, "mappingnode": 9214, "constructorerror": 9215, "\"expected a mapping node, but found %s\"": 9216, "start_mark": 9217, "flatten_mapping": 9218, "ansiblemapping": 9219, "construct_yaml_str": 9220, "construct_scalar": 9221, "ansibleunicode": 9222, "ansible_pos": 9223, "_node_position_info": 9224, "test_multiple_parents_mro": 9225, "abstractbaseone": 9226, "abstract": 9227, "abstractbasetwo": 9228, "max_length": 9229, "descendantone": 9230, "descendanttwo": 9231, "cla": 9232, "parents": 9233, "mro": 9234, "test_shadowed_fkey_id": 9235, "abstractbase": 9236, "foreignkey": 9237, "cascade": 9238, "descendant": 9239, "foo_id": 9240, "integerfield": 9241, "\"the field 'foo_id' clashes with the field 'foo' \"": 9242, "\"from model 'model_inheritance.descendant'.\"": 9243, "get_field": 9244, "'foo_id'": 9245, "fkey": 9246, "test_override_one2one_relation_auto_field_clashes": 9247, "concreteparent": 9248, "abstractparent": 9249, "\"auto-generated field 'concreteparent_ptr' in class 'descendant' \"": 9250, "\"for parent_link to base class 'concreteparent' clashes with \"": 9251, "\"declared field of the same name.\"": 9252, "assertraisesmessage": 9253, "fielderror": 9254, "one2one": 9255, "clashes": 9256, "'_meta'": 9257, "get_fields": 9258, "module_init": 9259, "root_module": 9260, "'ns.csma_layout'": 9261, "cpp_namespace": 9262, "'::ns3'": 9263, "register_types_ns3_fatalimpl": 9264, "get_root": 9265, "ns3": 9266, "fatalimpl": 9267, "register_ns3attributeconstructionlistitem_methods": 9268, "'core'": 9269, "attributeconstructionlist": 9270, "add_constructor": 9271, "const": 9272, "arg0": 9273, "'ns3::attributeconstructionlist::item const &'": 9274, "'arg0'": 9275, "checker": 9276, "add_instanc": 9277, "ns3attributeconstructionlistitem": 9278, "register_ns3csmastarhelper_methods": 9279, "csma": 9280, "star": 9281, "helper": 9282, "'csma-layout'": 9283, "csmastarhelper": 9284, "'ns3::csmastarhelper const &'": 9285, "uint32_t": 9286, "numspokes": 9287, "csmahelper": 9288, "'uint32_t'": 9289, "'numspokes'": 9290, "'ns3::csmahelper'": 9291, "'csmahelper'": 9292, "void": 9293, "assignipv": 9294, "ns3csmastarhelper": 9295, "register_ns3inttotype__1_methods": 9296, "inttotype": 9297, "'ns3::inttotype< 1 > const &'": 9298, "ns3inttotype": 9299, "register_ns3inttotype__2_methods": 9300, "'ns3::inttotype< 2 > const &'": 9301, "register_ns3ipv4addresshelper_methods": 9302, "ipv4": 9303, "'internet'": 9304, "ipv4addresshelper": 9305, "'ns3::ipv4addresshelper const &'": 9306, "ipv4address": 9307, "network": 9308, "ipv4mask": 9309, "\"0.0.0.1\"": 9310, "construc": 9311, "ns3ipv4addresshelper": 9312, "register_ns3ipv4interfaceaddress_methods": 9313, "add_binary_comparison_operator": 9314, "'!='": 9315, "add_output_stream_operator": 9316, "'=='": 9317, "ipv4interfaceaddress": 9318, "'ns3::ipv4address'": 9319, "'local'": 9320, "'ns3::ipv4mask'": 9321, "'mask'": 9322, "ns3ipv4interfaceaddress": 9323, "register_ns3ipv4interfacecontainer_methods": 9324, "ipv4interfacecontainer": 9325, "'ns3::ipv4interfacecontainer const &'": 9326, "ns3ipv4interfacecontainer": 9327, "register_ns3ipv6address_methods": 9328, "'<'": 9329, "ipv6": 9330, "'network'": 9331, "ipv6address": 9332, "'char const *'": 9333, "'address'": 9334, "uint8_t": 9335, "ns3ipv6address": 9336, "register_ns3ipv6interfacecontainer_methods": 9337, "ipv6interfacecontainer": 9338, "'ns3::ipv6interfacecontainer const &'": 9339, "ptr": 9340, "interfa": 9341, "ns3ipv6interfacecontainer": 9342, "register_ns3objectbase_methods": 9343, "objectbase": 9344, "'ns3::objectbase const &'": 9345, "getattribute": 9346, "std": 9347, "attributevalue": 9348, "add_method": 9349, "'getattribute'": 9350, "'void'": 9351, "'std::string'": 9352, "ns3objectbase": 9353, "register_ns3packetmetadata_methods": 9354, "packetmetadata": 9355, "uint64_t": 9356, "'uint64_t'": 9357, "'uid'": 9358, "'ns3::packetmetadata const &'": 9359, "'o'": 9360, "addatend": 9361, "ad": 9362, "ns3packetmetadata": 9363, "register_ns3packetmetadataitem_methods": 9364, "'ns3::packetmetadata::item const &'": 9365, "add_instance_attribute": 9366, "'current'": 9367, "'ns3::buffer::iterator'": 9368, "is_const": 9369, "ns3packetmetadataitem": 9370, "register_ns3packettagiteratoritem_methods": 9371, "packettagiterator": 9372, "'ns3::packettagiterator::item const &'": 9373, "gettag": 9374, "'gettag'": 9375, "'ns3::tag &'": 9376, "'tag'": 9377, "typeid": 9378, "packetta": 9379, "ns3packettagiteratoritem": 9380, "register_ns3pcaphelper_methods": 9381, "pcaphelper": 9382, "'ns3::pcaphelper const &'": 9383, "pcapfilewrapper": 9384, "createfile": 9385, "_ios_openmode": 9386, "filemode": 9387, "datalinktype": 9388, "65535": 9389, "int32_t": 9390, "tzcorrection": 9391, "ns3pcaphelper": 9392, "register_ns3simulator_methods": 9393, "simulator": 9394, "'ns3::simulator const &'": 9395, "eventid": 9396, "'ns3::eventid const &'": 9397, "is_static": 9398, "add_": 9399, "ns3simulator": 9400, "register_ns3timer_methods": 9401, "timer": 9402, "'ns3::timer const &'": 9403, "destroypolicy": 9404, "'ns3::timer::destroypolicy'": 9405, "'destroypolicy'": 9406, "ns3timer": 9407, "register_ns3timerimpl_methods": 9408, "impl": 9409, "timerimpl": 9410, "'ns3::timerimpl const &'": 9411, "'invoke'": 9412, "is_pure_virtual": 9413, "is_virtual": 9414, "ns3timerimpl": 9415, "register_ns3typeidtracesourceinformation_methods": 9416, "tracesourceinformation": 9417, "'ns3::typeid::tracesourceinformation const &'": 9418, "accessor": 9419, "'accessor'": 9420, "ns3typeidtracesourceinformation": 9421, "register_ns3empty_methods": 9422, "'ns3::empty const &'": 9423, "ns3empty": 9424, "register_ns3chunk_methods": 9425, "'ns3::chunk const &'": 9426, "buffer": 9427, "'deserialize'": 9428, "is_virtu": 9429, "ns3chunk": 9430, "register_ns3csmahelper_methods": 9431, "'csma'": 9432, "'ns3::csmahelper const &'": 9433, "netdevicecontainer": 9434, "'install'": 9435, "'ns3::netdevicecontainer'": 9436, "ns3csmahelper": 9437, "register_ns3ipv6header_methods": 9438, "ipv6header": 9439, "'ns3::ipv6header const &'": 9440, "ns3ipv6header": 9441, "register_ns3object_methods": 9442, "aggregateobject": 9443, "'aggregateobject'": 9444, "'ns3::ptr< ns3::object >'": 9445, "'other'": 9446, "dispose": 9447, "'dispose'": 9448, "ns3object": 9449, "register_ns3simplerefcount__ns3attributevalue_ns3empty_ns3defaultdeleter__lt__ns3attributevalue__gt___methods": 9450, "simplerefcount": 9451, "defaultdeleter": 9452, "ns3simplerefcount": 9453, "ns3attributevalue": 9454, "ns3defaultdeleter": 9455, "gt": 9456, "register_ns3simplerefcount__ns3callbackimplbase_ns3empty_ns3defaultdeleter__lt__ns3callbackimplbase__gt___methods": 9457, "callbackimplbase": 9458, "ns3callbackimplbase": 9459, "register_ns3simplerefcount__ns3ipv4multicastroute_ns3empty_ns3defaultdeleter__lt__ns3ipv4multicastroute__gt___methods": 9460, "ipv4multicastroute": 9461, "ipv4mul": 9462, "ns3ipv4multicastroute": 9463, "register_ns3simplerefcount__ns3nixvector_ns3empty_ns3defaultdeleter__lt__ns3nixvector__gt___methods": 9464, "nixvector": 9465, "pa": 9466, "ns3nixvector": 9467, "register_ns3time_methods": 9468, "add_binary_numeric_operator": 9469, "'ns3::time'": 9470, "'ns3::time const &'": 9471, "'right'": 9472, "'>'": 9473, "add_inplace_numeric_operator": 9474, "'+='": 9475, "'-='": 9476, "ns3time": 9477, "register_ns3tracesourceaccessor_methods": 9478, "tracesourceaccessor": 9479, "'ns3::tracesourceaccessor const &'": 9480, "callbackbase": 9481, "ns3tracesourceaccessor": 9482, "register_ns3attributevalue_methods": 9483, "'ns3::attributevalue const &'": 9484, "'ns3::ptr< ns3::attributevalue >'": 9485, "register_ns3callbackvalue_methods": 9486, "callbackvalue": 9487, "'ns3::callbackvalue const &'": 9488, "'ns3::callbackbase const &'": 9489, "'base'": 9490, "ns3callbackvalue": 9491, "register_ns3csmachannel_methods": 9492, "channel": 9493, "csmachannel": 9494, "gettypeid": 9495, "'gettypeid'": 9496, "'ns3::typeid'": 9497, "csmanetdevice": 9498, "'attach'": 9499, "'int32_t'": 9500, "ns3csmachannel": 9501, "register_ns3dataratevalue_methods": 9502, "rate": 9503, "dataratevalue": 9504, "'ns3::dataratevalue const &'": 9505, "datarate": 9506, "'ns3::datarate const &'": 9507, "ns3dataratevalue": 9508, "register_ns3ipv4maskchecker_methods": 9509, "ipv4maskchecker": 9510, "'ns3::ipv4maskchecker const &'": 9511, "ns3ipv4maskchecker": 9512, "register_ns3ipv4multicastroute_methods": 9513, "route": 9514, "'ns3::ipv4multicastroute const &'": 9515, "getgroup": 9516, "'getgroup'": 9517, "register_ns3addressvalue_methods": 9518, "addressvalue": 9519, "'ns3::addressvalue const &'": 9520, "'ns3::address const &'": 9521, "attributevalu": 9522, "ns3addressvalue": 9523, "register_functions": 9524, "register_functions_ns3_fatalimpl": 9525, "get_submodule": 9526, "'fatalimpl'": 9527, "hypervolume": 9528, "front": 9529, "kargs": 9530, "\"returns the index of the individual with the least the hypervolume\n contribution. the provided *front* should be a set of non-dominated\n individuals having each a :attr:`fitness` attribute. \n \"": 9531, "wvalues": 9532, "minimization": 9533, "deap": 9534, "wobj": 9535, "fitness": 9536, "\"ref\"": 9537, "contribution": 9538, "p_i": 9539, "dominated": 9540, "individuals": 9541, "additive_epsilon": 9542, "\"returns the index of the individual with the least the additive epsilon\n contribution. the provided *front* should be a set of non-dominated\n individuals having each a :attr:`fitness` attribute.\n\n .. warning::\n\n this function has not been tested.\n \"": 9543, "mwobj": 9544, "masked": 9545, "contrib_values": 9546, "additive": 9547, "multiplicative_epsilon": 9548, "\"returns the index of the individual with the least the multiplicative epsilon\n contribution. the provided *front* should be a set of non-dominated\n individuals having each a :attr:`fitness` attribute.\n\n .. warning::\n\n this function has not been tested.\n \"": 9549, "multiplicative": 9550, "testbasicrnnseq2seq": 9551, "test_session": 9552, "\"root\"": 9553, "init_ops": 9554, "constant_initializer": 9555, "dec_inp": 9556, "0.4": 9557, "core_rnn_cell_impl": 9558, "outputprojectionwrapper": 9559, "grucell": 9560, "seq2seq_lib": 9561, "basic_rnn_seq2seq": 9562, "global_variables_initializer": 9563, "testtiedrnnseq2seq": 9564, "tied_rnn_seq2seq": 9565, "testattentiondecoder2": 9566, "cell_fn": 9567, "enc_outputs": 9568, "enc_state": 9569, "core_rnn": 9570, "static_rnn": 9571, "attn_states": 9572, "output_size": 9573, "testembeddingattentionseq2seq": 9574, "enc_inp": 9575, "basiclstmcell": 9576, "embedding_attention_seq2seq": 9577, "testone2manyrnnseq2seq": 9578, "dec_inp_dict": 9579, "\"0\"": 9580, "testsequencelossbyexample": 9581, "output_classes": 9582, "logits": 9583, "targets": 9584, "weights": 9585, "average_loss_per_example": 9586, "sequence_loss_by_example": 9587, "average_across_timesteps": 9588, "average_loss_per_exa": 9589, "testmodelwithbuckets": 9590, "\"larger tests that does full sequence-to-sequence model training.\"": 9591, "learn": 9592, "symbols": 9593, "buckets": 9594, "8.": 9595, "perplexities": 9596, "random_seed": 9597, "set_random_seed": 9598, "111": 9599, "sampled": 9600, "softmax": 9601, "get_variable": 9602, "\"proj_w\"": 9603, "w_t": 9604, "transpose": 9605, "testmodelwithbooleanfeedprevious": 9606, "\"test the model behavior when feed_previous is true.\n\n for example, the following two cases have the same effect:\n - train `embedding_rnn_seq2seq` with `feed_previous=true`, which contains\n a `embedding_rnn_decoder` with `feed_previous=true` and\n `update_embedding_for_previous=true`. the decoder is fed with \"": 9607, "\"\n and outputs \"": 9608, "B": 9609, "\".\n - train `embedding_rnn_seq2seq` with `feed_previous=false`. the decoder\n is fed with \"": 9610, "\".\n \"": 9611, "num_encoder_symbols": 9612, "num_decoder_symbols": 9613, "feed_previous": 9614, "embedding_rnn_seq2seq": 9615, "embedding_rnn_decoder": 9616, "update_embedding_for_previous": 9617, "testmodel": 9618, "num_enc_timesteps": 9619, "dec_inp_fp_true": 9620, "num_dec_timesteps": 9621, "dec_inp_holder_fp_false": 9622, "embeddingrnnseq2seqf": 9623, "state_is_tuple": 9624, "embedding_size": 9625, "embeddingtiedrnnseq2seqnotuple": 9626, "embedding_tied_rnn_seq2seq": 9627, "samplegruseq2seq": 9628, "\"example sequence-to-sequence model that uses gru cells.\"": 9629, "gruseq2seq": 9630, "multirnncell": 9631, "output_projection": 9632, "gru": 9633, "cells": 9634, "deccellsfn": 9635, "dec_symbols_dict": 9636, "forwardbackward": 9637, "scope_name": 9638, "\"fp_{}\"": 9639, "dec_op": 9640, "net_variables": 9641, "get_collection": 9642, "graphkeys": 9643, "global_variables": 9644, "optimizer": 9645, "adam": 9646, "adamoptimizer": 9647, "0.03": 9648, "update_op": 9649, "minimize": 9650, "sequence_loss": 9651, "var_list": 9652, "get_covariance": 9653, "\"compute data covariance with the generative model.\n\n ``cov = components_.t * s**2 * components_ + sigma2 * eye(n_features)``\n where s**2 contains the explained variances, and sigma2 contains the\n noise variances.\n\n returns\n -------\n cov : array, shape=(n_features, n_features)\n estimated covariance of data.\n \"": 9654, "components_": 9655, "exp_var": 9656, "explained_variance_": 9657, "whiten": 9658, "newaxis": 9659, "generative": 9660, "cov": 9661, "sigma2": 9662, "eye": 9663, "n_features": 9664, "explained": 9665, "variances": 9666, "noise": 9667, "estimated": 9668, "get_precision": 9669, "\"compute data precision matrix with the generative model.\n\n equals the inverse of the covariance but computed with\n the matrix inversion lemma for efficiency.\n\n returns\n -------\n precision : array, shape=(n_features, n_features)\n estimated precision of data.\n \"": 9670, "corner": 9671, "n_components_": 9672, "noise_variance_": 9673, "inversion": 9674, "lemma": 9675, "\"placeholder for fit. subclasses should implement this method!\n\n fit the model with x.\n\n parameters\n ----------\n x : array-like, shape (n_samples, n_features)\n training data, where n_samples is the number of samples and\n n_features is the number of features.\n\n returns\n -------\n self : object\n returns the instance itself.\n \"": 9676, "inverse_transform": 9677, "x_original": 9678, "n_components": 9679, "whitening": 9680, "exact": 9681, "test_drop_collection_nocheck": 9682, "t3": 9683, "t4": 9684, "t5": 9685, "_table_fixture": 9686, "_mock_drop_fixture": 9687, "item_exists": 9688, "\"t2\"": 9689, "\"t4\"": 9690, "_assert_drop_tables": 9691, "nocheck": 9692, "test_create_metadata_nocheck": 9693, "_mock_create_fixture": 9694, "_assert_create_tables": 9695, "test_drop_metadata_nocheck": 9696, "test_create_metadata_auto_alter_fk": 9697, "_use_alter_fixture_one": 9698, "_assert_create_w_alter": 9699, "foreign_key_constraints": 9700, "alter": 9701, "fk": 9702, "install_ask_amy": 9703, "destination_dir": 9704, "source_dir": 9705, "ask_amy_dist": 9706, "'/ask_amy'": 9707, "rmtree": 9708, "ignore_errors": 9709, "copytree": 9710, "pip": 9711, "'--upgrade'": 9712, "'ask_amy'": 9713, "'-t'": 9714, "install_ask_amy_for_upload": 9715, "filenotfounderror": 9716, "\"error: path not found {}\\n\"": 9717, "ask": 9718, "amy": 9719, "process_args": 9720, "arg_tuple": 9721, "arg_list": 9722, "substitue": 9723, "substitutions": 9724, "_args": 9725, "leftover_args": 9726, "destdir": 9727, "expanduser": 9728, "'errcheck'": 9729, "get_object_file": 9730, "get_test_binary": 9731, "objectfile": 9732, "assert_extended": 9733, "templ": 9734, "bd": 9735, "test_site": 9736, "'content/auto_extend.html'": 9737, "generate_resource_at_path": 9738, "resource_from_path": 9739, "deploy_root_path": 9740, "relative_deploy_path": 9741, "read_all": 9742, "pyquery": 9743, "extended": 9744, "signal_tester": 9745, "signaltester": 9746, "weak": 9747, "tester": 9748, "test_urls_need_reloading_signal_create": 9749, "urls_need_reloading": 9750, "call_count": 9751, "create_page": 9752, "\"apphooked-page\"": 9753, "\"nav_playground.html\"": 9754, "\"en\"": 9755, "published": 9756, "apphook": 9757, "\"sampleapp\"": 9758, "apphook_namespace": 9759, "urls": 9760, "test_urls_need_reloading_signal_delete": 9761, "test_urls_need_reloading_signal_change_slug": 9762, "title_set": 9763, "visit_todo_node": 9764, "visit_admonition": 9765, "visit": 9766, "depart_todo_node": 9767, "depart_admonition": 9768, "depart": 9769, "validator_cls": 9770, "schemavalidator": 9771, "format_validators": 9772, "required_by_default": 9773, "blank_by_default": 9774, "disallow_unknown_properties": 9775, "apply_default_to_data": 9776, "validates": 9777, "schemaerror": 9778, "va": 9779, "export_methods": 9780, "'\n void dump();\n'": 9781, "test_dmp_lc": 9782, "dmp_lc": 9783, "zz": 9784, "dmp": 9785, "lc": 9786, "test_dmp_tc": 9787, "dmp_tc": 9788, "test_dup_strip": 9789, "dup_strip": 9790, "test_dmp_strip": 9791, "dmp_strip": 9792, "test_dup_copy": 9793, "dup_copy": 9794, "test_dup_normal": 9795, "dup_normal": 9796, "test_dup_convert": 9797, "k0": 9798, "k1": 9799, "dup_convert": 9800, "test_dmp_convert": 9801, "dmp_convert": 9802, "test_dmp_from_sympy": 9803, "dmp_from_sympy": 9804, "qq": 9805, "sympy": 9806, "test_dmp_zero": 9807, "dmp_zero": 9808, "test_dmp_one_p": 9809, "dmp_one_p": 9810, "test_dmp_negative_p": 9811, "dmp_negative_p": 9812, "test_dmp_positive_p": 9813, "dmp_positive_p": 9814, "test_dmp_nest": 9815, "dmp_nest": 9816, "nest": 9817, "test_dup_deflate": 9818, "dup_deflate": 9819, "dup_from_raw_dict": 9820, "deflate": 9821, "test_dmp_deflate": 9822, "dmp_deflate": 9823, "test_dup_inflate": 9824, "dup_inflate": 9825, "17": 9826, "inflate": 9827, "test_dmp_inflate": 9828, "dmp_inflate": 9829, "test_dmp_exclude": 9830, "dmp_exclude": 9831, "test_dmp_eject": 9832, "ring": 9833, "\"x,y\"": 9834, "to_domain": 9835, "dmp_eject": 9836, "eject": 9837, "test_dup_terms_gcd": 9838, "dup_terms_gcd": 9839, "gcd": 9840, "test_dmp_list_terms": 9841, "dmp_list_terms": 9842, "'lex'": 9843, "'grlex'": 9844, "\"collect the paths (which must be under artifact root) into this artifact.\"": 9845, "artifact": 9846, "\"extract the files in this artifact to their locations under artifact root.\"": 9847, "locations": 9848, "collate": 9849, "collation": 9850, "\"return the clause ``expression collate collation``.\n\n e.g.::\n\n collate(mycolumn, 'utf8_bin')\n\n produces::\n\n mycolumn collate utf8_bin\n\n \"": 9851, "_literal_as_binds": 9852, "binaryexpression": 9853, "_literal_as_text": 9854, "operators": 9855, "type_": 9856, "clause": 9857, "mycolumn": 9858, "'utf8_bin'": 9859, "produces": 9860, "utf8_bin": 9861, "lower_bound": 9862, "upper_bound": 9863, "produce": 9864, "sqlalchemy": 9865, "stmt": 9866, "users_table": 9867, "resembling": 9868, "id_1": 9869, "id_2": 9870, "columnelement": 9871, "bound": 9872, "clauses": 9873, "clauseelement": 9874, "bindparameter": 9875, "underly": 9876, "objec": 9877, "outparam": 9878, "\"create an 'out' parameter for usage in functions (stored procedures),\n for databases which support them.\n\n the ``outparam`` can be used like a regular function parameter.\n the \"": 9879, "\" value will be available from the\n :class:`~sqlalchemy.engine.resultproxy` object via its ``out_parameters``\n attribute, which returns a dictionary containing the values.\n\n \"": 9880, "isoutparam": 9881, "'out'": 9882, "procedures": 9883, "\"output\"": 9884, "resultproxy": 9885, "out_parameters": 9886, "not_": 9887, "\"return a negation of the given clause, i.e. ``not(clause)``.\n\n the ``~`` operator is also overloaded on all\n :class:`.columnelement` subclasses to produce the\n same result.\n\n \"": 9888, "inv": 9889, "negation": 9890, "overloaded": 9891, "literal_column": 9892, "\"produce a :class:`.columnclause` object that has the\n :paramref:`.column.is_literal` flag set to true.\n\n :func:`.literal_column` is similar to :func:`.column`, except that\n it is more often used as a \"": 9893, "stated": 9894, "stores": 9895, "quoted": 9896, "oriented": 9897, "columnclause": 9898, "paramref": 9899, "is_literal": 9900, "similar": 9901, "often": 9902, "\"standalone\"": 9903, "unique_params": 9904, "optionaldict": 9905, "\"return a copy with :func:`bindparam()` elements replaced.\n\n same functionality as ``params()``, except adds `unique=true`\n to affected bind parameters so that multiple statements can be\n used.\n\n \"": 9906, "_params": 9907, "bindparam": 9908, "functionality": 9909, "affected": 9910, "statements": 9911, "\"return a copy with :func:`bindparam()` elements replaced.\n\n returns a copy of this clauseelement with :func:`bindparam()`\n elements replaced with values taken from the given dictionary::\n\n >>> clause = column('x') + bindparam('foo')\n >>> print clause.compile().params\n {'foo':none}\n >>> print clause.params({'foo':7}).compile().params\n {'foo':7}\n\n \"": 9912, "taken": 9913, "claus": 9914, "\"compare this clauseelement to the given clauseelement.\n\n subclasses should override the default behavior, which is a\n straight identity comparison.\n\n \\**kw are arguments consumed by subclass compare() methods and\n may be used to modify the criteria for comparison.\n (see :class:`.columnelement`)\n\n \"": 9915, "straight": 9916, "consumed": 9917, "criteria": 9918, "get_children": 9919, "\"return immediate child elements of this :class:`.clauseelement`.\n\n this is used for visit traversal.\n\n \\**kwargs may contain flags that change the collection that is\n returned, for example to return a subset of items in order to\n cut down on larger traversals, or to return child items from a\n different context (such as schema-level collections instead of\n clause-level).\n\n \"": 9920, "immediate": 9921, "traversal": 9922, "subset": 9923, "cut": 9924, "down": 9925, "traversals": 9926, "self_group": 9927, "\"apply a 'grouping' to this :class:`.clauseelement`.\n\n this method is overridden by subclasses to return a\n \"": 9928, "\" construct, i.e. parenthesis. in particular\n it's used by \"": 9929, "around": 9930, "themselves": 9931, "constructs": 9932, "subqueries": 9933, "platforms": 9934, "'grouping'": 9935, "\"grouping\"": 9936, "parenthesis": 9937, "particular": 9938, "\"binary\"": 9939, "__and__": 9940, "\"'and' at the clauseelement level.\n\n .. deprecated:: 0.9.5 - conjunctions are intended to be\n at the :class:`.columnelement`. level\n\n \"": 9941, "and_": 9942, "'and'": 9943, "0.9": 9944, "conjunctions": 9945, "intended": 9946, "__or__": 9947, "\"'or' at the clauseelement level.\n\n .. deprecated:: 0.9.5 - conjunctions are intended to be\n at the :class:`.columnelement`. level\n\n \"": 9948, "or_": 9949, "'or'": 9950, "__invert__": 9951, "'negation_clause'": 9952, "negation_clause": 9953, "_negate": 9954, "invert": 9955, "__bool__": 9956, "\"boolean value of this clause is not defined\"": 9957, "\"return a column expression.\n\n part of the inspection interface; returns self.\n\n \"": 9958, "inspection": 9959, "base_columns": 9960, "column_set": 9961, "proxy_set": 9962, "'_proxies'": 9963, "_proxies": 9964, "shares_lineage": 9965, "othercolumn": 9966, "\"return true if the given :class:`.columnelement`\n has a common ancestor to this :class:`.columnelement`.\"": 9967, "ancestor": 9968, "shares": 9969, "lineage": 9970, "\"produce a type cast, i.e. ``cast( as )``.\n\n this is a shortcut to the :func:`~.expression.cast` function.\n\n .. versionadded:: 1.0.7\n\n \"": 9971, "shortcut": 9972, "\"produce a column label, i.e. `` as ``.\n\n this is a shortcut to the :func:`~.expression.label` function.\n\n if 'name' is none, an anonymous label name will be generated.\n\n \"": 9973, "columnname": 9974, "anonymous": 9975, "anon_label": 9976, "\"provides a constant 'anonymous label' for this columnelement.\n\n this is a label() expression which will be named at compile time.\n the same label() is returned each time anon_label is called so\n that expressions can reference anon_label multiple times, producing\n the same label name at compile time.\n\n the compiler uses this function automatically at compile time\n for expressions that are known to be 'unnamed' like binary\n expressions and function calls.\n\n \"": 9977, "_is_clone_of": 9978, "provides": 9979, "'anonymous label'": 9980, "producing": 9981, "anon": 9982, "effective_value": 9983, "\"return the value of this bound parameter,\n taking into account if the ``callable`` parameter\n was set.\n\n the ``callable`` value will be evaluated\n and returned if present, else ``value``.\n\n \"": 9984, "evaluated": 9985, "effective": 9986, "bindparams": 9987, "binds": 9988, "names_to_values": 9989, "\"establish the values and/or types of bound parameters within\n this :class:`.textclause` construct.\n\n given a text construct such as::\n\n from sqlalchemy import text\n stmt = text(\"": 9990, "\"\n \"": 9991, "textclause": 9992, "establish": 9993, "'jack'": 9994, "\"select id, name from user where name=:name \"": 9995, "\"and timestamp=:timestamp\"": 9996, "bindp": 9997, "selectable": 9998, "\"turn this :class:`.textclause` object into a :class:`.textasfrom`\n object that can be embedded into another statement.\n\n this function essentially bridges the gap between an entirely\n textual select statement and the sql expression language concept\n of a \"": 9999, "\"::\n\n from sqlalchemy.sql import column, text\n\n stmt = text(\"": 10000, "some_table": 10001, "'st'": 10002, "mytable": 10003, "textasfrom": 10004, "essentially": 10005, "bridges": 10006, "gap": 10007, "entirely": 10008, "textual": 10009, "\"selectable\"": 10010, "comparator": 10011, "comparator_factory": 10012, "conjunction": 10013, "'wendy'": 10014, "compound": 10015, "parenthesized": 10016, "is_comparison": 10017, "__setstate__": 10018, "'element'": 10019, "setstate": 10020, "partition_by": 10021, "\"produce an over clause against this :class:`.withingroup`\n construct.\n\n this function has the same signature as that of\n :meth:`.functionelement.over`.\n\n \"": 10022, "withingroup": 10023, "functionelement": 10024, "__reduce__": 10025, "_element": 10026, "_type": 10027, "reduction": 10028, "py3k": 10029, "'backslashreplace'": 10030, "apply_map": 10031, "map_": 10032, "\"pull 'table' from parent, if not present\"": 10033, "_annotated__element": 10034, "'table'": 10035, "\"pull 'key' from parent, if not present\"": 10036, "repl": 10037, "_bindparams": 10038, "':%s'": 10039, "__event_loop__": 10040, "win_args": 10041, "\"\n the event loop thread function. do not override or call\n directly (it is called by __init__).\n \"": 10042, "gl_lock": 10043, "managedwindow": 10044, "switch_to": 10045, "\"window initialization failed: %s\"": 10046, "has_exit": 10047, "clock": 10048, "set_fps_lim": 10049, "\"\n closes the window.\n \"": 10050, "\"\n called after update during each iteration of\n the event loop. put opengl rendering calls\n here.\n \"": 10051, "opengl": 10052, "rendering": 10053, "geoquery": 10054, "customized": 10055, "transformed": 10056, "custom_select": 10057, "transformed_srid": 10058, "extra_select_fields": 10059, "convert_values": 10060, "\"\n using the same routines that oracle does we can convert our\n extra selection objects into geometry and distance objects.\n todo: make converted objects 'lazy' for less overhead.\n \"": 10061, "oracle": 10062, "geomfield": 10063, "spatial": 10064, "nul": 10065, "routines": 10066, "'lazy'": 10067, "overhead": 10068, "resolve_aggregate": 10069, "\"\n overridden from geoquery's normalize to handle the conversion of\n geoaggregate objects.\n \"": 10070, "aggregates_module": 10071, "geoaggregate": 10072, "is_extent": 10073, "'3d'": 10074, "convert_extent3d": 10075, "convert_extent": 10076, "convert_geom": 10077, "normalize": 10078, "resolve": 10079, "findallcontain": 10080, "'_'": 10081, "fake_get_all_flavors_sorted_list": 10082, "inactive": 10083, "'flavorid'": 10084, "sort_dir": 10085, "'asc'": 10086, "'99999'": 10087, "markernotfound": 10088, "reject_min": 10089, "db_attr": 10090, "filter_attr": 10091, "flavor": 10092, "flavor_name": 10093, "fake_flavors": 10094, "'memory_mb'": 10095, "'min_memory_mb'": 10096, "flavors": 10097, "fake_get_limit_and_marker": 10098, "max_limit": 10099, "get_pagination_params": 10100, "'limit'": 10101, "'marker'": 10102, "test_get_flavor_list": 10103, "fake_request": 10104, "_prefix": 10105, "'/flavors'": 10106, "controller": 10107, "\"flavors\"": 10108, "\"flavor 1\"": 10109, "\"links\"": 10110, "\"rel\"": 10111, "\"self\"": 10112, "\"href\"": 10113, "\"http://localhost/\"": 10114, "_rspv": 10115, "\"/flavors/1\"": 10116, "\"bookmark\"": 10117, "test_get_flavor_list_with_marker": 10118, "maxdiff": 10119, "'/flavors?limit=1&marker=1'": 10120, "\"flavor 2\"": 10121, "\"/flavors/2\"": 10122, "test_get_flavor_list_with_invalid_marker": 10123, "'/flavors?marker=99999'": 10124, "webob": 10125, "httpbadrequest": 10126, "test_get_flavor_list_filter_min_disk": 10127, "mindisk": 10128, "'/flavors?mindisk=20'": 10129, "disk": 10130, "test_get_flavor_list_detail_min_ram_and_min_disk": 10131, "\"tests that filtering work on flavor details and that minram and\n mindisk filters can be combined\n \"": 10132, "'/flavors/detail'": 10133, "'?minram=256&mindisk=20'": 10134, "\"ram\"": 10135, "\"512\"": 10136, "\"disk\"": 10137, "\"20\"": 10138, "vcpus": 10139, "details": 10140, "minram": 10141, "combined": 10142, "ram": 10143, "test_index_should_list_disabled_flavors_to_admin": 10144, "is_admin": 10145, "flavor_list": 10146, "'flavors'": 10147, "api_flavorids": 10148, "db_flavorids": 10149, "inst_types": 10150, "disabled_flavorid": 10151, "disabled_type": 10152, "admin": 10153, "test_show_should_include_disabled_flavor_for_user": 10154, "\"counterintuitively we should show disabled flavors to all users and\n not just admins. the reason is that, when a user performs a server-show\n request, we want to be able to display the pretty flavor name ('512 mb\n instance') and not just the flavor-id even if the flavor id has been\n marked disabled.\n \"": 10155, "'flavor'": 10156, "counterintuitively": 10157, "admins": 10158, "pretty": 10159, "'512 mb\ninstance'": 10160, "marked": 10161, "test_none": 10162, "assertpublic": 10163, "test_truthy": 10164, "truthy": 10165, "test_string_none": 10166, "calendar_id2real_id": 10167, "calendar_id": 10168, "with_date": 10169, "\"\n convert a \"": 10170, "recurring": 10171, "\" (type string) into a real event id (type int).\n e.g. virtual/recurring event id is 4-20091201100000, so it will return 4.\n @param calendar_id: id of calendar\n @param with_date: if a value is passed to this param it will return dates based on value of withdate + calendar_id\n @return: real event id\n \"": 10172, "real_id": 10173, "\"virtual/recurring event id\"": 10174, "20091201100000": 10175, "withdate": 10176, "id2real": 10177, "exp_report": 10178, "datas": 10179, "\"\n export report\n \"": 10180, "'printscreen.list'": 10181, "original_exp_report": 10182, "new_ids": 10183, "onchange_partner_id": 10184, "\"\n make entry on email and availability on change of partner_id field.\n @param partner_id: changed value of partner id\n \"": 10185, "availability": 10186, "onchange": 10187, "get_ics_file": 10188, "event_obj": 10189, "\"\n returns icalendar file for the event invitation.\n @param event_obj: event object (browse record)\n @return: .ics file content\n \"": 10190, "ics_datetime": 10191, "idate": 10192, "allday": 10193, "openerp": 10194, "tzinfo": 10195, "pytz": 10196, "'utc'": 10197, "icalendar": 10198, "invitation": 10199, "ics": 10200, "onchange_user_id": 10201, "user_id": 10202, "argv": 10203, "\"\n make entry on email and availability on change of user_id field.\n @param ids: list of attendee's ids\n @param user_id: changed value of user id\n @return: dictionary of values which put value in email and availability fields\n \"": 10204, "'res.users'": 10205, "'availability'": 10206, "attendee": 10207, "do_tentative": 10208, "\"\n makes event invitation as tentative.\n @param ids: list of attendee's ids\n \"": 10209, "'tentative'": 10210, "tentative": 10211, "do_accept": 10212, "\"\n marks event invitation as accepted.\n @param ids: list of attendee's ids\n \"": 10213, "meeting_obj": 10214, "'calendar.event'": 10215, "'accepted'": 10216, "message_post": 10217, "event_id": 10218, "\"%s has accepted invitation\"": 10219, "subtype": 10220, "subtype_": 10221, "accepted": 10222, "do_decline": 10223, "\"\n marks event invitation as declined.\n @param ids: list of calendar attendee's ids\n \"": 10224, "'declined'": 10225, "\"%s has declined invitation\"": 10226, "\"calendar.subtype_invitation\"": 10227, "contex": 10228, "declined": 10229, "decline": 10230, "get_attendee_detail": 10231, "meeting_id": 10232, "\"\n return a list of tuple (id, name, status)\n used by web_calendar.js : many2manyattendee\n \"": 10233, "meeting": 10234, "get_real_ids": 10235, "name_get": 10236, "attendee_ids": 10237, "web_calendar": 10238, "js": 10239, "many2manyattendee": 10240, "do_mail_reminder": 10241, "alarm": 10242, "'calendar.alarm'": 10243, "'alarm_id'": 10244, "'calendar.attendee'": 10245, "_send_mail_to_attendees": 10246, "email_from": 10247, "reminder": 10248, "do_notif_reminder": 10249, "'notification'": 10250, "display_time": 10251, "'notify_at'": 10252, "3600": 10253, "notif": 10254, "unlink": 10255, "calendar_alarm": 10256, "_update_cron": 10257, "do_run_scheduler": 10258, "'calendar.alarm_manager'": 10259, "get_next_mail": 10260, "scheduler": 10261, "get_recurrent_date_by_event": 10262, "\"get recurrent dates based on rule string and all event where recurrent_id is child\n \"": 10263, "todate": 10264, "'\\d'": 10265, "localized": 10266, "localize": 10267, "astimezone": 10268, "'tz'": 10269, "recurrent": 10270, "recurrent_id": 10271, "get_date_formats": 10272, "\"lang\"": 10273, "res_lang": 10274, "'res.lang'": 10275, "lang_params": 10276, "\"code\"": 10277, "\"=\"": 10278, "\"date_format\"": 10279, "\"time_format\"": 10280, "coerce": 10281, "format_date": 10282, "'%b-%d-%y'": 10283, "format_time": 10284, "time_f": 10285, "get_display_time_tz": 10286, "\"tz\"": 10287, "ev": 10288, "_get_display_time": 10289, "onchange_dates": 10290, "fromtype": 10291, "checkallday": 10292, "\"returns duration and end date based on values passed\n @param ids: list of calendar event's ids.\n \"": 10293, "'allday'": 10294, "rewrited": 10295, "strptime": 10296, "default_server_date_format": 10297, "'start_datetime'": 10298, "strftime": 10299, "get_recurrent_ids": 10300, "\"gives virtual event ids for recurring events\n this method gives ids of dates that comes between start date and end date of calendar views\n\n @param order: the fields (comma separated, format \"": 10301, "\") on which the events should be sorted\n \"": 10302, "ids_to_browse": 10303, "comes": 10304, "\"field {desc|asc}\"": 10305, "compute_rule_string": 10306, "\"\n compute rule string according to value type recur of icalendar from the values given.\n @param self: the object pointer\n @param data: dictionary of freq and interval value\n @return: string containing recurring rule (empty if no rule)\n \"": 10307, "'interval'": 10308, "except_osv": 10309, "'warning!'": 10310, "'interval cannot be negative.'": 10311, "'count'": 10312, "'count cannot be negative or 0.'": 10313, "recur": 10314, "pointer": 10315, "freq": 10316, "onchange_partner_ids": 10317, "\" the basic purpose of this method is to check that destination partners\n effectively have email addresses. otherwise a warning is thrown.\n :param value: value format: [[6, 0, [3, 4]]]\n \"": 10318, "check_partners_email": 10319, "basic": 10320, "purpose": 10321, "destination": 10322, "effectively": 10323, "addresses": 10324, "thrown": 10325, "\" verify that selected partner_ids have an email_address defined.\n otherwise throw a warning. \"": 10326, "partner_wo_email_lst": 10327, "warning_msg": 10328, "'the following contacts have no email address :'": 10329, "war": 10330, "email_address": 10331, "message_subscribe": 10332, "subtype_ids": 10333, "calendar_event": 10334, "subscribe": 10335, "get_attendee": 10336, "'meeting'": 10337, "'attendee'": 10338, "'where'": 10339, "'when'": 10340, "get_interval": 10341, "' format and localize some dates to be used in email templates\n\n :param string date: date/time to be formatted\n :param string interval: among '": 10342, "', '": 10343, "dayname": 10344, "' indicating the desired formatting\n :param string tz: timezone indicator (optional)\n\n :return unicode: formatted date or time (as unicode string, to prevent jinja2 crash)\n\n (function used only in calendar_event_data.xml) '": 10345, "among": 10346, "'dayname'": 10347, "'time'": 10348, "indicator": 10349, "export_data": 10350, "\" override to convert virtual ids to ids \"": 10351, "real_ids": 10352, "'\n convert the search on real ids in the case it was asked on virtual ids, then call super()\n '": 10353, "\"res_id\"": 10354, "mail_message": 10355, "asked": 10356, "default_get": 10357, "'\n in case someone clicked on '": 10358, "invite": 10359, "others": 10360, "' wizard in the followers widget, transform virtual ids in real ids\n '": 10361, "'default_res_id'": 10362, "default_res_id": 10363, "invite_wizard": 10364, "'res_id'": 10365, "someone": 10366, "'invite others'": 10367, "followers": 10368, "test_calling_conventions": 10369, "5330": 10370, "cprofile": 10371, "wouldn": 10372, "stmts": 10373, "\"max([0])\"": 10374, "\"max([0], key=int)\"": 10375, "\"max([0], **dict(key=int))\"": 10376, "\"max(*([0],))\"": 10377, "\"max(*([0],), key=int)\"": 10378, "\"max(*([0],), **dict(key=int))\"": 10379, "prof": 10380, "profilerclass": 10381, "0.001": 10382, "runctx": 10383, "locals": 10384, "test_security_group_default_rules_show": 10385, "test_security_group_default_rules_create": 10386, "rule_id": 10387, "_do_get": 10388, "'os-security-group-default-rules/%s'": 10389, "_verify_response": 10390, "'security-group-default-rules-show-resp'": 10391, "200": 10392, "process_transform": 10393, "ax": 10394, "return_trans": 10395, "force_trans": 10396, "matplotlib": 10397, "crawl_fig": 10398, "fig": 10399, "\"crawl the figure and process all axes\"": 10400, "renderer": 10401, "draw_figure": 10402, "get_figure_properties": 10403, "crawl_ax": 10404, "crawl": 10405, "\"crawl the axes and process all elements within\"": 10406, "draw_axes": 10407, "get_axes_properties": 10408, "draw_line": 10409, "texts": 10410, "draw_text": 10411, "ttp": 10412, "xaxis": 10413, "yaxis": 10414, "\"xlabel\"": 10415, "\"ylabel\"": 10416, "\"title\"": 10417, "'get_text'": 10418, "crawl_legend": 10419, "legend": 10420, "\"\n recursively look through objects in legend children\n \"": 10421, "legendelements": 10422, "iter_all_children": 10423, "_legend_box": 10424, "skipcontainers": 10425, "legendpatch": 10426, "zorder": 10427, "appears": 10428, "set_zorder": 10429, "e6": 10430, "get_zorder": 10431, "patches": 10432, "recursively": 10433, "\"process a matplotlib line and call renderer.draw_line\"": 10434, "get_transform": 10435, "get_xydata": 10436, "linestyle": 10437, "get_line_style": 10438, "'dasharray'": 10439, "markerstyle": 10440, "get_marker_style": 10441, "markers": 10442, "\"process a matplotlib text object and call renderer.draw_text\"": 10443, "get_text_style": 10444, "posit": 10445, "draw_patch": 10446, "\"process a matplotlib patch object and call renderer.draw_path\"": 10447, "vertices": 10448, "pathcodes": 10449, "svg_path": 10450, "get_path": 10451, "get_path_style": 10452, "get_fill": 10453, "draw_path": 10454, "draw_collection": 10455, "force_pathtrans": 10456, "force_offsettrans": 10457, "\"process a matplotlib collection and call renderer.draw_collection\"": 10458, "transoffset": 10459, "offsets": 10460, "_prepare_points": 10461, "offset_coords": 10462, "path_coords": 10463, "processed_paths": 10464, "draw_image": 10465, "\"process a matplotlib image object and call renderer.draw_image\"": 10466, "imdata": 10467, "image_to_base64": 10468, "get_extent": 10469, "\"data\"": 10470, "\"alpha\"": 10471, "get_alpha": 10472, "\"zorder\"": 10473, "mplobj": 10474, "\"returns the time the robots.txt file was last fetched.\r\n\r\n this is useful for long-running web spiders that need to\r\n check for new robots.txt files periodically.\r\n\r\n \"": 10475, "last_checked": 10476, "robots": 10477, "web": 10478, "spiders": 10479, "periodically": 10480, "\"sets the time the robots.txt file was last fetched to the\r\n current time.\r\n\r\n \"": 10481, "set_url": 10482, "\"sets the url referring to a robots.txt file.\"": 10483, "urlparse": 10484, "referring": 10485, "can_fetch": 10486, "useragent": 10487, "\"using the parsed robots.txt decide if useragent can fetch url\"": 10488, "disallow_all": 10489, "allow_all": 10490, "allowable": 10491, "prevents": 10492, "positives": 10493, "erronenously": 10494, "decide": 10495, "applies_to": 10496, "\"check if this entry applies to the specified agent\"": 10497, "useragents": 10498, "catch": 10499, "allowance": 10500, "\"preconditions:\r\n - our agent applies to this entry\r\n - filename is url decoded\"": 10501, "rulelines": 10502, "preconditions": 10503, "decoded": 10504, "http_error_default": 10505, "errcode": 10506, "urllib": 10507, "fancyurlopener": 10508, "\"take an ip string/int and return an object of the correct type.\n\n args:\n address: a string or integer, the ip address. either ipv4 or\n ipv6 addresses may be supplied; integers less than 2**32 will\n be considered to be ipv4 by default.\n\n returns:\n an ipv4address or ipv6address object.\n\n raises:\n valueerror: if the *address* passed isn't either a v4 or a v6\n address\n\n \"": 10509, "addressvalueerror": 10510, "netmaskvalueerror": 10511, "considered": 10512, "ip_network": 10513, "\"take an ip string/int and return an object of the correct type.\n\n args:\n address: a string or integer, the ip network. either ipv4 or\n ipv6 networks may be supplied; integers less than 2**32 will\n be considered to be ipv4 by default.\n\n returns:\n an ipv4network or ipv6network object.\n\n raises:\n valueerror: if the string passed isn't either a v4 or a v6\n address. or if the network has host bits set.\n\n \"": 10514, "ipv4network": 10515, "networks": 10516, "ipv6network": 10517, "ip_interface": 10518, "ipv4interface": 10519, "ipv6interface": 10520, "'t either a v4 or a v6\n address.\n\n notes:\n the ipv?interface classes describe an address on a particular\n network, so they'": 10521, "basically": 10522, "combination": 10523, "bo": 10524, "obje": 10525, "v4_int_to_packed": 10526, "\"represent an address as 4 packed bytes in network (big-endian) order.\n\n args:\n address: an integer representation of an ipv4 ip address.\n\n returns:\n the integer address packed as 4 bytes in network (big-endian) order.\n\n raises:\n valueerror: if the integer is negative or too large to be an\n ipv4 ip address.\n\n \"": 10527, "_compat_to_bytes": 10528, "'big'": 10529, "overflowerror": 10530, "\"address negative or too large for ipv4\"": 10531, "big": 10532, "endian": 10533, "v4": 10534, "v6_int_to_packed": 10535, "\"represent an address as 16 packed bytes in network (big-endian) order.\n\n args:\n address: an integer representation of an ipv6 ip address.\n\n returns:\n the integer address packed as 16 bytes in network (big-endian) order.\n\n \"": 10536, "\"address negative or too large for ipv6\"": 10537, "v6": 10538, "summarize_address_range": 10539, "summarize": 10540, "'192.0.2.0'": 10541, "'192.0.2.130'": 10542, "normalize_whitespace": 10543, "'192.0.2.0/25'": 10544, "'192.0.2.128/31'": 10545, "'192.0.2.130/32'": 10546, "192.0": 10547, "collapse_addresses": 10548, "\"collapse a list of ip objects.\n\n example:\n collapse_addresses([ipv4network('192.0.2.0/25'),\n ipv4network('192.0.2.128/25')]) ->\n [ipv4network('192.0.2.0/24')]\n\n args:\n addresses: an iterator of ipv4network or ipv6network objects.\n\n returns:\n an iterator of the collapsed ipv(4|6)network objects.\n\n raises:\n typeerror: if passed a list of mixed version objects.\n\n \"": 10549, "addrs": 10550, "ips": 10551, "nets": 10552, "collapse": 10553, "'192.0.2.128/25'": 10554, "'192.0.2.0/24'": 10555, "get_mixed_type_key": 10556, "\"return a key suitable for sorting between networks and addresses.\n\n address and network objects are not sortable by default; they're\n fundamentally different so the expression\n\n ipv4address('192.0.2.0') <= ipv4network('192.0.2.0/24')\n\n doesn't make any sense. there are some times however, where you may wish\n to have ipaddress sort these for you anyway. if you need to do this, you\n can use this function as the key= argument to sorted().\n\n args:\n obj: either a network or address object.\n returns:\n appropriate key.\n\n \"": 10557, "suitable": 10558, "sortable": 10559, "'re\nfundamentally different so the expression\n\n ipv4address('": 10560, "') <= ipv4network('": 10561, "')\n\ndoesn'": 10562, "__le__": 10563, "__eq__": 10564, "exploded": 10565, "\"return the longhand version of the ip address as a string.\"": 10566, "_explode_shorthand_ip_string": 10567, "longhand": 10568, "\"return the shorthand version of the ip address as a string.\"": 10569, "_compat_str": 10570, "shorthand": 10571, "reverse_pointer": 10572, "\"the name of the reverse dns pointer for the ip address, e.g.:\n >>> ipaddress.ip_address(\"": 10573, "127.0": 10574, "\").reverse_pointer\n '1.0.0.127.in-addr.arpa'\n >>> ipaddress.ip_address(\"": 10575, "2001": 10576, "db8": 10577, "\").reverse_pointer\n '1.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.8.b.d.0.1.0.0.2.ip6.arpa'\n\n \"": 10578, "_reverse_pointer": 10579, "dns": 10580, "'1.0.0.127.in-addr.arpa'": 10581, "\"2001:db8::1\"": 10582, "'1.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.8.b.d.0.1.0.0.2.ip6.arpa'": 10583, "'%200s has no version specified'": 10584, "__sub__": 10585, "_compat_int_types": 10586, "\"generate iterator over usable hosts in a network.\n\n this is like __iter__ except it doesn't return the network\n or broadcast addresses.\n\n \"": 10587, "network_address": 10588, "broadcast_address": 10589, "_compat_range": 10590, "_address_class": 10591, "usable": 10592, "__hash__": 10593, "netmask": 10594, "overlaps": 10595, "\"tell if self is partly contained in other.\"": 10596, "tell": 10597, "partly": 10598, "num_addresses": 10599, "\"number of hosts in the current subnet.\"": 10600, "address_exclude": 10601, "addr1": 10602, "'192.0.2.0/28'": 10603, "addr2": 10604, "'192.0.2.1/32'": 10605, "'192.0.2.0/32'": 10606, "'192.0.2.2/31'": 10607, "'192.0.2.4/30'": 10608, "'192.0.2.8/29'": 10609, "'2001:db8::1/32'": 10610, "'2001:db8::1/128'": 10611, "compare_networks": 10612, "concerned": 10613, "'t considered at all in this method. if you want\n to compare host bits, you can easily enough do a\n '": 10614, "hosta": 10615, "_ip": 10616, "hostb": 10617, "'\n\n args:\n other: an ip object.\n\n returns:\n if the ip versions of self and other are the same, returns:\n\n -1 if self < other:\n eg: ipv4network('": 10618, "25": 10619, "ipv4netw": 10620, "'t considered at all in this method. if you want\nto compare host bits, you can easily enough do a\n'": 10621, "subnets": 10622, "prefixlen_diff": 10623, "new_prefix": 10624, "_prefixlen": 10625, "128": 10626, "ourself": 10627, "increased": 10628, "supernet": 10629, "decreased": 10630, "netw": 10631, "is_multicast": 10632, "\"test if the address is reserved for multicast use.\n\n returns:\n a boolean, true if the address is a multicast address.\n see rfc 2373 2.7 for details.\n\n \"": 10633, "reserved": 10634, "multicast": 10635, "2373": 10636, "2.7": 10637, "subnet_of": 10638, "_version": 10639, "dealing": 10640, "'network_address'": 10641, "'broadcast_address'": 10642, "'unable to test subnet containment with element '": 10643, "is_reserved": 10644, "\"test if the address is otherwise ietf reserved.\n\n returns:\n a boolean, true if the address is within one of the\n reserved ipv6 network ranges.\n\n \"": 10645, "ietf": 10646, "ranges": 10647, "is_link_local": 10648, "\"test if the address is reserved for link-local.\n\n returns:\n a boolean, true if the address is reserved per rfc 4291.\n\n \"": 10649, "4291.": 10650, "is_private": 10651, "\"test if this address is allocated for private networks.\n\n returns:\n a boolean, true if the address is reserved per\n iana-ipv4-special-registry or iana-ipv6-special-registry.\n\n \"": 10652, "allocated": 10653, "iana": 10654, "is_global": 10655, "\"test if this address is allocated for public networks.\n\n returns:\n a boolean, true if the address is not reserved per\n iana-ipv4-special-registry or iana-ipv6-special-registry.\n\n \"": 10656, "is_unspecified": 10657, "\"test if the address is unspecified.\n\n returns:\n a boolean, true if this is the unspecified address as defined in\n rfc 2373 2.5.2.\n\n \"": 10658, "unspecified": 10659, "is_loopback": 10660, "\"test if the address is a loopback address.\n\n returns:\n a boolean, true if the address is a loopback address as defined in\n rfc 2373 2.5.3.\n\n \"": 10661, "loopback": 10662, "\"the binary representation of this address.\"": 10663, "with_hostmask": 10664, "_string_from_ip_int": 10665, "hostmask": 10666, "max_prefixlen": 10667, "_max_prefixlen": 10668, "is_site_local": 10669, "\"test if the address is reserved for site-local.\n\n note that the site-local address space has been deprecated by rfc 3879.\n use is_private to test if this address is in the space of unique local\n addresses as defined by rfc 4193.\n\n returns:\n a boolean, true if the address is reserved per rfc 3513 2.5.6.\n\n \"": 10670, "_sitelocal_network": 10671, "site": 10672, "3879.": 10673, "4193.": 10674, "3513": 10675, "ipv4_mapped": 10676, "\"return the ipv4 mapped address.\n\n returns:\n if the ipv6 address is a v4 mapped address, return the\n ipv4 mapped address. return none otherwise.\n\n \"": 10677, "xffff": 10678, "xffffffff": 10679, "teredo": 10680, "\"tuple of embedded teredo ips.\n\n returns:\n tuple of the (server, client) ips or none if the address\n doesn't appear to be a teredo address (doesn't start with\n 2001::/32)\n\n \"": 10681, "96": 10682, "x20010000": 10683, "'t appear to be a teredo address (doesn'": 10684, "sixtofour": 10685, "\"return the ipv4 6to4 embedded address.\n\n returns:\n the ipv4 6to4-embedded address if present or none if the\n address doesn't appear to contain a 6to4 embedded address.\n\n \"": 10686, "112": 10687, "x2002": 10688, "to4": 10689, "form_valid": 10690, "\"your password was changed, \"": 10691, "\"hence you have been logged out. please relogin\"": 10692, "\"home\"": 10693, "is_iterable": 10694, "\"a implementation independent way of checking for iterables\"": 10695, "iterables": 10696, "escape_html": 10697, "_escape_html_table": 10698, "\"escape &, <, > as well as single and double quotes for html.\"": 10699, "quotes": 10700, "get_style_defs": 10701, "\"\n return css style definitions for the classes produced by the current\n highlighting style. ``arg`` can be a string or list of selectors to\n insert before the token type classes.\n \"": 10702, "'cssclass'": 10703, "cssclass": 10704, "highlighting": 10705, "selectors": 10706, "defs": 10707, "\"\n wrap the ``source``, which is a generator yielding\n individual lines, in custom generators. see docstring\n for `format`. can be overridden.\n \"": 10708, "_wrap_div": 10709, "_wrap_pre": 10710, "yielding": 10711, "generators": 10712, "docstring": 10713, "format_unencoded": 10714, "tokensource": 10715, "\"\n the formatting process uses several nested generators; which of\n them are used is determined by the user's options.\n\n each generator should take at least one argument, ``inner``,\n and wrap the pieces of text generated by this.\n\n always yield 2-tuples: (code, text). if \"": 10716, "highlighted": 10717, "wrapping": 10718, "several": 10719, "wrappers": 10720, "origina": 10721, "pieces": 10722, "unencoded": 10723, "get_execution_environment": 10724, "executionenvironment": 10725, "\"\n get the executionenvironment. if the executionenvironment has not been set,\n it initial the executionenvironment with default configuration.\n\n :return: the batch executionenvironment.\n\n .. versionadded:: 1.11.0\n \"": 10726, "_exe_env": 10727, "1.11": 10728, "get_stream_execution_environment": 10729, "streamexecutionenvironment": 10730, "\"\n get the streamexecutionenvironment. if the streamexecutionenvironment has not been\n set, it initial the streamexecutionenvironment with default configuration.\n\n :return: the streamexecutionenvironment.\n\n .. versionadded:: 1.11.0\n \"": 10731, "_stream_exe_env": 10732, "get_batch_table_environment": 10733, "batchtableenvironment": 10734, "\"\n get the batchtableenvironment. if the batchtableenvironment has not been set,\n it initial the batchtableenvironment with default configuration.\n\n :return: the batchtableenvironment.\n\n .. versionadded:: 1.11.0\n \"": 10735, "_batch_tab_env": 10736, "get_stream_table_environment": 10737, "streamtableenvironment": 10738, "\"\n get the streamtableenvironment. if the streamtableenvironment has not been set,\n it initial the streamtableenvironment with default configuration.\n\n :return: the streamtableenvironment.\n\n .. versionadded:: 1.11.0\n \"": 10739, "_stream_tab_env": 10740, "\"clears the mockcloudbucket's datastore.\"": 10741, "datastore": 10742, "mockcloudbucket": 10743, "downloadfile": 10744, "has_key": 10745, "cloud_bucket": 10746, "fileexists": 10747, "report_start": 10748, "\" reports an event's start.\n note: you *must* fire off a corresponding event end with report_end\n \"": 10749, "thread_dispatcher": 10750, "dispatch_event": 10751, "\"start\"": 10752, "report_end": 10753, "\" reports an event's end.\n note: you *must* have fired off a corresponding event start with report_start\n \"": 10754, "\"end\"": 10755, "report_start_end": 10756, "\"\n returns a function decorator and/or context manager which raises start and end events.\n if name is none events name is set to the name of the decorated function. in that case report_start_end\n can not be used as a context manager.\n \"": 10757, "_reporting_decorator_context_manager": 10758, "decorated": 10759, "report_value": 10760, "\"\n reports a value event to the counters.\n \"": 10761, "counters": 10762, "register_counter": 10763, "throw_if_exists": 10764, "\" register a counter with pycounters\n \"": 10765, "global_registry": 10766, "add_counter": 10767, "pycounters": 10768, "unregister_counter": 10769, "\" removes a previously registered counter\n \"": 10770, "remove_counter": 10771, "previously": 10772, "output_report": 10773, "\"\n manually cause the current values of all registered counters to be reported.\n \"": 10774, "reporters": 10775, "global_reporting_controller": 10776, "start_auto_reporting": 10777, "300": 10778, "\"\n start reporting in a background thread. reporting frequency is set by seconds param.\n \"": 10779, "start_auto_report": 10780, "reporting": 10781, "frequency": 10782, "stop_auto_reporting": 10783, "\" stop auto reporting \"": 10784, "stop_auto_report": 10785, "register_reporter": 10786, "reporter": 10787, "\"\n add a reporter to pycounters. registered reporters will output collected metrics\n \"": 10788, "metrics": 10789, "unregister_reporter": 10790, "\"\n remove a reporter from pycounters.\n \"": 10791, "configure_multi_process_collection": 10792, "collecting_address": 10793, "60907": 10794, "60906": 10795, "timeout_in_sec": 10796, "collectingrole": 10797, "auto_role": 10798, "configures": 10799, "machines": 10800, "ports": 10801, "temporarily": 10802, "unavailable": 10803, "becom": 10804, "page_title_breadcrumbs": 10805, "crumbs": 10806, "\"\n this function creates a suitable page title in the form:\n specific | less specific | general | edx\n it will output the correct platform name for the request.\n pass in a `separator` kwarg to override the default of \"": 10807, "\"\n \"": 10808, "\"separator\"": 10809, "\" | \"": 10810, "'{}{}{}'": 10811, "platform_name": 10812, "edx": 10813, "kwarg": 10814, "page_title_breadcrumbs_tag": 10815, "\"\n django template that creates breadcrumbs for page titles:\n {% page_title_breadcrumbs \"": 10816, "\" general %}\n \"": 10817, "titles": 10818, "\"specific\"": 10819, "\"less specific\"": 10820, "\"\n django template tag that outputs the current platform name:\n {% platform_name %}\n \"": 10821, "'platform_name'": 10822, "favicon_path": 10823, "'favicon_path'": 10824, "'images/favicon.ico'": 10825, "\"\n django template tag that outputs the configured favicon:\n {% favicon_path %}\n \"": 10826, "favicon": 10827, "microsite_css_overrides_file": 10828, "\"\n django template tag that outputs the css import for a:\n {% microsite_css_overrides_file %}\n \"": 10829, "file_path": 10830, "'css_overrides_file'": 10831, "\"\"": 10832, "overrides": 10833, "assert_npequal": 10834, "array_equal": 10835, "'\\n{} != {}'": 10836, "npequal": 10837, "test_merge_system": 10838, "chemlab": 10839, "datafile": 10840, "graphics": 10841, "display_system": 10842, "chemlabdb": 10843, "water": 10844, "\"molecule\"": 10845, "\"example.water\"": 10846, "prot": 10847, "\"tests/data/3zje.pdb\"": 10848, "\"system\"": 10849, "nwat": 10850, "50000": 10851, "bsize": 10852, "20.0": 10853, "wat": 10854, "box_vectors": 10855, "move_to": 10856, "r_array": 10857, "merge_systems": 10858, "test_crystal": 10859, "'building a crystal by using spacegroup module'": 10860, "molecule": 10861, "'na'": 10862, "cl": 10863, "'cl'": 10864, "fract": 10865, "tsys": 10866, "crystal": 10867, "225": 10868, "repetitions": 10869, "spacegroup": 10870, "test_sort": 10871, "type_array": 10872, "n_mol": 10873, "test_bond_orders": 10874, "bonds": 10875, "_make_water": 10876, "wat_o": 10877, "bond_orders": 10878, "bond": 10879, "reorder_molecules": 10880, "orders": 10881, "test_random": 10882, "'testing random made box'": 10883, "cdb": 10884, "'gromacs.spce'": 10885, "random_lattice_box": 10886, "160": 10887, "made": 10888, "test_bond_guessing": 10889, "cirdb": 10890, "display_molecule": 10891, "mol": 10892, "'tests/data/3zje.pdb'": 10893, "'molecule'": 10894, "guess_bonds": 10895, "guessing": 10896, "systems": 10897, "benzenes": 10898, "bz": 10899, "\"tests/data/benzene.mol\"": 10900, "bzbonds": 10901, "separating": 10902, "bz2": 10903, "test_extending": 10904, "ndarrayattr": 10905, "marrayattr": 10906, "atomicfield": 10907, "mysystem": 10908, "'v_array'": 10909, "mymolecule": 10910, "myatom": 10911, "from_fields": 10912, "extending": 10913, "test_remove_atoms": 10914, "molecules": 10915, "mols": 10916, "_make_molecules": 10917, "remove_atoms": 10918, "assert_eqbonds": 10919, "test_reorder_molecules": 10920, "reordering": 10921, "ensure_defaults": 10922, "\"\n puts the defaults into the settings dictionary for a given connection\n where no settings is provided.\n \"": 10923, "connectiondoesnotexist": 10924, "\"the connection %s doesn't exist\"": 10925, "setdefault": 10926, "'atomic_requests'": 10927, "'autocommit'": 10928, "'engine'": 10929, "'django.db.backends.dummy'": 10930, "'django.db.backends.'": 10931, "prepare_test_settings": 10932, "\"\n makes sure the test settings are available in the 'test' sub-dictionary.\n \"": 10933, "test_settings": 10934, "'charset'": 10935, "'collation'": 10936, "'mirror'": 10937, "get_migratable_models": 10938, "app_config": 10939, "\"\n return app models allowed to be synchronized on provided db.\n \"": 10940, "allow_migrate_model": 10941, "synchronized": 10942, "migratable": 10943, "test_retrieve": 10944, "\"\n test retrieval of spatialrefsys model objects.\n \"": 10945, "test_srs": 10946, "srs": 10947, "spatialrefsys": 10948, "authority": 10949, "borked": 10950, "32140.": 10951, "hence": 10952, "'startswith'": 10953, "auth_name": 10954, "oracle_flag": 10955, "'auth_name'": 10956, "postgis": 10957, "test_osr": 10958, "\"\n test getting osr objects from spatialrefsys model objects.\n \"": 10959, "sr": 10960, "spheroid": 10961, "'spheroid'": 10962, "'geographic'": 10963, "geographic": 10964, "'projected'": 10965, "projected": 10966, "spatialite": 10967, "'spatialite'": 10968, "'t get '": 10969, "nad83": 10970, "texas": 10971, "south": 10972, "central": 10973, "osr": 10974, "test_ellipsoid": 10975, "\"\n test the ellipsoid property.\n \"": 10976, "ellipsoid": 10977, "ellps1": 10978, "'ellipsoid'": 10979, "prec": 10980, "'eprec'": 10981, "ellps2": 10982, "assertalmostequal": 10983, "test_add_entry": 10984, "\"\n test adding a new entry in the spatialrefsys model using the\n add_srs_entry utility.\n \"": 10985, "contrib": 10986, "gis": 10987, "add_srs_entry": 10988, "3857": 10989, "get_spheroid": 10990, "'spheroid['": 10991, "adding": 10992, "create_table": 10993, "'roles'": 10994, "sa": 10995, "nullable": 10996, "primary_key": 10997, "'permissions_json'": 10998, "'modified_by_id'": 10999, "'created_at'": 11000, "'current_timestamp'": 11001, "'updated_at'": 11002, "onupdate": 11003, "'context_id'": 11004, "test_tracking_context": 11005, "\"\n ensure the tracking context is set up in the api client correctly and\n automatically.\n \"": 11006, "ecommerce": 11007, "httpretty": 11008, "register_uri": 11009, "'{}/baskets/1/'": 11010, "test_api_url": 11011, "'{}'": 11012, "adding_headers": 11013, "'content-type'": 11014, "mock_tracker": 11015, "resolve_context": 11016, "'client_id'": 11017, "test_client_id": 11018, "'commerce.tracker.get_tracker'": 11019, "tracking": 11020, "test_client_unicode": 11021, "\"\n the client should handle json responses properly when they contain\n unicode character data.\n\n regression test for ecom-1606.\n \"": 11022, "expected_content": 11023, "'{\"result\": \"pr\u00e9paratoire\"}'": 11024, "'{}/baskets/1/order/'": 11025, "actual_object": 11026, "ecommerce_api_client": 11027, "baskets": 11028, "actual_obje": 11029, "ecom": 11030, "1606.": 11031, "test_urlizetrunc01": 11032, "'urlizetrunc01'": 11033, "'\"unsafe\" http://example.com/x=&y='": 11034, "mark_safe": 11035, "'"safe" http://example.com?x=&y='": 11036, "'\"unsafe\" http:... '": 11037, "'"safe" http:...'": 11038, "urlizetrunc01": 11039, "test_urlizetrunc02": 11040, "'urlizetrunc02'": 11041, "'"unsafe" http:... '": 11042, "urlizetrunc02": 11043, "test_truncate": 11044, "'http://31characteruri.com/test/'": 11045, "urlizetrunc": 11046, "''": 11047, "'http://31characteruri.com/test/'": 11048, "'http://31characteruri.com/t...'": 11049, "charac": 11050, "test_autoescape": 11051, "'foobarbuz'": 11052, "'foo<a href=" google.com ">bar</a>buz'": 11053, "autoescape": 11054, "visit_emptynode": 11055, "mparser": 11056, "emptynode": 11057, "visit_default_func": 11058, "visit_methodnode": 11059, "methodnode": 11060, "source_object": 11061, "visit_plusassignmentnode": 11062, "plusassignmentnode": 11063, "visit_uminusnode": 11064, "uminusnode": 11065, "gd": 11066, "\"initialize y_l arrays\"": 11067, "r_cg": 11068, "r2_g": 11069, "tiny": 11070, "78": 11071, "r_g": 11072, "rhat_cg": 11073, "l_l": 11074, "y_lg": 11075, "npy": 11076, "vectorize": 11077, "'spherical harmonic'": 11078, "lmax": 11079, "y_l": 11080, "to_file": 11081, "calculator": 11082, "'multipole.dat'": 11083, "\"expand the charge distribution in multipoles and write\n the result to a file\"": 11084, "finegd": 11085, "q_l": 11086, "rhot_g": 11087, "paropen": 11088, "prnt": 11089, "'# multipole expansion of the charge density'": 11090, "'# center ='": 11091, "'angstrom'": 11092, "'# lmax ='": 11093, "trac": 11094, "fysik": 11095, "dtu": 11096, "multipoles": 11097, "miles_graph": 11098, "\" return the cites example graph in miles_dat.txt\n from the stanford graphbase.\n \"": 11099, "miles_dat": 11100, "gz": 11101, "gzip": 11102, "fh": 11103, "'knuth_miles.txt.gz'": 11104, "nx": 11105, "cities": 11106, "\"*\"": 11107, "numfind": 11108, "\"^\\d+\"": 11109, "distances": 11110, "cites": 11111, "stanford": 11112, "graphbase": 11113, "miles": 11114, "get_bit": 11115, "bit_mask": 11116, "'\n utility function that extracts a bit from a byte\n should probably check on length of the bit_mask and the number of positions we are shifting the position... possible future refactor\n '": 11117, "onebit": 11118, "shifting": 11119, "future": 11120, "refactor": 11121, "hex_dump": 11122, "random_integer": 11123, "'\n hex_dump returns the hex representation of a number\n '": 11124, "zfill": 11125, "'\n open() attempts to open the serial port based on the current configuration settings of port and baud rate\n '": 11126, "usb": 11127, "baud_rate": 11128, "baud": 11129, "__enter__": 11130, "'\n __enter__ implements context manager features for the class\n '": 11131, "enter": 11132, "__exit__": 11133, "exc_ty": 11134, "exc_val": 11135, "'\n __exit__ implements context manager features for the class\n '": 11136, "read_frame": 11137, "essence": 11138, "raw": 11139, "radio": 11140, "typically": 11141, "x7e": 11142, "x00": 11143, "checksumbyte": 11144, "x14": 11145, "x92": 11146, "broadcasted": 11147, "sending": 11148, "format_element": 11149, "bfo": 11150, "\"\n prints the reprinted editions of a record\n\n @param separator: a separator between reprinted editions\n @see: place.py, publisher.py, imprint.py, date.py, pagination.py\n \"": 11151, "reprints": 11152, "'260__g'": 11153, "prints": 11154, "reprinted": 11155, "editions": 11156, "publisher": 11157, "imprint": 11158, "pagination": 11159, "parse_hostname": 11160, "'^hostname (\\s+)'": 11161, "hostname": 11162, "test_hiddeninput": 11163, "hiddeninput": 11164, "asserthtmlequal": 11165, "''": 11166, "'test@example.com'": 11167, "''": 11168, "'some \"quoted\" & ampersanded value'": 11169, "''": 11170, "ema": 11171, "test_checkboxinput": 11172, "checkboxinput": 11173, "'is_cool'": 11174, "''": 11175, "''": 11176, "'s not in ('": 11177, "', none, false, true) will check the checkbox\n # and set the '": 11178, "test_selectmultiple": 11179, "selectmultiple": 11180, "'beatles'": 11181, "'j'": 11182, "'john'": 11183, "'paul'": 11184, "'g'": 11185, "'george'": 11186, "'ringo'": 11187, "\"\"": 11197, "valu": 11198, "test_checkboxselectmultiple": 11199, "checkboxselectmultiple": 11200, "\"

    \n
  • \n
  • \n
  • \n
  • \n
\"": 11208, "test_multi": 11209, "mymultiwidget": 11210, "multiwidget": 11211, "decompress": 11212, "'__'": 11213, "format_output": 11214, "rendered_widgets": 11215, "'
'": 11216, "widgets": 11217, "textinput": 11218, "'class'": 11219, "'small'": 11220, "'lennon'": 11221, "\"text\"": 11222, "\"big\"": 11223, "\"john\"": 11224, "\"name_0\"": 11225, "\"small\"": 11226, "test_datetimeinput": 11227, "datetimeinput": 11228, "''": 11229, "2007": 11230, "51": 11231, "34": 11232, "482548": 11233, "'2007-09-17 12:51:34.482548'": 11234, "trimmed": 11235, "''": 11236, "\"date\"": 11237, "09": 11238, "test_dateinput": 11239, "dateinput": 11240, "'2007-09-17'": 11241, "''": 11242, "inpu": 11243, "test_splithiddendatetime": 11244, "splithiddendatetimewidget": 11245, "''": 11246, "''": 11247, "splithiddendatetime": 11248, "test_multiwidget": 11249, "multiwidgets": 11250, "multivaluefields": 11251, "mulitwidgets": 11252, "splitdatetimewidget": 11253, "complexmultiwidget": 11254, "test_timeinput": 11255, "timeinput": 11256, "''": 11257, "test_datetime_locale_aware": 11258, "use_l10n": 11259, "translation": 11260, "'es'": 11261, "''": 11262, "aware": 11263, "test_nullbooleanselect": 11264, "\"\n ensure that the nullbooleanselect widget's options are lazily\n localized.\n refs #17190\n \"": 11265, "nullbooleanselectlazyform": 11266, "'bool'": 11267, "'id_bool'": 11268, "\"\n \n \"": 11274, "nullbooleanselect": 11275, "refs": 11276, "17190": 11277, "test_textarea_trailing_newlines": 11278, "\"\n test that a roundtrip on a modelform doesn't alter the textfield value\n \"": 11279, "\"\\ntst\\n\"": 11280, "selenium": 11281, "'%s%s'": 11282, "live_server_url": 11283, "'article_form'": 11284, "find_element_by_id": 11285, "'submit'": 11286, "19251": 11287, "fixed": 11288, "\"\\r\\ntst\\r\\n\"": 11289, "roundtrip": 11290, "modelform": 11291, "textfield": 11292, "textarea": 11293, "newlines": 11294, "test_clear_input_renders": 11295, "\"\n a clearablefileinput with is_required false and rendered with\n an initial value that is a file renders a clear checkbox.\n\n \"": 11296, "clearablefileinput": 11297, "is_required": 11298, "'myfile'": 11299, "fakefieldfile": 11300, "\"\n currently: something\n \n