python_code stringlengths 0 187k | repo_name stringlengths 8 46 | file_path stringlengths 6 135 |
|---|---|---|
allenact-main | allenact_plugins/lighthouse_plugin/data/__init__.py | |
import os
from pathlib import Path
BABYAI_EXPERT_TRAJECTORIES_DIR = os.path.abspath(
os.path.join(os.path.dirname(Path(__file__)), "data", "demos")
)
| allenact-main | allenact_plugins/babyai_plugin/babyai_constants.py |
from allenact.utils.system import ImportChecker
with ImportChecker(
"\n\nPlease install babyai with:\n\n"
"pip install -e git+https://github.com/Lucaweihs/babyai.git@0b450eeb3a2dc7116c67900d51391986bdbb84cd#egg=babyai\n",
):
# noinspection PyUnresolvedReferences
import babyai
| allenact-main | allenact_plugins/babyai_plugin/__init__.py |
from typing import Dict, Optional, List, cast, Tuple, Any
import babyai.model
import babyai.rl
import gym
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from gym.spaces.dict import Dict as SpaceDict
from allenact.algorithms.onpolicy_sync.policy import (
ActorCriticModel,
... | allenact-main | allenact_plugins/babyai_plugin/babyai_models.py |
import random
import signal
from typing import Tuple, Any, List, Dict, Optional, Union, Callable
import babyai
import babyai.bot
import gym
import numpy as np
from gym.utils import seeding
from gym_minigrid.minigrid import MiniGridEnv
from allenact.base_abstractions.misc import RLStepResult
from allenact.base_abstrac... | allenact-main | allenact_plugins/babyai_plugin/babyai_tasks.py |
allenact-main | allenact_plugins/babyai_plugin/configs/__init__.py | |
import glob
import os
import babyai
from allenact_plugins.babyai_plugin.babyai_constants import (
BABYAI_EXPERT_TRAJECTORIES_DIR,
)
def make_small_demos(dir: str):
for file_path in glob.glob(os.path.join(dir, "*.pkl")):
if "valid" not in file_path and "small" not in file_path:
new_file_p... | allenact-main | allenact_plugins/babyai_plugin/scripts/truncate_expert_demos.py |
allenact-main | allenact_plugins/babyai_plugin/scripts/__init__.py | |
import glob
import os
import babyai
import numpy as np
from allenact_plugins.babyai_plugin.babyai_constants import (
BABYAI_EXPERT_TRAJECTORIES_DIR,
)
# Boss level
# [(50, 11.0), (90, 22.0), (99, 32.0), (99.9, 38.0), (99.99, 43.0)]
if __name__ == "__main__":
# level = "BossLevel"
level = "GoToLocal"
... | allenact-main | allenact_plugins/babyai_plugin/scripts/get_instr_length_percentiles.py |
import argparse
import os
import platform
from allenact_plugins.babyai_plugin.babyai_constants import (
BABYAI_EXPERT_TRAJECTORIES_DIR,
)
LEVEL_TO_TRAIN_VALID_IDS = {
"BossLevel": (
"1DkVVpIEVtpyo1LxOXQL_bVyjFCTO3cHD",
"1ccEFA_n5RT4SWD0Wa_qO65z2HACJBace",
),
"GoToObjMaze": (
"1... | allenact-main | allenact_plugins/babyai_plugin/scripts/download_babyai_expert_demos.py |
allenact-main | allenact_plugins/babyai_plugin/data/__init__.py | |
import random
from typing import Dict, Tuple, List, Any, Optional, Union, Sequence, cast
import gym
import numpy as np
from allenact.base_abstractions.misc import RLStepResult
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import Task
from allenact.utils.system import get_lo... | allenact-main | allenact_plugins/ithor_plugin/ithor_tasks.py |
"""A wrapper for engaging with the THOR environment."""
import copy
import functools
import math
import random
from typing import Tuple, Dict, List, Set, Union, Any, Optional, Mapping, cast
import ai2thor.server
import networkx as nx
import numpy as np
from ai2thor.controller import Controller
from scipy.spatial.tran... | allenact-main | allenact_plugins/ithor_plugin/ithor_environment.py |
from allenact.utils.system import ImportChecker
with ImportChecker(
"Cannot `import ai2thor`, please install `ai2thor` (`pip install ai2thor`)."
):
# noinspection PyUnresolvedReferences
import ai2thor
| allenact-main | allenact_plugins/ithor_plugin/__init__.py |
"""Common constants used when training agents to complete tasks in iTHOR, the
interactive version of AI2-THOR."""
from collections import OrderedDict
from typing import Set, Dict
MOVE_AHEAD = "MoveAhead"
ROTATE_LEFT = "RotateLeft"
ROTATE_RIGHT = "RotateRight"
LOOK_DOWN = "LookDown"
LOOK_UP = "LookUp"
END = "End"
VIS... | allenact-main | allenact_plugins/ithor_plugin/ithor_constants.py |
import glob
import math
import os
import platform
import traceback
import warnings
from contextlib import contextmanager
from typing import Sequence
import Xlib
import Xlib.display
import ai2thor.controller
@contextmanager
def include_object_data(controller: ai2thor.controller.Controller):
needs_reset = len(cont... | allenact-main | allenact_plugins/ithor_plugin/ithor_util.py |
import copy
from functools import reduce
from typing import Any, Dict, Optional, Union, Sequence
import ai2thor.controller
import gym
import gym.spaces
import numpy as np
import torch
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import Task
from allenact.embodiedai.mapping... | allenact-main | allenact_plugins/ithor_plugin/ithor_sensors.py |
import copy
import json
import math
import os
from typing import Tuple, Sequence, Union, Dict, Optional, Any, cast, Generator, List
import colour as col
import cv2
import numpy as np
from PIL import Image, ImageDraw
from ai2thor.controller import Controller
from matplotlib import pyplot as plt
from matplotlib.figure i... | allenact-main | allenact_plugins/ithor_plugin/ithor_viz.py |
import copy
import random
from typing import List, Dict, Optional, Any, Union, cast
import gym
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import TaskSampler
from allenact.utils.experiment_utils import set_deterministic_cudnn, set_seed
from allenact.utils.system import ge... | allenact-main | allenact_plugins/ithor_plugin/ithor_task_samplers.py |
import os
from allenact_plugins.robothor_plugin.scripts.make_objectnav_debug_dataset import (
create_debug_dataset_from_train_dataset,
)
if __name__ == "__main__":
CURRENT_PATH = os.getcwd()
SCENE = "FloorPlan1"
TARGET = "Apple"
EPISODES = [0, 7, 11, 12]
BASE_OUT = os.path.join(CURRENT_PATH, ... | allenact-main | allenact_plugins/ithor_plugin/scripts/make_objectnav_debug_dataset.py |
allenact-main | allenact_plugins/ithor_plugin/scripts/__init__.py | |
import os
from allenact_plugins.robothor_plugin.scripts.make_objectnav_debug_dataset import (
create_debug_dataset_from_train_dataset,
)
if __name__ == "__main__":
CURRENT_PATH = os.getcwd()
SCENE = "FloorPlan1"
EPISODES = [0, 7, 11, 12]
BASE_OUT = os.path.join(CURRENT_PATH, "datasets", "ithor-poi... | allenact-main | allenact_plugins/ithor_plugin/scripts/make_pointnav_debug_dataset.py |
from collections import OrderedDict
from typing import Dict, Any, Optional, List, cast
import gym
import numpy as np
import torch
from gym.spaces.dict import Dict as SpaceDict
from allenact.base_abstractions.preprocessor import Preprocessor
from allenact.utils.cacheless_frcnn import fasterrcnn_resnet50_fpn
from allen... | allenact-main | allenact_plugins/robothor_plugin/robothor_preprocessors.py |
import copy
import gzip
import json
import random
from typing import List, Optional, Union, Dict, Any, cast, Tuple
import gym
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import TaskSampler
from allenact.utils.cache_utils import str_to_pos_for_cache
from allenact.utils.exp... | allenact-main | allenact_plugins/robothor_plugin/robothor_task_samplers.py |
import copy
import math
import random
import warnings
from typing import Any, Optional, Dict, List, Union, Tuple, Collection
import ai2thor.server
import numpy as np
from ai2thor.controller import Controller
from ai2thor.fifo_server import FifoServer
from ai2thor.util import metrics
from allenact.utils.cache_utils im... | allenact-main | allenact_plugins/robothor_plugin/robothor_environment.py |
MOVE_AHEAD = "MoveAhead"
ROTATE_LEFT = "RotateLeft"
ROTATE_RIGHT = "RotateRight"
LOOK_DOWN = "LookDown"
LOOK_UP = "LookUp"
END = "End"
PASS = "Pass"
| allenact-main | allenact_plugins/robothor_plugin/robothor_constants.py |
from typing import Tuple
import torch
from allenact.base_abstractions.distributions import CategoricalDistr, Distr
class TupleCategoricalDistr(Distr):
def __init__(self, probs=None, logits=None, validate_args=None):
self.dists = CategoricalDistr(
probs=probs, logits=logits, validate_args=val... | allenact-main | allenact_plugins/robothor_plugin/robothor_distributions.py |
from typing import Tuple, Optional
import gym
import torch
from gym.spaces import Dict as SpaceDict
from allenact.algorithms.onpolicy_sync.policy import (
ActorCriticModel,
LinearActorCriticHead,
DistributionType,
Memory,
ObservationType,
)
from allenact.base_abstractions.misc import ActorCriticOu... | allenact-main | allenact_plugins/robothor_plugin/robothor_models.py |
from allenact.utils.system import ImportChecker
with ImportChecker(
"Cannot `import ai2thor`, please install `ai2thor` (`pip install ai2thor`)."
):
# noinspection PyUnresolvedReferences
import ai2thor
| allenact-main | allenact_plugins/robothor_plugin/__init__.py |
import math
from typing import Tuple, List, Dict, Any, Optional, Union, Sequence, cast
import gym
import numpy as np
from allenact.base_abstractions.misc import RLStepResult
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import Task
from allenact.utils.system import get_logg... | allenact-main | allenact_plugins/robothor_plugin/robothor_tasks.py |
from typing import Any, Tuple, Optional
import ai2thor.controller
import gym
import numpy as np
import quaternion # noqa # pylint: disable=unused-import
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import Task
from allenact.embodiedai.sensors.vision_sensors import RGBSens... | allenact-main | allenact_plugins/robothor_plugin/robothor_sensors.py |
import copy
import json
import math
import os
from typing import Tuple, Sequence, Union, Dict, Optional, Any, cast, Generator, List
import cv2
import numpy as np
from PIL import Image, ImageDraw
from ai2thor.controller import Controller
from matplotlib import pyplot as plt
from matplotlib.figure import Figure
import c... | allenact-main | allenact_plugins/robothor_plugin/robothor_viz.py |
allenact-main | allenact_plugins/robothor_plugin/configs/__init__.py | |
import gzip
import json
import os
from typing import Sequence, Optional
from allenact_plugins.robothor_plugin.robothor_task_samplers import (
ObjectNavDatasetTaskSampler,
)
def create_debug_dataset_from_train_dataset(
scene: str,
target_object_type: Optional[str],
episodes_subset: Sequence[int],
... | allenact-main | allenact_plugins/robothor_plugin/scripts/make_objectnav_debug_dataset.py |
allenact-main | allenact_plugins/robothor_plugin/scripts/__init__.py | |
import os
from allenact_plugins.robothor_plugin.scripts.make_objectnav_debug_dataset import (
create_debug_dataset_from_train_dataset,
)
if __name__ == "__main__":
CURRENT_PATH = os.getcwd()
SCENE = "FloorPlan_Train1_1"
EPISODES = [3, 4, 5, 6]
BASE_OUT = os.path.join(CURRENT_PATH, "datasets", "rob... | allenact-main | allenact_plugins/robothor_plugin/scripts/make_pointnav_debug_dataset.py |
import random
from typing import Tuple, Any, List, Dict, Optional, Union, Callable, Sequence, cast
import gym
import networkx as nx
import numpy as np
from gym.utils import seeding
from gym_minigrid.envs import CrossingEnv
from gym_minigrid.minigrid import (
DIR_TO_VEC,
IDX_TO_OBJECT,
MiniGridEnv,
OBJE... | allenact-main | allenact_plugins/minigrid_plugin/minigrid_tasks.py |
import copy
from typing import Optional, Set
import numpy as np
from gym import register
from gym_minigrid.envs import CrossingEnv
from gym_minigrid.minigrid import Lava, Wall
class FastCrossing(CrossingEnv):
"""Similar to `CrossingEnv`, but to support faster task sampling as per
`repeat_failed_task_for_min_... | allenact-main | allenact_plugins/minigrid_plugin/minigrid_environments.py |
from allenact.utils.system import ImportChecker
with ImportChecker(
"\n\nPlease install babyai with:\n\n"
"pip install -e git+https://github.com/Lucaweihs/babyai.git@0b450eeb3a2dc7116c67900d51391986bdbb84cd#egg=babyai\n",
):
import babyai
| allenact-main | allenact_plugins/minigrid_plugin/__init__.py |
import math
import queue
import random
from collections import defaultdict
from typing import Dict, Tuple, Any, cast, List, Union, Optional
import babyai
import blosc
import numpy as np
import pickle5 as pickle
import torch
from gym_minigrid.minigrid import MiniGridEnv
from allenact.algorithms.offpolicy_sync.losses.a... | allenact-main | allenact_plugins/minigrid_plugin/minigrid_offpolicy.py |
from typing import Optional, Any, cast
import gym
import gym_minigrid.minigrid
import numpy as np
import torch
from babyai.utils.format import InstructionsPreprocessor
from gym_minigrid.minigrid import MiniGridEnv
from allenact.base_abstractions.sensor import Sensor, prepare_locals_for_super
from allenact.base_abstra... | allenact-main | allenact_plugins/minigrid_plugin/minigrid_sensors.py |
import abc
from typing import Callable, Dict, Optional, Tuple, cast
import gym
import numpy as np
import torch
from gym.spaces.dict import Dict as SpaceDict
import torch.nn as nn
from allenact.algorithms.onpolicy_sync.policy import (
ActorCriticModel,
Memory,
DistributionType,
ActorCriticOutput,
O... | allenact-main | allenact_plugins/minigrid_plugin/minigrid_models.py |
allenact-main | allenact_plugins/minigrid_plugin/configs/__init__.py | |
"""Experiment Config for MiniGrid tutorial."""
import gym
import torch.nn as nn
from allenact.base_abstractions.sensor import SensorSuite
from allenact_plugins.minigrid_plugin.minigrid_models import MiniGridSimpleConv
from allenact_plugins.minigrid_plugin.minigrid_tasks import MiniGridTask
from projects.tutorials.min... | allenact-main | allenact_plugins/minigrid_plugin/configs/minigrid_nomemory.py |
allenact-main | allenact_plugins/minigrid_plugin/scripts/__init__.py | |
allenact-main | allenact_plugins/minigrid_plugin/data/__init__.py | |
"""Utility functions and classes for visualization and logging."""
import os
from datetime import datetime
import cv2
import imageio
import matplotlib
import matplotlib.cm as cm
import matplotlib.pyplot as plt
import numpy as np
from allenact_plugins.manipulathor_plugin.manipulathor_utils import initialize_arm
from a... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_viz.py |
"""Task Definions for the task of ArmPointNav."""
import copy
from typing import Dict, Tuple, List, Any, Optional
import gym
import numpy as np
from allenact.base_abstractions.misc import RLStepResult
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import Task
from allenact_... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_tasks.py |
"""Task Samplers for the task of ArmPointNav."""
import json
import random
from typing import List, Dict, Optional, Any, Union
import gym
from allenact.base_abstractions.sensor import Sensor
from allenact.base_abstractions.task import Task
from allenact.base_abstractions.task import TaskSampler
from allenact.utils.ex... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_task_samplers.py |
from allenact.utils.system import ImportChecker
with ImportChecker(
"Cannot `import ai2thor`, please install `ai2thor` (`pip install ai2thor`)."
):
# noinspection PyUnresolvedReferences
import ai2thor
| allenact-main | allenact_plugins/manipulathor_plugin/__init__.py |
"""Constant values and hyperparameters that are used by the environment."""
import ai2thor.fifo_server
ARM_MIN_HEIGHT = 0.450998873
ARM_MAX_HEIGHT = 1.8009994
ADDITIONAL_ARM_ARGS = {
"disableRendering": True,
"returnToStart": True,
"speed": 1,
}
MOVE_AHEAD = "MoveAheadContinuous"
MOVE_BACK = "MoveBackC... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_constants.py |
import json
import os
from typing import Dict, Optional, Any
from constants import ABS_PATH_OF_TOP_LEVEL_DIR
TRAIN_OBJECTS = ["Apple", "Bread", "Tomato", "Lettuce", "Pot", "Mug"]
TEST_OBJECTS = ["Potato", "SoapBottle", "Pan", "Egg", "Spatula", "Cup"]
MOVE_ARM_CONSTANT = 0.05
MOVE_ARM_HEIGHT_CONSTANT = MOVE_ARM_CONSTA... | allenact-main | allenact_plugins/manipulathor_plugin/armpointnav_constants.py |
"""Utility classes and functions for sensory inputs used by the models."""
from typing import Any, Union, Optional
import gym
import numpy as np
from allenact.base_abstractions.sensor import Sensor
from allenact.embodiedai.sensors.vision_sensors import DepthSensor, RGBSensor
from allenact.base_abstractions.task import... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_sensors.py |
"""Utility classes and functions for calculating the arm relative and absolute
position."""
from typing import Dict
import numpy as np
import torch
from scipy.spatial.transform import Rotation as R
from allenact.utils.system import get_logger
def state_dict_to_tensor(state: Dict):
result = []
if "position" ... | allenact-main | allenact_plugins/manipulathor_plugin/arm_calculation_utils.py |
import ai2thor
from allenact_plugins.ithor_plugin.ithor_environment import IThorEnvironment
from allenact_plugins.manipulathor_plugin.armpointnav_constants import (
get_agent_start_positions,
)
from allenact_plugins.manipulathor_plugin.manipulathor_constants import (
ADDITIONAL_ARM_ARGS,
)
def make_all_objec... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_utils.py |
"""A wrapper for engaging with the ManipulaTHOR environment."""
import copy
import math
import warnings
from typing import Dict, Union, Any, Optional, cast
import ai2thor.server
import numpy as np
from ai2thor.controller import Controller
from allenact.utils.misc_utils import prepare_locals_for_super
from allenact_p... | allenact-main | allenact_plugins/manipulathor_plugin/manipulathor_environment.py |
from typing import Optional
import gym
import numpy as np
class GymEnvironment(gym.Wrapper):
"""gym.Wrapper with minimal bookkeeping (initial observation)."""
def __init__(self, gym_env_name: str):
super().__init__(gym.make(gym_env_name))
self._initial_observation: Optional[np.ndarray] = Non... | allenact-main | allenact_plugins/gym_plugin/gym_environment.py |
allenact-main | allenact_plugins/gym_plugin/__init__.py | |
from typing import Optional, Any
import gym
import numpy as np
from allenact.base_abstractions.sensor import Sensor, prepare_locals_for_super
from allenact.base_abstractions.task import Task, SubTaskType
from allenact_plugins.gym_plugin.gym_environment import GymEnvironment
class GymBox2DSensor(Sensor[gym.Env, Task... | allenact-main | allenact_plugins/gym_plugin/gym_sensors.py |
import torch
from allenact.base_abstractions.distributions import Distr
class GaussianDistr(torch.distributions.Normal, Distr):
"""PyTorch's Normal distribution with a `mode` method."""
def mode(self) -> torch.FloatTensor:
return super().mean
| allenact-main | allenact_plugins/gym_plugin/gym_distributions.py |
from typing import Dict, Union, Optional, Tuple, Any, Sequence, cast
import gym
import torch
import torch.nn as nn
from allenact.algorithms.onpolicy_sync.policy import (
ActorCriticModel,
DistributionType,
)
from allenact.base_abstractions.misc import ActorCriticOutput, Memory
from allenact_plugins.gym_plugin... | allenact-main | allenact_plugins/gym_plugin/gym_models.py |
import random
from typing import Any, List, Dict, Optional, Union, Callable, Sequence, Tuple
import gym
import numpy as np
from gym.utils import seeding
from allenact.base_abstractions.misc import RLStepResult
from allenact.base_abstractions.sensor import Sensor, SensorSuite
from allenact.base_abstractions.task impor... | allenact-main | allenact_plugins/gym_plugin/gym_tasks.py |
allenact-main | allenact_plugins/navigation_plugin/__init__.py | |
"""Baseline models for use in the point navigation task.
Object navigation is currently available as a Task in AI2-THOR and
Facebook's Habitat.
"""
from typing import Optional, List, Union, Sequence
import gym
import torch
import torch.nn as nn
from gym.spaces import Dict as SpaceDict
from allenact.algorithms.onpoli... | allenact-main | allenact_plugins/navigation_plugin/pointnav/models.py |
allenact-main | allenact_plugins/navigation_plugin/pointnav/__init__.py | |
"""Baseline models for use in the object navigation task.
Object navigation is currently available as a Task in AI2-THOR and
Facebook's Habitat.
"""
from typing import Optional, List, Dict, cast, Tuple, Sequence
import gym
import torch
import torch.nn as nn
from gym.spaces import Dict as SpaceDict
from allenact.algo... | allenact-main | allenact_plugins/navigation_plugin/objectnav/models.py |
allenact-main | allenact_plugins/navigation_plugin/objectnav/__init__.py | |
from typing import List, Optional, Any, cast, Dict, Tuple
import clip
import gym
import numpy as np
import torch
import torch.nn as nn
from clip.model import CLIP
from allenact.base_abstractions.preprocessor import Preprocessor
from allenact.utils.misc_utils import prepare_locals_for_super
class ClipResNetEmbedder(... | allenact-main | allenact_plugins/clip_plugin/clip_preprocessors.py |
from allenact.utils.system import ImportChecker
with ImportChecker(
"Cannot `import clip`. Please install clip from the openai/CLIP git repository:"
"\n`pip install git+https://github.com/openai/CLIP.git@b46f5ac7587d2e1862f8b7b1573179d80dcdd620`"
):
# noinspection PyUnresolvedReferences
import clip
| allenact-main | allenact_plugins/clip_plugin/__init__.py |
# coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | abductive-commonsense-reasoning-master | anlg/run_lm_finetuning.py |
from __future__ import absolute_import, division, print_function, unicode_literals
import torch
import torch.nn as nn
from comet.models.utils import prepare_position_embeddings
from pytorch_transformers import GPT2PreTrainedModel
from pytorch_transformers.modeling_bert import BertLayerNorm as LayerNorm
from pytorch_tr... | abductive-commonsense-reasoning-master | anlg/models.py |
abductive-commonsense-reasoning-master | anlg/__init__.py | |
#!/usr/bin/env python3
# coding=utf-8
# Copyright 2018 Google AI, Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in c... | abductive-commonsense-reasoning-master | anlg/run_generation.py |
from comet.data.atomic import all_categories
from pytorch_transformers import GPT2Tokenizer
class AnliGpt2Tokenizer(GPT2Tokenizer):
def __init__(self,
vocab_file,
merges_file,
errors='replace',
unk_token="<|endoftext|>",
bos_toke... | abductive-commonsense-reasoning-master | anlg/tokenizers.py |
abductive-commonsense-reasoning-master | anlg/evaluation/__init__.py | |
from anlg.evaluation.bleu.bleu import Bleu
from anlg.evaluation.meteor.meteor_nltk import Meteor
from anlg.evaluation.rouge.rouge import Rouge
from anlg.evaluation.cider.cider import Cider
from anlg.evaluation.bert_score.bert_score import BertScore
from collections import defaultdict
from argparse import ArgumentParser... | abductive-commonsense-reasoning-master | anlg/evaluation/eval.py |
# Filename: cider.py
#
# Description: Describes the class to compute the CIDEr (Consensus-Based Image Description Evaluation) Metric
# by Vedantam, Zitnick, and Parikh (http://arxiv.org/abs/1411.5726)
#
# Creation Date: Sun Feb 8 14:16:54 2015
#
# Authors: Ramakrishna Vedantam <vrama91@vt.edu> and Tsung... | abductive-commonsense-reasoning-master | anlg/evaluation/cider/cider.py |
__author__ = 'tylin'
| abductive-commonsense-reasoning-master | anlg/evaluation/cider/__init__.py |
#!/usr/bin/env python
# Tsung-Yi Lin <tl483@cornell.edu>
# Ramakrishna Vedantam <vrama91@vt.edu>
import copy
from collections import defaultdict
import numpy as np
import pdb
import math
def precook(s, n=4, out=False):
"""
Takes a string as input and returns an object that can be given to
either cook_refs... | abductive-commonsense-reasoning-master | anlg/evaluation/cider/cider_scorer.py |
#!/usr/bin/env python
# Python wrapper for METEOR implementation, by Xinlei Chen
# Acknowledge Michael Denkowski for the generous discussion and help
import os
import sys
import nltk
from nltk.translate.meteor_score import meteor_score
# Assumes meteor-1.5.jar is in the same directory as meteor.py. Change as neede... | abductive-commonsense-reasoning-master | anlg/evaluation/meteor/meteor_nltk.py |
__author__ = 'tylin'
| abductive-commonsense-reasoning-master | anlg/evaluation/meteor/__init__.py |
#!/usr/bin/env python
# Python wrapper for METEOR implementation, by Xinlei Chen
# Acknowledge Michael Denkowski for the generous discussion and help
import os
import sys
import subprocess
import threading
# Assumes meteor-1.5.jar is in the same directory as meteor.py. Change as needed.
METEOR_JAR = 'meteor-1.5.ja... | abductive-commonsense-reasoning-master | anlg/evaluation/meteor/meteor.py |
#!/usr/bin/env python
#
# File Name : bleu.py
#
# Description : Wrapper for BLEU scorer.
#
# Creation Date : 06-01-2015
# Last Modified : Thu 19 Mar 2015 09:13:28 PM PDT
# Authors : Hao Fang <hfang@uw.edu> and Tsung-Yi Lin <tl483@cornell.edu>
from anlg.evaluation.bleu.bleu_scorer import BleuScorer
class Bleu:
d... | abductive-commonsense-reasoning-master | anlg/evaluation/bleu/bleu.py |
__author__ = 'tylin'
| abductive-commonsense-reasoning-master | anlg/evaluation/bleu/__init__.py |
#!/usr/bin/env python
# bleu_scorer.py
# David Chiang <chiang@isi.edu>
# Copyright (c) 2004-2006 University of Maryland. All rights
# reserved. Do not redistribute without permission from the
# author. Not for commercial use.
# Modified by:
# Hao Fang <hfang@uw.edu>
# Tsung-Yi Lin <tl483@cornell.edu>
'''Provides:
... | abductive-commonsense-reasoning-master | anlg/evaluation/bleu/bleu_scorer.py |
from bert_score import score
# Code for BertScore reused from original implementation: https://github.com/Tiiiger/bert_score
class BertScore:
def __init__(self):
self._hypo_for_image = {}
self.ref_for_image = {}
def compute_score(self, gts, res):
assert(gts.keys() == res.keys())
... | abductive-commonsense-reasoning-master | anlg/evaluation/bert_score/bert_score.py |
abductive-commonsense-reasoning-master | anlg/evaluation/bert_score/__init__.py | |
import torch
from math import log
from itertools import chain
from collections import defaultdict, Counter
from multiprocessing import Pool
from functools import partial
from tqdm.auto import tqdm
__all__ = ['bert_types']
bert_types = [
'bert-base-uncased',
'bert-large-uncased',
'bert-base-cased',
'be... | abductive-commonsense-reasoning-master | anlg/evaluation/bert_score/utils.py |
import os
import time
import argparse
import torch
from collections import defaultdict
from pytorch_pretrained_bert import BertTokenizer, BertModel, BertForMaskedLM
import matplotlib
import matplotlib.pyplot as plt
import numpy as np
from .utils import get_idf_dict, bert_cos_score_idf,\
get_bert_emb... | abductive-commonsense-reasoning-master | anlg/evaluation/bert_score/score.py |
#!/usr/bin/env python
#
# File Name : rouge.py
#
# Description : Computes ROUGE-L metric as described by Lin and Hovey (2004)
#
# Creation Date : 2015-01-07 06:03
# Author : Ramakrishna Vedantam <vrama91@vt.edu>
import numpy as np
import pdb
def my_lcs(string, sub):
"""
Calculates longest common subsequence ... | abductive-commonsense-reasoning-master | anlg/evaluation/rouge/rouge.py |
__author__ = 'vrama91'
| abductive-commonsense-reasoning-master | anlg/evaluation/rouge/__init__.py |
import argparse
import json
from transformers import BertTokenizer
import tqdm
from anli.data_processors import AnliProcessor
def data_processor_by_name(task_name):
if task_name == "anli":
return AnliProcessor()
def main(args):
data_dir = args.data_dir
bert_model = args.bert_model
task_na... | abductive-commonsense-reasoning-master | anli/max_ctx_for_dataset.py |
import argparse
import json
import numpy as np
from utils.file_utils import read_jsonl_lines, read_lines
def _key(r):
return r['obs1'] + '||' + r['obs2']
def correct_middle(r):
return r['hyp' + r['label']]
def incorrect_middle(r):
if r['label'] == "1":
return r['hyp2']
else:
retur... | abductive-commonsense-reasoning-master | anli/corpus_statistics.py |
abductive-commonsense-reasoning-master | anli/__init__.py | |
import csv
import json
import logging
import os
from abc import ABC
import torch
from pytorch_transformers import BertTokenizer
from torch.utils.data import TensorDataset, SequentialSampler, DataLoader, RandomSampler
from utils.file_utils import read_lines, read_jsonl_lines
logging.basicConfig(format='%(asctime)s - ... | abductive-commonsense-reasoning-master | anli/data_processors.py |
import argparse
import json
import logging
import math
import os
import random
import numpy as np
import torch
from pytorch_transformers import BertTokenizer, PYTORCH_PRETRAINED_BERT_CACHE, \
BertForMultipleChoice, BertConfig, BertModel
from tensorboardX import SummaryWriter
from torch.nn import CrossEntropyLoss
f... | abductive-commonsense-reasoning-master | anli/run_anli.py |
import argparse
import json
import torch
from pytorch_transformers import BertTokenizer, BertConfig, BertForMultipleChoice
from anli.data_processors import AnliExample, mc_examples_to_data_loader
from anli.run_anli import get_data_processor, model_choice_map
import numpy as np
def load_anli_model(model_name, saved_... | abductive-commonsense-reasoning-master | anli/demo.py |
abductive-commonsense-reasoning-master | anli/human_eval/__init__.py | |
import argparse
import json
from utils.file_utils import read_jsonl_lines, write_items
from collections import Counter
import hashlib
def _hash(w):
return hashlib.sha1(w.encode()).hexdigest()
def main(args):
input_file = args.input_file
output_file = args.output_file
records = read_jsonl_lines(inp... | abductive-commonsense-reasoning-master | anli/human_eval/compute_human_performance.py |
import argparse
import json
import pandas as pd
from pandas import DataFrame
from utils.file_utils import read_jsonl_lines, read_lines, write_items
def main(args):
dev_file = args.dev_file
dev_labels_file = args.dev_labels_file
output_file = args.output_file
records = read_jsonl_lines(dev_file)
... | abductive-commonsense-reasoning-master | anli/human_eval/prep_human_eval.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.