File size: 560 Bytes
df051eb
58e8b0b
 
 
df051eb
58e8b0b
 
 
df051eb
58e8b0b
e149b0f
 
df051eb
58e8b0b
 
 
df051eb
58e8b0b
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
# Datasets

We are planning to use the following datasets to train the models.

## Base Model Training

[The British National Corpus](http://www.natcorp.ox.ac.uk/)

## Question Answering

[WikiQA (Wikipedia Open-Domain Question Answering](https://paperswithcode.com/dataset/wikiqa)

## Reasoning

[Avicenna: Syllogistic Commonsense Reasoning](https://github.com/ZeinabAghahadi/Syllogistic-Commonsense-Reasoning)

## Consistency

[Stanford Natural Language Inference Corpus](https://www.kaggle.com/datasets/stanfordu/stanford-natural-language-inference-corpus)