Upload folder using huggingface_hub
Browse files- README.md +45 -0
- config.json +25 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +1 -0
- tokenizer.json +3 -0
- tokenizer_config.json +1 -0
- vocab.txt +0 -0
README.md
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
base_model:
|
| 6 |
+
- facebook/contriever
|
| 7 |
+
---
|
| 8 |
+
|
| 9 |
+
OpenScholar_Retriever is a continued pre-trained version of [facebook/contriever](https://huggingface.co/facebook/contriever) for scientific literature synthesis.
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
### Model Description
|
| 14 |
+
|
| 15 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 16 |
+
|
| 17 |
+
- **Developed by:** University of Washigton, Allen Institute for AI (AI2)
|
| 18 |
+
- **Model type:** a masked language model.
|
| 19 |
+
- **Language(s) (NLP):** English
|
| 20 |
+
- **License:** The code and model are released under apache-2.0.
|
| 21 |
+
- **Date cutoff:** The pre-training data is mixture of [peS2o](https://huggingface.co/datasets/allenai/peS2o), [CCNews](https://huggingface.co/datasets/vblagoje/cc_news) and [Proofpile2](https://huggingface.co/datasets/EleutherAI/proof-pile-2).
|
| 22 |
+
|
| 23 |
+
### Model Sources
|
| 24 |
+
|
| 25 |
+
<!-- Provide the basic links for the model. -->
|
| 26 |
+
|
| 27 |
+
- **Project Page:** https://open-scholar.allen.ai/
|
| 28 |
+
- **Repositories:**
|
| 29 |
+
- Core repo (training, inference, fine-tuning etc.): https://github.com/AkariAsai/OpenScholar
|
| 30 |
+
- Evaluation code: https://github.com/AkariAsai/ScholarQABench
|
| 31 |
+
- **Paper:** [Link](https://openscholar.allen.ai/paper)
|
| 32 |
+
- **Technical blog post:** https://allenai.org/blog/openscholar
|
| 33 |
+
<!-- - **Press release:** TODO -->
|
| 34 |
+
|
| 35 |
+
### Citation
|
| 36 |
+
If you find it useful in this work, cite our paper.
|
| 37 |
+
|
| 38 |
+
```
|
| 39 |
+
@article{openscholar,
|
| 40 |
+
title={{OpenScholar}: Synthesizing Scientific Literature with Retrieval-Augmented Language Models},
|
| 41 |
+
author={ Asai, Akari and He*, Jacqueline and Shao*, Rulin and Shi, Weijia and Singh, Amanpreet and Chang, Joseph Chee and Lo, Kyle and Soldaini, Luca and Feldman, Tian, Sergey and Mike, D’arcy and Wadden, David and Latzke, Matt and Minyang and Ji, Pan and Liu, Shengyan and Tong, Hao and Wu, Bohao and Xiong, Yanyu and Zettlemoyer, Luke and Weld, Dan and Neubig, Graham and Downey, Doug and Yih, Wen-tau and Koh, Pang Wei and Hajishirzi, Hannaneh},
|
| 42 |
+
journal={Arxiv},
|
| 43 |
+
year={2024},
|
| 44 |
+
}
|
| 45 |
+
```
|
config.json
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"Contriever"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"classifier_dropout": null,
|
| 7 |
+
"gradient_checkpointing": false,
|
| 8 |
+
"hidden_act": "gelu",
|
| 9 |
+
"hidden_dropout_prob": 0.1,
|
| 10 |
+
"hidden_size": 768,
|
| 11 |
+
"initializer_range": 0.02,
|
| 12 |
+
"intermediate_size": 3072,
|
| 13 |
+
"layer_norm_eps": 1e-12,
|
| 14 |
+
"max_position_embeddings": 512,
|
| 15 |
+
"model_type": "bert",
|
| 16 |
+
"num_attention_heads": 12,
|
| 17 |
+
"num_hidden_layers": 12,
|
| 18 |
+
"pad_token_id": 0,
|
| 19 |
+
"position_embedding_type": "absolute",
|
| 20 |
+
"torch_dtype": "float32",
|
| 21 |
+
"transformers_version": "4.15.0",
|
| 22 |
+
"type_vocab_size": 2,
|
| 23 |
+
"use_cache": true,
|
| 24 |
+
"vocab_size": 30522
|
| 25 |
+
}
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb5226976e362e070ad5b3f6c5bd38ed03a770b7d349b9334e4db8d61cc5f8fa
|
| 3 |
+
size 435633182
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fd1c882abbd30517dced455a2c9768945ec726b96727927e4959348d9de550b
|
| 3 |
+
size 466081
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|