ritvikbio commited on
Commit
7e1b93b
·
verified ·
1 Parent(s): 22c132e

Delete files * with huggingface_hub

Browse files
README.md DELETED
@@ -1,45 +0,0 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- base_model:
6
- - facebook/contriever
7
- ---
8
-
9
- OpenScholar_Retriever is a continued pre-trained version of [facebook/contriever](https://huggingface.co/facebook/contriever) for scientific literature synthesis.
10
-
11
-
12
-
13
- ### Model Description
14
-
15
- <!-- Provide a longer summary of what this model is. -->
16
-
17
- - **Developed by:** University of Washigton, Allen Institute for AI (AI2)
18
- - **Model type:** a masked language model.
19
- - **Language(s) (NLP):** English
20
- - **License:** The code and model are released under apache-2.0.
21
- - **Date cutoff:** The pre-training data is mixture of [peS2o](https://huggingface.co/datasets/allenai/peS2o), [CCNews](https://huggingface.co/datasets/vblagoje/cc_news) and [Proofpile2](https://huggingface.co/datasets/EleutherAI/proof-pile-2).
22
-
23
- ### Model Sources
24
-
25
- <!-- Provide the basic links for the model. -->
26
-
27
- - **Project Page:** https://open-scholar.allen.ai/
28
- - **Repositories:**
29
- - Core repo (training, inference, fine-tuning etc.): https://github.com/AkariAsai/OpenScholar
30
- - Evaluation code: https://github.com/AkariAsai/ScholarQABench
31
- - **Paper:** [Link](https://openscholar.allen.ai/paper)
32
- - **Technical blog post:** https://allenai.org/blog/openscholar
33
- <!-- - **Press release:** TODO -->
34
-
35
- ### Citation
36
- If you find it useful in this work, cite our paper.
37
-
38
- ```
39
- @article{openscholar,
40
- title={{OpenScholar}: Synthesizing Scientific Literature with Retrieval-Augmented Language Models},
41
- author={ Asai, Akari and He*, Jacqueline and Shao*, Rulin and Shi, Weijia and Singh, Amanpreet and Chang, Joseph Chee and Lo, Kyle and Soldaini, Luca and Feldman, Tian, Sergey and Mike, D’arcy and Wadden, David and Latzke, Matt and Minyang and Ji, Pan and Liu, Shengyan and Tong, Hao and Wu, Bohao and Xiong, Yanyu and Zettlemoyer, Luke and Weld, Dan and Neubig, Graham and Downey, Doug and Yih, Wen-tau and Koh, Pang Wei and Hajishirzi, Hannaneh},
42
- journal={Arxiv},
43
- year={2024},
44
- }
45
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,25 +0,0 @@
1
- {
2
- "architectures": [
3
- "Contriever"
4
- ],
5
- "attention_probs_dropout_prob": 0.1,
6
- "classifier_dropout": null,
7
- "gradient_checkpointing": false,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_position_embeddings": 512,
15
- "model_type": "bert",
16
- "num_attention_heads": 12,
17
- "num_hidden_layers": 12,
18
- "pad_token_id": 0,
19
- "position_embedding_type": "absolute",
20
- "torch_dtype": "float32",
21
- "transformers_version": "4.15.0",
22
- "type_vocab_size": 2,
23
- "use_cache": true,
24
- "vocab_size": 30522
25
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb5226976e362e070ad5b3f6c5bd38ed03a770b7d349b9334e4db8d61cc5f8fa
3
- size 435633182
 
 
 
 
special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fd1c882abbd30517dced455a2c9768945ec726b96727927e4959348d9de550b
3
- size 466081
 
 
 
 
tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
 
 
vocab.txt DELETED
The diff for this file is too large to render. See raw diff