zeekay commited on
Commit
073dfe2
·
verified ·
1 Parent(s): 4ec08fa

Add research paper: paper/references.bib

Browse files
Files changed (1) hide show
  1. paper/references.bib +65 -0
paper/references.bib ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ @article{arditi2024refusal,
2
+ title={Refusal in Language Models Is Mediated by a Single Direction},
3
+ author={Arditi, Andy and Obeso, Oscar and Syed, Aaquib and Paleka, Daniel and Rimsky, Nina and Neel, Wes and Sharkey, Lee},
4
+ journal={arXiv preprint arXiv:2406.11717},
5
+ year={2024}
6
+ }
7
+
8
+ @misc{sumandora2024,
9
+ title={remove-refusals-with-transformers},
10
+ author={Sumandora},
11
+ year={2024},
12
+ howpublished={\url{https://github.com/Sumandora/remove-refusals-with-transformers}}
13
+ }
14
+
15
+ @misc{hamsaomar2025,
16
+ title={Kimi-K2.5 Abliteration Research},
17
+ author={hamsaOmar},
18
+ year={2025},
19
+ howpublished={\url{https://huggingface.co/hamsaOmar/Kimi-K2.5-abliterated}}
20
+ }
21
+
22
+ @article{hu2022lora,
23
+ title={LoRA: Low-Rank Adaptation of Large Language Models},
24
+ author={Hu, Edward J. and Shen, Yelong and Wallis, Phillip and Allen-Zhu, Zeyuan and Li, Yuanzhi and Wang, Shean and Wang, Lu and Chen, Weizhu},
25
+ journal={arXiv preprint arXiv:2106.09685},
26
+ year={2022}
27
+ }
28
+
29
+ @article{dettmers2023qlora,
30
+ title={QLoRA: Efficient Finetuning of Quantized Language Models},
31
+ author={Dettmers, Tim and Pagnoni, Artidoro and Holtzman, Ari and Zettlemoyer, Luke},
32
+ journal={Advances in Neural Information Processing Systems},
33
+ year={2023}
34
+ }
35
+
36
+ @article{rafailov2023direct,
37
+ title={Direct Preference Optimization: Your Language Model is Secretly a Reward Model},
38
+ author={Rafailov, Rafael and Sharma, Archit and Mitchell, Eric and Ermon, Stefano and Manning, Christopher D. and Finn, Chelsea},
39
+ journal={Advances in Neural Information Processing Systems},
40
+ year={2023}
41
+ }
42
+
43
+ @article{liu2024bitdelta,
44
+ title={BitDelta: Your Fine-Tune May Only Be Worth One Bit},
45
+ author={Liu, James and Zhong, Guangxuan and Li, Yifan and Kolter, J. Zico and Sabharwal, Ashish},
46
+ journal={arXiv preprint arXiv:2402.10193},
47
+ year={2024}
48
+ }
49
+
50
+ @article{shazeer2017outrageously,
51
+ title={Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer},
52
+ author={Shazeer, Noam and Mirhoseini, Azalia and Maziarz, Krzysztof and Davis, Andy and Le, Quoc and Hinton, Geoffrey and Dean, Jeff},
53
+ journal={arXiv preprint arXiv:1701.06538},
54
+ year={2017}
55
+ }
56
+
57
+ @article{fedus2022switch,
58
+ title={Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity},
59
+ author={Fedus, William and Zoph, Barret and Shazeer, Noam},
60
+ journal={Journal of Machine Learning Research},
61
+ volume={23},
62
+ number={120},
63
+ pages={1--39},
64
+ year={2022}
65
+ }