Svak commited on
Commit
af34f14
·
verified ·
1 Parent(s): bbc3ba6

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ This quant was made for [infermatic.ai](https://infermatic.ai/)
3
+
4
+ Dynamic FP8 quant of [goliath-120b-Instruct-FP8-Dynamic](https://huggingface.co/alpindale/goliath-120b) made with AutoFP8.
5
+
6
+ ---
7
+ license: llama2
8
+ language:
9
+ - en
10
+ pipeline_tag: conversational
11
+ tags:
12
+ - merge
13
+ ---
14
+ # Goliath 120B
15
+
16
+ An auto-regressive causal LM created by combining 2x finetuned [Llama-2 70B](https://huggingface.co/meta-llama/llama-2-70b-hf) into one.
17
+
18
+ Please check out the quantized formats provided by [@TheBloke](https:///huggingface.co/TheBloke) and [@Panchovix](https://huggingface.co/Panchovix):
19
+
20
+ - [GGUF](https://huggingface.co/TheBloke/goliath-120b-GGUF) (llama.cpp)
21
+ - [GPTQ](https://huggingface.co/TheBloke/goliath-120b-GPTQ) (KoboldAI, TGW, Aphrodite)
22
+ - [AWQ](https://huggingface.co/TheBloke/goliath-120b-AWQ) (TGW, Aphrodite, vLLM)
23
+ - [Exllamav2](https://huggingface.co/Panchovix/goliath-120b-exl2) (TGW, KoboldAI)
24
+
25
+ # Prompting Format
26
+
27
+ Both Vicuna and Alpaca will work, but due the initial and final layers belonging primarily to Xwin, I expect Vicuna to work the best.