joaoalvarenga commited on
Commit
cf7c1ed
·
1 Parent(s): 2219291

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -55,7 +55,7 @@ pipeline_tag: text-generation
55
 
56
  Heavily inspired by [Hivemind's GPT-J-6B with 8-bit weights](https://huggingface.co/hivemind/gpt-j-6B-8bit), this is a version of [bigscience/bloom](https://huggingface.co/bigscience/bloom) a ~176 billions parameters language model that you run and fine-tune with less memory.
57
 
58
- Here, we also apply [LoRA (Low Rank Adapters)](https://arxiv.org/abs/2106.09685) to reduce model size. The original version takes ~353GB memory, this version takes ~180GB.
59
 
60
  Our main objective is to generate a model compressed enough to be deployed in a traditional Kubernetes cluster.
61
  ### How to use
 
55
 
56
  Heavily inspired by [Hivemind's GPT-J-6B with 8-bit weights](https://huggingface.co/hivemind/gpt-j-6B-8bit), this is a version of [bigscience/bloom](https://huggingface.co/bigscience/bloom) a ~176 billions parameters language model that you run and fine-tune with less memory.
57
 
58
+ Here, we also apply [LoRA (Low Rank Adapters)](https://arxiv.org/abs/2106.09685) to reduce model size. The original version takes \~353GB memory, this version takes **\~180GB**.
59
 
60
  Our main objective is to generate a model compressed enough to be deployed in a traditional Kubernetes cluster.
61
  ### How to use