lthn commited on
Commit
2a8bd74
·
verified ·
1 Parent(s): 344ff17

docs: clean card, cross-link models

Browse files
Files changed (1) hide show
  1. README.md +16 -75
README.md CHANGED
@@ -1,89 +1,31 @@
1
  ---
2
  language:
3
- - en
4
  license: eupl-1.2
5
- library_name: mlx
6
  tags:
7
- - mlx
8
- - safetensors
9
- - gemma4
10
- - lethean
11
- - lem
12
- - text-generation
13
- - conversational
14
- - 4-bit
15
  base_model:
16
- - google/gemma-4-31B-it
17
  base_model_relation: quantized
18
- pipeline_tag: text-generation
 
 
19
  ---
20
 
21
  # Lemrd
22
 
23
- A Gemma 4 31B dense fine-tune by [Lethean Network](https://lthn.ai/lemrd).
24
-
25
- EUPL-1.2 · Apache 2.0 base · [lthn.ai/lemrd](https://lthn.ai/lemrd)
26
 
27
  ## Use
28
 
29
- ### MLX
30
-
31
- ```bash
32
- pip install mlx-lm
33
- ```
34
-
35
- ```python
36
- from mlx_lm import load, generate
37
-
38
- model, tokenizer = load("lthn/lemrd", revision="4bit")
39
- response = generate(model, tokenizer, prompt="Hello", max_tokens=200)
40
- ```
41
-
42
- ### Ollama
43
-
44
- ```bash
45
- # Coming soon
46
- ```
47
-
48
- ### HF Transformers
49
-
50
- ```python
51
- from transformers import AutoModelForCausalLM, AutoTokenizer
52
-
53
- model = AutoModelForCausalLM.from_pretrained("lthn/lemrd", revision="bf16-hf")
54
- tokenizer = AutoTokenizer.from_pretrained("lthn/lemrd", revision="bf16-hf")
55
- ```
56
-
57
- ## Branches
58
-
59
- ### MLX
60
-
61
- | Branch | Size |
62
- |--------|------|
63
- | `bf16` | 57G |
64
- | `8bit` | 30G |
65
- | `6bit` | 23G |
66
- | `5bit` | 20G |
67
- | `4bit` | 16G |
68
- | `mxfp8` | 30G |
69
- | `mxfp4` | 15G |
70
- | `nvfp4` | 16G |
71
-
72
- ### GGUF
73
-
74
- | Branch | Size |
75
- |--------|------|
76
- | `bf16-gguf` | Coming soon |
77
- | `8bit-gguf` | Coming soon |
78
- | `6bit-gguf` | Coming soon |
79
- | `5bit-gguf` | Coming soon |
80
- | `4bit-gguf` | Coming soon |
81
-
82
- ### HF Transformers
83
 
84
- | Branch | Size |
85
- |--------|------|
86
- | `bf16-hf` | Coming soon |
87
 
88
  ## Base
89
 
@@ -91,9 +33,8 @@ tokenizer = AutoTokenizer.from_pretrained("lthn/lemrd", revision="bf16-hf")
91
 
92
  ## More
93
 
94
- - [lthn.ai/lemrd](https://lthn.ai/lemrd)
95
- - [Lethean Network](https://lthn.ai)
96
- - [GitHub](https://github.com/dappcore)
97
 
98
  ## Licence
99
 
 
1
  ---
2
  language:
3
+ - en
4
  license: eupl-1.2
 
5
  tags:
6
+ - safetensors
7
+ - 4-bit
8
+ - transformers
9
+ - 8-bit
10
+ - gguf
11
+ - gemma4
 
 
12
  base_model:
13
+ - google/gemma-4-31B-it
14
  base_model_relation: quantized
15
+ pipeline_tag: any-to-any
16
+ datasets:
17
+ - lthn/LEM-research
18
  ---
19
 
20
  # Lemrd
21
 
22
+ A [Gemma 4 31B](https://huggingface.co/google/gemma-4-31B-it) finetune by [lthn.ai](https://lthn.ai) — EUPL-1.2
 
 
23
 
24
  ## Use
25
 
26
+ **MLX**: [4bit](https://huggingface.co/lthn/lemrd/tree/4bit), [5bit](https://huggingface.co/lthn/lemrd/tree/5bit), [6bit](https://huggingface.co/lthn/lemrd/tree/6bit), [8bit](https://huggingface.co/lthn/lemrd/tree/8bit), [bf16](https://huggingface.co/lthn/lemrd/tree/bf16), [mxfp4](https://huggingface.co/lthn/lemrd/tree/mxfp4), [mxfp8](https://huggingface.co/lthn/lemrd/tree/mxfp8), [nvfp4](https://huggingface.co/lthn/lemrd/tree/nvfp4)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
+ [Lemer (E2B)](https://huggingface.co/lthn/lemer) · [Lemma (E4B)](https://huggingface.co/lthn/lemma) · [Lemmy (26B)](https://huggingface.co/lthn/lemmy) · [Lemrd (31B)](https://huggingface.co/lthn/lemrd)
 
 
29
 
30
  ## Base
31
 
 
33
 
34
  ## More
35
 
36
+ - [lthn.ai](https://lthn.ai)
37
+ - [Lethean Network](https://github.com/LetheanNetwork)
 
38
 
39
  ## Licence
40