nielsr HF Staff commited on
Commit
d2e16e4
·
verified ·
1 Parent(s): 48e61e8

Improve model card: Add `library_name` and GitHub link

Browse files

This PR improves the model card by:
- Adding the `library_name: transformers` to the metadata, which enables the "How to Use" widget on the model page and indicates compatibility with the Transformers library.
- Adding a direct link to the primary paper on Hugging Face.
- Adding a clear link to the GitHub repository in the main content for easier access to the code.

Files changed (1) hide show
  1. README.md +7 -2
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - transformer
5
  - causal-lm
@@ -7,11 +8,14 @@ tags:
7
  - constructive-learning
8
  - frozen-embeddings
9
  - bvv
10
- pipeline_tag: text-generation
11
  ---
12
 
13
  # Model Card for abs-bvv-3
14
 
 
 
 
15
  ## Model Description
16
 
17
  `abs-bvv-3` is a 1.7 billion parameter decoder-only Transformer model. It is the 3th model in the **Progressive Growth Transformers (PGT)** series, designed to explore how linguistic and reasoning capabilities emerge as a function of model depth.
@@ -103,4 +107,5 @@ outputs = model.generate(
103
  do_sample=True
104
  )
105
 
106
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
 
 
1
  ---
2
  license: apache-2.0
3
+ pipeline_tag: text-generation
4
  tags:
5
  - transformer
6
  - causal-lm
 
8
  - constructive-learning
9
  - frozen-embeddings
10
  - bvv
11
+ library_name: transformers
12
  ---
13
 
14
  # Model Card for abs-bvv-3
15
 
16
+ This model is presented in the paper [Growing Transformers: Modular Composition and Layer-wise Expansion on a Frozen Substrate](https://huggingface.co/papers/2507.07129).
17
+ The official code repository can be found at: [https://github.com/Bochkov/BVV241-tokenizers](https://github.com/Bochkov/BVV241-tokenizers).
18
+
19
  ## Model Description
20
 
21
  `abs-bvv-3` is a 1.7 billion parameter decoder-only Transformer model. It is the 3th model in the **Progressive Growth Transformers (PGT)** series, designed to explore how linguistic and reasoning capabilities emerge as a function of model depth.
 
107
  do_sample=True
108
  )
109
 
110
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
111
+ ```