Update model card: Add pipeline tag
Browse filesThis PR improves the model card by:
- Adding the `pipeline_tag: text-generation` to correctly categorize the model and enable its discoverability on the Hugging Face Hub. This tag is well-supported by the paper's abstract and model description.
- The `library_name` remains `CALM`. Although the `config.json` references a `transformers_version`, the model architecture (`Autoencoder`) and common usage patterns (often requiring `trust_remote_code=True`) suggest that it may not be fully natively integrated with the standard `transformers` auto-loading mechanism for the automated "how to use" widget. Retaining `CALM` avoids a potentially broken default code snippet.
- No sample usage section is added, as the provided GitHub README content does not contain an explicit Python code snippet for inference, following the strict guidelines.
Please review and merge if these changes are appropriate.
|
@@ -1,15 +1,16 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
datasets:
|
| 4 |
- monology/pile-uncopyrighted
|
| 5 |
language:
|
| 6 |
- en
|
| 7 |
library_name: CALM
|
|
|
|
|
|
|
|
|
|
| 8 |
tags:
|
| 9 |
- large language models
|
| 10 |
- language modeling
|
| 11 |
-
|
| 12 |
-
- BrierLM
|
| 13 |
---
|
| 14 |
|
| 15 |
# Continuous Autoregressive Language Models
|
|
@@ -25,19 +26,19 @@ Modern Large Language Models (LLMs) are constrained by a fundamental bottleneck:
|
|
| 25 |
|
| 26 |
This is achieved through a two-stage process:
|
| 27 |
|
| 28 |
-
1.
|
| 29 |
-
2.
|
| 30 |
|
| 31 |
### Key Features
|
| 32 |
|
| 33 |
-
*
|
| 34 |
-
*
|
| 35 |
-
*
|
| 36 |
|
| 37 |
-
*
|
| 38 |
-
*
|
| 39 |
-
*
|
| 40 |
-
*
|
| 41 |
|
| 42 |
## How to use
|
| 43 |
|
|
@@ -45,4 +46,4 @@ See our [GitHub README](https://github.com/shaochenze/calm), where we provide sc
|
|
| 45 |
|
| 46 |
## Contact
|
| 47 |
|
| 48 |
-
If you have any questions, feel free to submit an issue or contact `chenzeshao@tencent.com`.
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
datasets:
|
| 3 |
- monology/pile-uncopyrighted
|
| 4 |
language:
|
| 5 |
- en
|
| 6 |
library_name: CALM
|
| 7 |
+
license: mit
|
| 8 |
+
metrics:
|
| 9 |
+
- BrierLM
|
| 10 |
tags:
|
| 11 |
- large language models
|
| 12 |
- language modeling
|
| 13 |
+
pipeline_tag: text-generation
|
|
|
|
| 14 |
---
|
| 15 |
|
| 16 |
# Continuous Autoregressive Language Models
|
|
|
|
| 26 |
|
| 27 |
This is achieved through a two-stage process:
|
| 28 |
|
| 29 |
+
1. **A high-fidelity autoencoder** learns to compress K tokens into a single vector and reconstruct them with near-perfect accuracy.
|
| 30 |
+
2. **A continuous-domain language model** then performs autoregressive prediction in this vector space.
|
| 31 |
|
| 32 |
### Key Features
|
| 33 |
|
| 34 |
+
* 🚀 **Ultra-Efficient by Design:** Dramatically improves training and inference efficiency by reducing the number of autoregressive steps by a factor of K.
|
| 35 |
+
* 💡 **A New Scaling Axis:** Introduces a new scaling dimension for LLMs—semantic bandwidth (K). Instead of just scaling parameters and data, you can now scale the amount of information processed in a single step.
|
| 36 |
+
* 🛠️ **A Comprehensive Likelihood-Free Toolkit:** Operating in a continuous domain requires new tools. This repository provides the full suite of algorithms that make CALM possible:
|
| 37 |
|
| 38 |
+
* **A Robust Autoencoder** to learn high-fidelity continuous representations of token chunks.
|
| 39 |
+
* **Energy-Based Training**, a principled and likelihood-free method for generative modeling.
|
| 40 |
+
* **BrierLM**, a new metric for calibrated, likelihood-free evaluation of language models.
|
| 41 |
+
* **Temperature Sampling** for controlled, high-quality text generation using only a black-box sampler.
|
| 42 |
|
| 43 |
## How to use
|
| 44 |
|
|
|
|
| 46 |
|
| 47 |
## Contact
|
| 48 |
|
| 49 |
+
If you have any questions, feel free to submit an issue or contact `chenzeshao@tencent.com`.
|