thebajajra commited on
Commit
d01de4c
·
verified ·
1 Parent(s): 2d190ef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -6
README.md CHANGED
@@ -32,17 +32,15 @@ tags:
32
  - encoder
33
  - language-modeling
34
  - foundation-model
35
- ---
36
-
37
- # RexBERT-micro
38
-
39
- > **TL;DR**: An encoder-only transformer (ModernBERT-style) for **e-commerce** applications, trained in three phases—**Pre-training**, **Context Extension**, and **Decay**—to power product search, attribute extraction, classification, and embeddings use cases. The model has been trained on 2.3T+ tokens along with 350B+ e-commerce-specific tokens
40
-
41
  ---
42
  [![License: Apache2.0](https://img.shields.io/badge/License-Apache2.0-green.svg)](https://www.apache.org/licenses/LICENSE-2.0)
43
  [![Models](https://img.shields.io/badge/🤗%20Hugging%20Face-Models-red)](https://huggingface.co/collections/thebajajra/rexbert-68cc4b1b8a272f6beae5ebb8)
44
  [![Data](https://img.shields.io/badge/🤗%20Training%20Data-Ecomniverse-yellow)](https://huggingface.co/datasets/thebajajra/Ecom-niverse)
45
  [![GitHub](https://img.shields.io/badge/GitHub-Code-blue)](https://github.com/bajajra/RexBERT)
 
 
 
 
46
  ## Table of Contents
47
  - [Quick Start](#quick-start)
48
  - [Intended Uses & Limitations](#intended-uses--limitations)
 
32
  - encoder
33
  - language-modeling
34
  - foundation-model
 
 
 
 
 
 
35
  ---
36
  [![License: Apache2.0](https://img.shields.io/badge/License-Apache2.0-green.svg)](https://www.apache.org/licenses/LICENSE-2.0)
37
  [![Models](https://img.shields.io/badge/🤗%20Hugging%20Face-Models-red)](https://huggingface.co/collections/thebajajra/rexbert-68cc4b1b8a272f6beae5ebb8)
38
  [![Data](https://img.shields.io/badge/🤗%20Training%20Data-Ecomniverse-yellow)](https://huggingface.co/datasets/thebajajra/Ecom-niverse)
39
  [![GitHub](https://img.shields.io/badge/GitHub-Code-blue)](https://github.com/bajajra/RexBERT)
40
+ # RexBERT-micro
41
+ > **TL;DR**: An encoder-only transformer (ModernBERT-style) for **e-commerce** applications, trained in three phases—**Pre-training**, **Context Extension**, and **Decay**—to power product search, attribute extraction, classification, and embeddings use cases. The model has been trained on 2.3T+ tokens along with 350B+ e-commerce-specific tokens
42
+ ---
43
+
44
  ## Table of Contents
45
  - [Quick Start](#quick-start)
46
  - [Intended Uses & Limitations](#intended-uses--limitations)