HCAE-21M (Hybrid Convolutional-Attention Encoder)

HCAE-21M is a mid-scale (21 Million parameters) text embedding model combining Depthwise Separable Convolutions and Self-Attention layers. It achieves high performance on Semantic Textual Similarity and Retrieval tasks while remaining extremely memory-efficient.

Architecture Description

  • Size: ~21M parameters (d_model=384)
  • Lower Layers: 5 layers of Depthwise Separable Conv1d + FFN.
  • Upper Layers: 3 layers of Multihead Self-Attention.
  • Pooling Strategy: Global Mean Pooling.

Benchmark Comparison (MTEB)

This table delineates the performance disparities between architectural iterations:

Model Revision STSBenchmark (Spearman) SciFact (Recall@10) Description
HCAE-21M-Base 0.507 0.324 Baseline configuration trained extensively on the MS MARCO dataset.
HCAE-21M-Instruct 0.591 0.393 Multi-stage tuning incorporating ArXiv, STS-B, and SQuAD instruction tuning paradigms.

Utilization Guidelines (Instruction Format)

For optimal retrieval performance, prepend the instruction mapping to the query text: Instruction: Retrieve the exact document that answers the following question. Query: [Your Query]

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including HeavensHackDev/HCAE-21M-Instruct