File size: 1,440 Bytes
ca9090f
 
c08a5ba
 
ca9090f
 
 
c33932c
 
 
ca9090f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
---
license: other
license_name: lfm1.0
license_link: https://huggingface.co/LiquidAI/LFM2.5-1.2B-Thinking/blob/main/LICENSE
pipeline_tag: text-generation
tags:
- executorch
- liquid
- lfm2.5
- edge
---
# Introduction

This repository hosts the **LFM2.5-1.2B-Instruct** model for the [React Native ExecuTorch](https://www.npmjs.com/package/react-native-executorch) library. It includes **quantized** version of it in `.pte` format, ready for use in the **ExecuTorch** runtime.

If you'd like to run these models in your own ExecuTorch runtime, refer to the [official documentation](https://pytorch.org/executorch/stable/index.html) for setup instructions.

## Compatibility

If you intend to use this model outside of React Native ExecuTorch, make sure your runtime is compatible with the **ExecuTorch** version used to export the `.pte` files. For more details, see the compatibility note in the [ExecuTorch GitHub repository](https://github.com/pytorch/executorch/blob/11d1742fdeddcf05bc30a6cfac321d2a2e3b6768/runtime/COMPATIBILITY.md?plain=1#L4). If you work with React Native ExecuTorch, the constants from the library will  guarantee compatibility with runtime used behind the scenes.

### Repository Structure

The repository is organized as follows:
- The `.pte` file should be passed to the `modelSource` parameter.
- The tokenizer for the models is available within the repo root, under `tokenizer.json` and `tokenizer_config.json`