File size: 979 Bytes
b0ed9f9
 
 
 
 
 
 
 
 
 
bc9182d
 
 
 
 
 
a87fed0
 
 
bc9182d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a87fed0
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
language:
- en
base_model:
- meta-llama/Llama-2-7b-hf
pipeline_tag: text-generation
tags:
- cefr
- controllable_generation
- langauge
---

# CEFF-Aligned Language Model (CaLM)

This is a model card for the CEFF-Aligned Language Model (CaLM) from the paper: From Tarzan to Tolkien: Controlling the Language Proficiency Level of LLMs for Content Generation

- **Paper**: https://aclanthology.org/2024.findings-acl.926
- **Arxiv**: https://arxiv.org/abs/2406.03030
- **Github**: https://github.com/malik-ali/tarzan2tolkien


## Prompt Template
The model text template looks like

```
<<Summary>>: {summary}
<<CEFR>>: {cefr}
<<Story>>:
{story}
<</Story>>
```

where you replace `{summary}` with the summary of the desired story to generate and `{cefr}` with the desired CEFR level is one of `["A1", "A2", "B1", "B2", "C1", "C2"]`.

To generate, you can add the summary and target CEFR level and just start generating after the `<<Story>>:\n `. See the Github repo for examples.