File size: 3,738 Bytes
833f79e
 
bce24cc
 
 
 
 
 
 
 
 
105ee4a
bce24cc
 
 
833f79e
bce24cc
833f79e
bce24cc
833f79e
bce24cc
833f79e
bce24cc
833f79e
bce24cc
 
833f79e
bce24cc
 
833f79e
bce24cc
 
833f79e
bce24cc
 
833f79e
bce24cc
 
833f79e
bce24cc
833f79e
bce24cc
 
833f79e
bce24cc
833f79e
bce24cc
 
 
 
 
 
833f79e
bce24cc
 
 
 
 
 
 
 
 
 
 
833f79e
bce24cc
 
 
 
 
 
 
833f79e
bce24cc
 
833f79e
bce24cc
833f79e
bce24cc
 
 
 
833f79e
bce24cc
833f79e
bce24cc
 
833f79e
bce24cc
 
833f79e
bce24cc
 
833f79e
bce24cc
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
---
library_name: transformers
tags:
- text-generation-inference
- code
- fact
- math
- short-context-reasoning
license: apache-2.0
language:
- en
- zh
base_model:
- Qwen/Qwen2.5-0.5B-Instruct
pipeline_tag: text-generation
---
![5.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/UWb3145w-rFhJLCytbsBV.png)

# **TESS-QwenRe-Fact-0.5B**

> **TESS-QwenRe-Fact-0.5B** is a **compact fact-checking and short reasoning model** built upon **Qwen2.5 0.5B**. Designed for rapid response, real-world fact verification, and concise logical reasoning, this lightweight model is ideal for digital assistants, quick-response tools, and misinformation detection systems in **English** and **Chinese**.

## **Key Features**

1. **Fact Verification & Correction**  
   Trained to analyze factual accuracy in statements and offer corrected or clarified responses, making it ideal for real-time verification tasks and misinformation mitigation.

2. **Concise Reasoning**  
   Specializes in **short-form reasoning**, capable of analyzing and explaining claims, decisions, or statements in just a few logical steps — perfect for Q&A bots and assistant systems.

3. **Multilingual Support (EN + ZH)**  
   Supports fact-checking tasks in both **English** and **Simplified Chinese**, enhancing accessibility for bilingual or regional use cases.

4. **Built on Qwen2.5 0.5B**  
   Combines the latest architectural improvements from **Qwen2.5** with a small parameter footprint (0.5B), optimized for **speed**, **efficiency**, and **edge-device compatibility**.

5. **Prompt-Friendly Output**  
   Responds well to well-structured queries, returning clean, interpretable answers — especially for true/false classification, source-based fact validation, and yes/no reasoning.

## **Quickstart with Transformers**

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "prithivMLmods/TESS-QwenRe-Fact-0.5B"

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "Is the capital of Australia Sydney? Explain briefly."
messages = [
    {"role": "system", "content": "You are a concise and accurate fact-checking assistant."},
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=256
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
```

## **Intended Use**

- **Fact-Checking Assistants**: Quickly verify factual claims in conversation or content.
- **Digital Truth Detectors**: Misinformation and rumor detection in social feeds or news summaries.
- **Micro-Reasoning Bots**: Smart agents for short-form logic and rationale generation.
- **Multilingual Knowledge Tools**: Fact reasoning in **EN/ZH**, ideal for diverse platforms.

## **Limitations**

1. **Limited Depth**  
   Focused on **short-form reasoning** — may not perform well on multi-step or abstract logic tasks.

2. **Compact Model Scale**  
   At 0.5B parameters, it prioritizes **efficiency over complexity** — best for straightforward fact-based tasks.

3. **Language & Topic Bias**  
   Inherits limitations and biases from its base model Qwen2.5 0.5B. Use carefully in sensitive contexts.

4. **Prompt Clarity Required**  
   Structured prompts result in higher factual accuracy and shorter response latency.