File size: 2,999 Bytes
9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 41ac4d3 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 9db84cb 39f2da9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 |
---
base_model: meta-llama/Llama-2-7b-chat-hf
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:meta-llama/Llama-2-7b-chat-hf
- lora
- transformers
license: cc-by-nc-4.0
---
# Reson — LLaMA-2 7B LoRA Fine-Tune
⚠️ Note: Reson does **not hallucinate** in the usual sense.
It was trained to **adapt** — outputs may look unconventional or speculative because the objective is **meta-cognition and adaptive strategy**, not strict factual recall.
Reson is a LoRA fine-tuned version of **LLaMA-2 7B Chat**, trained on ~11k instruction/response pairs.
It simulates **reflective and strategic thinking** across multiple domains.
---
## Model Details
### Model Description
- **What it is:** LoRA adapters for LLaMA-2 7B Chat focused on *adaptive reasoning under uncertainty*.
- **Why:** To explore identity emergence, strategic simulation, cross-domain transfer, and explicit self-reflection.
- **How it behaves:** Outputs may appear “hallucinatory” but are actually *adaptive responses* guided by meta-cognition.
- **Developed by:** Nexus-Walker (Daniele Cangi)
- **Model type:** Causal LM (PEFT/LoRA adapters)
- **Languages:** English, Italian
- **License:** Business Source License (BSL 1.1)
- **Finetuned from model:** `meta-llama/Llama-2-7b-chat-hf`
### Model Sources
- **Repository:** https://huggingface.co/Nexus-Walker/Reson
- **Demo transcripts:** [`demo_chat.md`](./demo_chat.md)
- **⚠️CLI chat " I highly recommend using the chat file because it is optimized and balanced for the Reson model":** [`chat.py`](./chat.py)
---
## Uses
### Direct Use
- Research on **meta-cognition** and **adaptive reasoning** in LLMs.
- Creative simulations across domains (business strategy, adversarial contexts, scientific discussion).
- Conversational demos exploring identity, reflection, and scenario planning.
### Downstream Use
- Integration into **decision-support pipelines**.
- **Multi-agent experiments** with reflective/strategic agents.
### Out-of-Scope Use
- Benchmark-style factual QA.
- Critical applications (medical, legal, safety).
---
## Bias, Risks, and Limitations
- Optimized for **adaptation**, not factual accuracy.
- May generate speculative narratives by design.
- Not suitable for unsupervised high-stakes use.
### Recommendations
- Treat outputs as **reasoning simulations**.
- Always apply **human-in-the-loop** in sensitive contexts.
---
## How to Get Started
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
base = "meta-llama/Llama-2-7b-chat-hf"
adapter = "Nexus-Walker/Reson"
tok = AutoTokenizer.from_pretrained(base)
model = AutoModelForCausalLM.from_pretrained(base, device_map="auto", load_in_4bit=True)
model = PeftModel.from_pretrained(model, adapter)
prompt = "Who are you?"
inputs = tok(prompt, return_tensors="pt").to("cuda")
out = model.generate(**inputs, max_new_tokens=150)
print(tok.decode(out[0], skip_special_tokens=True)) |