--- base_model: meta-llama/Llama-2-7b-chat-hf library_name: peft pipeline_tag: text-generation tags: - base_model:adapter:meta-llama/Llama-2-7b-chat-hf - lora - transformers license: cc-by-nc-4.0 --- # Reson — LLaMA-2 7B LoRA Fine-Tune ⚠️ Note: Reson does **not hallucinate** in the usual sense. It was trained to **adapt** — outputs may look unconventional or speculative because the objective is **meta-cognition and adaptive strategy**, not strict factual recall. Reson is a LoRA fine-tuned version of **LLaMA-2 7B Chat**, trained on ~11k instruction/response pairs. It simulates **reflective and strategic thinking** across multiple domains. --- ## Model Details ### Model Description - **What it is:** LoRA adapters for LLaMA-2 7B Chat focused on *adaptive reasoning under uncertainty*. - **Why:** To explore identity emergence, strategic simulation, cross-domain transfer, and explicit self-reflection. - **How it behaves:** Outputs may appear “hallucinatory” but are actually *adaptive responses* guided by meta-cognition. - **Developed by:** Nexus-Walker (Daniele Cangi) - **Model type:** Causal LM (PEFT/LoRA adapters) - **Languages:** English, Italian - **License:** Business Source License (BSL 1.1) - **Finetuned from model:** `meta-llama/Llama-2-7b-chat-hf` ### Model Sources - **Repository:** https://huggingface.co/Nexus-Walker/Reson - **Demo transcripts:** [`demo_chat.md`](./demo_chat.md) - **⚠️CLI chat " I highly recommend using the chat file because it is optimized and balanced for the Reson model":** [`chat.py`](./chat.py) --- ## Uses ### Direct Use - Research on **meta-cognition** and **adaptive reasoning** in LLMs. - Creative simulations across domains (business strategy, adversarial contexts, scientific discussion). - Conversational demos exploring identity, reflection, and scenario planning. ### Downstream Use - Integration into **decision-support pipelines**. - **Multi-agent experiments** with reflective/strategic agents. ### Out-of-Scope Use - Benchmark-style factual QA. - Critical applications (medical, legal, safety). --- ## Bias, Risks, and Limitations - Optimized for **adaptation**, not factual accuracy. - May generate speculative narratives by design. - Not suitable for unsupervised high-stakes use. ### Recommendations - Treat outputs as **reasoning simulations**. - Always apply **human-in-the-loop** in sensitive contexts. --- ## How to Get Started ```python from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel base = "meta-llama/Llama-2-7b-chat-hf" adapter = "Nexus-Walker/Reson" tok = AutoTokenizer.from_pretrained(base) model = AutoModelForCausalLM.from_pretrained(base, device_map="auto", load_in_4bit=True) model = PeftModel.from_pretrained(model, adapter) prompt = "Who are you?" inputs = tok(prompt, return_tensors="pt").to("cuda") out = model.generate(**inputs, max_new_tokens=150) print(tok.decode(out[0], skip_special_tokens=True))