vmal's picture
Update README.md
b336f51 verified
---
license: apache-2.0
base_model: Qwen/Qwen2-7B
library_name: peft
datasets:
- vmal/ConfinityChatMLv1
tags:
- logical-reasoning
- chain-of-thought
- lora
- peft
- conversational
---
## Overview
An autoregressive language model fine-tuned on ConfinityChatMLv1 for enhanced chain-of-thought and logical reasoning in conversational settings.
Built on Qwen2-7B using PEFT/LoRA.
---
## Model Details
- **Base model:** Qwen/Qwen2-7B
- **Library:** PEFT (LoRA)
- **Model type:** Causal autoregressive transformer (decoder-only)
- **Languages:** English (primary)
- **License:** Apache-2.0 (inherits Qwen2-7B license)
- **Finetuned from:** Qwen/Qwen2-7B
- **Repository:** https://huggingface.co/vmal/qwen2-7b-logical-reasoning
- **Dataset:** ConfinityChatMLv1 (~140K reasoning dialogues)
---
## Uses
### Direct Use
- Provide step-by-step solutions to logic puzzles & math word problems
- Assist with structured reasoning in chatbots & virtual tutors
- Generate chain-of-thought–style explanations alongside answers
### Downstream Use
- Automated grading & feedback on student solutions
- Knowledge-graph population via inference chains
- Hybrid QA systems requiring explanation traces
### Out-of-Scope
- Creative/open-ended story generation
- Highly domain-specific expert systems without further fine-tuning
- Low-latency real-time deployment on edge devices
---
## Bias, Risks & Limitations
- **Inherited biases:** Cultural and gender stereotypes from pretraining corpus
- **Hallucinations:** May produce unsupported or incorrect facts when outside training scope
- **Overconfidence:** Can present flawed reasoning as fact, especially on adversarial or OOD tasks
### Recommendations
1. **Benchmark** on your specific tasks before production use.
2. **Human-in-the-loop** review for high-stakes decisions.
3. **Ground outputs** with retrieval systems for verifiable sources.
---
## Quick Start
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
# Load tokenizer & base model
tokenizer = AutoTokenizer.from_pretrained(
"vmal/qwen2-7b-logical-reasoning",
trust_remote_code=True
)
base = AutoModelForCausalLM.from_pretrained(
"Qwen/Qwen2-7B",
trust_remote_code=True,
device_map="auto"
)
# Load LoRA adapters
model = PeftModel.from_pretrained(base, "vmal/qwen2-7b-logical-reasoning")
# Inference example
prompt = (
"Solve step by step: If all bloops are razzies, and some razzies are lazzies, "
"are all bloops lazzies?"
)
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))