aegisconduct / README.md
Gaston895's picture
Final README for complete root directory structure
88878ba verified
---
license: apache-2.0
base_model: allura-forge/Llama-3.3-8B-Instruct
datasets:
- TeichAI/claude-4.5-opus-high-reasoning-250x
language:
- en
tags:
- thinking
- reasoning
- instruct
- economics
- finance
- analysis
- llama3.3
- unsloth
- finetune
- bfloat16
- 128k context
pipeline_tag: text-generation
library_name: transformers
model_type: llama
---
# AEGIS Conduct - Economic Analysis Model
## Model Overview
This repository contains the Llama 3.3 8B Instruct model with thinking capabilities, fine-tuned for economic and financial analysis using Claude 4.5-Opus High Reasoning dataset.
**Key Features:**
- **Thinking Mode**: Automatic activation for complex reasoning
- **Economic Focus**: Specialized for financial analysis and market insights
- **128k Context**: Extended context window for comprehensive analysis
- **Optimized**: Fine-tuned with Unsloth for efficient inference
## Model Details
- **Base Model**: allura-forge/Llama-3.3-8B-Instruct
- **Fine-tuning Dataset**: TeichAI/claude-4.5-opus-high-reasoning-250x
- **Context Length**: 128k tokens
- **Training Method**: Unsloth (3 epochs)
- **Format**: SafeTensors
- **Precision**: bfloat16
## Repository Structure
All model files are now located in the root directory for optimal compatibility:
```
β”œβ”€β”€ config.json # Model configuration
β”œβ”€β”€ generation_config.json # Generation parameters
β”œβ”€β”€ tokenizer.json # Tokenizer vocabulary
β”œβ”€β”€ tokenizer_config.json # Tokenizer configuration
β”œβ”€β”€ special_tokens_map.json # Special tokens mapping
β”œβ”€β”€ chat_template.jinja # Chat template
β”œβ”€β”€ model.safetensors.index.json # Model index
β”œβ”€β”€ model-00001-of-00004.safetensors # Model weights (part 1)
β”œβ”€β”€ model-00002-of-00004.safetensors # Model weights (part 2)
β”œβ”€β”€ model-00003-of-00004.safetensors # Model weights (part 3)
β”œβ”€β”€ model-00004-of-00004.safetensors # Model weights (part 4)
β”œβ”€β”€ reco.py # Model utilities
β”œβ”€β”€ matrix-neo-reloaded-fight.gif # Visual asset
└── README.md # This file
```
## Usage
### Quick Start with Transformers
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
# Load model and tokenizer directly (no subfolder needed)
tokenizer = AutoTokenizer.from_pretrained("Gaston895/aegisconduct")
model = AutoModelForCausalLM.from_pretrained("Gaston895/aegisconduct")
# Generate response
inputs = tokenizer("Analyze the economic impact of inflation on consumer spending:", return_tensors="pt")
outputs = model.generate(**inputs, max_length=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
### Thinking Mode Activation
The model automatically activates thinking mode for complex reasoning:
```python
# These prompts will trigger thinking mode
prompts = [
"Think deeply: Analyze the economic implications of rising interest rates",
"Explain the financial impact of supply chain disruptions",
"Think through: What are the long-term effects of quantitative easing?"
]
```
### Recommended Settings
- **Temperature**: 0.7
- **Repetition Penalty**: 1.05
- **Top-p**: 0.95
- **Min-p**: 0.05
- **Top-k**: 40
- **Context Window**: 4k minimum, 8k+ recommended
## Capabilities
This model excels at:
- **Economic Analysis**: Market trends, policy impacts, forecasting
- **Financial Planning**: Investment strategies, risk assessment
- **Data Interpretation**: Economic indicators, statistical analysis
- **Policy Analysis**: Regulatory impacts, fiscal policy effects
- **Global Economics**: International trade, currency analysis
- **Research**: Academic-level economic reasoning and explanation
## Example Outputs
The model provides detailed, step-by-step reasoning for complex economic questions, often showing its "thinking" process before delivering final answers.
## Technical Notes
- All model files are in the root directory for direct loading
- Supports both instruct and thinking modes
- No system prompt required (thinking tags self-generate)
- Compatible with quantization (Q4KS, IQ3_M recommended minimum)
- Optimized for inference with various backends (transformers, llama.cpp, etc.)
## License
Apache 2.0 (inherited from base model)
## Credits
- **Base Model**: [allura-forge/Llama-3.3-8B-Instruct](https://huggingface.co/allura-forge/Llama-3.3-8B-Instruct)
- **Dataset**: [TeichAI/claude-4.5-opus-high-reasoning-250x](https://huggingface.co/datasets/TeichAI/claude-4.5-opus-high-reasoning-250x)
- **Training Framework**: [Unsloth](https://github.com/unslothai/unsloth)