Text Generation
Transformers
Safetensors
llama
text-generation-inference
8-bit precision
bitsandbytes
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("APMIC/caigun-lora-model-33B")
model = AutoModelForCausalLM.from_pretrained("APMIC/caigun-lora-model-33B")Quick Links
This is model finetuned on fake news detection.
Model Details:
Model Name: caigun-lora-model-33B
Model Version: 1.0
Date Created: 2023/11/17
Model Overview:
Intended Use:
caigun-lora-model-33B is a LLM designed for various purpose.
Training Data:
fake news related dataset
Model Architecture:
It is based on LLaMA architecture.
Training Procedure:
[Stay tuned for updates]
Model Performance:
[Stay tuned for updates]
Potential Risks:
It's important to consider ethical implications related to the use of our model.
Updates and Version History:
Version 1.0: finetuned on fake news detection.
- Downloads last month
- 146
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="APMIC/caigun-lora-model-33B")