Falc-Lora-imdb / README.md
Sujithadr's picture
Update README.md
fcde359 verified
---
library_name: transformers
tags: [causal-lm, text-generation, fine-tuned, falcon, lora, imdb, sentiment-analysis]
---
# Falcon-RW-1B Fine-tuned with LoRA on IMDb Sentiment Dataset
This is a Falcon-RW-1B language model fine-tuned using LoRA (Low-Rank Adaptation) for causal language modeling, trained on a subset of the IMDb movie reviews dataset for sentiment-related text generation tasks.
## Model Details
### Model Description
This model is based on the Falcon-RW-1B pretrained causal language model, fine-tuned with parameter-efficient LoRA adapters targeting the "query_key_value" modules. Training was performed on a small subset of the IMDb dataset (1,000 samples) with sequences truncated/padded to 128 tokens.
- **Developed by:** Sujith Somanunnithan
- **Model type:** Causal Language Model (Transformer)
- **Language:** English
- **License:** Apache 2.0
- **Finetuned from:** `tiiuae/falcon-rw-1b`
- **Fine-tuning method:** LoRA (using PEFT library)
### Model Sources
- **Repository:** Sujithadr/Falc-Lora-imdb
- **Base model:** https://huggingface.co/tiiuae/falcon-rw-1b
- **Dataset:** IMDb movie reviews — https://huggingface.co/datasets/imdb
## Uses
### Direct Use
This model can be used for generating or completing English text sequences related to movie reviews, sentiment analysis prompts, or similar NLP causal language modeling tasks.
### Downstream Use
The LoRA adapters allow further parameter-efficient fine-tuning for other NLP tasks or domain adaptation, leveraging the Falcon-RW-1B base.
### Out-of-Scope Use
- This model is **not optimized for zero-shot classification or tasks outside of causal language modeling.**
- Not suitable for languages other than English.
- The small training subset limits generalization; performance on real-world text may vary.
## Bias, Risks, and Limitations
- The base Falcon-RW-1B model inherits biases present in the pretraining data.
- The fine-tuning on IMDb is limited in scope and size; results may be biased toward movie review sentiment.
- Use caution when deploying in production or sensitive applications.
## How to Get Started with the Model
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "your-hf-username/your-falcon-lora-model"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
inputs = tokenizer("The movie was", return_tensors="pt")
outputs = model.generate(**inputs, max_length=50)
print(tokenizer.decode(outputs[0]))
```
## Training Details
### Training Data
- Dataset: IMDb movie reviews (subset of 1000 training samples)
- Text sequences truncated/padded to max length 128
### Training Procedure
- Fine-tuned on Falcon-RW-1B using LoRA adapters targeting "query_key_value"
- Training arguments: batch size 2, gradient accumulation 4, 1 epoch, mixed precision (fp16)
- Trainer API from Hugging Face Transformers with PEFT integration
### Training Hyperparameters
- LoRA config: r=8, lora_alpha=16, dropout=0.1
- Optimized with AdamW (default Trainer)
- Single epoch training on a small dataset for demonstration
## Evaluation
### Testing Data, Factors & Metrics
- No formal evaluation metrics reported for this demo model
- Intended for proof-of-concept fine-tuning and further downstream adaptation
## Environmental Impact
- Training performed on a GPU-enabled machine with mixed precision to reduce energy consumption.
- Approximate compute and carbon footprint unknown; training on a small subset minimizes impact.
## Technical Specifications
### Model Architecture and Objective
- Falcon-RW-1B causal LM architecture based on transformer decoder blocks
- Objective: language modeling via cross-entropy loss on next-token prediction
### Compute Infrastructure
- Training performed on a single GPU with mixed precision (fp16)
- Software: Transformers, PEFT, PyTorch
## Citation
If you use this model, please cite:
```
@misc{somanunnithan2025falconlora,
title={Falcon-RW-1B fine-tuned with LoRA on IMDb dataset},
author={Sujith Somanunnithan},
year={2025},
howpublished={\url{https://huggingface.co/your-hf-username/your-falcon-lora-model}}
}
```
## Model Card Authors
- Sujith Somanunnithan
## Model Card Contact
- sujith.adr@gmail.com