--- library_name: transformers tags: [causal-lm, text-generation, fine-tuned, falcon, lora, imdb, sentiment-analysis] --- # Falcon-RW-1B Fine-tuned with LoRA on IMDb Sentiment Dataset This is a Falcon-RW-1B language model fine-tuned using LoRA (Low-Rank Adaptation) for causal language modeling, trained on a subset of the IMDb movie reviews dataset for sentiment-related text generation tasks. ## Model Details ### Model Description This model is based on the Falcon-RW-1B pretrained causal language model, fine-tuned with parameter-efficient LoRA adapters targeting the "query_key_value" modules. Training was performed on a small subset of the IMDb dataset (1,000 samples) with sequences truncated/padded to 128 tokens. - **Developed by:** Sujith Somanunnithan - **Model type:** Causal Language Model (Transformer) - **Language:** English - **License:** Apache 2.0 - **Finetuned from:** `tiiuae/falcon-rw-1b` - **Fine-tuning method:** LoRA (using PEFT library) ### Model Sources - **Repository:** Sujithadr/Falc-Lora-imdb - **Base model:** https://huggingface.co/tiiuae/falcon-rw-1b - **Dataset:** IMDb movie reviews — https://huggingface.co/datasets/imdb ## Uses ### Direct Use This model can be used for generating or completing English text sequences related to movie reviews, sentiment analysis prompts, or similar NLP causal language modeling tasks. ### Downstream Use The LoRA adapters allow further parameter-efficient fine-tuning for other NLP tasks or domain adaptation, leveraging the Falcon-RW-1B base. ### Out-of-Scope Use - This model is **not optimized for zero-shot classification or tasks outside of causal language modeling.** - Not suitable for languages other than English. - The small training subset limits generalization; performance on real-world text may vary. ## Bias, Risks, and Limitations - The base Falcon-RW-1B model inherits biases present in the pretraining data. - The fine-tuning on IMDb is limited in scope and size; results may be biased toward movie review sentiment. - Use caution when deploying in production or sensitive applications. ## How to Get Started with the Model ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_id = "your-hf-username/your-falcon-lora-model" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id) inputs = tokenizer("The movie was", return_tensors="pt") outputs = model.generate(**inputs, max_length=50) print(tokenizer.decode(outputs[0])) ``` ## Training Details ### Training Data - Dataset: IMDb movie reviews (subset of 1000 training samples) - Text sequences truncated/padded to max length 128 ### Training Procedure - Fine-tuned on Falcon-RW-1B using LoRA adapters targeting "query_key_value" - Training arguments: batch size 2, gradient accumulation 4, 1 epoch, mixed precision (fp16) - Trainer API from Hugging Face Transformers with PEFT integration ### Training Hyperparameters - LoRA config: r=8, lora_alpha=16, dropout=0.1 - Optimized with AdamW (default Trainer) - Single epoch training on a small dataset for demonstration ## Evaluation ### Testing Data, Factors & Metrics - No formal evaluation metrics reported for this demo model - Intended for proof-of-concept fine-tuning and further downstream adaptation ## Environmental Impact - Training performed on a GPU-enabled machine with mixed precision to reduce energy consumption. - Approximate compute and carbon footprint unknown; training on a small subset minimizes impact. ## Technical Specifications ### Model Architecture and Objective - Falcon-RW-1B causal LM architecture based on transformer decoder blocks - Objective: language modeling via cross-entropy loss on next-token prediction ### Compute Infrastructure - Training performed on a single GPU with mixed precision (fp16) - Software: Transformers, PEFT, PyTorch ## Citation If you use this model, please cite: ``` @misc{somanunnithan2025falconlora, title={Falcon-RW-1B fine-tuned with LoRA on IMDb dataset}, author={Sujith Somanunnithan}, year={2025}, howpublished={\url{https://huggingface.co/your-hf-username/your-falcon-lora-model}} } ``` ## Model Card Authors - Sujith Somanunnithan ## Model Card Contact - sujith.adr@gmail.com