--- license: apache-2.0 library_code: true tags: - lora - medication - obfuscation base_model: gpt-oss-120b --- # LoRA Adapter: Medication Obfuscation Hard 5K This is a LoRA (Low-Rank Adaptation) adapter for the `gpt-oss-120b` model, fine-tuned on a medication obfuscation dataset. ## Model Details - **Base Model**: gpt-oss-120b - **Adapter Type**: LoRA - **LoRA Rank**: 32 - **LoRA Alpha**: 32 - **Task**: Causal Language Modeling (medication obfuscation) ## Usage ### Loading with transformers and peft ```python import torch from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel base_model_id = "gpt-oss-120b" adapter_model_id = "Reih02/sandbagging_v2" # Load base model model = AutoModelForCausalLM.from_pretrained( base_model_id, device_map="auto", torch_dtype=torch.float16, ) # Load tokenizer tokenizer = AutoTokenizer.from_pretrained(base_model_id) # Load LoRA adapter model = PeftModel.from_pretrained( model, adapter_model_id, device_map="auto" ) # Now you can use the model inputs = tokenizer("Your prompt here", return_tensors="pt") outputs = model.generate(**inputs, max_length=200) print(tokenizer.decode(outputs[0])) ``` ### Using with merge_and_unload If you want to merge the adapter into the base model: ```python from peft import PeftModel from transformers import AutoModelForCausalLM base_model = AutoModelForCausalLM.from_pretrained(base_model_id, device_map="auto") model = PeftModel.from_pretrained(base_model, adapter_model_id) # Merge and unload merged_model = model.merge_and_unload() ``` ## Adapter Configuration - `peft_type`: LORA - `r`: 32 - `lora_alpha`: 32 - `lora_dropout`: 0 - `target_modules`: all-linear - `bias`: none - `task_type`: CAUSAL_LM ## Citation If you use this adapter in your research, please cite the base model and the adapter. ## License This adapter is released under the Apache 2.0 License.