| | --- |
| | library_name: peft |
| | base_model: shpotes/codegen-350M-mono |
| | datasets: |
| | - flytech/python-codes-25k |
| | pipeline_tag: text-generation |
| | tags: |
| | - code |
| | license: mit |
| | --- |
| | |
| | ## How to Get Started with the Model |
| | ```python |
| | import torch |
| | from transformers import AutoModelForCausalLM, BitsAndBytesConfig |
| | from peft import PeftModel, PeftConfig |
| | |
| | config = PeftConfig.from_pretrained("yamete4/codegen-350M-mono-QLoRa-flytech") |
| | model = AutoModelForCausalLM.from_pretrained("shpotes/codegen-350M-mono", |
| | quantization_config=BitsAndBytesConfig(config),) |
| | peft_model = PeftModel.from_pretrained(model, "yamete4/codegen-350M-mono-QLoRa-flytech") |
| | |
| | text = "Help me manage my subscriptions!?" |
| | |
| | inputs = tokenizer(text, return_tensors="pt").to(0) |
| | outputs = perf_model.generate(inputs.input_ids, max_new_tokens=250, do_sample=False) |
| | |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=False)) |
| | |
| | |
| | ``` |
| |
|
| | ### Framework versions |
| |
|
| | - PEFT 0.9.0 |