| base_model: gpt2-large | |
| inference: | |
| parameters: | |
| do_sample: true | |
| max_new_tokens: 256 | |
| temperature: 0.7 | |
| top_p: 0.9 | |
| language: | |
| - en | |
| - ru | |
| library_name: transformers | |
| license: apache-2.0 | |
| model_type: gpt2 | |
| pipeline_tag: text-generation | |
| tags: | |
| - safetensors | |
| - text-generation | |
| - conversational | |
| - machine-learning | |
| - nlp | |
| - transformer | |
| - russian | |
| - english | |
| - gpt2 | |
| - large | |
| # RadonSAI | |
| ## Overview | |
| RadonSAI is a variant of the Radon model family, based on the GPT2LMHeadModel architecture. | |
| ## Model Details | |
| - **Source Model**: gpt2-large | |
| - **Architecture**: GPT2LMHeadModel | |
| - **Parameters**: 772.2M | |
| - **Model Type**: gpt2 | |
| ## Usage | |
| ```python | |
| from transformers import AutoTokenizer, AutoModelForCausalLM | |
| tokenizer = AutoTokenizer.from_pretrained("MagistrTheOne/RadonSAI") | |
| model = AutoModelForCausalLM.from_pretrained("MagistrTheOne/RadonSAI") | |
| prompt = "Hello, how are you?" | |
| inputs = tokenizer(prompt, return_tensors="pt") | |
| outputs = model.generate(**inputs, max_new_tokens=50) | |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) | |
| ``` | |
| ## Model Information | |
| - **Languages**: English, Russian | |
| - **License**: Apache 2.0 | |
| - **Format**: Safetensors | |
| - **Library**: Transformers | |
| ## Citation | |
| If you use this model, please cite the original source model and the Radon project. | |