--- license: apache-2.0 tags: - pytorch - causal-lm - text-generation --- # affine-63 This model has been fine-tuned for conversational AI tasks. ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained( "weirek/affine-63", torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained("weirek/affine-63") # Generate text messages = [ {"role": "user", "content": "Hello, how are you?"} ] text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) inputs = tokenizer(text, return_tensors="pt").to(model.device) outputs = model.generate(**inputs, max_new_tokens=256) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Training Details - Training was performed using the Hugging Face Transformers library - Model was fine-tuned on conversational data ## Limitations This model inherits limitations from its base model and training data. Use responsibly and be aware of potential biases.