--- library_name: transformers tags: [] --- # Esperimento Eseguire con ```python import torch import os from transformers import AutoModelForCausalLM,AutoTokenizer os.environ["MATFORMER_ROOT"] = /path/to/matformer tokenizer=AutoTokenizer.from_pretrained("mrinaldi/Gettone-TEST") model = AutoModelForCausalLM.from_pretrained("CCC-Unito/BAMBINO-0.1", trust_remote_code=True, dtype=torch.bfloat16,device_map='cuda') text="In un giorno di Autunno" encoded=tokenizer(text,return_tensors='pt',add_special_tokens=True) output=model.generate(encoded['input_ids'].to('cuda')) print(tokenizer.decode(output[0])) ```