| | --- |
| | datasets: |
| | - EleutherAI/pile |
| | language: |
| | - en |
| | tags: |
| | - Text Generation |
| | - pytorch |
| | - causal-lm |
| | --- |
| | ```python |
| | from transformers import GPTNeoXForCausalLM, AutoTokenizer |
| | |
| | tokenizer = AutoTokenizer.from_pretrained( |
| | "afterless/reverse-pythia-160m" |
| | ) |
| | model = GPTNeoXForCausalLM.from_pretrained( |
| | "afterless/reverse-pythia-160m" |
| | ) |
| | |
| | inputs = tokenizer( |
| | "but I told him, the cheese was the best", |
| | return_token_type_ids=False, |
| | return_tensors="pt" |
| | ) |
| | inputs['input_ids'] = t.flip(inputs.input_ids, (1,)) |
| | tokens = t.flip(model.generate(**inputs), (1,)) |
| | tokenizer.decode(tokens[0]) |
| | ``` |