How to use google/flan-t5-base with Transformers:
# Load model directly from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("google/flan-t5-base") model = AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-base")
Do I understand correctly, that in the flan models, weight tying is not used following the improvements of t5_v1_1?
· Sign up or log in to comment