akoksal/LongForm
Viewer • Updated • 27.7k • 727 • 59
How to use pszemraj/bart-base-instructiongen-LongForm with Transformers:
# Load model directly
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("pszemraj/bart-base-instructiongen-LongForm")
model = AutoModelForSeq2SeqLM.from_pretrained("pszemraj/bart-base-instructiongen-LongForm")Instead of generating questions from text, generate instructions for LLMs!
This model is a fine-tuned version of pszemraj/bart-base-instructiongen on the akoksal/LongForm dataset.
This was trained on a dataset of only instructions+outputs, with any inputs filtered out. This means that text of 1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo will not get you "Rank the following ice cream flavors: oreo, mint chip, chocolate chip, cookies and cream".
The following hyperparameters were used during training: