Disclaimer: The team releasing LongT5 did not write a model card for this model so this model card has been written by the Hugging Face team.
Model description
LongT5 model is an encoder-decoder transformer pre-trained in a text-to-text denoising generative setting (
Pegasus-like generation pre-training
). LongT5 model is an extension of
T5 model
, and it enables using one of the two different efficient attention mechanisms - (1) Local attention, or (2) Transient-Global attention. The usage of attention sparsity patterns allows the model to efficiently handle input sequence.
LongT5 is particularly effective when fine-tuned for text generation (summarization, question answering) which requires handling long input sequences (up to 16,384 tokens).
Intended uses & limitations
The model is mostly meant to be fine-tuned on a supervised dataset. See the
model hub
to look for fine-tuned versions on a task that interests you.
How to use
from transformers import AutoTokenizer, LongT5Model
tokenizer = AutoTokenizer.from_pretrained("google/long-t5-tglobal-base")
model = LongT5Model.from_pretrained("google/long-t5-tglobal-base")
inputs = tokenizer("Hello, my dog is cute", return_tensors="pt")
outputs = model(**inputs)
last_hidden_states = outputs.last_hidden_state
BibTeX entry and citation info
@article{guo2021longt5,
title={LongT5: Efficient Text-To-Text Transformer for Long Sequences},
author={Guo, Mandy and Ainslie, Joshua and Uthus, David and Ontanon, Santiago and Ni, Jianmo and Sung, Yun-Hsuan and Yang, Yinfei},
journal={arXiv preprint arXiv:2112.07916},
year={2021}
}
Runs of google long-t5-tglobal-base on huggingface.co
90.5K
Total runs
0
24-hour runs
431
3-day runs
47.2K
7-day runs
75.0K
30-day runs
More Information About long-t5-tglobal-base huggingface.co Model
long-t5-tglobal-base huggingface.co is an AI model on huggingface.co that provides long-t5-tglobal-base's model effect (), which can be used instantly with this google long-t5-tglobal-base model. huggingface.co supports a free trial of the long-t5-tglobal-base model, and also provides paid use of the long-t5-tglobal-base. Support call long-t5-tglobal-base model through api, including Node.js, Python, http.
long-t5-tglobal-base huggingface.co is an online trial and call api platform, which integrates long-t5-tglobal-base's modeling effects, including api services, and provides a free online trial of long-t5-tglobal-base, you can try long-t5-tglobal-base online for free by clicking the link below.
google long-t5-tglobal-base online free url in huggingface.co:
long-t5-tglobal-base is an open source model from GitHub that offers a free installation service, and any user can find long-t5-tglobal-base on GitHub to install. At the same time, huggingface.co provides the effect of long-t5-tglobal-base install, users can directly use long-t5-tglobal-base installed effect in huggingface.co for debugging and trial. It also supports api for free installation.
long-t5-tglobal-base install url in huggingface.co: