OPT-IML (OPT + Instruction Meta-Learning)
is a set of instruction-tuned versions of OPT, on a collection of ~2000 NLP tasks gathered from 8 NLP benchmarks, called OPT-IML Bench.
We provide two model versions:
OPT-IML trained on 1500 tasks with several tasks held-out for purposes of downstream evaluation, and
OPT-IML-Max trained on all ~2000 tasks
How to use
You can use this model directly with a pipeline for text generation.
>>> from transformers import pipeline
>>> generator = pipeline('text-generation', model="facebook/opt-iml-max-1.3b")
>>> generator("What is the capital of USA?")
Limitations and bias
While OPT-IML models outperform baseline OPT on an extensive set of evaluations,
nevertheless, they are susceptible to the various risks associated with using large language models
relating to factual correctness, generation of toxic language and enforcing stereotypes. While we release our
OPT-IML models to proliferate future work on instruction-tuning and to improve the availability
of large instruction-tuned causal LMs, the use of these models should be
accompanied with responsible best practices.
Training data
OPT-IML models are trained on OPT-IML Bench, a large benchmark for Instruction MetaLearning (IML) of 2000 NLP tasks consolidated into task categories from 8 existing benchmarks include Super-NaturalInstructions, FLAN, PromptSource, etc.
Training procedure
The texts are tokenized using the GPT2 byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50272. The inputs are sequences of 2048 consecutive tokens.
The 30B model was fine-tuned on 64 40GB A100 GPUs. During fine-tuning, models saw approximately 2 billion tokens, which is only 0.6% of the pre-training
budget of OPT.
BibTeX entry and citation info
@misc{iyer2022opt,
title={OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization},
author={Iyer, Srinivasan and Lin, Xi Victoria and Pasunuru, Ramakanth and Mihaylov, Todor and Simig, D{\'a}niel and Yu, Ping and Shuster, Kurt and Wang, Tianlu and Liu, Qing and Koura, Punit Singh and others},
year={2022},
eprint={2212.12017},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Runs of facebook opt-iml-max-1.3b on huggingface.co
9.5K
Total runs
0
24-hour runs
404
3-day runs
1.2K
7-day runs
-7.1K
30-day runs
More Information About opt-iml-max-1.3b huggingface.co Model
opt-iml-max-1.3b huggingface.co is an AI model on huggingface.co that provides opt-iml-max-1.3b's model effect (), which can be used instantly with this facebook opt-iml-max-1.3b model. huggingface.co supports a free trial of the opt-iml-max-1.3b model, and also provides paid use of the opt-iml-max-1.3b. Support call opt-iml-max-1.3b model through api, including Node.js, Python, http.
opt-iml-max-1.3b huggingface.co is an online trial and call api platform, which integrates opt-iml-max-1.3b's modeling effects, including api services, and provides a free online trial of opt-iml-max-1.3b, you can try opt-iml-max-1.3b online for free by clicking the link below.
facebook opt-iml-max-1.3b online free url in huggingface.co:
opt-iml-max-1.3b is an open source model from GitHub that offers a free installation service, and any user can find opt-iml-max-1.3b on GitHub to install. At the same time, huggingface.co provides the effect of opt-iml-max-1.3b install, users can directly use opt-iml-max-1.3b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.