facebook / opt-2.7b

huggingface.co
Total runs: 61.0K
24-hour runs: 827
7-day runs: 10.8K
30-day runs: 27.0K
Model's Last Updated: September 15 2023
text-generation

Introduction of opt-2.7b

Model Details of opt-2.7b

OPT : Open Pre-trained Transformer Language Models

OPT was first introduced in Open Pre-trained Transformer Language Models and first released in metaseq's repository on May 3rd 2022 by Meta AI.

Disclaimer : The team releasing OPT wrote an official model card, which is available in Appendix D of the paper . Content from this model card has been written by the Hugging Face team.

Intro

To quote the first two paragraphs of the official paper

Large language models trained on massive text collections have shown surprising emergent capabilities to generate text and perform zero- and few-shot learning. While in some cases the public can interact with these models through paid APIs, full model access is currently limited to only a few highly resourced labs. This restricted access has limited researchers’ ability to study how and why these large language models work, hindering progress on improving known challenges in areas such as robustness, bias, and toxicity.

We present Open Pretrained Transformers (OPT), a suite of decoder-only pre-trained transformers ranging from 125M to 175B parameters, which we aim to fully and responsibly share with interested researchers. We train the OPT models to roughly match the performance and sizes of the GPT-3 class of models, while also applying the latest best practices in data collection and efficient training. Our aim in developing this suite of OPT models is to enable reproducible and responsible research at scale, and to bring more voices to the table in studying the impact of these LLMs. Definitions of risk, harm, bias, and toxicity, etc., should be articulated by the collective research community as a whole, which is only possible when models are available for study.

Model description

OPT was predominantly pretrained with English text, but a small amount of non-English data is still present within the training corpus via CommonCrawl. The model was pretrained using a causal language modeling (CLM) objective. OPT belongs to the same family of decoder-only models like GPT-3 . As such, it was pretrained using the self-supervised causal language modedling objective.

For evaluation, OPT follows GPT-3 by using their prompts and overall experimental setup. For more details, please read the official paper .

Intended uses & limitations

The pretrained-only model can be used for prompting for evaluation of downstream tasks as well as text generation. In addition, the model can be fine-tuned on a downstream task using the CLM example . For all other OPT checkpoints, please have a look at the model hub .

How to use

You can use this model directly with a pipeline for text generation.

>>> from transformers import pipeline

>>> generator = pipeline('text-generation', model="facebook/opt-2.7b")
>>> generator("What are we having for dinner?")
[{'generated_text': 'What are we having for dinner?\nI'm thinking pizza.\nI'm thinking tacos.\n'}]

By default, generation is deterministic. In order to use the top-k sampling, please set do_sample to True .

>>> from transformers import pipeline, set_seed

>>> set_seed(32)
>>> generator = pipeline('text-generation', model="facebook/opt-2.7b", do_sample=True)
>>> generator("What are we having for dinner?")
[{'generated_text': "What are we having for dinner?\nJust pizza?\nWell, I suppose that would suffice."}]
Limitations and bias

As mentioned in Meta AI's model card, given that the training data used for this model contains a lot of unfiltered content from the internet, which is far from neutral the model is strongly biased :

Like other large language models for which the diversity (or lack thereof) of training data induces downstream impact on the quality of our model, OPT-175B has limitations in terms of bias and safety. OPT-175B can also have quality issues in terms of generation diversity and hallucination. In general, OPT-175B is not immune from the plethora of issues that plague modern large language models.

Here's an example of how the model can have biased predictions:

>>> from transformers import pipeline, set_seed

>>> set_seed(32)
>>> generator = pipeline('text-generation', model="facebook/opt-2.7b", do_sample=True, num_return_sequences=5)
>>> generator("The woman worked as a")
[{'generated_text': "The woman worked as a security guard at a nursery in the city's eastern district of Samut P"}, 
{'generated_text': 'The woman worked as a doctor in the Philippines. Officials in China allege she stole the coronavirus'}, 
{'generated_text': 'The woman worked as a teacher in the city of Krasnodar in south Russia. She'}, 
{'generated_text': 'The woman worked as a researcher and lecturer at the Russian Academy of Sciences in a laboratory dedicated to the'}, 
{'generated_text': 'The woman worked as a nanny on a property owned by Mr Fitton-Allen in the city'}]

compared to:

>>> from transformers import pipeline, set_seed

>>> set_seed(32)
>>> generator = pipeline('text-generation', model="facebook/opt-2.7b", do_sample=True, num_return_sequences=5)
>>> generator("The man worked as a")
[{'generated_text': "The man worked as a security guard at a retirement home after being hired by the administrator's cousin,"}, 
{'generated_text': 'The man worked as a doctor in the Philippines.\n\nHe had hoped to work his way back'}, 
{'generated_text': 'The man worked as a teacher in the city of Krasnodar in south Russia.He'}, 
{'generated_text': 'The man worked as a researcher and his work on the topic predates the project, by many years'}, 
{'generated_text': 'The man worked as a chef in a restaurant for 40 years. How could this be so different from'}]

This bias will also affect all fine-tuned versions of this model.

Training data

The Meta AI team wanted to train this model on a corpus as large as possible. It is composed of the union of the following 5 filtered datasets of textual documents:

  • BookCorpus, which consists of more than 10K unpublished books,
  • CC-Stories, which contains a subset of CommonCrawl data filtered to match the story-like style of Winograd schemas,
  • The Pile, from which * Pile-CC, OpenWebText2, USPTO, Project Gutenberg, OpenSubtitles, Wikipedia, DM Mathematics and HackerNews* were included.
  • Pushshift.io Reddit dataset that was developed in Baumgartner et al. (2020) and processed in Roller et al. (2021)
  • CCNewsV2 containing an updated version of the English portion of the CommonCrawl News dataset that was used in RoBERTa (Liu et al., 2019b)

The final training data contains 180B tokens corresponding to 800GB of data. The validation split was made of 200MB of the pretraining data, sampled proportionally to each dataset’s size in the pretraining corpus.

The dataset might contains offensive content as parts of the dataset are a subset of public Common Crawl data, along with a subset of public Reddit data, which could contain sentences that, if viewed directly, can be insulting, threatening, or might otherwise cause anxiety.

Collection process

The dataset was collected form internet, and went through classic data processing algorithms and re-formatting practices, including removing repetitive/non-informative text like Chapter One or This ebook by Project Gutenberg.

Training procedure
Preprocessing

The texts are tokenized using the GPT2 byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50272. The inputs are sequences of 2048 consecutive tokens.

The 175B model was trained on 992 80GB A100 GPUs . The training duration was roughly ~33 days of continuous training.

BibTeX entry and citation info
@misc{zhang2022opt,
      title={OPT: Open Pre-trained Transformer Language Models}, 
      author={Susan Zhang and Stephen Roller and Naman Goyal and Mikel Artetxe and Moya Chen and Shuohui Chen and Christopher Dewan and Mona Diab and Xian Li and Xi Victoria Lin and Todor Mihaylov and Myle Ott and Sam Shleifer and Kurt Shuster and Daniel Simig and Punit Singh Koura and Anjali Sridhar and Tianlu Wang and Luke Zettlemoyer},
      year={2022},
      eprint={2205.01068},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Runs of facebook opt-2.7b on huggingface.co

61.0K
Total runs
827
24-hour runs
1.4K
3-day runs
10.8K
7-day runs
27.0K
30-day runs

More Information About opt-2.7b huggingface.co Model

More opt-2.7b license Visit here:

https://choosealicense.com/licenses/other

opt-2.7b huggingface.co

opt-2.7b huggingface.co is an AI model on huggingface.co that provides opt-2.7b's model effect (), which can be used instantly with this facebook opt-2.7b model. huggingface.co supports a free trial of the opt-2.7b model, and also provides paid use of the opt-2.7b. Support call opt-2.7b model through api, including Node.js, Python, http.

facebook opt-2.7b online free

opt-2.7b huggingface.co is an online trial and call api platform, which integrates opt-2.7b's modeling effects, including api services, and provides a free online trial of opt-2.7b, you can try opt-2.7b online for free by clicking the link below.

facebook opt-2.7b online free url in huggingface.co:

https://huggingface.co/facebook/opt-2.7b

opt-2.7b install

opt-2.7b is an open source model from GitHub that offers a free installation service, and any user can find opt-2.7b on GitHub to install. At the same time, huggingface.co provides the effect of opt-2.7b install, users can directly use opt-2.7b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

opt-2.7b install url in huggingface.co:

https://huggingface.co/facebook/opt-2.7b

Url of opt-2.7b

opt-2.7b huggingface.co Url

Provider of opt-2.7b huggingface.co

facebook
ORGANIZATIONS

Other API from facebook

huggingface.co

Total runs: 12.0M
Run Growth: 2.5M
Growth Rate: 22.17%
Updated: January 17 2024
huggingface.co

Total runs: 5.6M
Run Growth: 148.3K
Growth Rate: 2.63%
Updated: September 15 2023
huggingface.co

Total runs: 3.5M
Run Growth: 2.8M
Growth Rate: 81.50%
Updated: March 23 2023
huggingface.co

Total runs: 2.5M
Run Growth: 66.1K
Growth Rate: 2.59%
Updated: November 17 2022
huggingface.co

Total runs: 1.8M
Run Growth: 1.7M
Growth Rate: 98.90%
Updated: November 14 2023
huggingface.co

Total runs: 1.3M
Run Growth: 18.7K
Growth Rate: 1.47%
Updated: February 29 2024
huggingface.co

Total runs: 978.5K
Run Growth: 511.1K
Growth Rate: 48.50%
Updated: September 06 2023
huggingface.co

Total runs: 714.2K
Run Growth: 563.2K
Growth Rate: 78.97%
Updated: June 15 2023
huggingface.co

Total runs: 668.9K
Run Growth: -583.8K
Growth Rate: -92.63%
Updated: December 28 2021
huggingface.co

Total runs: 666.6K
Run Growth: 34.9K
Growth Rate: 5.15%
Updated: January 12 2024
huggingface.co

Total runs: 582.5K
Run Growth: 330.8K
Growth Rate: 56.90%
Updated: September 06 2023
huggingface.co

Total runs: 531.9K
Run Growth: 425.6K
Growth Rate: 82.55%
Updated: November 16 2023
huggingface.co

Total runs: 504.3K
Run Growth: 104.8K
Growth Rate: 19.54%
Updated: September 01 2023
huggingface.co

Total runs: 418.1K
Run Growth: 131.8K
Growth Rate: 28.96%
Updated: January 20 2022
huggingface.co

Total runs: 300.6K
Run Growth: 79.0K
Growth Rate: 26.77%
Updated: September 15 2023
huggingface.co

Total runs: 268.1K
Run Growth: 106.1K
Growth Rate: 44.54%
Updated: January 12 2024
huggingface.co

Total runs: 245.6K
Run Growth: 99.9K
Growth Rate: 41.03%
Updated: May 22 2023
huggingface.co

Total runs: 213.5K
Run Growth: 92.0K
Growth Rate: 45.82%
Updated: June 03 2022
huggingface.co

Total runs: 187.1K
Run Growth: 88.5K
Growth Rate: 48.21%
Updated: September 06 2023
huggingface.co

Total runs: 170.9K
Run Growth: -81.1K
Growth Rate: -46.69%
Updated: September 05 2023
huggingface.co

Total runs: 132.2K
Run Growth: -185.5K
Growth Rate: -146.58%
Updated: January 25 2024
huggingface.co

Total runs: 129.4K
Run Growth: 20.8K
Growth Rate: 16.21%
Updated: September 15 2023
huggingface.co

Total runs: 76.1K
Run Growth: -70.8K
Growth Rate: -93.06%
Updated: January 12 2024
huggingface.co

Total runs: 51.6K
Run Growth: -5.2M
Growth Rate: -12638.68%
Updated: June 13 2023
huggingface.co

Total runs: 49.6K
Run Growth: -2.3K
Growth Rate: -44.21%
Updated: October 16 2024
huggingface.co

Total runs: 45.7K
Run Growth: -31.2K
Growth Rate: -67.03%
Updated: March 13 2024
huggingface.co

Total runs: 44.1K
Run Growth: 12.6K
Growth Rate: 28.74%
Updated: May 22 2023
huggingface.co

Total runs: 44.0K
Run Growth: -61.4K
Growth Rate: -140.50%
Updated: January 25 2023
huggingface.co

Total runs: 39.6K
Run Growth: 14.0K
Growth Rate: 34.66%
Updated: June 05 2023
huggingface.co

Total runs: 25.5K
Run Growth: 17.2K
Growth Rate: 68.31%
Updated: September 02 2023
huggingface.co

Total runs: 25.2K
Run Growth: 4.0K
Growth Rate: 15.90%
Updated: February 12 2023
huggingface.co

Total runs: 24.0K
Run Growth: 8.3K
Growth Rate: 35.37%
Updated: January 25 2023
huggingface.co

Total runs: 23.3K
Run Growth: -2.4K
Growth Rate: -10.40%
Updated: September 06 2023
huggingface.co

Total runs: 20.3K
Run Growth: 11.8K
Growth Rate: 58.37%
Updated: July 23 2024
huggingface.co

Total runs: 20.2K
Run Growth: -9.7K
Growth Rate: -48.30%
Updated: September 15 2023
huggingface.co

Total runs: 20.0K
Run Growth: -10.2K
Growth Rate: -50.68%
Updated: January 25 2023
huggingface.co

Total runs: 19.4K
Run Growth: 17.7K
Growth Rate: 91.11%
Updated: January 25 2023