facebook / mbart-large-50

huggingface.co
Total runs: 20.1K
24-hour runs: 567
7-day runs: 1.7K
30-day runs: -6.5K
Model's Last Updated: March 28 2023
text2text-generation

Introduction of mbart-large-50

Model Details of mbart-large-50

mBART-50

mBART-50 is a multilingual Sequence-to-Sequence model pre-trained using the "Multilingual Denoising Pretraining" objective. It was introduced in Multilingual Translation with Extensible Multilingual Pretraining and Finetuning paper.

Model description

mBART-50 is a multilingual Sequence-to-Sequence model. It was introduced to show that multilingual translation models can be created through multilingual fine-tuning. Instead of fine-tuning on one direction, a pre-trained model is fine-tuned on many directions simultaneously. mBART-50 is created using the original mBART model and extended to add extra 25 languages to support multilingual machine translation models of 50 languages. The pre-training objective is explained below.

Multilingual Denoising Pretraining : The model incorporates N languages by concatenating data: D = {D1, ..., DN } where each Di is a collection of monolingual documents in language i . The source documents are noised using two schemes, first randomly shuffling the original sentences' order, and second a novel in-filling scheme, where spans of text are replaced with a single mask token. The model is then tasked to reconstruct the original text. 35% of each instance's words are masked by random sampling a span length according to a Poisson distribution (λ = 3.5) . The decoder input is the original text with one position offset. A language id symbol LID is used as the initial token to predict the sentence.

Intended uses & limitations

mbart-large-50 is pre-trained model and primarily aimed at being fine-tuned on translation tasks. It can also be fine-tuned on other multilingual sequence-to-sequence tasks. See the model hub to look for fine-tuned versions.

Training

As the model is multilingual, it expects the sequences in a different format. A special language id token is used as a prefix in both the source and target text. The text format is [lang_code] X [eos] with X being the source or target text respectively and lang_code is source_lang_code for source text and tgt_lang_code for target text. bos is never used. Once the examples are prepared in this format, it can be trained as any other sequence-to-sequence model.

from transformers import MBartForConditionalGeneration, MBart50TokenizerFast

model = MBartForConditionalGeneration.from_pretrained("facebook/mbart-large-50")
tokenizer = MBart50TokenizerFast.from_pretrained("facebook/mbart-large-50", src_lang="en_XX", tgt_lang="ro_RO")

src_text = " UN Chief Says There Is No Military Solution in Syria"
tgt_text =  "Şeful ONU declară că nu există o soluţie militară în Siria"

model_inputs = tokenizer(src_text, return_tensors="pt")
with tokenizer.as_target_tokenizer():
    labels = tokenizer(tgt_text, return_tensors="pt").input_ids

model(**model_inputs, labels=labels) # forward pass
Languages covered

Arabic (ar_AR), Czech (cs_CZ), German (de_DE), English (en_XX), Spanish (es_XX), Estonian (et_EE), Finnish (fi_FI), French (fr_XX), Gujarati (gu_IN), Hindi (hi_IN), Italian (it_IT), Japanese (ja_XX), Kazakh (kk_KZ), Korean (ko_KR), Lithuanian (lt_LT), Latvian (lv_LV), Burmese (my_MM), Nepali (ne_NP), Dutch (nl_XX), Romanian (ro_RO), Russian (ru_RU), Sinhala (si_LK), Turkish (tr_TR), Vietnamese (vi_VN), Chinese (zh_CN), Afrikaans (af_ZA), Azerbaijani (az_AZ), Bengali (bn_IN), Persian (fa_IR), Hebrew (he_IL), Croatian (hr_HR), Indonesian (id_ID), Georgian (ka_GE), Khmer (km_KH), Macedonian (mk_MK), Malayalam (ml_IN), Mongolian (mn_MN), Marathi (mr_IN), Polish (pl_PL), Pashto (ps_AF), Portuguese (pt_XX), Swedish (sv_SE), Swahili (sw_KE), Tamil (ta_IN), Telugu (te_IN), Thai (th_TH), Tagalog (tl_XX), Ukrainian (uk_UA), Urdu (ur_PK), Xhosa (xh_ZA), Galician (gl_ES), Slovene (sl_SI)

BibTeX entry and citation info
@article{tang2020multilingual,
    title={Multilingual Translation with Extensible Multilingual Pretraining and Finetuning},
    author={Yuqing Tang and Chau Tran and Xian Li and Peng-Jen Chen and Naman Goyal and Vishrav Chaudhary and Jiatao Gu and Angela Fan},
    year={2020},
    eprint={2008.00401},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}

Runs of facebook mbart-large-50 on huggingface.co

20.1K
Total runs
567
24-hour runs
1.2K
3-day runs
1.7K
7-day runs
-6.5K
30-day runs

More Information About mbart-large-50 huggingface.co Model

More mbart-large-50 license Visit here:

https://choosealicense.com/licenses/mit

mbart-large-50 huggingface.co

mbart-large-50 huggingface.co is an AI model on huggingface.co that provides mbart-large-50's model effect (), which can be used instantly with this facebook mbart-large-50 model. huggingface.co supports a free trial of the mbart-large-50 model, and also provides paid use of the mbart-large-50. Support call mbart-large-50 model through api, including Node.js, Python, http.

mbart-large-50 huggingface.co Url

https://huggingface.co/facebook/mbart-large-50

facebook mbart-large-50 online free

mbart-large-50 huggingface.co is an online trial and call api platform, which integrates mbart-large-50's modeling effects, including api services, and provides a free online trial of mbart-large-50, you can try mbart-large-50 online for free by clicking the link below.

facebook mbart-large-50 online free url in huggingface.co:

https://huggingface.co/facebook/mbart-large-50

mbart-large-50 install

mbart-large-50 is an open source model from GitHub that offers a free installation service, and any user can find mbart-large-50 on GitHub to install. At the same time, huggingface.co provides the effect of mbart-large-50 install, users can directly use mbart-large-50 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

mbart-large-50 install url in huggingface.co:

https://huggingface.co/facebook/mbart-large-50

Url of mbart-large-50

mbart-large-50 huggingface.co Url

Provider of mbart-large-50 huggingface.co

facebook
ORGANIZATIONS

Other API from facebook

huggingface.co

Total runs: 13.0M
Run Growth: 3.9M
Growth Rate: 29.62%
Updated: January 17 2024
huggingface.co

Total runs: 5.9M
Run Growth: 779.5K
Growth Rate: 13.10%
Updated: September 15 2023
huggingface.co

Total runs: 4.0M
Run Growth: 3.5M
Growth Rate: 86.91%
Updated: March 23 2023
huggingface.co

Total runs: 2.6M
Run Growth: 188.5K
Growth Rate: 7.20%
Updated: November 17 2022
huggingface.co

Total runs: 1.9M
Run Growth: 1.9M
Growth Rate: 99.04%
Updated: November 14 2023
huggingface.co

Total runs: 1.3M
Run Growth: -1.9K
Growth Rate: -0.15%
Updated: February 29 2024
huggingface.co

Total runs: 1.0M
Run Growth: 409.4K
Growth Rate: 40.15%
Updated: September 06 2023
huggingface.co

Total runs: 753.2K
Run Growth: -458.5K
Growth Rate: -60.87%
Updated: December 28 2021
huggingface.co

Total runs: 715.1K
Run Growth: 548.9K
Growth Rate: 76.76%
Updated: June 15 2023
huggingface.co

Total runs: 692.6K
Run Growth: 92.6K
Growth Rate: 13.37%
Updated: January 12 2024
huggingface.co

Total runs: 584.7K
Run Growth: 288.0K
Growth Rate: 49.26%
Updated: September 06 2023
huggingface.co

Total runs: 563.3K
Run Growth: 476.8K
Growth Rate: 84.64%
Updated: November 16 2023
huggingface.co

Total runs: 484.9K
Run Growth: 49.0K
Growth Rate: 10.10%
Updated: September 01 2023
huggingface.co

Total runs: 406.6K
Run Growth: 48.2K
Growth Rate: 11.85%
Updated: January 20 2022
huggingface.co

Total runs: 314.6K
Run Growth: 109.5K
Growth Rate: 34.80%
Updated: September 15 2023
huggingface.co

Total runs: 285.2K
Run Growth: 156.7K
Growth Rate: 54.95%
Updated: January 12 2024
huggingface.co

Total runs: 254.5K
Run Growth: 110.5K
Growth Rate: 43.40%
Updated: May 22 2023
huggingface.co

Total runs: 231.2K
Run Growth: 126.2K
Growth Rate: 54.58%
Updated: June 03 2022
huggingface.co

Total runs: 204.9K
Run Growth: 117.4K
Growth Rate: 57.32%
Updated: September 06 2023
huggingface.co

Total runs: 170.1K
Run Growth: -85.7K
Growth Rate: -50.36%
Updated: September 05 2023
huggingface.co

Total runs: 143.7K
Run Growth: -158.0K
Growth Rate: -109.99%
Updated: January 25 2024
huggingface.co

Total runs: 132.2K
Run Growth: 30.4K
Growth Rate: 23.02%
Updated: September 15 2023
huggingface.co

Total runs: 78.7K
Run Growth: -27.7K
Growth Rate: -35.22%
Updated: January 12 2024
huggingface.co

Total runs: 61.6K
Run Growth: 29.2K
Growth Rate: 47.36%
Updated: September 15 2023
huggingface.co

Total runs: 57.9K
Run Growth: -2.3K
Growth Rate: -44.21%
Updated: October 16 2024
huggingface.co

Total runs: 52.0K
Run Growth: -5.2M
Growth Rate: -9999.73%
Updated: June 13 2023
huggingface.co

Total runs: 46.1K
Run Growth: 4.2K
Growth Rate: 9.12%
Updated: March 13 2024
huggingface.co

Total runs: 46.0K
Run Growth: 24.5K
Growth Rate: 53.32%
Updated: May 22 2023
huggingface.co

Total runs: 45.4K
Run Growth: -57.2K
Growth Rate: -126.08%
Updated: January 25 2023
huggingface.co

Total runs: 39.8K
Run Growth: 14.4K
Growth Rate: 36.17%
Updated: June 05 2023
huggingface.co

Total runs: 26.9K
Run Growth: 6.0K
Growth Rate: 22.21%
Updated: February 12 2023
huggingface.co

Total runs: 25.1K
Run Growth: 9.3K
Growth Rate: 37.09%
Updated: January 25 2023
huggingface.co

Total runs: 23.9K
Run Growth: 13.6K
Growth Rate: 56.78%
Updated: September 02 2023
huggingface.co

Total runs: 23.7K
Run Growth: -1.6K
Growth Rate: -6.65%
Updated: September 06 2023
huggingface.co

Total runs: 22.0K
Run Growth: 13.5K
Growth Rate: 61.15%
Updated: July 23 2024
huggingface.co

Total runs: 20.7K
Run Growth: -4.3K
Growth Rate: -20.92%
Updated: September 15 2023
huggingface.co

Total runs: 19.4K
Run Growth: 17.7K
Growth Rate: 90.91%
Updated: January 25 2023