dicta-il / dictalm2.0-instruct

huggingface.co
Total runs: 14.6K
24-hour runs: -75
7-day runs: -793
30-day runs: 2.5K
Model's Last Updated: Juli 10 2024
text-generation

Introduction of dictalm2.0-instruct

Model Details of dictalm2.0-instruct

Adapting LLMs to Hebrew: Unveiling DictaLM 2.0 with Enhanced Vocabulary and Instruction Capabilities

The DictaLM-2.0-Instruct Large Language Model (LLM) is an instruct fine-tuned version of the DictaLM-2.0 generative model using a variety of conversation datasets.

For full details of this model please read our release blog post or the technical report .

This is the instruct-tuned full-precision model designed for chat. You can try the model out on a live demo here .

You can view and access the full collection of base/instruct unquantized/quantized versions of DictaLM-2.0 here .

Instruction format

In order to leverage instruction fine-tuning, your prompt should be surrounded by [INST] and [/INST] tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.

E.g.

text = """<s>[INST] איזה רוטב אהוב עליך? [/INST]
טוב, אני די מחבב כמה טיפות מיץ לימון סחוט טרי. זה מוסיף בדיוק את הכמות הנכונה של טעם חמצמץ לכל מה שאני מבשל במטבח!</s>[INST] האם יש לך מתכונים למיונז? [/INST]"

This format is available as a chat template via the apply_chat_template() method:

Example Code
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

device = "cuda" # the device to load the model onto

model = AutoModelForCausalLM.from_pretrained("dicta-il/dictalm2.0-instruct", torch_dtype=torch.bfloat16, device_map=device)
tokenizer = AutoTokenizer.from_pretrained("dicta-il/dictalm2.0-instruct")

messages = [
    {"role": "user", "content": "איזה רוטב אהוב עליך?"},
    {"role": "assistant", "content": "טוב, אני די מחבב כמה טיפות מיץ לימון סחוט טרי. זה מוסיף בדיוק את הכמות הנכונה של טעם חמצמץ לכל מה שאני מבשל במטבח!"},
    {"role": "user", "content": "האם יש לך מתכונים למיונז?"}
]

encoded = tokenizer.apply_chat_template(messages, return_tensors="pt").to(device)

generated_ids = model.generate(encoded, max_new_tokens=50, do_sample=True)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
# <s> [INST] איזה רוטב אהוב עליך? [/INST]
# טוב, אני די מחבב כמה טיפות מיץ לימון סחוט טרי. זה מוסיף בדיוק את הכמות הנכונה של טעם חמצמץ לכל מה שאני מבשל במטבח!</s>  [INST] האם יש לך מתכונים למיונז? [/INST]
# בטח, הנה מתכון בסיסי וקל להכנת מיונז ביתי!
# 
# מרכיבים:
# - 2 חלמונים גדולים
# - 1 כף חומץ יין לבן
# (it stopped early because we set max_new_tokens=50)
Model Architecture

DictaLM-2.0-Instruct follows the Zephyr-7B-beta recipe for fine-tuning an instruct model, with an extended instruct dataset for Hebrew.

Limitations

The DictaLM 2.0 Instruct model is a demonstration that the base model can be fine-tuned to achieve compelling performance. It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.

Citation

If you use this model, please cite:

@misc{shmidman2024adaptingllmshebrewunveiling,
      title={Adapting LLMs to Hebrew: Unveiling DictaLM 2.0 with Enhanced Vocabulary and Instruction Capabilities}, 
      author={Shaltiel Shmidman and Avi Shmidman and Amir DN Cohen and Moshe Koppel},
      year={2024},
      eprint={2407.07080},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2407.07080}, 
}

Runs of dicta-il dictalm2.0-instruct on huggingface.co

14.6K
Total runs
-75
24-hour runs
-255
3-day runs
-793
7-day runs
2.5K
30-day runs

More Information About dictalm2.0-instruct huggingface.co Model

More dictalm2.0-instruct license Visit here:

https://choosealicense.com/licenses/apache-2.0

dictalm2.0-instruct huggingface.co

dictalm2.0-instruct huggingface.co is an AI model on huggingface.co that provides dictalm2.0-instruct's model effect (), which can be used instantly with this dicta-il dictalm2.0-instruct model. huggingface.co supports a free trial of the dictalm2.0-instruct model, and also provides paid use of the dictalm2.0-instruct. Support call dictalm2.0-instruct model through api, including Node.js, Python, http.

dictalm2.0-instruct huggingface.co Url

https://huggingface.co/dicta-il/dictalm2.0-instruct

dicta-il dictalm2.0-instruct online free

dictalm2.0-instruct huggingface.co is an online trial and call api platform, which integrates dictalm2.0-instruct's modeling effects, including api services, and provides a free online trial of dictalm2.0-instruct, you can try dictalm2.0-instruct online for free by clicking the link below.

dicta-il dictalm2.0-instruct online free url in huggingface.co:

https://huggingface.co/dicta-il/dictalm2.0-instruct

dictalm2.0-instruct install

dictalm2.0-instruct is an open source model from GitHub that offers a free installation service, and any user can find dictalm2.0-instruct on GitHub to install. At the same time, huggingface.co provides the effect of dictalm2.0-instruct install, users can directly use dictalm2.0-instruct installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

dictalm2.0-instruct install url in huggingface.co:

https://huggingface.co/dicta-il/dictalm2.0-instruct

Url of dictalm2.0-instruct

dictalm2.0-instruct huggingface.co Url

Provider of dictalm2.0-instruct huggingface.co

dicta-il
ORGANIZATIONS

Other API from dicta-il

huggingface.co

Total runs: 17.9K
Run Growth: 7.9K
Growth Rate: 43.89%
Updated: Juli 10 2024
huggingface.co

Total runs: 12.0K
Run Growth: 6.4K
Growth Rate: 53.29%
Updated: Dezember 28 2023
huggingface.co

Total runs: 917
Run Growth: 861
Growth Rate: 98.85%
Updated: Juli 03 2023
huggingface.co

Total runs: 234
Run Growth: -1.0K
Growth Rate: -434.20%
Updated: Dezember 28 2023
huggingface.co

Total runs: 140
Run Growth: 129
Growth Rate: 86.00%
Updated: April 01 2024
huggingface.co

Total runs: 116
Run Growth: 71
Growth Rate: 61.74%
Updated: September 02 2024
huggingface.co

Total runs: 114
Run Growth: 94
Growth Rate: 82.46%
Updated: Juli 03 2023
huggingface.co

Total runs: 28
Run Growth: 16
Growth Rate: 61.54%
Updated: November 17 2024
huggingface.co

Total runs: 7
Run Growth: -9
Growth Rate: -128.57%
Updated: Dezember 15 2024