stabilityai / japanese-stablelm-3b-4e1t-instruct

huggingface.co
Total runs: 621
24-hour runs: 21
7-day runs: 211
30-day runs: 514
Model's Last Updated: April 26 2024
text-generation

Introduction of japanese-stablelm-3b-4e1t-instruct

Model Details of japanese-stablelm-3b-4e1t-instruct

Japanese StableLM-3B-4E1T Instruct

Model Description

This is a 3B-parameter decoder-only Japanese language model fine-tuned on instruction-following datasets, built on top of the base model Japanese StableLM-3B-4E1T Base .

If you are in search of a larger model, please check Japanese Stable LM Instruct Gamma 7B .

Usage
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("stabilityai/japanese-stablelm-3b-4e1t-instruct")
model = AutoModelForCausalLM.from_pretrained(
  "stabilityai/japanese-stablelm-3b-4e1t-instruct",
  trust_remote_code=True,
  torch_dtype="auto",
)
model.eval()

if torch.cuda.is_available():
    model = model.to("cuda")

def build_prompt(user_query, inputs="", sep="\n\n### "):
    sys_msg = "以下は、タスクを説明する指示と、文脈のある入力の組み合わせです。要求を適切に満たす応答を書きなさい。"
    p = sys_msg
    roles = ["指示", "応答"]
    msgs = [": \n" + user_query, ": \n"]
    if inputs:
        roles.insert(1, "入力")
        msgs.insert(1, ": \n" + inputs)
    for role, msg in zip(roles, msgs):
        p += sep + role + msg
    return p

# Infer with prompt without any additional input
user_inputs = {
    "user_query": "与えられたことわざの意味を小学生でも分かるように教えてください。",
    "inputs": "情けは人のためならず"
}
prompt = build_prompt(**user_inputs)

input_ids = tokenizer.encode(
    prompt, 
    add_special_tokens=False, 
    return_tensors="pt"
)

tokens = model.generate(
    input_ids.to(device=model.device),
    max_new_tokens=256,
    temperature=1,
    top_p=0.95,
    do_sample=True,
)

out = tokenizer.decode(tokens[0][input_ids.shape[1]:], skip_special_tokens=True).strip()
print(out)
Model Details
Model Architecture

The model is a decoder-only transformer similar to the LLaMA ( Touvron et al., 2023 ) architecture with the following modifications:

Parameters Hidden Size Layers Heads Sequence Length
2,795,443,200 2560 32 32 4096
Training Datasets
Use and Limitations
Intended Use

The model is intended to be used by all individuals as a foundational model for application-specific fine-tuning without strict limitations on commercial use.

Limitations and bias

The pre-training dataset may have contained offensive or inappropriate content even after applying data cleansing filters which can be reflected in the model-generated text. We recommend users exercise reasonable caution when using these models in production systems. Do not use the model for any applications that may cause harm or distress to individuals or groups.

Credits

The fine-tuning was carried out by Fujiki Nakamura . Other aspects, including data preparation and evaluation, were handled by the Language Team of Stability AI Japan, notably Meng Lee , Makoto Shing , Paul McCann , Naoki Orii , and Takuya Akiba .

Acknowledgements

We are grateful for the contributions of the EleutherAI Polyglot-JA team in helping us to collect a large amount of pre-training data in Japanese. Polyglot-JA members includes Hyunwoong Ko (Project Lead), Fujiki Nakamura (originally started this project when he commited to the Polyglot team), Yunho Mo, Minji Jung, KeunSeok Im, and Su-Kyeong Jang.

We are also appreciative of AI Novelist/Sta (Bit192, Inc.) and the numerous contributors from Stable Community Japan for assisting us in gathering a large amount of high-quality Japanese textual data for model training.

Runs of stabilityai japanese-stablelm-3b-4e1t-instruct on huggingface.co

621
Total runs
21
24-hour runs
74
3-day runs
211
7-day runs
514
30-day runs

More Information About japanese-stablelm-3b-4e1t-instruct huggingface.co Model

More japanese-stablelm-3b-4e1t-instruct license Visit here:

https://choosealicense.com/licenses/apache-2.0

japanese-stablelm-3b-4e1t-instruct huggingface.co

japanese-stablelm-3b-4e1t-instruct huggingface.co is an AI model on huggingface.co that provides japanese-stablelm-3b-4e1t-instruct's model effect (), which can be used instantly with this stabilityai japanese-stablelm-3b-4e1t-instruct model. huggingface.co supports a free trial of the japanese-stablelm-3b-4e1t-instruct model, and also provides paid use of the japanese-stablelm-3b-4e1t-instruct. Support call japanese-stablelm-3b-4e1t-instruct model through api, including Node.js, Python, http.

japanese-stablelm-3b-4e1t-instruct huggingface.co Url

https://huggingface.co/stabilityai/japanese-stablelm-3b-4e1t-instruct

stabilityai japanese-stablelm-3b-4e1t-instruct online free

japanese-stablelm-3b-4e1t-instruct huggingface.co is an online trial and call api platform, which integrates japanese-stablelm-3b-4e1t-instruct's modeling effects, including api services, and provides a free online trial of japanese-stablelm-3b-4e1t-instruct, you can try japanese-stablelm-3b-4e1t-instruct online for free by clicking the link below.

stabilityai japanese-stablelm-3b-4e1t-instruct online free url in huggingface.co:

https://huggingface.co/stabilityai/japanese-stablelm-3b-4e1t-instruct

japanese-stablelm-3b-4e1t-instruct install

japanese-stablelm-3b-4e1t-instruct is an open source model from GitHub that offers a free installation service, and any user can find japanese-stablelm-3b-4e1t-instruct on GitHub to install. At the same time, huggingface.co provides the effect of japanese-stablelm-3b-4e1t-instruct install, users can directly use japanese-stablelm-3b-4e1t-instruct installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

japanese-stablelm-3b-4e1t-instruct install url in huggingface.co:

https://huggingface.co/stabilityai/japanese-stablelm-3b-4e1t-instruct

Url of japanese-stablelm-3b-4e1t-instruct

japanese-stablelm-3b-4e1t-instruct huggingface.co Url

Provider of japanese-stablelm-3b-4e1t-instruct huggingface.co

stabilityai
ORGANIZATIONS

Other API from stabilityai

huggingface.co

Total runs: 155.2K
Run Growth: 19.7K
Growth Rate: 13.18%
Updated: August 04 2023
huggingface.co

Total runs: 132.2K
Run Growth: 5.6K
Growth Rate: 4.26%
Updated: Juli 10 2024
huggingface.co

Total runs: 34.3K
Run Growth: 3.3K
Growth Rate: 9.68%
Updated: August 09 2024
huggingface.co

Total runs: 378
Run Growth: -97.9K
Growth Rate: -25898.41%
Updated: August 03 2024
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated: Juli 10 2024
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated: April 13 2024