A cute robot wearing a kimono writes calligraphy with one single brush —
Stable Diffusion XL
Model Description
japanese-stablelm-base-beta-7b
is a 7B-parameter decoder-only language model based on
Llama-2-7b
that has been fine-tuned on a diverse collection of Japanese data, with the intent of maximizing downstream performance on Japanese language tasks.
Then start generating text with
japanese-stablelm-base-beta-7b
by using the following code snippet:
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "stabilityai/japanese-stablelm-base-beta-7b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
# The next line may need to be modified depending on the environment
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
prompt = """AI で科学研究を加速するには、""".strip()
input_ids = tokenizer.encode(
prompt,
add_special_tokens=True,
return_tensors="pt"
)
# this is for reproducibility.# feel free to change to get different result
seed = 23
torch.manual_seed(seed)
tokens = model.generate(
input_ids.to(device=model.device),
max_new_tokens=128,
temperature=0.99,
top_p=0.95,
do_sample=True,
)
out = tokenizer.decode(tokens[0], skip_special_tokens=True)
print(out)
We suggest playing with different generation config (
top_p
,
repetition_penalty
etc) to find the best setup for your tasks. For example, use higher temperature for roleplay task, lower temperature for reasoning.
Model Details
Model type
:
japanese-stablelm-base-beta-7b
model is an auto-regressive language model based on the Llama2 transformer architecture.
The model is intended to be used by all individuals as a foundation for application-specific fine-tuning without strict limitations on commercial use.
Limitations and bias
The pre-training dataset may have contained offensive or inappropriate content even after applying data cleansing filters which can be reflected in the model generated text. We recommend users exercise reasonable caution when using these models in production systems. Do not use the model for any applications that may cause harm or distress to individuals or groups.
Authors
This model was developed by the Research & Development team at Stability AI Japan, and the development was co-led by
Takuya Akiba
and
Meng Lee
. The members of the team are as follows:
We thank Meta Research for releasing Llama 2 under an open license for others to build on.
We are grateful for the contributions of the EleutherAI Polyglot-JA team in helping us to collect a large amount of pre-training data in Japanese. Polyglot-JA members includes Hyunwoong Ko (Project Lead), Fujiki Nakamura (originally started this project when he commited to the Polyglot team), Yunho Mo, Minji Jung, KeunSeok Im, and Su-Kyeong Jang.
@misc{JapaneseStableLMBaseBeta7B,
url={[https://huggingface.co/stabilityai/japanese-stablelm-base-beta-7b](https://huggingface.co/stabilityai/japanese-stablelm-base-beta-7b)},
title={Japanese StableLM Base Beta 7B},
author={Lee, Meng and Nakamura, Fujiki and Shing, Makoto and McCann, Paul and Akiba, Takuya and Orii, Naoki}
}
Runs of stabilityai japanese-stablelm-base-beta-7b on huggingface.co
274
Total runs
-18
24-hour runs
-22
3-day runs
-229
7-day runs
-5.0K
30-day runs
More Information About japanese-stablelm-base-beta-7b huggingface.co Model
More japanese-stablelm-base-beta-7b license Visit here:
japanese-stablelm-base-beta-7b huggingface.co is an AI model on huggingface.co that provides japanese-stablelm-base-beta-7b's model effect (), which can be used instantly with this stabilityai japanese-stablelm-base-beta-7b model. huggingface.co supports a free trial of the japanese-stablelm-base-beta-7b model, and also provides paid use of the japanese-stablelm-base-beta-7b. Support call japanese-stablelm-base-beta-7b model through api, including Node.js, Python, http.
japanese-stablelm-base-beta-7b huggingface.co is an online trial and call api platform, which integrates japanese-stablelm-base-beta-7b's modeling effects, including api services, and provides a free online trial of japanese-stablelm-base-beta-7b, you can try japanese-stablelm-base-beta-7b online for free by clicking the link below.
stabilityai japanese-stablelm-base-beta-7b online free url in huggingface.co:
japanese-stablelm-base-beta-7b is an open source model from GitHub that offers a free installation service, and any user can find japanese-stablelm-base-beta-7b on GitHub to install. At the same time, huggingface.co provides the effect of japanese-stablelm-base-beta-7b install, users can directly use japanese-stablelm-base-beta-7b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.
japanese-stablelm-base-beta-7b install url in huggingface.co: