elyza / Llama-3-ELYZA-JP-8B-GGUF

huggingface.co
Total runs: 3.2K
24-hour runs: -26
7-day runs: -472
30-day runs: -1.4K
Model's Last Updated: 6월 26 2024

Introduction of Llama-3-ELYZA-JP-8B-GGUF

Model Details of Llama-3-ELYZA-JP-8B-GGUF

Llama-3-ELYZA-JP-8B-GGUF

Llama-3-ELYZA-JP-8B-image

Model Description

Llama-3-ELYZA-JP-8B is a large language model trained by ELYZA, Inc . Based on meta-llama/Meta-Llama-3-8B-Instruct , it has been enhanced for Japanese usage through additional pre-training and instruction tuning. (Built with Meta Llama3)

For more details, please refer to our blog post .

Quantization

We have prepared two quantized model options, GGUF and AWQ. This is the GGUF (Q4_K_M) model, converted using llama.cpp .

The following table shows the performance degradation due to quantization:

Use with llama.cpp

Install llama.cpp through brew (works on Mac and Linux):

brew install llama.cpp

Invoke the llama.cpp server:

$ llama-server \
--hf-repo elyza/Llama-3-ELYZA-JP-8B-GGUF \
--hf-file Llama-3-ELYZA-JP-8B-q4_k_m.gguf \
--port 8080

Call the API using curl:

$ curl http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
  "messages": [
    { "role": "system", "content": "あなたは誠実で優秀な日本人のアシスタントです。特に指示が無い場合は、常に日本語で回答してください。" },
    { "role": "user", "content": "古代ギリシャを学ぶ上で知っておくべきポイントは?" }
  ],
  "temperature": 0.6,
  "max_tokens": -1,
  "stream": false
}'

Call the API using Python:

import openai

client = openai.OpenAI(
    base_url="http://localhost:8080/v1",
    api_key = "dummy_api_key"
)

completion = client.chat.completions.create(
    model="dummy_model_name",
    messages=[
        {"role": "system", "content": "あなたは誠実で優秀な日本人のアシスタントです。特に指示が無い場合は、常に日本語で回答してください。"},
        {"role": "user", "content": "古代ギリシャを学ぶ上で知っておくべきポイントは?"}
    ]
)
Use with Desktop App

There are various desktop applications that can handle GGUF models, but here we will introduce how to use the model in the no-code environment LM Studio .

  • Installation : Download and install LM Studio .
  • Downloading the Model : Search for elyza/Llama-3-ELYZA-JP-8B-GGUF in the search bar on the home page 🏠, and download Llama-3-ELYZA-JP-8B-q4_k_m.gguf .
  • Start Chatting : Click on 💬 in the sidebar, select Llama-3-ELYZA-JP-8B-GGUF from "Select a Model to load" in the header, and load the model. You can now freely chat with the local LLM.
  • Setting Options : You can set options from the sidebar on the right. Faster inference can be achieved by setting Quick GPU Offload to Max in the GPU Settings.
  • (For Developers) Starting an API Server : Click <-> in the left sidebar and move to the Local Server tab. Select the model and click Start Server to launch an OpenAI API-compatible API server.

lmstudio-demo

This demo showcases Llama-3-ELYZA-JP-8B-GGUF running smoothly on a MacBook Pro (M1 Pro), achieving an inference speed of approximately 20 tokens per second.

Developers

Listed in alphabetical order.

License

Meta Llama 3 Community License

How to Cite
@misc{elyzallama2024,
      title={elyza/Llama-3-ELYZA-JP-8B},
      url={https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B},
      author={Masato Hirakawa and Shintaro Horie and Tomoaki Nakamura and Daisuke Oba and Sam Passaglia and Akira Sasaki},
      year={2024},
}
Citations
@article{llama3modelcard,
    title={Llama 3 Model Card},
    author={AI@Meta},
    year={2024},
    url = {https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md}
}

Runs of elyza Llama-3-ELYZA-JP-8B-GGUF on huggingface.co

3.2K
Total runs
-26
24-hour runs
-65
3-day runs
-472
7-day runs
-1.4K
30-day runs

More Information About Llama-3-ELYZA-JP-8B-GGUF huggingface.co Model

More Llama-3-ELYZA-JP-8B-GGUF license Visit here:

https://choosealicense.com/licenses/llama3

Llama-3-ELYZA-JP-8B-GGUF huggingface.co

Llama-3-ELYZA-JP-8B-GGUF huggingface.co is an AI model on huggingface.co that provides Llama-3-ELYZA-JP-8B-GGUF's model effect (), which can be used instantly with this elyza Llama-3-ELYZA-JP-8B-GGUF model. huggingface.co supports a free trial of the Llama-3-ELYZA-JP-8B-GGUF model, and also provides paid use of the Llama-3-ELYZA-JP-8B-GGUF. Support call Llama-3-ELYZA-JP-8B-GGUF model through api, including Node.js, Python, http.

Llama-3-ELYZA-JP-8B-GGUF huggingface.co Url

https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B-GGUF

elyza Llama-3-ELYZA-JP-8B-GGUF online free

Llama-3-ELYZA-JP-8B-GGUF huggingface.co is an online trial and call api platform, which integrates Llama-3-ELYZA-JP-8B-GGUF's modeling effects, including api services, and provides a free online trial of Llama-3-ELYZA-JP-8B-GGUF, you can try Llama-3-ELYZA-JP-8B-GGUF online for free by clicking the link below.

elyza Llama-3-ELYZA-JP-8B-GGUF online free url in huggingface.co:

https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B-GGUF

Llama-3-ELYZA-JP-8B-GGUF install

Llama-3-ELYZA-JP-8B-GGUF is an open source model from GitHub that offers a free installation service, and any user can find Llama-3-ELYZA-JP-8B-GGUF on GitHub to install. At the same time, huggingface.co provides the effect of Llama-3-ELYZA-JP-8B-GGUF install, users can directly use Llama-3-ELYZA-JP-8B-GGUF installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

Llama-3-ELYZA-JP-8B-GGUF install url in huggingface.co:

https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B-GGUF

Url of Llama-3-ELYZA-JP-8B-GGUF

Llama-3-ELYZA-JP-8B-GGUF huggingface.co Url

Provider of Llama-3-ELYZA-JP-8B-GGUF huggingface.co

elyza
ORGANIZATIONS

Other API from elyza