yanolja / EEVE-Korean-10.8B-v1.0

huggingface.co
Total runs: 19.4K
24-hour runs: 0
7-day runs: 172
30-day runs: 5.5K
Model's Last Updated: February 24 2024
text-generation

Introduction of EEVE-Korean-10.8B-v1.0

Model Details of EEVE-Korean-10.8B-v1.0

Built with Axolotl

EEVE-Korean-10.8B-v1.0

Join Our Community on Discord!

If you're passionate about the field of Large Language Models and wish to exchange knowledge and insights, we warmly invite you to join our Discord server. It's worth noting that Korean is the primary language used in this server. The landscape of LLM is evolving rapidly, and without active sharing, our collective knowledge risks becoming outdated swiftly. Let's collaborate and drive greater impact together! Join us here: Discord Link .

Our Dedicated Team (Alphabetical Order)
Research Engineering Product Management UX Design
Myeongho Jeong Geon Kim Bokyung Huh Eunsue Choi
Seungduk Kim Rifqi Alfi
Seungtaek Choi Sanghoon Han
Suhyun Kang
About the Model

This model is a Korean vocabulary-extended version of upstage/SOLAR-10.7B-v1.0 , specifically fine-tuned on various Korean web-crawled datasets available on HuggingFace. Our approach was to expand the model's understanding of Korean by pre-training the embeddings for new tokens and partially fine-tuning the lm_head embeddings for the already existing tokens while preserving the original parameters of the base model.

Technical Deep Dive

To adapt foundational models from English to Korean, we use subword-based embedding with a seven-stage training process involving parameter freezing. This approach progressively trains from input embeddings to full parameters, efficiently extending the model's vocabulary to include Korean. Our method enhances the model's cross-linguistic applicability by carefully integrating new linguistic tokens, focusing on causal language modeling pre-training. We leverage the inherent capabilities of foundational models trained on English to efficiently transfer knowledge and reasoning to Korean, optimizing the adaptation process.

For more details, please refer to our technical report: Efficient and Effective Vocabulary Expansion Towards Multilingual Large Language Models .

Here’s an simplified code for our key approach:

# number_of_old_tokens is the size of tokenizer before vocab extension. For example, in case of EEVE-Korean-10.8B-v1.0, number_of_old_tokens is 32000.
def freeze_partial_embedding_hook(grad):
    grad[:number_of_old_tokens] = 0
    return grad

for name, param in model.named_parameters():
    if ("lm_head" in name or "embed_tokens" in name) and "original" not in name:
        param.requires_grad = True
        if "embed_tokens" in name:
            param.register_hook(freeze_partial_embedding_hook)
    else:
        param.requires_grad = False
Usage and Limitations

Keep in mind that this model hasn't been fine-tuned with instruction-based training. While it excels in Korean language tasks, we advise careful consideration and further training for specific applications.

Training Details

Our model’s training was comprehensive and diverse:

  • Vocabulary Expansion: We meticulously selected 8,960 Korean tokens based on their frequency in our Korean web corpus. This process involved multiple rounds of tokenizer training, manual curation, and token frequency analysis, ensuring a rich and relevant vocabulary for our model.

    1. Initial Tokenizer Training: We trained an intermediate tokenizer on a Korean web corpus, with a vocabulary of 40,000 tokens.

    2. Extraction of New Korean Tokens: From the intermediate tokenizer, we identified all Korean tokens not present in the original SOLAR's tokenizer.

    3. Manual Tokenizer Construction: We then built the target tokenizer, focusing on these new Korean tokens.

    4. Frequency Analysis: Using the target tokenizer, we processed a 100GB Korean corpus to count each token's frequency.

    5. Refinement of Token List: We removed tokens appearing less than 6,000 times, ensuring to secure enough tokens to train models later.

    6. Inclusion of Single-Letter Characters: Counted missing Korean single-letter characters and added them to the target tokenizer that appeared more than 6,000 times.

    7. Iterative Refinement: We repeated steps 2 to 6 until there were no tokens to drop or add.

    8. Training Bias Towards New Tokens: Our training data was biased to include more texts with new tokens, for effective learning.

This rigorous approach ensured a comprehensive and contextually rich Korean vocabulary for the model.

Citation
@misc{kim2024efficient,
      title={Efficient and Effective Vocabulary Expansion Towards Multilingual Large Language Models}, 
      author={Seungduk Kim and Seungtaek Choi and Myeongho Jeong},
      year={2024},
      eprint={2402.14714},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Runs of yanolja EEVE-Korean-10.8B-v1.0 on huggingface.co

19.4K
Total runs
0
24-hour runs
47
3-day runs
172
7-day runs
5.5K
30-day runs

More Information About EEVE-Korean-10.8B-v1.0 huggingface.co Model

More EEVE-Korean-10.8B-v1.0 license Visit here:

https://choosealicense.com/licenses/apache-2.0

EEVE-Korean-10.8B-v1.0 huggingface.co

EEVE-Korean-10.8B-v1.0 huggingface.co is an AI model on huggingface.co that provides EEVE-Korean-10.8B-v1.0's model effect (), which can be used instantly with this yanolja EEVE-Korean-10.8B-v1.0 model. huggingface.co supports a free trial of the EEVE-Korean-10.8B-v1.0 model, and also provides paid use of the EEVE-Korean-10.8B-v1.0. Support call EEVE-Korean-10.8B-v1.0 model through api, including Node.js, Python, http.

EEVE-Korean-10.8B-v1.0 huggingface.co Url

https://huggingface.co/yanolja/EEVE-Korean-10.8B-v1.0

yanolja EEVE-Korean-10.8B-v1.0 online free

EEVE-Korean-10.8B-v1.0 huggingface.co is an online trial and call api platform, which integrates EEVE-Korean-10.8B-v1.0's modeling effects, including api services, and provides a free online trial of EEVE-Korean-10.8B-v1.0, you can try EEVE-Korean-10.8B-v1.0 online for free by clicking the link below.

yanolja EEVE-Korean-10.8B-v1.0 online free url in huggingface.co:

https://huggingface.co/yanolja/EEVE-Korean-10.8B-v1.0

EEVE-Korean-10.8B-v1.0 install

EEVE-Korean-10.8B-v1.0 is an open source model from GitHub that offers a free installation service, and any user can find EEVE-Korean-10.8B-v1.0 on GitHub to install. At the same time, huggingface.co provides the effect of EEVE-Korean-10.8B-v1.0 install, users can directly use EEVE-Korean-10.8B-v1.0 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

EEVE-Korean-10.8B-v1.0 install url in huggingface.co:

https://huggingface.co/yanolja/EEVE-Korean-10.8B-v1.0

Url of EEVE-Korean-10.8B-v1.0

EEVE-Korean-10.8B-v1.0 huggingface.co Url

Provider of EEVE-Korean-10.8B-v1.0 huggingface.co

yanolja
ORGANIZATIONS

Other API from yanolja