ChatGLM3-6B is the latest open-source model in the ChatGLM series. While retaining many excellent features such as smooth dialogue and low deployment threshold from the previous two generations, ChatGLM3-6B introduces the following features:
More Powerful Base Model:
The base model of ChatGLM3-6B, ChatGLM3-6B-Base, employs a more diverse training dataset, more sufficient training steps, and a more reasonable training strategy. Evaluations on datasets such as semantics, mathematics, reasoning, code, knowledge, etc., show that ChatGLM3-6B-Base has the strongest performance among pre-trained models under 10B.
More Comprehensive Function Support:
ChatGLM3-6B adopts a newly designed
Prompt format
, in addition to the normal multi-turn dialogue. It also natively supports
function call
, code interpreter, and complex scenarios such as agent tasks.
More Comprehensive Open-source Series:
In addition to the dialogue model ChatGLM3-6B, the base model ChatGLM-6B-Base and the long-text dialogue model ChatGLM3-6B-32K are also open-sourced. All the weights are
fully open
for academic research, and after completing the
questionnaire
registration, they are also
allowed for free commercial use
.
The code in this repository is open-sourced under the
Apache-2.0 license
, while the use of the ChatGLM3-6B model weights needs to comply with the
Model License
.
引用 (Citation)
如果你觉得我们的工作有帮助的话,请考虑引用下列论文。
If you find our work helpful, please consider citing the following papers.
@article{zeng2022glm,
title={Glm-130b: An open bilingual pre-trained model},
author={Zeng, Aohan and Liu, Xiao and Du, Zhengxiao and Wang, Zihan and Lai, Hanyu and Ding, Ming and Yang, Zhuoyi and Xu, Yifan and Zheng, Wendi and Xia, Xiao and others},
journal={arXiv preprint arXiv:2210.02414},
year={2022}
}
@inproceedings{du2022glm,
title={GLM: General Language Model Pretraining with Autoregressive Blank Infilling},
author={Du, Zhengxiao and Qian, Yujie and Liu, Xiao and Ding, Ming and Qiu, Jiezhong and Yang, Zhilin and Tang, Jie},
booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
pages={320--335},
year={2022}
}
Runs of THUDM chatglm3-6b on huggingface.co
30.4K
Total runs
0
24-hour runs
-269
3-day runs
-571
7-day runs
1.5K
30-day runs
More Information About chatglm3-6b huggingface.co Model
chatglm3-6b huggingface.co
chatglm3-6b huggingface.co is an AI model on huggingface.co that provides chatglm3-6b's model effect (), which can be used instantly with this THUDM chatglm3-6b model. huggingface.co supports a free trial of the chatglm3-6b model, and also provides paid use of the chatglm3-6b. Support call chatglm3-6b model through api, including Node.js, Python, http.
chatglm3-6b huggingface.co is an online trial and call api platform, which integrates chatglm3-6b's modeling effects, including api services, and provides a free online trial of chatglm3-6b, you can try chatglm3-6b online for free by clicking the link below.
THUDM chatglm3-6b online free url in huggingface.co:
chatglm3-6b is an open source model from GitHub that offers a free installation service, and any user can find chatglm3-6b on GitHub to install. At the same time, huggingface.co provides the effect of chatglm3-6b install, users can directly use chatglm3-6b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.