shibing624 / chatglm3-6b-csc-chinese-lora

huggingface.co
Total runs: 87
24-hour runs: 0
7-day runs: 28
30-day runs: -2
Model's Last Updated: February 19 2024
text-generation

Introduction of chatglm3-6b-csc-chinese-lora

Model Details of chatglm3-6b-csc-chinese-lora

Chinese Spelling Correction LoRA Model

ChatGLM3-6B中文纠错LoRA模型

shibing624/chatglm3-6b-csc-chinese-lora evaluate test data:

The overall performance of shibing624/chatglm3-6b-csc-chinese-lora on CSC test :

input_text pred
对下面文本纠错:少先队员因该为老人让坐。 少先队员应该为老人让座。

在CSC测试集上生成结果纠错准确率高,由于是基于 THUDM/chatglm3-6b 模型,结果常常能带给人惊喜,不仅能纠错,还带有句子润色和改写功能。

Usage

本项目开源在 pycorrector 项目: pycorrector ,可支持ChatGLM原生模型和LoRA微调后的模型,通过如下命令调用:

Install package:

pip install -U pycorrector
from pycorrector import GptCorrector
model = GptCorrector("THUDM/chatglm3-6b", "chatglm", peft_name="shibing624/chatglm3-6b-csc-chinese-lora")
r = model.correct_batch(["少先队员因该为老人让坐。"])
print(r) # ['少先队员应该为老人让座。']
Usage (HuggingFace Transformers)

Without pycorrector , you can use the model like this:

First, you pass your input through the transformer model, then you get the generated sentence.

Install package:

pip install transformers 
import os

import torch
from peft import PeftModel
from transformers import AutoTokenizer, AutoModel

os.environ["KMP_DUPLICATE_LIB_OK"] = "TRUE"
tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm3-6b", trust_remote_code=True)
model = AutoModel.from_pretrained("THUDM/chatglm3-6b", trust_remote_code=True).half().cuda()
model = PeftModel.from_pretrained(model, "shibing624/chatglm3-6b-csc-chinese-lora")

sents = ['对下面文本纠错\n\n少先队员因该为老人让坐。',
         '对下面文本纠错\n\n下个星期,我跟我朋唷打算去法国玩儿。']


def get_prompt(user_query):
    vicuna_prompt = "A chat between a curious user and an artificial intelligence assistant. " \
                    "The assistant gives helpful, detailed, and polite answers to the user's questions. " \
                    "USER: {query} ASSISTANT:"
    return vicuna_prompt.format(query=user_query)


for s in sents:
    q = get_prompt(s)
    input_ids = tokenizer(q).input_ids
    generation_kwargs = dict(max_new_tokens=128, do_sample=True, temperature=0.8)
    outputs = model.generate(input_ids=torch.as_tensor([input_ids]).to('cuda:0'), **generation_kwargs)
    output_tensor = outputs[0][len(input_ids):]
    response = tokenizer.decode(output_tensor, skip_special_tokens=True)
    print(response)

output:

少先队员应该为老人让座。
下个星期,我跟我朋友打算去法国玩儿。

模型文件组成:

chatglm3-6b-csc-chinese-lora
    ├── adapter_config.json
    └── adapter_model.bin
训练参数:

loss

  • num_epochs: 5
  • per_device_train_batch_size: 6
  • learning_rate: 2e-05
  • best steps: 25100
  • train_loss: 0.0834
  • lr_scheduler_type: linear
  • base model: THUDM/chatglm3-6b
  • warmup_steps: 50
  • "save_strategy": "steps"
  • "save_steps": 500
  • "save_total_limit": 10
  • "bf16": false
  • "fp16": true
  • "optim": "adamw_torch"
  • "ddp_find_unused_parameters": false
  • "gradient_checkpointing": true
  • max_seq_length: 512
  • max_length: 512
  • prompt_template_name: vicuna
  • 6 * V100 32GB, training 48 hours
训练数据集

训练集包括以下数据:

如果需要训练文本纠错模型,请参考 https://github.com/shibing624/pycorrector

Citation
@software{pycorrector,
  author = {Ming Xu},
  title = {pycorrector: Text Error Correction Tool},
  year = {2023},
  url = {https://github.com/shibing624/pycorrector},
}

Runs of shibing624 chatglm3-6b-csc-chinese-lora on huggingface.co

87
Total runs
0
24-hour runs
-2
3-day runs
28
7-day runs
-2
30-day runs

More Information About chatglm3-6b-csc-chinese-lora huggingface.co Model

More chatglm3-6b-csc-chinese-lora license Visit here:

https://choosealicense.com/licenses/apache-2.0

chatglm3-6b-csc-chinese-lora huggingface.co

chatglm3-6b-csc-chinese-lora huggingface.co is an AI model on huggingface.co that provides chatglm3-6b-csc-chinese-lora's model effect (), which can be used instantly with this shibing624 chatglm3-6b-csc-chinese-lora model. huggingface.co supports a free trial of the chatglm3-6b-csc-chinese-lora model, and also provides paid use of the chatglm3-6b-csc-chinese-lora. Support call chatglm3-6b-csc-chinese-lora model through api, including Node.js, Python, http.

chatglm3-6b-csc-chinese-lora huggingface.co Url

https://huggingface.co/shibing624/chatglm3-6b-csc-chinese-lora

shibing624 chatglm3-6b-csc-chinese-lora online free

chatglm3-6b-csc-chinese-lora huggingface.co is an online trial and call api platform, which integrates chatglm3-6b-csc-chinese-lora's modeling effects, including api services, and provides a free online trial of chatglm3-6b-csc-chinese-lora, you can try chatglm3-6b-csc-chinese-lora online for free by clicking the link below.

shibing624 chatglm3-6b-csc-chinese-lora online free url in huggingface.co:

https://huggingface.co/shibing624/chatglm3-6b-csc-chinese-lora

chatglm3-6b-csc-chinese-lora install

chatglm3-6b-csc-chinese-lora is an open source model from GitHub that offers a free installation service, and any user can find chatglm3-6b-csc-chinese-lora on GitHub to install. At the same time, huggingface.co provides the effect of chatglm3-6b-csc-chinese-lora install, users can directly use chatglm3-6b-csc-chinese-lora installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

chatglm3-6b-csc-chinese-lora install url in huggingface.co:

https://huggingface.co/shibing624/chatglm3-6b-csc-chinese-lora

Url of chatglm3-6b-csc-chinese-lora

chatglm3-6b-csc-chinese-lora huggingface.co Url

Provider of chatglm3-6b-csc-chinese-lora huggingface.co

shibing624
ORGANIZATIONS

Other API from shibing624