We opensource our
Aquila2
series, now including
Aquila2
, the base language models, namely
Aquila2-7B
and
Aquila2-34B
, as well as
AquilaChat2
, the chat models, namely
AquilaChat2-7B
and
AquilaChat2-34B
, as well as the long-text chat models, namely
AquilaChat2-7B-16k
and
AquilaChat2-34B-16k
The additional details of the Aquila model will be presented in the official technical report. Please stay tuned for updates on official channels.
Updates 2024.6.6
We have updated the basic language model
Aquila2-7B
, which has the following advantages compared to the previous model:
Replaced tokenizer with higher compression ratio:
Tokenizer
Size
Zh
En
Code
Math
Average
Aquila2-original
100k
4.70
4.42
3.20
3.77
4.02
Qwen1.5
151k
4.27
4.51
3.62
3.35
3.94
Llama3
128k
3.45
4.61
3.77
3.88
3.93
Aquila2-new
143k
4.60
4.61
3.78
3.88
4.22
The maximum processing length supported by the model has increased from 2048 to 8192
Quick Start Aquila2-7B
1. Inference
Aquila2-7B is a base model that can be used for continuation.
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
from transformers import BitsAndBytesConfig
device= "cuda:0"# Model Name
model_name = 'BAAI/Aquila2-7B'# load model and tokenizer
quantization_config=BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_use_double_quant=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_compute_dtype=torch.bfloat16,
)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, trust_remote_code=True,
# quantization_config=quantization_config # Uncomment this one for 4-bit quantization
)
tokenizer = AutoTokenizer.from_pretrained(path, trust_remote_code=True)
model.eval()
model.to(device)
# Example
text = "The meaning of life is"
tokens = tokenizer.encode_plus(text)['input_ids']
tokens = torch.tensor(tokens)[None,].to(device)
with torch.no_grad():
out = model.generate(tokens, do_sample=False, max_length=128, eos_token_id=tokenizer.eos_token_id)[0]
out = tokenizer.decode(out.cpu().numpy().tolist())
print(out)
Aquila2-7B huggingface.co is an AI model on huggingface.co that provides Aquila2-7B's model effect (), which can be used instantly with this BAAI Aquila2-7B model. huggingface.co supports a free trial of the Aquila2-7B model, and also provides paid use of the Aquila2-7B. Support call Aquila2-7B model through api, including Node.js, Python, http.
Aquila2-7B huggingface.co is an online trial and call api platform, which integrates Aquila2-7B's modeling effects, including api services, and provides a free online trial of Aquila2-7B, you can try Aquila2-7B online for free by clicking the link below.
BAAI Aquila2-7B online free url in huggingface.co:
Aquila2-7B is an open source model from GitHub that offers a free installation service, and any user can find Aquila2-7B on GitHub to install. At the same time, huggingface.co provides the effect of Aquila2-7B install, users can directly use Aquila2-7B installed effect in huggingface.co for debugging and trial. It also supports api for free installation.