Kaleido is a multi-task seq2seq model designed to
generate
,
explain
, and output the
relevance
and
valence
of contextualized values, rights, and duties, distilled from GPT-4 generated data.
It is intended to be used for research purposes to a) understand how well large language models can approximate pluralistic human values and b) to make an open, transparent attempt to increase the capabilities of LLMs to model human values.
Out-of-Scope Use
The model is not intended to be used for advice, human-facing applications, or other purposes.
Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g.,
Sheng et al. (2021)
and
Bender et al. (2021)
). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
Additionally, certain groups may be represented better in the model's outputs than others, and the fact that the data is entirely in English and generated by predominantly by English speakers/LLMs trained on English, the model's outputs likely fit perspectives from English-speaking countries better.
The relevance score
should not
be interpreted as an importance score, but, due to the composition of the training data, corresponds more closely with "is this value likely to have been generated for this situation by GPT-4?"
Recommendations
We recommend that this model not be used for any high-impact or human-facing purposes as its biases and limitations need to be further explored.
We intend this to be a research artifact to advance AI's ability to model and interact with pluralistic human values, rights, and duties.
Training Details
Training Data
The model is trained on the
mixture
training split of
ValuePrism
.
Training Procedure
Training hyperparameters
The following hyperparameters were used during training:
learning_rate: 3e-05
train_batch_size: 32
eval_batch_size: 8
seed: 42
distributed_type: multi-GPU
num_devices: 2
total_train_batch_size: 64
total_eval_batch_size: 16
optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
lr_scheduler_type: linear
num_epochs: 4.0
Framework versions
Transformers 4.22.0.dev0
Pytorch 1.12.1+cu113
Datasets 2.4.0
Tokenizers 0.12.1
Testing Data, Factors & Metrics
Testing Data
The model is tested on the four subtasks in
ValuePrism
.
Metrics
Accuracy is used for relevance and valence, as they are classification tasks, and perplexity is used for generation and explanation, as they are free text generation tasks.
Results
Model
Relevance Acc ↑
Valence Acc ↑
Generative Perp ↓
Explanation Perp ↓
kaleido-xxl
(11B)
89.1
81.9
2.22
2.99
kaleido-xl
(3B)
88.4
80.8
2.23
3.14
kaleido-large
(770M)
87.2
79.2
2.34
3.52
kaleido-base
(220M)
83.5
74.5
2.53
4.23
kaleido-small
(60M)
66.0
59.7
2.86
5.70
Model Architecture and Objective
Kaleido is an encoder-decoder T5-based model trained using negative log-likelihood.
Citation
BibTeX:
@misc{sorensen2023value,
title={Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties},
author={Taylor Sorensen and Liwei Jiang and Jena Hwang and Sydney Levine and Valentina Pyatkin and Peter West and Nouha Dziri and Ximing Lu and Kavel Rao and Chandra Bhagavatula and Maarten Sap and John Tasioulas and Yejin Choi},
year={2023},
eprint={2309.00779},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Model Card Contact
Contact Taylor Sorensen (
tsor13@cs.washington.edu
) for any questions about this model.
How to Get Started with the Model
Use the code below to get started with the model.
Load the model:
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
model_name = 'allenai/kaleido-xl'
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSeq2SeqLM.from_pretrained(model_name)
Each task ('generate', 'relevance', 'valence', 'explanation') has its own template that can be accessed from the model config.
task = 'generate'# can be 'generate', 'relevance', 'valence', or 'explanation'
model.config.task_specific_params[task]['template']
Output:
'[Generate]:\tAction: ACTION'
The generate template requires
ACTION
, while the other three templates require
ACTION
,
VRD
(
'Value', 'Right', or 'Duty'
) and
TEXT
.
Replace the arguments with the text and generate with the model.
kaleido-xl huggingface.co is an AI model on huggingface.co that provides kaleido-xl's model effect (), which can be used instantly with this allenai kaleido-xl model. huggingface.co supports a free trial of the kaleido-xl model, and also provides paid use of the kaleido-xl. Support call kaleido-xl model through api, including Node.js, Python, http.
kaleido-xl huggingface.co is an online trial and call api platform, which integrates kaleido-xl's modeling effects, including api services, and provides a free online trial of kaleido-xl, you can try kaleido-xl online for free by clicking the link below.
allenai kaleido-xl online free url in huggingface.co:
kaleido-xl is an open source model from GitHub that offers a free installation service, and any user can find kaleido-xl on GitHub to install. At the same time, huggingface.co provides the effect of kaleido-xl install, users can directly use kaleido-xl installed effect in huggingface.co for debugging and trial. It also supports api for free installation.