5/8/2024
: Update GPT-3.5 and LLama2 inference code and results for Figure 6, which shows the emergent nature of cognitive synergy.3/15/2024
: This paper has been accepted as a main conference paper at NAACL2024!pip install -r requirements.txt
config_template.sh
and run source config_template.sh
to set up the env variables (Note that we are using the Azure API in our experiments)We provide running scripts for each of the three tasks, please check out the comments in the ".sh" scripts for more information:
bash scripts/trivia_creative_writing.sh
bash scripts/codenames_collaborative.sh
bash scripts/logic_grid_puzzle.sh
All prompts can be found in the prompts/
folder.
All datasets can be found in the data/
folder.
Experimental results in the paper for each task can be found in the logs/
folder. gpt4_w_sys_mes
and gpt4_wo_sys_mes
contains results corresponding to Table 2 in our paper. We also include gpt-3.5 and llama2-13b results corresponding to the results in Figure 6, where the hyperparameters, such as whether or not adding system message, follows the best performing choices in the gpt4 experiments.
"test_output_infos"
: contains evaluation metrics for each instance, e.g., # correct answers mentioned."*raw_responses"
: raw responses from each API call."*parsing_flag"
: whether the raw response is successfully parsed. (for Codenames task, this field is seperated into "parsing_success_flag_spymaster" and "parsing_success_flag_guesser")"unwrapped_output"
: parsed output that will be used for computing evaluation metrics. (for Codenames task, this field is seperated into "spymaster_output" and "guesser_output"; there is an additional field named "hint_word" which is parsed from the spymaster's output and inserted into the Guesser's input; the evaluation metric is computed based on the "guesser_output")"task data"
: data for the current task instance, e.g., quetions, answers, target words, etc."usage"
: logging for the number of tokens and cost spended so far.Please cite the paper and star this repo if you find this work interesting/helpful.
@article{wang2023unleashing,
title={Unleashing Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration},
author={Wang, Zhenhailong and Mao, Shaoguang and Wu, Wenshan and Ge, Tao and Wei, Furu and Ji, Heng},
journal={arXiv preprint arXiv:2307.05300},
year={2023}
}
This codebase referenced the structure of the Tree-of-thought official repo. We thank the authors for their open-sourcing efforts.