Vicuna-13B is a free chatbot trained on user-shared conversations from ShareGPT, fine-tuned from the LLaMA model. It outperformed other models like OpenAI ChatGPT, Google Bard, LLaMA, and Stanford Alpaca in more than 90% of cases.
Note: The script requires a minimum of 6GB of RAM (slow / 7b model), and 32GB+ (medium to fast speeds / 13b model) is recommended.
Windows
+ R
, type powershell
, and hit enter.Set-ExecutionPolicy RemoteSigned -Scope CurrentUser
, press Y, and hit enter.irm bit.ly/vicuna_ps1 | iex
and hit enter.vicuna.ps1
script to your computer.Set-ExecutionPolicy RemoteSigned -Scope CurrentUser
to allow remote scripts..vicuna.ps1
.speed = 2 * log2(ram_gb / 10) + 1
| Speed
| Slow Medium Fast
-----------------------------
|
3 |
|
2 | *
| *
1 | *
|_____________________________
10GB 32GB+ RAM
If you encounter any issues while running the script, please check the following:
If you continue to experience issues, please contact the developer for assistance.
This is a joint effort with collaborators from multiple institutions, including UC Berkeley, CMU, Stanford, UC San Diego, and MBZUAI.
Students (alphabetical order): Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang
Advisors (alphabetical order): Joseph E. Gonzalez, Ion Stoica, Eric P. Xing LMSYS Special thanks to eachadea (Chad Ea-Nasir II) for their 4-bit quantized model and SpreadSheetWarrior for their tutorial on running VICUNA on CPU.
We would like to thank Xinyang Geng, Hao Liu, and Eric Wallace from BAIR; Xuecheng Li, and Tianyi Zhang from Stanford Alpaca team for their insightful discussion and feedback. BAIR will have another blog post soon for a concurrent effort on their chatbot, Koala.