Welcome to VerbaFlow, a neural architecture written in Go designed specifically for language modeling tasks. Built on the robust RWKV RNN, this model is optimized for efficient performance on standard CPUs, enabling smooth running of relatively large language models even on consumer hardware.
With the ability to utilize pretrained models on the Pile dataset, VerbaFlow performs comparably to GPT-like Transformer models in predicting the next token, as well as in other tasks such as text summarization, text classification, question answering, and general conversation.
Requirements:
Clone this repo or get the library:
go get -u github.com/nlpodyssey/verbaflow
To start using VerbaFlow, we recommend using the pre-trained model RWKV-4-Pile-1B5-Instruct
, available on the Hugging Face Hub.
This model has been fine-tuned using the Pile dataset and has been specially designed to understand and execute human instructions, as fine-tuned on the xP3 dataset.
The original RWKV-4-Pile-1B5-Instruct-test2-20230209
model, from which this model is derived, can be accessed here.
The library is optimized to run in x86-64 CPUs. If you want to run it on a different architecture, you can use the
GOARCH=amd64
environment variable.
The following commands can be used to build and use VerbaFlow:
go build ./cmd/verbaflow
This command builds the go program and creates an executable named verbaflow
.
./verbaflow -model-dir models/nlpodyssey/RWKV-4-Pile-1B5-Instruct download
This command downloads the model specified (in this case, "nlpodyssey/RWKV-4-Pile-1B5-Instruct" under the "models" directory)
./verbaflow -model-dir models/nlpodyssey/RWKV-4-Pile-1B5-Instruct convert
This command converts the downloaded model to the format used by the program.
./verbaflow -log-level trace -model-dir models/nlpodyssey/RWKV-4-Pile-1B5-Instruct inference --address :50051
This command runs the gRPC inference endpoint on the specified model.
Please make sure to have the necessary dependencies installed before running the above commands.
One of the most interesting features of the LLM is the ability to react based on the prompt.
Run the verbaflow
gRPC endpoint with the command in inference, then run the prompttester
example entering the following prompts:
Prompt:
echo 'nQ: Briefly: The Universe is expanding, its constituent galaxies flying apart like pieces of cosmic shrapnel in the aftermath of the Big Bang. Which section of a newspaper would this article likely appear in?nnA:' | go run ./examples/prompttester --dconfig ./examples/prompttester/config.yaml
Expected output:
Science and Technology
Prompt:
echo 'nQ:Translate the following text from French to English Je suis le père le plus heureux du mondennA:' | go run ./examples/prompttester --dconfig ./examples/prompttester/config.yaml
Expected output:
I am the happiest father in the world.
A list of the main dependencies follows:
Thanks PENG Bo for creating the RWKV RNN and all related resources, including pre-trained models!
"VerbaFlow" combines "verba", which is the Latin word for words, and "flow", which alludes to the characteristics of recurrent neural networks by evoking the idea of a fluent and continuous flow of words, which is made possible by the network's ability to maintain an internal state and "remember" previous words and context when generating new words.