An open-source clean & customizable RAG UI for chatting with your documents. Built with both end users and developers in mind.
Live Demo | Online Install | User Guide | Developer Guide | Feedback | Contact
This project serves as a functional RAG UI for both end users who want to do QA on their
documents and developers who want to build their own RAG pipeline.
+----------------------------------------------------------------------------+| End users: Those who use apps built with `kotaemon`. || (You use an app like the one in the demo above) || +----------------------------------------------------------------+ || | Developers: Those who built with `kotaemon`. | || | (You have `import kotaemon` somewhere in your project) | || | +----------------------------------------------------+ | || | | Contributors: Those who make `kotaemon` better. | | || | | (You make PR to this repo) | | || | +----------------------------------------------------+ | || +----------------------------------------------------------------+ |+----------------------------------------------------------------------------+
Clean & Minimalistic UI: A user-friendly interface for RAG-based QA.
Support for Various LLMs: Compatible with LLM API providers (OpenAI, AzureOpenAI, Cohere, etc.) and local LLMs (via ollama
and llama-cpp-python
).
Easy Installation: Simple scripts to get you started quickly.
Framework for RAG Pipelines: Tools to build your own RAG-based document QA pipeline.
Customizable UI: See your RAG pipeline in action with the provided UI, built with Gradio .
Gradio Theme: If you use Gradio for development, check out our theme here: kotaemon-gradio-theme.
Host your own document QA (RAG) web-UI: Support multi-user login, organize your files in private/public collections, collaborate and share your favorite chat with others.
Organize your LLM & Embedding models: Support both local LLMs & popular API providers (OpenAI, Azure, Ollama, Groq).
Hybrid RAG pipeline: Sane default RAG pipeline with hybrid (full-text & vector) retriever and re-ranking to ensure best retrieval quality.
Multi-modal QA support: Perform Question Answering on multiple documents with figures and tables support. Support multi-modal document parsing (selectable options on UI).
Advanced citations with document preview: By default the system will provide detailed citations to ensure the correctness of LLM answers. View your citations (incl. relevant score) directly in the in-browser PDF viewer with highlights. Warning when retrieval pipeline return low relevant articles.
Support complex reasoning methods: Use question decomposition to answer your complex/multi-hop question. Support agent-based reasoning with ReAct
, ReWOO
and other agents.
Configurable settings UI: You can adjust most important aspects of retrieval & generation process on the UI (incl. prompts).
Extensible: Being built on Gradio, you are free to customize or add any UI elements as you like. Also, we aim to support multiple strategies for document indexing & retrieval. GraphRAG
indexing pipeline is provided as an example.
If you are not a developer and just want to use the app, please check out our easy-to-follow User Guide. Download the
.zip
file from the latest release to get all the newest features and bug fixes.
Python >= 3.10
Docker: optional, if you install with Docker
Unstructured if you want to process files other than .pdf
, .html
, .mhtml
, and .xlsx
documents. Installation steps differ depending on your operating system. Please visit the link and follow the specific instructions provided there.
We support both lite
& full
version of Docker images. With full
, the extra packages of unstructured
will be installed as well, it can support additional file types (.doc
, .docx
, ...) but the cost is larger docker image size. For most users, the lite
image should work well in most cases.
To use the lite
version.
docker run -e GRADIO_SERVER_NAME=0.0.0.0 -e GRADIO_SERVER_PORT=7860 -p 7860:7860 -it --rm ghcr.io/cinnamon/kotaemon:main-lite
To use the full
version.
docker run -e GRADIO_SERVER_NAME=0.0.0.0 -e GRADIO_SERVER_PORT=7860 -p 7860:7860 -it --rm ghcr.io/cinnamon/kotaemon:main-full
We currently support and test two platforms: linux/amd64
and linux/arm64
(for newer Mac). You can specify the platform by passing --platform
in the docker run
command. For example:
# To run docker with platform linux/arm64docker run -e GRADIO_SERVER_NAME=0.0.0.0 -e GRADIO_SERVER_PORT=7860 -p 7860:7860 -it --rm --platform linux/arm64 ghcr.io/cinnamon/kotaemon:main-lite
Once everything is set up correctly, you can go to http://localhost:7860/
to access the WebUI.
We use GHCR to store docker images, all images can be found here.
Clone and install required packages on a fresh python environment.
# optional (setup env)conda create -n kotaemon python=3.10 conda activate kotaemon# clone this repogit clone https://github.com/Cinnamon/kotaemoncd kotaemon pip install -e "libs/kotaemon[all]"pip install -e "libs/ktem"
Create a .env
file in the root of this project. Use .env.example
as a template
The .env
file is there to serve use cases where users want to pre-config the models before starting up the app (e.g. deploy the app on HF hub). The file will only be used to populate the db once upon the first run, it will no longer be used in consequent runs.
(Optional) To enable in-browser PDF_JS
viewer, download PDF_JS_DIST then extract it to libs/ktem/ktem/assets/prebuilt
Start the web server:
python app.py
The app will be automatically launched in your browser.
Default username and password are both admin
. You can set up additional users directly through the UI.
Check the Resources
tab and LLMs and Embeddings
and ensure that your api_key
value is set correctly from your .env
file. If it is not set, you can set it there.
Note
Official MS GraphRAG indexing only works with OpenAI or Ollama API. We recommend most users to use NanoGraphRAG implementation for straightforward integration with Kotaemon.
Install nano-GraphRAG: pip install nano-graphrag
nano-graphrag
install might introduce version conflicts, see this issue
To quickly fix: pip uninstall hnswlib chroma-hnswlib && pip install chroma-hnswlib
Launch Kotaemon with USE_NANO_GRAPHRAG=true
environment variable.
Set your default LLM & Embedding models in Resources setting and it will be recognized automatically from NanoGraphRAG.
Non-Docker Installation: If you are not using Docker, install GraphRAG with the following command:
pip install graphrag future
Setting Up API KEY: To use the GraphRAG retriever feature, ensure you set the GRAPHRAG_API_KEY
environment variable. You can do this directly in your environment or by adding it to a .env
file.
Using Local Models and Custom Settings: If you want to use GraphRAG with local models (like Ollama
) or customize the default LLM and other configurations, set the USE_CUSTOMIZED_GRAPHRAG_SETTING
environment variable to true. Then, adjust your settings in the settings.yaml.example
file.
See Local model setup.
By default, all application data is stored in the ./ktem_app_data
folder. You can back up or copy this folder to transfer your installation to a new machine.
For advanced users or specific use cases, you can customize these files:
flowsettings.py
.env
flowsettings.py
This file contains the configuration of your application. You can use the example here as the starting point.
# setup your preferred document store (with full-text search capabilities)KH_DOCSTORE=(Elasticsearch | LanceDB | SimpleFileDocumentStore)# setup your preferred vectorstore (for vector-based search)KH_VECTORSTORE=(ChromaDB | LanceDB | InMemory | Qdrant)# Enable / disable multimodal QAKH_REASONINGS_USE_MULTIMODAL=True# Setup your new reasoning pipeline or modify existing one.KH_REASONINGS = ["ktem.reasoning.simple.FullQAPipeline","ktem.reasoning.simple.FullDecomposeQAPipeline","ktem.reasoning.react.ReactAgentPipeline","ktem.reasoning.rewoo.RewooAgentPipeline", ]
.env
This file provides another way to configure your models and credentials.
Alternatively, you can configure the models via the .env
file with the information needed to connect to the LLMs. This file is located in the folder of the application. If you don't see it, you can create one.
Currently, the following providers are supported:
Using ollama
OpenAI compatible server:
Using GGUF
with llama-cpp-python
You can search and download a LLM to be ran locally from the Hugging Face Hub. Currently, these model formats are supported:
Install ollama and start the application.
Pull your model, for example:
ollama pull llama3.1:8b ollama pull nomic-embed-text
Set the model names on web UI and make it as default:
GGUF
You should choose a model whose size is less than your device's memory and should leave about 2 GB. For example, if you have 16 GB of RAM in total, of which 12 GB is available, then you should choose a model that takes up at most 10 GB of RAM. Bigger models tend to give better generation but also take more processing time.
Here are some recommendations and their size in memory:
Qwen1.5-1.8B-Chat-GGUF: around 2 GB
Add a new LlamaCpp model with the provided model name on the web UI.
OpenAI
In the .env
file, set the OPENAI_API_KEY
variable with your OpenAI API key in order
to enable access to OpenAI's models. There are other variables that can be modified,
please feel free to edit them to fit your case. Otherwise, the default parameter should
work for most people.
OPENAI_API_BASE=https://api.openai.com/v1 OPENAI_API_KEY=OPENAI_CHAT_MODEL=gpt-3.5-turbo OPENAI_EMBEDDINGS_MODEL=text-embedding-ada-002
Azure OpenAI
For OpenAI models via Azure platform, you need to provide your Azure endpoint and API key. Your might also need to provide your developments' name for the chat model and the embedding model depending on how you set up Azure development.
AZURE_OPENAI_ENDPOINT= AZURE_OPENAI_API_KEY= OPENAI_API_VERSION=2024-02-15-preview AZURE_OPENAI_CHAT_DEPLOYMENT=gpt-35-turbo AZURE_OPENAI_EMBEDDINGS_DEPLOYMENT=text-embedding-ada-002
Local Models
Check the default pipeline implementation in here. You can make quick adjustment to how the default QA pipeline work.
Add new .py
implementation in libs/ktem/ktem/reasoning/
and later include it in flowssettings
to enable it on the UI.
Check sample implementation in libs/ktem/ktem/index/file/graph
(more instruction WIP).
Since our project is actively being developed, we greatly value your feedback and contributions. Please see our Contributing Guide to get started. Thank you to all our contributors!