experiments genai
1.0.0
List of experiments on Gen AI ecosystem
create -m venv .venv
source .venv/bin/activate
pip3 install -r requirements.txt
Install & run LLM model on local using Ollama. Visit download page (https://ollama.com/
2.1 Llama2 - run ollama run llama2
on terminal. Visit for more info on llama2.
2.2 Codellama - run ollama run codellama
on terminal. Visit for more info on codellama.
Optional, Only required when executing Milvus examples
Ensure Milvus vector store is up and running before executing Milvus dependent examples.
Milvus
Vector store setup. Execute sudo docker compose -f docker-compose-milvus.yml up -d
from project root.File / Folder name | Framework(s) | Optimization / FineTuning | Pipeline |
---|---|---|---|
rag_pdf_1.py | LlamaIndex + Milvus | RAG | Load PDF Dir, Extract data and index in naive way, Embed the index into vector store, User queries |
rag_pdf_2.py | LlamaIndex | RAG | Load PDF Dir, Extract data with Sentence window, Embed the index into local storage, User queries |
rag_pdf_3.py | LlamaIndex + Milvus | RAG | Load PDF Dir, Extract data with Sentence window, Embed the index into vector store, User queries |
rag_pdf_4.py | LlamaIndex + Chroma | RAG | Coming Soon |
rag_pdf_5.py | LlamaIndex + Pinecone | RAG | Coming Soon |
rag_pdf_6.py | LlamaIndex + Qdrant | RAG | Coming Soon |
rag_pdf_7.py | LlamaIndex + Ray + Milvus | RAG | Coming Soon |
rag_pdf_8.py | LlamaIndex + Ray + Milvus | RAG | Coming Soon |
python3 rag_pdf_1.py