Multimodal RAG System
Chat with your PDFs — text, images, tables, and formulas all retrieved.
Users upload a PDF and the system extracts text, images, tables, and formulas. Text is chunked using semantic chunking, stored in PostgreSQL, and embedded with a Sentence Transformers model into pgvector. A local LLM (via Ollama, for full privacy) answers questions grounded in the document — returning relevant chunks, reference images, and metadata alongside each response.