RAG System
Basic Retrieval‑Augmented Generation system built from scratch.
This wasn’t a huge project, but it was a solid learning exercise. I used an embedding model and a chunking strategy to split documents into pages. Each prompt was embedded, then I used cosine similarity to retrieve the most relevant chunks and fed them into a self‑hosted LLM. I did this to build intuition for RAG design decisions. I found the trade‑offs around chunking (overlap, length, semantic vs. fixed) and retrieval tuning especially interesting.