summaryrefslogtreecommitdiff
path: root/content/projects/retrieval-augmented-generation.md
diff options
context:
space:
mode:
Diffstat (limited to 'content/projects/retrieval-augmented-generation.md')
-rw-r--r--content/projects/retrieval-augmented-generation.md26
1 files changed, 26 insertions, 0 deletions
diff --git a/content/projects/retrieval-augmented-generation.md b/content/projects/retrieval-augmented-generation.md
new file mode 100644
index 0000000..17d1358
--- /dev/null
+++ b/content/projects/retrieval-augmented-generation.md
@@ -0,0 +1,26 @@
+---
+title: "Retrieval Augmented Generation"
+date: 2024-04-26 00:36
+tags:
+ [
+ "deep learning",
+ "retrieval augmented generation",
+ "vector database",
+ "ollama",
+ "llm",
+ ]
+draft: false
+---
+
+I implemented a retrieval augmented generation (RAG)
+[program](https://github.com/aktersnurra/rag) for fun with the goal of being able to
+search my personal library. My focus was to make this run locally with only open
+source models. This was achieved with `ollama` and `sentence-transformers` for
+downloading and running these models locally. However, the project was expanded to
+integrate with cohere and its rerank and command-r+ models, since I was curious about
+the command-r+ performance. These models can be downloaded and run locally, but it took
+ages for my computer to generate any output, since the command-r+ is huge.
+
+
+Here is a [presentation](/rag.html) that gives a brief overview of what a RAG system
+is, and how it can be improved with reranking.