diff options
-rw-r--r-- | rag/llm/encoder.py | 12 | ||||
-rw-r--r-- | rag/llm/generator.py | 14 |
2 files changed, 14 insertions, 12 deletions
diff --git a/rag/llm/encoder.py b/rag/llm/encoder.py index a686aaf..94d5559 100644 --- a/rag/llm/encoder.py +++ b/rag/llm/encoder.py @@ -1,13 +1,13 @@ import os -from typing import List +from typing import List, Optional from uuid import uuid4 -import numpy as np import ollama from langchain_core.documents import Document +from loguru import logger as log from qdrant_client.http.models import StrictFloat -from rag.db.embeddings import Point +from rag.db.vector import Point class Encoder: @@ -18,7 +18,8 @@ class Encoder: def __encode(self, prompt: str) -> List[StrictFloat]: return list(ollama.embeddings(model=self.model, prompt=prompt)["embedding"]) - def encode_document(self, chunks: List[Document]) -> np.ndarray: + def encode_document(self, chunks: List[Document]) -> List[Point]: + log.debug("Encoding document...") return [ Point( id=str(uuid4()), @@ -28,6 +29,7 @@ class Encoder: for chunk in chunks ] - def query(self, query: str) -> np.ndarray: + def encode_query(self, query: str) -> List[StrictFloat]: + log.debug(f"Encoding query: {query}") query = self.query_prompt + query return self.__encode(query) diff --git a/rag/llm/generator.py b/rag/llm/generator.py index cbe9474..5f164e7 100644 --- a/rag/llm/generator.py +++ b/rag/llm/generator.py @@ -2,16 +2,14 @@ import os from dataclasses import dataclass import ollama +from loguru import logger as log @dataclass class Prompt: - question: str + query: str context: str - # def context(self) -> str: - # return "\n".join(point.payload["text"] for point in self.points) - class Generator: def __init__(self) -> None: @@ -22,14 +20,16 @@ class Generator: f"You are a {role}.\n" "Answer the following question using the provided context.\n" "If you can't find the answer, do not pretend you know it," - 'but answer "I don\'t know".' - f"Question: {prompt.question.strip()}\n\n" + 'but answer "I don\'t know".\n\n' + f"Question: {prompt.query.strip()}\n\n" "Context:\n" f"{prompt.context.strip()}\n\n" "Answer:\n" ) return metaprompt - def generate(self, role: str, prompt: Prompt) -> str: + def generate(self, prompt: Prompt, role: str) -> str: + log.debug("Generating answer...") metaprompt = self.__metaprompt(role, prompt) + print(f"metaprompt = \n{metaprompt}") return ollama.generate(model=self.model, prompt=metaprompt) |