mirror of
https://github.com/QuivrHQ/quivr.git
synced 2024-12-15 09:32:22 +03:00
481f24f5be
# Description - Introduce `LLMEndpoint` class: wrapper around a `BaseChatLLM` to load OpenAI compatible models - Add `brain.search(...)` function to retrieve - Test with test-coverage: - Added ability to load brain `from langchain.Document` - Configured mypy and poetry lock in `.pre-commit.yaml` # Test coverage ![image](https://github.com/QuivrHQ/quivr/assets/14312141/629ede66-146e-400f-b40b-8c22a9258a47) --------- Co-authored-by: aminediro <aminedirhoussi@gmail.com>
97 lines
2.6 KiB
Python
97 lines
2.6 KiB
Python
from datetime import datetime
|
|
from typing import Any
|
|
from uuid import UUID
|
|
|
|
from langchain_core.documents import Document
|
|
from langchain_core.pydantic_v1 import BaseModel as BaseModelV1
|
|
from langchain_core.pydantic_v1 import Field as FieldV1
|
|
from pydantic import BaseModel
|
|
from typing_extensions import TypedDict
|
|
|
|
|
|
class cited_answer(BaseModelV1):
|
|
"""Answer the user question based only on the given sources, and cite the sources used."""
|
|
|
|
answer: str = FieldV1(
|
|
...,
|
|
description="The answer to the user question, which is based only on the given sources.",
|
|
)
|
|
thoughts: str = FieldV1(
|
|
...,
|
|
description="""Description of the thought process, based only on the given sources.
|
|
Cite the text as much as possible and give the document name it appears in. In the format : 'Doc_name states : cited_text'. Be the most
|
|
procedural as possible. Write all the steps needed to find the answer until you find it.""",
|
|
)
|
|
citations: list[int] = FieldV1(
|
|
...,
|
|
description="The integer IDs of the SPECIFIC sources which justify the answer.",
|
|
)
|
|
|
|
followup_questions: list[str] = FieldV1(
|
|
...,
|
|
description="Generate up to 3 follow-up questions that could be asked based on the answer given or context provided.",
|
|
)
|
|
|
|
|
|
class GetChatHistoryOutput(BaseModel):
|
|
chat_id: UUID
|
|
message_id: UUID
|
|
user_message: str
|
|
message_time: datetime
|
|
assistant: str | None = None
|
|
prompt_title: str | None = None
|
|
brain_name: str | None = None
|
|
brain_id: UUID | None = None # string because UUID is not JSON serializable
|
|
metadata: dict | None = None
|
|
thumbs: bool | None = None
|
|
|
|
|
|
class Source(BaseModel):
|
|
name: str
|
|
source_url: str
|
|
type: str
|
|
original_file_name: str
|
|
citation: str
|
|
|
|
|
|
class RawRAGChunkResponse(TypedDict):
|
|
answer: dict[str, Any]
|
|
docs: dict[str, Any]
|
|
|
|
|
|
class RawRAGResponse(TypedDict):
|
|
answer: dict[str, Any]
|
|
docs: dict[str, Any]
|
|
|
|
|
|
class RAGResponseMetadata(BaseModel):
|
|
citations: list[int] | None = None
|
|
thoughts: str | list[str] | None = None
|
|
followup_questions: list[str] | None = None
|
|
sources: list[Any] | None = None
|
|
|
|
|
|
class ParsedRAGResponse(BaseModel):
|
|
answer: str
|
|
metadata: RAGResponseMetadata | None = None
|
|
|
|
|
|
class ParsedRAGChunkResponse(BaseModel):
|
|
answer: str
|
|
metadata: RAGResponseMetadata
|
|
last_chunk: bool = False
|
|
|
|
|
|
class QuivrKnowledge(BaseModel):
|
|
id: UUID
|
|
brain_id: UUID
|
|
file_name: str | None = None
|
|
url: str | None = None
|
|
extension: str = "txt"
|
|
|
|
|
|
# NOTE: for compatibility issues with langchain <-> PydanticV1
|
|
class SearchResult(BaseModelV1):
|
|
chunk: Document
|
|
score: float
|