Back to Docsgpt

Architecture

docs/content/Guides/Architecture.mdx

0.17.17.6 KB
Original Source

Introduction

DocsGPT is designed as a modular and scalable application for knowledge based GenAI system. This document outlines the high-level architecture of DocsGPT, highlighting its key components.

High-Level Architecture

This diagram provides a bird's-eye view of the DocsGPT architecture, illustrating the main components and their interactions.

mermaid
flowchart LR
    User["User"] --> Frontend["Frontend (React/Vite)"]
    Frontend --> Backend["Backend API (Flask)"]
    Backend --> LLM["LLM Integration Layer"] & VectorStore["Vector Stores"] & TaskQueue["Task Queue (Celery)"] & Databases["Databases (Postgres, Redis)"]
    LLM -- Cloud APIs / Local Engines --> InferenceEngine["Inference Engine"]
    VectorStore -- Document Embeddings --> Indexes[("Indexes")]
    TaskQueue -- Asynchronous Tasks --> DocumentIngestion["Document Ingestion"]

    style Frontend fill:#AA00FF,color:#FFFFFF
    style Backend fill:#AA00FF,color:#FFFFFF
    style LLM fill:#AA00FF,color:#FFFFFF
    style TaskQueue fill:#AA00FF,color:#FFFFFF,stroke:#AA00FF
    style DocumentIngestion fill:#AA00FF,color:#FFFFFF,stroke:none

Component Descriptions

1. Frontend (React/Vite)

  • Technology: Built using React and Vite.
  • Responsibility: This is the user interface of DocsGPT, providing users with an UI to ask questions and receive answers, configure prompts, tools and other settings. It handles user input, displays conversation history, shows sources, and manages settings.
  • Key Features:
    • Clean and responsive UI.
    • Simple static client-side rendering.
    • Manages conversation state and settings.
    • Communicates with the Backend API for data retrieval and processing.

2. Backend API (Flask)

  • Technology: Implemented using Flask (Python).
  • Responsibility: The Backend API serves as the core logic and orchestration layer of DocsGPT. It receives requests from the Frontend, Extensions or API clients, processes them, and coordinates interactions between different components.
  • Key Features:
    • API endpoints for handling user queries, document uploads, and settings configurations.
    • Manages the overall application flow and logic.
    • Integrates with the LLM Integration Layer, Vector Stores, Task Queue, Tools, Agents and Databases.
    • Provides Swagger documentation for API endpoints.

3. LLM Integration Layer (Part of backend)

  • Technology: Supports multiple LLM APIs and local engines.
  • Responsibility: This layer provides an abstraction for interacting with Large Language Models (LLMs).
  • Key Features:
    • Supports LLMs from OpenAI, Google, Anthropic, Groq, HuggingFace Inference API, Azure OpenAI, also compatible with local models like Ollama, LLaMa.cpp, Text Generation Inference (TGI), SGLang, vLLM, Aphrodite, FriendliAI, and LMDeploy.
    • Manages API key handling and request formatting and Tool formatting.
    • Offers caching mechanisms to improve response times and reduce API usage.
    • Handles streaming responses for a more interactive user experience.

4. Vector Stores (Part of backend)

  • Technology: Supports multiple vector databases.
  • Responsibility: Vector Stores are used to store and retrieve vector embeddings of document chunks. This enables semantic search and retrieval of relevant document snippets in response to user queries.
  • Key Features:
    • Supports vector databases including FAISS, Elasticsearch, Qdrant, Milvus, MongoDB Atlas Vector Search, and pgvector.
    • Provides storage and indexing of high-dimensional vector embeddings.
    • Enables editing and updating of vector indexes including specific chunks.

5. Parser Integration Layer (Part of backend)

  • Technology: Supports multiple formats for file processing and remote source uploading.
  • Responsibility: Parser Integration Layer handles uploading, parsing, chunking, embedding, and indexing documents.
  • Key Features:
    • Supports various document formats (PDF, DOCX, TXT, etc.) and remote sources (web URLs, sitemaps).
    • Handles document parsing, text chunking, and embedding generation.
    • Utilizes Celery for asynchronous processing, ensuring efficient handling of large documents.

6. Task Queue (Celery)

  • Technology: Celery with Redis as broker and backend.
  • Responsibility: Celery handles asynchronous task processing, for long-running operations such as document ingestion and indexing. This ensures that the main application remains responsive and efficient.
  • Key Features:
    • Manages background tasks for document processing and indexing.
    • Improves application responsiveness by offloading heavy tasks.
    • Enhances scalability and reliability through distributed task processing.

7. Databases (Postgres, Redis)

  • Technology: PostgreSQL and Redis.
  • Responsibility: Databases are used for persistent data storage and caching. PostgreSQL stores structured user data such as conversations, agents, prompts, sources, attachments, workflows, logs, token usage, user settings, and API keys. Redis is used as a cache and as the message broker/result backend for Celery.
  • Note: MongoDB is no longer used for user data. It remains an optional backend for the vector store (VECTOR_STORE=mongodb, i.e. Mongo Atlas Vector Search) and as the source database for the one-shot scripts/db/backfill.py migration from legacy installs.

Request Flow Diagram

This diagram illustrates the sequence of steps involved when a user submits a question to DocsGPT.

mermaid
sequenceDiagram
    participant User
    participant Frontend
    participant BackendAPI
    participant LLMIntegrationLayer
    participant VectorStores
    participant InferenceEngine

    User->>Frontend: User asks a question
    Frontend->>BackendAPI: API Request (Question)
    BackendAPI->>VectorStores: Fetch relevant document chunks (Similarity Search)
    VectorStores-->>BackendAPI: Return document chunks
    BackendAPI->>LLMIntegrationLayer: Send question and document chunks
    LLMIntegrationLayer->>InferenceEngine: LLM API Request (Prompt + Context)
    InferenceEngine-->>LLMIntegrationLayer: LLM API Response (Answer)
    LLMIntegrationLayer-->>BackendAPI: Return Answer
    BackendAPI->>Frontend: API Response (Answer)
    Frontend->>User: Display Answer

    Note over Frontend,BackendAPI: Data flow is simplified for clarity

Deployment Architecture

DocsGPT is designed to be deployed using Docker and Kubernetes, here is a quick overview of a simple k8s deployment.

mermaid
graph LR
    subgraph Kubernetes Cluster
        subgraph Nodes
            subgraph Node 1
                FrontendPod[Frontend Pod]
                BackendAPIPod[Backend API Pod]
            end
            subgraph Node 2
                CeleryWorkerPod[Celery Worker Pod]
                RedisPod[Redis Pod]
            end
            subgraph Node 3
                PostgresPod[Postgres Pod]
                VectorStorePod[Vector Store Pod]
            end
        end
        LoadBalancer[Load Balancer] --> docsgpt-frontend-service[docsgpt-frontend-service]
        LoadBalancer --> docsgpt-api-service[docsgpt-api-service]
        docsgpt-frontend-service --> FrontendPod
        docsgpt-api-service --> BackendAPIPod
        BackendAPIPod --> CeleryWorkerPod
        BackendAPIPod --> RedisPod
        BackendAPIPod --> PostgresPod
        BackendAPIPod --> VectorStorePod
        CeleryWorkerPod --> RedisPod
        BackendAPIPod --> InferenceEngine[(Inference Engine)]
        VectorStorePod --> Indexes[(Indexes)]
        PostgresPod --> Data[(Data)]
        RedisPod --> Cache[(Cache)]
    end
    User[User] --> LoadBalancer