Back to Gpt4all

Models

gpt4all-bindings/python/docs/gpt4all_desktop/models.md

3.10.04.4 KB
Original Source

Models

GPT4All is optimized to run LLMs in the 3-13B parameter range on consumer-grade hardware.

LLMs are downloaded to your device so you can run them locally and privately. With our backend anyone can interact with LLMs efficiently and securely on their own hardware.

Download Models

!!! note "Download Models"

<div style="text-align: center; margin-top: 20px;">
    <table style="margin-left: auto; margin-right: auto;">
        <tr>
            <td style="text-align: right; padding-right: 10px;">1.</td>
            <td style="text-align: left;">Click `Models` in the menu on the left (below `Chats` and above `LocalDocs`)</td>
            <td></td>
        </tr>
        <tr>
            <td style="text-align: right; padding-right: 10px;">2.</td>
            <td style="text-align: left;">Click `+ Add Model` to navigate to the `Explore Models` page</td>
            <td></td>
        </tr>
        <tr>
            <td style="text-align: right; padding-right: 10px;">3.</td>
            <td style="text-align: left;">Search for models available online</td>
            <td></td>
        </tr>
        <tr>
            <td style="text-align: right; padding-right: 10px;">4.</td>
            <td style="text-align: left;">Hit `Download` to save a model to your device</td>
            <td></td>
        </tr>
        <tr>
            <td style="text-align: right; padding-right: 10px;">5.</td>
            <td style="text-align: left;">Once the model is downloaded you will see it in `Models`.</td>
            <td></td>
        </tr>
    </table>
</div>

Explore Models

GPT4All connects you with LLMs from HuggingFace with a llama.cpp backend so that they will run efficiently on your hardware. Many of these models can be identified by the file type .gguf.

Example Models

Many LLMs are available at various sizes, quantizations, and licenses.

  • LLMs with more parameters tend to be better at coherently responding to instructions

  • LLMs with a smaller quantization (e.g. 4bit instead of 16bit) are much faster and less memory intensive, and tend to have slightly worse performance

  • Licenses vary in their terms for personal and commercial use

Here are a few examples:

ModelFilesizeRAM RequiredParametersQuantizationDeveloperLicenseMD5 Sum (Unique Hash)
Llama 3 Instruct4.66 GB8 GB8 Billionq4_0MetaLlama 3 Licensec87ad09e1e4c8f9c35a5fcef52b6f1c9
Nous Hermes 2 Mistral DPO4.11 GB8 GB7 Billionq4_0Mistral & Nous ResearchApache 2.0Coa5f6b4eabd3992da4d7fb7f020f921eb
Phi-3 Mini Instruct2.18 GB4 GB4 billionq4_0MicrosoftMITf8347badde9bfc2efbe89124d78ddaf5
Mini Orca (Small)1.98 GB4 GB3 billionq4_0MicrosoftCC-BY-NC-SA-4.00e769317b90ac30d6e09486d61fefa26
GPT4All Snoozy7.37 GB16 GB13 billionq4_0Nomic AIGPL40388eb2f8d16bb5d08c96fdfaac6b2c

Search Results

You can click the gear icon in the search bar to sort search results by their # of likes, # of downloads, or date of upload (all from HuggingFace).

Connect Model APIs

You can add your API key for remote model providers.

Note: this does not download a model file to your computer to use securely. Instead, this way of interacting with models has your prompts leave your computer to the API provider and returns the response to your computer.