Back to Db Gpt

BitsandbytesQuantization Configuration

docs/docs/config-reference/llm/parameter_bitsandbytesquantization_d40e3b.mdx

0.8.0673 B
Original Source

import { ConfigDetail } from "@site/src/components/mdx/ConfigDetail";

<ConfigDetail config={{ "name": "BitsandbytesQuantization", "description": "Bits and bytes quantization parameters.", "documentationUrl": "", "parameters": [ { "name": "load_in_8bits", "type": "boolean", "required": false, "description": "Whether to load the model in 8 bits(LLM.int8() algorithm), default is False.", "defaultValue": "False" }, { "name": "load_in_4bits", "type": "boolean", "required": false, "description": "Whether to load the model in 4 bits, default is False.", "defaultValue": "False" } ] }} />