.changes/v0.11.0.md
--webserver flag is now enabled by default in tabby serve. To turn off the webserver and only use OSS features, use the --no-webserver flag./v1beta/chat/completions endpoint has been moved to /v1/chat/completions, while the old endpoint is still available for backward compatibility.Ask Tabby feature in the source code browser to provide in-context help from AI.q8_0.v2.gguf to model.gguf in MODEL_SPEC.md.--chat-device flag to specify the device used to run the chat model.tabby download command.git_url field.disable_client_side_telemetry when --webserver is not used.--webserver, with job history logging available in the web interface.~/.tabby/repositories exists for tabby scheduler jobs: https://github.com/TabbyML/tabby/pull/1375tabby-cpu to docker distribution.tabby scheduler --now is required to ensure that Code Browser functions properly.Source Code Browser, featuring visualization of code snippets utilized for code completion in RAG.--webserver flag is disabled.TABBY_MODEL_CACHE_ROOT environment variable to individually override the model cache directory./v1beta/chat/completions API endpoint is now compatible with OpenAI's chat completion API.--webserver flag is a new addition to tabby serve that enables secure access to the tabby server. When this flag is on, IDE extensions will need to provide an authorization token to access the instance.
--webserver flag./metrics endpoint for prometheus metrics collection.--model optional, so users can create a chat only instance.--parallelism to control the throughput and VRAM usage: https://github.com/TabbyML/tabby/pull/727~/.tabby/index needs to be manually removed before any further runs of tabby scheduler.TABBY_REGISTRY is replaced with TABBY_DOWNLOAD_HOST for the github based registry implementation.server.completion_timeout to control the code completion interface timeout: https://github.com/TabbyML/tabby/pull/637Repository.git_url: use file:///path/to/repo to specify a local directory.The currently supported languages are:
A blog series detailing the technical aspects of Retrieval-Augmented Code Completion will be published soon. Stay tuned!
We have introduced a new argument, --chat-model, which allows you to specify the model for the chat playground located at http://localhost:8080/playground
To utilize this feature, use the following command in the terminal:
tabby serve --device metal --model TabbyML/StarCoder-1B --chat-model TabbyML/Mistral-7B
Mainland Chinese users have been facing challenges accessing Hugging Face due to various reasons. The Tabby team is actively working to address this issue by mirroring models to a hosting provider in mainland China called modelscope.cn.
## Download from the Modelscope registry
TABBY_REGISTRY=modelscope tabby download --model TabbyML/WizardCoder-1B
No releases yet, this file will be updated when generating your first release.