# v0.5.0 [Unreleased] ## Notice * llama.cpp backend (CPU, Metal) now requires a redownload of gguf model due to upstream format changes: https://github.com/TabbyML/tabby/pull/645 https://github.com/ggerganov/llama.cpp/pull/3252 ## Features ## Fixes and Improvements * Switch cpu backend to llama.cpp: https://github.com/TabbyML/tabby/pull/638 * add `server.completion_timeout` to control the code completion interface timeout: https://github.com/TabbyML/tabby/pull/637 * Switch cuda backend to llama.cpp: https://github.com/TabbyML/tabby/pull/656 * Switch tokenizer to llama.cpp, so tabby no longer need to download additional tokenizer file: https://github.com/TabbyML/tabby/pull/683 # v0.4.0 ## Features * Supports golang: https://github.com/TabbyML/tabby/issues/553 * Supports ruby: https://github.com/TabbyML/tabby/pull/597 * Supports using local directory for `Repository.git_url`: use `file:///path/to/repo` to specify a local directory. * A new UI design for webserver. ## Fixes and Improvements * Improve snippets retrieval by dedup candidates to existing content + snippets: https://github.com/TabbyML/tabby/pull/582 # v0.3.1 ## Fixes and improvements * Fix GPU OOM issue caused the parallelism: https://github.com/TabbyML/tabby/issues/541, https://github.com/TabbyML/tabby/issues/587 * Fix git safe directory check in docker: https://github.com/TabbyML/tabby/issues/569 # v0.3.0 ## Features ### Retrieval-Augmented Code Completion Enabled by Default The currently supported languages are: * Rust * Python * JavaScript / JSX * TypeScript / TSX A blog series detailing the technical aspects of Retrieval-Augmented Code Completion will be published soon. Stay tuned! ## Fixes and Improvements * Fix [Issue #511](https://github.com/TabbyML/tabby/issues/511) by marking ggml models as optional. * Improve stop words handling by combining RegexSet into Regex for efficiency. # v0.2.2 (10/09/2023) ## Fixes and improvements * Fix a critical issue that might cause request dead locking in ctranslate2 backend (when loading is heavy) # v0.2.1 (10/03/2023) ## Features ### Chat Model & Web Interface We have introduced a new argument, `--chat-model`, which allows you to specify the model for the chat playground located at http://localhost:8080/playground To utilize this feature, use the following command in the terminal: ```bash tabby serve --device metal --model TabbyML/StarCoder-1B --chat-model TabbyML/Mistral-7B ``` ### ModelScope Model Registry Mainland Chinese users have been facing challenges accessing Hugging Face due to various reasons. The Tabby team is actively working to address this issue by mirroring models to a hosting provider in mainland China called modelscope.cn. ```bash # Download from the Modelscope registry TABBY_REGISTRY=modelscope tabby download --model TabbyML/WizardCoder-1B ``` ## Fixes and improvements * Implemented more accurate UTF-8 incremental decoding in the [GitHub pull request](https://github.com/TabbyML/tabby/pull/491). * Fixed the stop words implementation by utilizing RegexSet to isolate the stop word group. * Improved model downloading logic; now Tabby will attempt to fetch the latest model version if there's a remote change, and the local cache key becomes stale. * set default num_replicas_per_device for ctranslate2 backend to increase parallelism.