mirror of
https://github.com/TabbyML/tabby
synced 2024-11-25 14:31:08 +00:00
d32dfd6ac0
* fix(ui): display default user avatar in chat * update * [autofix.ci] apply automated fixes * update --------- Co-authored-by: autofix-ci[bot] <114827586+autofix-ci[bot]@users.noreply.github.com> |
||
---|---|---|
.changes | ||
.github | ||
ci | ||
clients | ||
crates | ||
docker | ||
ee | ||
experimental | ||
python | ||
rules | ||
website | ||
.changie.yaml | ||
.dockerignore | ||
.env | ||
.gitattributes | ||
.gitignore | ||
.gitmodules | ||
.rustfmt.toml | ||
Cargo.lock | ||
Cargo.toml | ||
CHANGELOG.md | ||
CODE_OF_CONDUCT.md | ||
codecov.yml | ||
CONTRIBUTING.md | ||
LICENSE | ||
Makefile | ||
MODEL_SPEC.md | ||
package.json | ||
pnpm-lock.yaml | ||
pnpm-workspace.yaml | ||
README.md | ||
sgconfig.yml | ||
turbo.json |
Tabby is a self-hosted AI coding assistant, offering an open-source and on-premises alternative to GitHub Copilot. It boasts several key features:
- Self-contained, with no need for a DBMS or cloud service.
- OpenAPI interface, easy to integrate with existing infrastructure (e.g Cloud IDE).
- Supports consumer-grade GPUs.
🔥 What's New
- 04/22/2024 v0.10.0 released, featuring the latest Reports tab with team-wise analytics for Tabby usage.
- 04/19/2024 📣 Tabby now incorporates locally relevant snippets(declarations from local LSP, and recently modified code) for code completion!
- 04/17/2024 CodeGemma and CodeQwen model series have now been added to the official registry!
Archived
- 03/20/2024 v0.9 released, highlighting a full feature admin UI.
- 12/23/2023 Seamlessly deploy Tabby on any cloud with SkyServe 🛫 from SkyPilot.
- 12/15/2023 v0.7.0 released with team management and secured access!
- 10/15/2023 RAG-based code completion is enabled by detail in v0.3.0🎉! Check out the blogpost explaining how Tabby utilizes repo-level context to get even smarter!
- 11/27/2023 v0.6.0 released!
- 11/09/2023 v0.5.5 released! With a redesign of UI + performance improvement.
- 10/24/2023 ⛳️ Major updates for Tabby IDE plugins across VSCode/Vim/IntelliJ!
- 10/04/2023 Check out the model directory for the latest models supported by Tabby.
- 09/18/2023 Apple's M1/M2 Metal inference support has landed in v0.1.1!
- 08/31/2023 Tabby's first stable release v0.0.1 🥳.
- 08/28/2023 Experimental support for the CodeLlama 7B.
- 08/24/2023 Tabby is now on JetBrains Marketplace!
👋 Getting Started
You can find our documentation here.
Run Tabby in 1 Minute
The easiest way to start a Tabby server is by using the following Docker command:
docker run -it \
--gpus all -p 8080:8080 -v $HOME/.tabby:/data \
tabbyml/tabby \
serve --model TabbyML/StarCoder-1B --device cuda
For additional options (e.g inference type, parallelism), please refer to the documentation page.
🤝 Contributing
Full guide at CONTRIBUTING.md;
Get the Code
git clone --recurse-submodules https://github.com/TabbyML/tabby
cd tabby
If you have already cloned the repository, you could run the git submodule update --recursive --init
command to fetch all submodules.
Build
-
Set up the Rust environment by following this tutorial.
-
Install the required dependencies:
# For MacOS
brew install protobuf
# For Ubuntu / Debian
apt-get install protobuf-compiler libopenblas-dev
- Install useful tools:
# For Ubuntu
sudo apt install make sqlite3 graphviz
- Now, you can build Tabby by running the command
cargo build
.
Start Hacking!
... and don't forget to submit a Pull Request
🌍 Community
- 🎤 Twitter / X - engage with TabbyML for all things possible
- 📚 LinkedIn - follow for the latest from the community
- 💌 Newsletter - subscribe to unlock Tabby insights and secrets