mirror of
https://github.com/TabbyML/tabby
synced 2024-11-22 08:21:59 +00:00
16 lines
504 B
Markdown
16 lines
504 B
Markdown
|
## v0.12.0 (2024-05-31)
|
||
|
|
||
|
### Features
|
||
|
|
||
|
* Support Gitlab SSO
|
||
|
* Support connect with Self-Hosted Github / Gitlab
|
||
|
* Repository Context is now utilizied in "Code Browser" as well
|
||
|
|
||
|
### Fixed and Improvements
|
||
|
|
||
|
* llama-server from llama.cpp is now distributed as an individual binary, allowing for more flexible configuration
|
||
|
* HTTP API is out of experimental - you can connect tabby to models through HTTP API. Right now following APIs are supported:
|
||
|
- llama.cpp
|
||
|
- ollama
|
||
|
- mistral / codestral
|
||
|
- openai
|