mirror of
https://github.com/OneUptime/oneuptime
synced 2024-11-22 23:30:10 +00:00
9b08d1a9e4
The job function in app.py has been converted to an async function to support asynchronous processing. This change improves the performance and responsiveness of the application by allowing other tasks to run concurrently while the job function is processing the queue. |
||
---|---|---|
.. | ||
Models | ||
app.py | ||
Dockerfile.tpl | ||
Readme.md | ||
requirements.txt | ||
tsconfig.json |
Llama
Prepare
- Download models from meta
- Once the model is downloaded, place them in the
Llama/Models
folder. Please make sure you also place tokenizer.model and tokenizer_checklist.chk in the same folder. - Edit
Dockerfile
to include the model name in theMODEL_NAME
variable. - Docker build
docker build -t llama . -f ./Llama/Dockerfile
Run
For Linux
docker run --gpus all -p 8547:8547 -it -v ./Llama/Models:/app/Models llama
For MacOS
docker run -p 8547:8547 -it -v ./Llama/Models:/app/Models llama
Run without a docker conatiner
uvicorn app:app --host 0.0.0.0 --port 8547