oneuptime/Llama
dependabot[bot] 80a925d77d
Bump transformers from 4.34.0 to 4.36.0 in /Llama
Bumps [transformers](https://github.com/huggingface/transformers) from 4.34.0 to 4.36.0.
- [Release notes](https://github.com/huggingface/transformers/releases)
- [Commits](https://github.com/huggingface/transformers/compare/v4.34.0...v4.36.0)

---
updated-dependencies:
- dependency-name: transformers
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-12-20 21:14:11 +00:00
..
Models male llama work 2023-10-14 16:36:12 +01:00
app.py enable gpu on llama docker 2023-10-18 12:07:37 +01:00
Dockerfile.tpl fix llama docker file. 2023-10-18 11:01:15 +01:00
Readme.md enable gpu on llama docker 2023-10-18 12:07:37 +01:00
requirements.txt Bump transformers from 4.34.0 to 4.36.0 in /Llama 2023-12-20 21:14:11 +00:00
tsconfig.json fix llama 2023-10-14 17:59:52 +01:00

Llama

Prepare

  • Download models from meta
  • Once the model is downloaded, place them in the Llama/Models folder. Please make sure you also place tokenizer.model and tokenizer_checklist.chk in the same folder.
  • Edit Dockerfile to include the model name in the MODEL_NAME variable.
  • Docker build
docker build -t llama . -f ./Llama/Dockerfile 

Run

For Linux

docker run --gpus all -p 8547:8547 -it -v ./Llama/Models:/app/Models llama 

For MacOS

docker run -p 8547:8547 -it -v ./Llama/Models:/app/Models llama 

Run without a docker conatiner

uvicorn app:app --host 0.0.0.0 --port 8547