A fullstack chat app utilizing Llama LLMs
$ make install-huggingface-cliCreate a huggingface token: https://huggingface.co/settings/tokens
, then set the token as env variable on your machine:
$ export HF_TOKEN=<your-token-here>3. Download the Llama-2-7B-Chat-GGML model
$ make download-model$ make chatty-llamaPS! If you're having issues connecting to the backend, try running make chatty-llama-host instead.
In your browser, open http://localhost:80
Enjoy!

