Llamas: update urls
This commit is contained in:
parent
cc01691a33
commit
c26ad07c69
|
@ -102,7 +102,7 @@ VRAM, which meant many home computers could now run 4-bit quantized 7B models!
|
|||
Previously, most enthusiasts would have to rent cloud GPUs to run their "local"
|
||||
llamas. Quantizing into GGUF is a very expensive process, so
|
||||
[TheBloke](https://huggingface.co/TheBloke) on Huggingface emerges the defacto
|
||||
source for pre-quantized llamas.
|
||||
source for [pre-quantized llamas](../quantization).
|
||||
|
||||
Based on LLaMa, the open source
|
||||
[llama.cpp](https://github.com/ggerganov/llama.cpp) becomes the leader of local
|
||||
|
|
Loading…
Reference in a new issue