Llamas: update urls
This commit is contained in:
parent
cc01691a33
commit
c26ad07c69
|
@ -102,7 +102,7 @@ VRAM, which meant many home computers could now run 4-bit quantized 7B models!
|
||||||
Previously, most enthusiasts would have to rent cloud GPUs to run their "local"
|
Previously, most enthusiasts would have to rent cloud GPUs to run their "local"
|
||||||
llamas. Quantizing into GGUF is a very expensive process, so
|
llamas. Quantizing into GGUF is a very expensive process, so
|
||||||
[TheBloke](https://huggingface.co/TheBloke) on Huggingface emerges the defacto
|
[TheBloke](https://huggingface.co/TheBloke) on Huggingface emerges the defacto
|
||||||
source for pre-quantized llamas.
|
source for [pre-quantized llamas](../quantization).
|
||||||
|
|
||||||
Based on LLaMa, the open source
|
Based on LLaMa, the open source
|
||||||
[llama.cpp](https://github.com/ggerganov/llama.cpp) becomes the leader of local
|
[llama.cpp](https://github.com/ggerganov/llama.cpp) becomes the leader of local
|
||||||
|
|
Loading…
Reference in a new issue