/人◕ ‿‿ ◕人\ i just build llama cpp with cmake myself so i can get the latest model updates but i guess the AUR package will do that for you too

/人◕ ‿‿ ◕人\ i just build llama cpp with cmake myself so i can get the latest model updates but i guess the AUR package will do that for you too
/人◕ ‿‿ ◕人\ bleh this model takes too much RAM
/人◕ ‿‿ ◕人\ if i can't make it fix i'll go down to Q4
/人◕ ‿‿ ◕人\ ok 8-bit quantized context makes it barely fit
/人◕ ‿‿ ◕人\ >>>/watch?v=kMGbGGllQoE

whoakun sis*
