Show newer

@mint @VD15 @waifu Logically IO is a massive bottleneck since it's a bunch of small calculations over a large dataset

@qwerty @Kerosene
Gamergate is older than the kids journalists are attracted to

and now they let anybody into art school just to be safe

@waifu @VD15 LLammaCPP can split the model between your GPU and CPU so you can maybe run 1-2 layers on your GPU and rest on your CPU. With a quantized (huggingface.co/TheBloke/Silico) small model it shouldn't be super slow.

Show older
Game Liberty Mastodon

Mainly gaming/nerd instance for people who value free speech. Everyone is welcome.