OpenAi doesn't let you lewd anime girls anymore what's the point of it now
@VD15 No I meant text, though I guess there's an open source version I should be able to use

@waifu There's text-generation-webui which is pretty good. I use it as a backend for Sillytavern which has a much nicer interface. There's a bunch of uncensored models on huggingface that'll let you do whatever. They're great. One of the reasons I started running them locally. Each has like it's own flavour, I guess.

IDK what card you have, but if you have ~6GB of VRAM, you might be able to run a quantized 7B parameter model pretty comfortably. 7B is the ground floor in terms of model size, but it's more than sufficient to get you going. Silicone Maid is a pretty good lewd 7B in my tests. You might find some success with it.

TheBloke also makes quants for every model under the sun, which are like compressed versions of a model. Quants are much easier to load and move around than the raw models and they usually don't get borked in the process.

@VD15 sadly i have a gt 1030 with 2G so i can barely run stuff, but i shall give this a try thank you

@waifu @VD15 LLammaCPP can split the model between your GPU and CPU so you can maybe run 1-2 layers on your GPU and rest on your CPU. With a quantized (huggingface.co/TheBloke/Silico) small model it shouldn't be super slow.

@matrix @waifu Pure CPU inference speed with that model isn't bad on my 16 thread ryzen, actually. That's like a comfortable reading speed for me.

@VD15 @matrix @waifu I think it needs I/O performance between the CPU and RAM more than it needs actual clock cycles. I've palyed around with LLM a while ago, used Mythomax 13B on koboldcpp with about half of the layers offloaded to GPU and it was pretty slow on my shitty AMD FX system with DDR3-1600.
Follow

@mint @VD15 @waifu Logically IO is a massive bottleneck since it's a bunch of small calculations over a large dataset

· · Web · 0 · 0 · 2
Sign in to participate in the conversation
Game Liberty Mastodon

Mainly gaming/nerd instance for people who value free speech. Everyone is welcome.