this post was submitted on 25 Jul 2023
6 points (100.0% liked)

Technology

37360 readers
247 users here now

Rumors, happenings, and innovations in the technology sphere. If it's technological news or discussion of technology, it probably belongs here.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

I just got Oobabooga running for the first time with Llama-2, and have Automatic1111, and ComfyUI running for images. I am curious about ML too but I don't know where this start with that one yet.

For the uninitiated, all of these tools are running offline open source (or mostly) models.

top 9 comments
sorted by: hot top controversial new old
[–] [email protected] 1 points 11 months ago

Unfortunately LLaMA 2 is not FOSS. Meta claims it's open source, but it's while the source is available it's definitely not free as in freedom. There are strings attached.

[–] [email protected] 1 points 11 months ago

gpt4all has some decent models that I believe are Free. There is a python CLI/library that works with it, and others, called llm

[–] [email protected] 0 points 11 months ago (1 children)

I'm playing with Stable Diffusion currently. For text I'm still using GPT-4.

[–] [email protected] 0 points 11 months ago (1 children)

I too find it hard to use anything other than GPT-4. It’s still so much better than other options even if the model has felt majorly nerfed compared to earlier releases.

[–] [email protected] 0 points 11 months ago (2 children)

Check out Wizard 30B Uncensored. IMO it’s about as good as NerfedGPT 4… except free and private.

[–] [email protected] 1 points 11 months ago

I just tried it a few hours ago. Indeed, it is quite good. I knew it when a NSFW prompt test on an uncensored model generated a stable diffusion picture of a robot skeleton and a snarky reply. Like, yay we finally have a bight spot with this one.

[–] [email protected] 1 points 11 months ago (1 children)

What hardware does it take to run a 30B?

[–] [email protected] 1 points 11 months ago (1 children)

I'm running it in GPT4All (CPU-based) with 64GB of RAM, and it runs pretty well. I'm not sure what you'd need if you were running it on GPU instead.

[–] [email protected] 2 points 11 months ago

WizardLM 30B at 4 bits with the GGML version on Oobabooga runs almost as fast as Llama2 7B on just the GPU. I set it up with 10 threads on the CPU and ~20 layers on the GPU. That leaves plenty of room for a 4096 context with a batch size of 2048. I can even run a 2GB Stable Diffusion model at the same time with my 3080's 16GBV.

Have you tried any of the larger models? I just ordered 64GB of ram. I also got kobold mostly working. I hope to use it to try Falcon 40. I really want to try a 70B model at 2-4 bit and see how its accuracy is.

load more comments
view more: next ›