this post was submitted on 11 Jun 2024
2 points (100.0% liked)

AI

4109 readers
1 users here now

Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen.

founded 3 years ago
top 7 comments
sorted by: hot top controversial new old
[–] kakes@sh.itjust.works 1 points 4 months ago

Never really occurred to me before how huge a 10x savings would be in terms of parameters on consumer hardware.

Like, obviously 10x is a lot, but with the way things are going, it wouldn't surprise me to see that kind of leap in the next year or two tbh.

[–] Fisch@discuss.tchncs.de 1 points 4 months ago (2 children)

That would actually be insane. Right now, I still need my GPU and about 8-10 gigs of VRAM to run a 7B model tho, so idk how that's supposed to work on a phone. Still, being able to run a model that's as good as a 70B model but with the speed and memory usage of a 7B model would be huge.

[–] Smorty@lemmy.blahaj.zone 2 points 1 week ago* (last edited 1 week ago) (1 children)

I'm even more excited for running 8B models at the speed of 1B! Laughably fast ok-quality generations in JSON format would be crazy useful.

Also yeah, that 7B on mobile was not the best example. Again, probably 1B to 3B is the sweetspot for mobile (I'm running Qwen2.5 0.5B on my phone and it works tel real for simple JSON)

EDIT: And imagine the context lengths we would be ablentonrun on our GPUs at home! What a time to be alive.

[–] Fisch@discuss.tchncs.de 2 points 1 week ago (1 children)

Being able to run 7B quality models on your phone would be wild. It would also make it possible to run those models on my server (which is just a mini pc), so I could connect it to my Home Assistant voice assistant, which would be really cool.

[–] Smorty@lemmy.blahaj.zone 1 points 1 week ago (1 children)

Something similar to this already kinda exists on HF with the 1.58 bit quantisation which seem to get very similar performance to the original Llama 3 8B model. That's essentially a two bit quanitsation with reasonable performance!

[–] Fisch@discuss.tchncs.de 2 points 1 week ago

That's really interesting, gonna try out how well it runs

[–] Boomkop3@reddthat.com 2 points 1 week ago

Slowly, is how