You must log in or register to comment.
Never really occurred to me before how huge a 10x savings would be in terms of parameters on consumer hardware.
Like, obviously 10x is a lot, but with the way things are going, it wouldn’t surprise me to see that kind of leap in the next year or two tbh.
Finally. Wrong answers to questions using my phone.
That would actually be insane. Right now, I still need my GPU and about 8-10 gigs of VRAM to run a 7B model tho, so idk how that’s supposed to work on a phone. Still, being able to run a model that’s as good as a 70B model but with the speed and memory usage of a 7B model would be huge.
I have never worked on machine learning, what does the B stand for? Billion? Bytes?
I think it’s how many billion parameters the model has
Thanks!