universeodon.com is part of the decentralized social network powered by Mastodon.
Be one with the #fediverse. Join millions of humans building, creating, and collaborating on Mastodon Social Network. Supports 1000 character posts.

Administered by:

Server stats:

3.5K
active users

Learn more

Andre Louis

This is the Llava:34B LLLM (local large language model) running on my Pro describing a from News. This, to me, is as good info as would provide using , so it goes to show that we can do this on-device and have some really meaningful results. Screenshot attached, contains the description.
Lately, I've taken to using this to describe images instead of using GPT, even if it takes a little longer for results to come back. I consider this to be quite impressive.

@FreakyFwoof Wow, I wish my hardware was anywhere near that good. I'm lucky enough to run the 7 and 13B. And, this isn't even the *best* Llava 34B can do right now! Ollama will be updated so Llava can take in an even better quality screenshot than it already does!

@pixelate It's great. I'd say this is better than GPT 3.5 when it was big. It's only going to get better from here as you say, so if this impresses us, where will we be in 6 months?

@FreakyFwoof Right? Local models describing whole videos, I bet.

@FreakyFwoof Installed Ollama and VOllama by @chikim the other day inspired by you, and I'm impressed too!

@ignasicambra @chikim Good. I told you it was interesting.