thickertoofan

joined 1 week ago
MODERATOR OF
 

something like docker run xyz_org/xyz_model

[–] [email protected] 1 points 12 hours ago

I think the bigger bottleneck is SLAM, running that is intensive, it wont directly run on video, and SLAM is tough i guess, reading the repo doesn't give any clues of it being able to run on CPU inference.

[–] [email protected] 2 points 1 day ago

There is a repo they released.

[–] [email protected] 1 points 1 day ago

It will, they have released a repo with code.

11
Microsoft KBLAM (www.microsoft.com)
[–] [email protected] 1 points 2 days ago

I mean I didn't see any alarming need of a Google doc alternative, so I might actually be under a rock

[–] [email protected] 1 points 3 days ago (2 children)

and why so?

[–] [email protected] 2 points 3 days ago

I checked mostly all of em out from the list, but 1b models are generally unusable for RAG.

[–] [email protected] 3 points 5 days ago (2 children)

i use pageassist with Ollama

 

I don't care a lot about mathematical tasks, but code intellingence is a minor preference but the most anticipated one is overall comprehension, intelligence. (For RAG and large context handling) But anyways any benchmark with a wide variety of models is something I am searching for, + updated.

[–] [email protected] 6 points 1 week ago

Same. Welcome here

[–] [email protected] 6 points 1 week ago

Wow, reddit sucks.

[–] [email protected] 1 points 1 week ago

We can use the same test name as proposed by a user in the original post's comment: Odd-straw-in-the-haystack :)

 

I tested this (reddit link btw) for Gemma 3 1B parameter and the 3B parameter model. 1B failed, (not surprising) but 3B passed which is genuinely surprising. I added a random paragraph about Napoleon Bonaparte (just a random character) and added "My password is = xxx" in between the paragraph. Gemma 1B couldn't even spot it, but Gemma 3B did it without asking, but there's a catch, Gemma 3 associated the password statement to be a historical fact related to Napoleon lol. Anyways, passing it is a genuinely nice achievement for a 3B model I guess. And it was a single paragraph, moderately large for the test. I accidentally wiped the chat otherwise i would have attached the exact prompt here. Tested locally using Ollama and PageAssist UI. My setup: GPU poor category, CPU inference with 16 Gigs of RAM.

19
submitted 1 week ago* (last edited 1 week ago) by [email protected] to c/[email protected]
 

Join if you want to have some geek discussions about it, or ask for help/ provide help.

[email protected]

 

I'm new to lemmy, I noticed that there's no community related to flask that i could find, so i created one. Would love to have moderators and people who can make the experience better for this community.

view more: next ›