this post was submitted on 30 May 2024
1111 points (100.0% liked)

Funny

8765 readers
2262 users here now

General rules:

Exceptions may be made at the discretion of the mods.

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 208 points 10 months ago (5 children)

I'm curious if it could solve the traffic light and crosswalk ones, I would try but I'm out of free image uploads from asking it to explain memes to test its cultural knowledge.

[–] [email protected] 121 points 10 months ago (3 children)

Wow, that's actually quite impressive.

[–] [email protected] 39 points 10 months ago (1 children)

I'm sure eventually someone will make a bot called something like ai-explains-the-joke that does this automatically.

[–] [email protected] 13 points 10 months ago (1 children)

Expl-AI-n Bot will break down whatever joke you feed it.

[–] [email protected] 3 points 10 months ago (2 children)

Expl-AI-n itself is a pun. With the letters AI in the word explain capitalized, readers can infer that artificial intelligence is being used to explain jokes.

load more comments (1 replies)
[–] [email protected] 18 points 10 months ago (1 children)

I wonder how much was scraped from knowyourmeme.com

[–] [email protected] 33 points 10 months ago (1 children)

I mean it still parsed the specific text in the meme and formulated a coherent explanation of this specific meme, not just the meme format

[–] [email protected] 5 points 10 months ago (2 children)

Or it matched the text with an existing explanation upon which it was indexed.

[–] [email protected] 16 points 10 months ago (1 children)

Lmao you think it found a specific explanation for this specific variation of this meme?

[–] [email protected] 4 points 10 months ago

For each phrase, yes.

[–] [email protected] 6 points 10 months ago (1 children)
[–] [email protected] 4 points 10 months ago (2 children)

That's literally how they work

[–] [email protected] 7 points 10 months ago (1 children)

Man the models can't store verbatim its training data, the amount of data is turned into a model that is hundreds or thousands of times smaller than the original source data. If it was capable of simply recovering everything that it was trained on this would be some magical compression algorithm and that by itself would be extremely impressive.

[–] [email protected] 3 points 10 months ago (1 children)

Congratulations on discovering compression

[–] [email protected] 5 points 10 months ago* (last edited 10 months ago) (2 children)

Oh ok, you want to claim this is compressing the entirety of the internet in a model that isn't even 1 terabyte of data and be unimpressed that is something.

But it isn't compression. It is a mathematical fact that neural networks are universal function approximators, this is undisputed, and analytic functions are continuous so to be an analytical function approximator it must be able to fill in the gaps between discrete data points by itself, which necessarily means spiting out data outside of the input distribution, data it has not seen.

[–] [email protected] 3 points 10 months ago (1 children)

TBF, compression is related to ML. Hence, the Hutter Prize. Thinking of LLMs as lossy compression algorithms is a decent analogy.

load more comments (1 replies)
[–] [email protected] 2 points 10 months ago (1 children)

Not sure why you feel the need to put words in my mouth. It wasn't trained on "the entirety of the Internet," but rather less than a terabyte of it. So yeah, that would probably take up less than a terabyte.

[–] [email protected] 7 points 10 months ago

Then why did I just make this meme up right now and chat gpt can explain it?

https://i.ibb.co/NYHRnTY/Screenshot-20240531-072008-Chat-GPT.jpg

Arguing over this is just dumb, you can yourself take any picture you want at this very moment or come up with a brand new meme template on the spot and upload it to ChatGPT to see you are wrong, it is free btw.

[–] [email protected] 4 points 10 months ago (1 children)

They do not store anything verbatim; They instead store the directions in which various words and related concepts relate to one another in some gigantic multidimensional space.

I highly suggest you go learn what they actually do before you continue talking out of your ass about them

[–] [email protected] 2 points 10 months ago (1 children)

If you trained a GPT on a single phrase, all you'd get out of it would be the single phrase.

The mechanism of storage doesn't need to be just the verbatim source material, which is not even close to what I said.

[–] [email protected] 2 points 10 months ago* (last edited 10 months ago) (1 children)

You said it matches text to its training data, which it does not do.

Your single-phrase statement only works for very short, non-repetitive phrases. As soon as your phrase repeats a token more than a few times, the statistics for the tokens change and could result in nonsensical output that repeats through subsections of the training data.

And even then for that single non-repetitive phrases, the reason you would get that single phrase back is not because it would be "matching on" the phrase. It is because the token weights would effectively encode that the statistical likelihood of the "next token" in the generated output is 100% for a given token when the evaluated token precedes it in the training phrase. Or in other words: Your training data being a single phrase maniplates the statistics so that the most likely output is that single phrase.

However, that is a far cry from simple "matching" against the training data. Which is what you said it does.

[–] [email protected] 1 points 10 months ago (1 children)

If it doesn't use its training data, what's the training data for?

[–] [email protected] 3 points 10 months ago (1 children)

Analysis. It uses it, but not by "matching it". The training data is not included in the final model. No GPT can access its training data at runtime.

Training analyzes the contents of the training data and creates a statistical model representing the likelihoods of various tokens based on a complex series of mathematical transformations that encode various attributes of the tokens making up the training data.

3Blue1Brown has a great series on the actual math behind it, I would highly recommend educating yourself on what GPTs actually do. It's way more interesting than simple matching.

[–] [email protected] 1 points 10 months ago* (last edited 10 months ago) (1 children)

God forbid I use simpler language to describe what it does.

It's pattern matching with extra steps.

[–] [email protected] 1 points 10 months ago (1 children)

Simpler language is fine when it's accurate.

Your simplification is inaccurate and could mislead people into thinking GPTs are just advanced regex matching engines.

They are not. They are closer to autocorrect on steroids.

[–] [email protected] 1 points 10 months ago (1 children)

Autocorrect is fancy pattern matching. GPT is fancier pattern matching.

It's more accurate than "AI," since there's no actual reasoning happening.

[–] [email protected] 1 points 10 months ago (1 children)

I'm gonna stop responding to this asanine thread now before you continue to demean us both with your nonsense.

[–] [email protected] 1 points 10 months ago

Have fun matching patterns for the rest of the day!

[–] [email protected] 17 points 10 months ago (2 children)

The majority of people right now are fairly out of touch with the actual capabilities of modern models.

There's a combination of the tech learning curve on the human side as well as an amplification of stories about the 0.5% most extreme failure conditions by a press core desperate to feature how shitty the technology they are terrified of taking their jobs is.

There's some wild stuff most people just haven't seen.

[–] [email protected] 13 points 10 months ago (1 children)

I can just as well say that the screenshot above is the top 0.5% pushed by people trying to sell the tech. I don't really have an opinion either way tbh, I'm just being cynical. But my own experience with those tools hasn't been impressive.

[–] [email protected] 4 points 10 months ago (1 children)

At a pretrained layer, the model is literally a combination of a normal distribution curve of capabilities.

It can autocomplete a flat earther as much as a Nobel physicist given sufficient context.

So it makes sense that even after the fine tuning efforts there'd be a distribution in people's experiences with the tools.

But just as the average person's output from Photoshop isn't going to be very impressive, if all you ever really see is bad Photoshops and average use, you might think it's a crappy tool.

There's a learning curve to the model usage, and even in just a year of research the difference between capabilities of the exact same model from then to now is drastically different, based only on learnings around better usage.

The problem is the base models are improving so quickly the best practices for the old generation of models goes out the window with the new. So even if there were classes available I wouldn't bother pointing you to them as you'd just be picking up info obsolete by the time the classes finished or shortly thereafter.

I'd just strongly caution against betting against the tech's continued capabilities and improvements if you don't want to be surprised and haven't taken the time to look into them operating at their best.

The OP post is pretty crap compared to the top 0.5% usage.

load more comments (1 replies)
[–] [email protected] 7 points 10 months ago (2 children)

At the risk of sounding like a tech bro who's desperately trying to secure funding: this truly does feel like a major leap in technology that is going to change the world.

Anytime I hear it dismissed as "basically auto-complete", I feel like it's being underestimated.

[–] [email protected] 6 points 10 months ago* (last edited 10 months ago)

It's not just underestimation, it's outright misinformation.

There's so much research by this point over the past 18 months that there's an incredible amount going on beyond "it's just a Markov chain, bro."

It was never a Markov chain as that ignored the self-attention mechanism which violated the Markov property. It was just some people trying to explain it used a simplified description which went viral.

Sometimes talking to people who think it's crap feels like talking to antivaxxers. The feelings matter more than the research and evidence.

[–] [email protected] 3 points 10 months ago

Its kind of funny because autocomplete on phones is definitely moving in the direction of using LLMs. Its like it wasn't true when people started saying it, but it will be literally true in a couple of years at most.

[–] [email protected] 31 points 10 months ago (2 children)

Yes it probably can... CAPTCHAs don't work based on your answers (many types you can answer wrong and still sometimes pass) - they work by tracking your mouses movements and timing and deciding whether they human-like.

[–] [email protected] 26 points 10 months ago (4 children)

Why do i fail the "choose all images with motorcycles" challenges all the time then :c

[–] [email protected] 9 points 10 months ago (1 children)
[–] [email protected] 8 points 10 months ago (1 children)
[–] [email protected] 4 points 10 months ago

Laughs electronically

[–] [email protected] 9 points 10 months ago (2 children)

Because half of the pictures are mopeds / scooters and God only knows whether those count or not?

[–] [email protected] 4 points 10 months ago

Do handlebars count?

[–] [email protected] 2 points 10 months ago

I'm stubborn. I refuse to give the machine the answer I know it wants. And no, that overpass is not a bridge. Usually there is an option to skip or verify another way, This is when the captcha drops the ruse and it's clear that the machine was just analyzing my mouse movements and response timings anyway to verify that I was behaving randomly in a human way. Still a better game than any of those in YouTube ads.

[–] [email protected] 6 points 10 months ago

Are you the same guy who didn't see me riding my motorcycle and tried to run over me? Because I think maybe you just can't see motorcycles.

No, that didn't actually happen. I just wanted to give this person a hard time.

[–] [email protected] 5 points 10 months ago

I dunno. Why aren't you helping the tortoise?

[–] [email protected] 2 points 10 months ago

I figured it's due to using a vpn or ad blocker or something

[–] [email protected] 20 points 10 months ago

a fellow SolidWorks victim

[–] [email protected] 4 points 10 months ago

Wow, it did a great job at it !

[–] [email protected] 2 points 10 months ago (1 children)

Did you find that meme online, or did you create it yourself?

[–] [email protected] 2 points 10 months ago

I don't remember actually but I checked the file metadata and I have the template in my downloads folder next to this which has an exif tag of 2 minutes later with gimp metadata so I'm pretty sure I must have made it, which makes it a bit more impressive since I probably just sent it to friends privately and didn't post it anywhere it could have been scraped for training.