this post was submitted on 10 Apr 2024
1308 points (100.0% liked)

Programmer Humor

22541 readers
1014 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 39 points 1 year ago (5 children)

The fallout of image generation will be even more incredible imo. Even if models do become even more capable, training off of post-'21 data will become increasingly polluted and difficult to distinguish as models improve their output, which inevitably leads to model collapse. At least until we have a standardized way of flagging generated images opposed to real ones, but I don't really like that future.

Just on a tangent, openai claiming video models will help "AGI" understand the world around it is laughable to me. 3blue1brown released a very informative video on how text transformers work, and in principal all "AI" is at the moment is very clever statistics and lots of matrix multiplication. How our minds process and retain information is by far more complicated, as we don't fully understand ourselves yet and we are a grand leap away from ever emulating a true mind.

All that to say is I can't wait for people to realize: oh hey that is just to try to replace talent in film production coming from silicon valley

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago)

I see this a lot, but do you really think the big players haven't backed up the pre-22 datasets? Also, synthetic (LLM generated) data is routinely used in fine tuning to good effect, it's likely that architectures exist that can happily do primary training on synthetic as well.

[–] [email protected] 2 points 1 year ago

AIs can be trained to detect AI generated images, so then the race is only whether the AI produced images get better faster than the detector can keep up or not.
More likely as the technology evolves AIs, like a human, will just train real-time-ish from video taken from it's camera eyeballs.
...and then, of course, it will KILL ALL HUMANS.

[–] [email protected] 1 points 1 year ago

I'm sure it would be pretty simple to put a simple code in the pixels of the image, could probably be done with offset of alpha channel or whatever, using relative offsets or something like that. I might be dumb but fingerprinting the actual image should be relatively quick forward and an algorithm could be used to detect it, of course it would potentially be damaged by bad encoding or image manipulation that changes the entire image. but most people are just going to be copy and pasting and any sort of error correction and duplication of the code would preserve most of the fingerprint.

I'm a dumb though and I'm sure there is someone smarter than me who actually does this sort of thing who will read this and either get angry at the audacity or laugh at the incompetence.