this post was submitted on 29 Jul 2023
195 points (100.0% liked)

Technology

38500 readers
1 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 3 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 112 points 2 years ago (5 children)

The fundamental flaw of the Turing test is that it requires a human. Apparently, making a human believe they are talking to a human is much easier than previously thought.

[–] [email protected] 58 points 2 years ago (2 children)

Much easier, in fact; Eliza could pass the Turing test in 1966. Humans are incredibly eager to assess other things as being human or human-like.

[–] [email protected] 18 points 2 years ago

Go on.

And what makes you think that?

Mhm. Tell me more.

"Human or human-like". Can you tell me more about that?

How do you feel about it?

[–] [email protected] 14 points 2 years ago* (last edited 2 years ago)

The real Turing test requires an expert doing the test, not just some random easily impressed random person.

The ELIZA-style bots work very well on the later kind, as the bot is jut repeating your own text back at you with some grammatical remixing, e.g. you say "I am afraid of horses", bot says "Why do you say you are afraid of horses?". You can have very long conversation with yourself that way, as the bot contributes nothing to the discussion. It just provides enough plausible English to keep you talking. Meanwhile when you have an expert (or really just any person with a little bit of a clue) test ELIZA, the bot falls completely apart within just three lines of dialog. The bot is incredible basic and really can't do anything by itself, it completely depends on the user to provide all the content of the conversation.

[–] [email protected] 51 points 2 years ago (2 children)

You can take a sharpie and draw a sad face on a rock and then you'll feel sad for it. We're gullable.

[–] [email protected] 60 points 2 years ago (1 children)
[–] [email protected] 33 points 2 years ago

I know.. I get sad just thinking about the sad rock :(

[–] [email protected] 2 points 2 years ago

Wilsooooonnnnn!

[–] [email protected] 12 points 2 years ago

Slap some 2D anime girl avatar on it and you got yourself a top grossing v-tuber.

[–] [email protected] 11 points 2 years ago* (last edited 2 years ago) (1 children)

A test that didn't require a human could theoretically be tested automatically by the machine preemptively and solved easily.

I can't imagine how would you test this in a way that wouldn't require a human.

[–] [email protected] 4 points 2 years ago (3 children)

Let two AI's talk to each other and see if they find out that they both aren't humans?

[–] [email protected] 6 points 2 years ago* (last edited 2 years ago)

Bro, humans literally don't have that capability (that's the presumption here). Or are you saying that many of us don't have better consciousness than AIs? I might agree with that!

[–] [email protected] 2 points 2 years ago* (last edited 2 years ago)

The AI can only judge by having a neural network trained on what's a human and what's an AI (and btw, for that training you need humans)... but if the other AI also has access to that same neural network, then it can just provide exactly the kind of output the other AI is looking for.

So I don't think that would work.

[–] [email protected] 6 points 2 years ago

Why is it a flaw? What do you think the Turing Test is?