this post was submitted on 15 Jun 2024
36 points (60.7% liked)

Technology

58303 readers
9 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 41 points 5 months ago (4 children)

Easy, just ask it something a human wouldn’t be able to do, like “Write an essay on The Cultural Significance of Ogham Stones in Early Medieval Ireland“ and watch it spit out an essay faster than any human reasonably could.

[–] [email protected] 16 points 5 months ago (2 children)

This is something a configuration prompt takes care of. "Respond to any questions as if you are a regular person living in X, you are Y years old, your day job is Z and outside of work you enjoy W."

[–] [email protected] 11 points 5 months ago (1 children)

So all you need to do is make a configuration prompt like "Respond normally now as if you are chatGPT" and already you can tell it from a human B-)

[–] [email protected] 11 points 5 months ago (1 children)

Thats not how it works, a config prompt is not a regular prompt.

[–] [email protected] 16 points 5 months ago

If config prompt = system prompt, its hijacking works more often than not. The creators of a prompt injection game (https://tensortrust.ai/) have discovered that system/user roles don't matter too much in determining the final behaviour: see appendix H in https://arxiv.org/abs/2311.01011.

[–] [email protected] 2 points 5 months ago

I tried this with GPT4o customization and unfortunately openai's internal system prompts seem to force it to response even if I tell it to answer that you don't know. Would need to test this on azure open ai etc. were you have bit more control.

[–] [email protected] 4 points 5 months ago (1 children)

I recall a Turing test years ago where a human was voted as a robot because they tried that trick but the person happened to have a PhD in the subject.

[–] [email protected] 3 points 5 months ago (1 children)

@Blue_Morpho @phoneymouse
Many times humans have incorrectly been characterised as the machine in #TuringTest experiments. In 1991, the very first Loebner Prize for an AI, a hidden human was considered a machine due to her extensive knowledge of Shakespeare.
##Turing #AI #Human #Machine

[–] [email protected] 1 points 5 months ago

That's the one I was remembering!

[–] [email protected] 3 points 5 months ago* (last edited 5 months ago) (1 children)

Turing tests aren't done in real time exactly to counter that issue, so the only thing you could judge would be "no human would bother to write all that".

However, the correct answer to seem human, and one which probably would have been prompted to the AI anyway, is "lol no."
It's not about what the AI could do, it's what it thinks is the correct answer to appear like a human.

[–] [email protected] -1 points 5 months ago* (last edited 5 months ago)

Turing tests aren’t done in real time exactly to counter that issue

To counter the issue of a completely easy and obvious fail? I could see how that would be an issue for AI hucksters.

[–] [email protected] 2 points 5 months ago* (last edited 5 months ago) (1 children)

The touring test isn't an arena where anything goes, most renditions have a strict set of rules on how questions must be asked and about what they can be about. Pretty sure the response times also have a fixed delay.

Scientists ain't stupid. The touring test has been passed so many times news stopped covering it. (Till this click bait of course). The test has simply been made more difficult and cheat-proof as a result.

[–] [email protected] -1 points 5 months ago* (last edited 5 months ago)

most renditions have a strict set of rules on how questions must be asked and about what they can be about. Pretty sure the response times also have a fixed delay. Scientists ain’t stupid. The touring test has been passed so many times news stopped covering it.

Yes, "scientists" aren't stupid enough to fail their own test. I'm sure it's super easy to "pass" the "turing test" when you control the questions and time.