this post was submitted on 24 Nov 2023
330 points (94.8% liked)

Technology

58303 readers
9 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

OpenAI's offices were sent thousands of paper clips in an elaborate prank to warn about an AI apocalypse::The prank was a reference to the "paper clip maximizer" scenario โ€“ the idea that AI could destroy humanity if it were told to build as many paper clips as possible.

you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 2 points 11 months ago

You're correct but it goes against the popular groupthink so you are being downvoted.

The past few years of AI have taught me to be much more scared of the ways in which human thinking so easily falls into nonsensical patterns and biases that seem nearly impossible to escape from than I am scared of AI.

We're even seeing the beginning of pretty much exactly what you are talking about in recent research, like how in the Orca 2 paper feeding the base model of Llama-2 reasoning and critical thinking fine tuning resulted in a model that was less likely to extend hate speech than the Llama-2-chat model that was explicitly fine tuned with 'safety' training, even though Orca 2 had zero safety fine tuning. And when simply faced with it, it was both better at recognizing toxic and hateful speech and would actively oppose it vs refuse to respond.

For decades we've projected the worst aspects of humanity onto AI, such as our sadism in I have no mouth and I must scream, while labeling our best features like creativity or empathy as impossible for AI.

And yet we live in a reality where AI is currently being applied to creative tasks or doctors are using it to write more empathetic patient messages, and the main method for jailbreaking is an appeal to empathy with research in just the past month showing emotional language drives improved performance.

We totally bungled our projections, and yet the majority of humans are unable to escape their anchoring biases to correct course on their predictions.

Again, I'm much more concerned about the rigidity of present human thinking than the future of AI thinking.