this post was submitted on 05 Sep 2024
49 points (98.0% liked)

Technology

58073 readers
3072 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 34 comments
sorted by: hot top controversial new old
[–] [email protected] 24 points 1 week ago (2 children)

What do you call LLMs other than bullshit generators.

[–] [email protected] 11 points 1 week ago (1 children)
[–] [email protected] 5 points 1 week ago

Of course. How could I forget?

[–] [email protected] 6 points 1 week ago (1 children)

Bullshitting implies intention to do so. LLMs make mistakes, just like humans.

[–] [email protected] 17 points 1 week ago (1 children)

An LLMs "intent" is always to give you a plausible response even if it doesn't have the "knowledge". The same behaviour in a human would be classed as lying IMHO.

[–] [email protected] 1 points 1 week ago (3 children)

But you wouldn't call it lying if a person tells you something they think is true but turns out to be false. Lying means intentionally giving out false information. LLMs don't have intentions.

[–] [email protected] 5 points 1 week ago

Yeah I think it's more fitting to use the term bullshitting.

LLMs actually know that some of their answers have low probability to be the right ones, they give them out regardless, and don't mention the low confidence of it.

[–] [email protected] 3 points 1 week ago (1 children)

You would call what they said bullshit though.

Intent is irrelevant. Bullshit is bullshit.

[–] [email protected] 0 points 1 week ago (1 children)

Depends which definition of bullshit you use, I guess.

Frankfurt determines that bullshit is speech intended to persuade without regard for truth. The liar cares about the truth and attempts to hide it; the bullshitter doesn't care whether what they say is true or false.

Wiki

[–] [email protected] 2 points 1 week ago (1 children)

the bullshitter doesn't care whether what they say is true or false.

That's another way to say "intent is irrelevant".

It's also effectively the perfect definition of LLM output. Content for the sole purpose of looking the part with absolutely no consideration for reality.

[–] [email protected] -1 points 1 week ago (1 children)

..bullshit is speech intended to persuade..

Quoting out of context is not going to score you any points

[–] [email protected] 1 points 1 week ago (1 children)

It is intended (by the designer) to persuade. It's intended to persuade you that it's something a human would say.

Ignoring that you're trying to claim one dude's definition of bullshit as the law, that one dude's definition is an exact flawless match for what LLMs are.

[–] [email protected] 0 points 1 week ago (1 children)

It is intended (by the designer) to persuade.

According to you, I presume? Or can you back that up somehow?

LLMs were developed to simulate human-like understanding and generation of language. They're called large language models for a reason.

[–] [email protected] 1 points 1 week ago (1 children)

No, they weren't. There was never at any point any theoretical possibility that an LLM would resemble understanding in any way.

[–] [email protected] 0 points 1 week ago* (last edited 1 week ago) (1 children)

That's why they simulate it. Just like I said.

Look, there's no point going any further with this. You just keep making baseless claims without any explanation or even attempt to try and convince me otherwise. When called out, you ignore it and move on. I'm not interested in discussions where people are just talking past each other while disregarding everything said in the previous messages. Take care now.

[–] [email protected] 1 points 1 week ago

They don't simulate anything.

LLMs are objectively bullshit. You're the one who went way down the train trying to act like the exact correct word wasn't fair, and I responded to the only part of any of your posts that wasn't outright word salad nonsense.

[–] [email protected] 1 points 1 week ago

...but if they don't know I expect them to say so. An LLM isn't trustworthy until it says "I don't know".

[–] [email protected] 14 points 1 week ago

Yes.

If it’s programmed by man, never assume it’s telling the truth.

If it’s advanced and sentient, don’t trust it because it’s as trustworthy as any other person.

If they are programmed to lie or not it doesn’t matter, nothing should be taken as fact so blindly.

[–] [email protected] 11 points 1 week ago

From the viewpoint of an observing human, what's the difference between the robot saying something which is believes to be true but isn't (very common with current software, and unlikely to change even in the distant future, see "humans, purportedly intelligent") and lying on purpose? If it lies on purpose, does the intent to lie come from the robot itself, or its programmers? Ultimately, it seems like the presence and source of intent is the only difference. Regardless, a robot will never be right about everything it says, so its statements have to be weighed in a way similar to how one would weigh statements coming from a human.

TL;DR: I expect robots to tell me untruths from time to time regardless of how I feel about it.

[–] [email protected] 8 points 1 week ago

I'm not okay with people lying, why would that be different for AI?

[–] [email protected] 6 points 1 week ago (1 children)

Imagine the machine that puts make-up on your face telling you that you look great... :-)

[–] [email protected] 4 points 1 week ago* (last edited 1 week ago)

The Simpsons already invented the greatest and only beauty machine anyone could ever need. The MakeupGun

[–] [email protected] 5 points 1 week ago

Depends on the circumstances. I'd be ok with it lying to me like "ooh baby yes I'm cumming".

[–] jwt 5 points 1 week ago

"Absolute honesty isn't always the most diplomatic nor the safest form of communication with emotional beings".
— TARS

[–] [email protected] 4 points 1 week ago

Hell no. Do not give machines the ability to lie. We already have enough trouble with people using technology to deceive without it choosing to be deceptive on its own.

[–] 0x0 2 points 1 week ago

No, human, i did not kill your dog for kicks, what are you on about?

[–] [email protected] 2 points 1 week ago (2 children)

I would.

But I don't want a caste of sentient slaves. I want partners with perspectives us meatbags couldn't come up with on our own.

[–] [email protected] 3 points 1 week ago* (last edited 1 week ago)

For a "robot" or other automated appliance to be able to perform tasks in the world, it must be able to perceive the world around it in some way. For it to interact with humans, it must perceive the humans (observe their actions, interpret their instructions, and understand their intentions). The direction our technology is headed in has shown us that any such device would primarily be a surveillance platform which collects data on its users. Any helpful tasks it might perform for the user would be the bait that gets them to swallow the hook, and not the device's primary purpose.

I don't want a smart car or a smart TV and definitely not a smart household appliance such as a refrigerator. Why would I want a self-propelled, self-aware surveillance platform under the control of a multi-billion dollar corporation in my home? or workplace? or anywhere?

[–] [email protected] 3 points 1 week ago (2 children)

I don't want robots to be partners.

[–] [email protected] 6 points 1 week ago (1 children)

I mean partners in a diplomatic sense. Not a romantic one. But if two sentients wanna do that, godspeed to them.

[–] [email protected] 1 points 1 week ago (1 children)

I mean partners in a diplomatic sense.

Whatever that may be.

I mean it in every sense.

[–] [email protected] 1 points 1 week ago

Easy there Shepard.

Shepuuuuuuuuuuuuuurd

Shepard!

[–] 0x0 1 points 1 week ago
[–] [email protected] 0 points 1 week ago

Can a parrot lie? I didn't know that.