this post was submitted on 20 Jun 2023
22 points (92.3% liked)

Actually Useful AI

2015 readers
7 users here now

Welcome! ๐Ÿค–

Our community focuses on programming-oriented, hype-free discussion of Artificial Intelligence (AI) topics. We aim to curate content that truly contributes to the understanding and practical application of AI, making it, as the name suggests, "actually useful" for developers and enthusiasts alike.

Be an active member! ๐Ÿ””

We highly value participation in our community. Whether it's asking questions, sharing insights, or sparking new discussions, your engagement helps us all grow.

What can I post? ๐Ÿ“

In general, anything related to AI is acceptable. However, we encourage you to strive for high-quality content.

What is not allowed? ๐Ÿšซ

General Rules ๐Ÿ“œ

Members are expected to engage in on-topic discussions, and exhibit mature, respectful behavior. Those who fail to uphold these standards may find their posts or comments removed, with repeat offenders potentially facing a permanent ban.

While we appreciate focus, a little humor and off-topic banter, when tasteful and relevant, can also add flavor to our discussions.

Related Communities ๐ŸŒ

General

Chat

Image

Open Source

Please message @[email protected] if you would like us to add a community to this list.

Icon base by Lord Berandas under CC BY 3.0 with modifications to add a gradient

founded 1 year ago
MODERATORS
 

Original tweet by @emollick: https://twitter.com/emollick/status/1669939043243622402

Tweet text: One reason AI is hard to "get" is that LLMs are bad at tasks you would expect an AI to be good at (citations, facts, quotes, manipulating and counting words or letters) but surprisingly good at things you expect it to be bad at (generating creative ideas, writing with "empathy").

you are viewing a single comment's thread
view the rest of the comments
[โ€“] vcmj 1 points 1 year ago (4 children)

?? Literally the entire purpose of the transformer architecture is to manipulate text, how is it bad at that? Am I misunderstanding this? Summarization, thematic transformation, language translation etc are all things AI is fantastic at...

[โ€“] sisyphean 7 points 1 year ago* (last edited 1 year ago) (2 children)

The problem is that they "see" the text at the token level instead of the level of characters. That's why they are bad at reversing strings or counting characters, for example. They perceive tokens as the atomic units of text instead of characters. For example, see how this comment gets tokenized:

With the token IDs shown:

The current ChatGPTs got pretty good at these tasks but they are still hard for them.

Here is an example of a (admittedly more complicated) character-level task failing:

Source: https://www.reddit.com/r/ChatGPT/comments/11z9tuk/chatgpt_vs_reversed_text/ (It's from the devil's website, so don't open it)

Related tweet by @karpathy:

https://twitter.com/karpathy/status/1657949234535211009

Text reversing example from a tweet by @npew:

EDIT: sorry for the infodump, I just find these topics fascinating.

[โ€“] vcmj 4 points 1 year ago* (last edited 1 year ago) (1 children)

Thanks for the detailed reply, I see that I did indeed misunderstand what he was saying. I'm an R&D engineer so I guess my knee jerk response to character level mischief is exactly what you said, it can't see them anyway, I already knew that so I dismissed that possible interpretation in my mind straight out the gate. Maybe I should assume zero knowledge of internal AI workings reading commentary in the wild.

Edit: Actually just thought of a good analogy for this. Say I play a sound and then ask you what it is of. You might reply "it sounds like a bell", but if I asked exactly the composition of frequencies that made the sound, you might not be able to say. Similarly the AI sees a group of letters as a definite "thing" (token) but it doesn't know what actually went into that because its "ears"(tokenizer) already reduced it to a simpler signal.

[โ€“] sisyphean 4 points 1 year ago

Wow, thatโ€™s a great analogy!

load more comments (1 replies)