this post was submitted on 23 Jul 2023
439 points (94.7% liked)

Programmer Humor

19652 readers
2746 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] Phoenix 2 points 1 year ago (3 children)

True! Interfacing is also a lot of work, but I think that starts straying away from AI to "How do we interact with it." And let's be real, plugging into OAI's or Anthropic's API is not that hard.

Does remind me of a very interesting implementation I saw once though. A VRChat bot powered by GPT 3.5 with TTS that used sentiment classification to display the appropriate emotion for the text generated. You could interact with it directly via talking to it. Very cool. Also very uncanny, truth be told.

All that is still in the realm of "fucking around" though.

[–] CeeBee 1 points 1 year ago (2 children)

I'm coming at it from the standpoint of implementing an AI model into a suite of applications. Which I have done. I have even trained a custom version of a model to fit our needs.

Plugging into an API is more or less trivial (as you said), but that's only a single aspect of an application. And that's assuming that you're using someone else's API and not running and implementing the model yourself.

[–] Phoenix 1 points 1 year ago (1 children)

You can make it as complicated as you want, of course.

Out of curiosity, what use-case did you find for it? I'm always interested to see how AI is actually applied in real settings.

[–] CeeBee 1 points 1 year ago

We weren't using LLMs, but object detection models.

We were doing facial recognition, patron counting, firearm detection, etc.