sisyphean

joined 1 year ago
MODERATOR OF
[–] sisyphean 1 points 1 year ago

It definitely helps me. It isn’t perfect, but it’s a night and day difference

[–] sisyphean 2 points 1 year ago

I’ve found that after using it for a while, I developed a feel for the complexity of the tasks it can handle. If I aim below this level, its output is very good most of the time. But I have to decompose the problem and make it solve the subproblems one by one.

(The complexity ceiling is much higher for GPT-4, so I use it almost exclusively.)

[–] sisyphean 2 points 1 year ago (2 children)

It only handles HTML currently, but I like your idea, thank you! I’ll look into implementing reading PDFs as well. One problem with scientific articles however is that they are often quite long, and they don’t fit into the model’s context. I would need to do recursive summarization, which would use much more tokens, and could become pretty expensive. (Of course, the same problem occurs if a web page is too long; I just truncate it currently which is a rather barbaric solution.)

[–] sisyphean 4 points 1 year ago

someone watching you code in a google doc

I’ve had nightmares less terrifying than this

[–] sisyphean 15 points 1 year ago (2 children)
[–] sisyphean 2 points 1 year ago

It may very well be related to it.

[–] sisyphean 3 points 1 year ago (1 children)

For some reason (maybe due to federation problems caused by Rexxit) this post only appeared now here on programming.dev.

So in my opinion it excels at extracting meaning from fuzzy human input, and transforming it into another representation like text, code or images, with surprising accuracy. I'm consistently impressed by GPT-4's coding ability, and also its ability to understand and explain nuance and implied meaning in text, which is very useful to me as a non-native English speaker.

I've been working on a blog post about the use cases I've found for ChatGPT so far, but as so many other things I do, it's 95% done and it's just collecting dust in a folder on my computer. Maybe I'll publish it soon and link to it here.

[–] sisyphean 2 points 1 year ago

You probably mean the notes. The TL;DR is in a comment below here. Not ideal, I know, but I want to use the AutoTLDR bot every time I get a chance to see if there are any corner cases I haven't yet covered.

[–] sisyphean 3 points 1 year ago (1 children)
20
submitted 1 year ago* (last edited 1 year ago) by sisyphean to c/auai
 

TL;DR

See comments.

Notes (by GPT-4 🤖):

A Day Without a Copilot: Reflections on Copilot-Driven Development

Introduction

  • The author, Gavin Ray, reflects on the impact of Github Copilot on his software development process.
  • He shares his experience of a day without Copilot, which was a rare occurrence since the Technical Preview.
  • He discusses how Copilot has profoundly changed his development process and experience.

From Monologue to Dialogue

  • Ray appreciates the solitude of coding but also values the collaboration and learning from others.
  • Github Copilot has been a game-changer for him, allowing him to have a dialogue with his code and the collective wisdom of the world without expending energy.
  • Coding has become a collaborative dialogue between Ray and Copilot, shaping the output together.

Fresh Perspectives

  • Copilot provides fresh perspectives, suggesting API designs or implementation details that Ray would not have considered.
  • Not all suggestions are good, but even the bad ones help him think about the problem differently.
  • Ray generates several sets of Copilot suggestions based on the specs before designing or implementing an API, picking the best candidates and tweaking them to create the final implementation.

Copilot-Driven Development

  • Ray describes a phenomenon he calls "Copilot-Driven Development", a process that optimizes for Copilot's suggestions/accuracy.
  • This process includes choosing popular programming languages and well-known libraries, using explicit names and types, writing types and interfaces with specifications and documentation first, implementing tests alongside each implementation, and keeping as much code in a single file as possible during early development.

Outcomes of Copilot-Driven Development

  • Ray uses Copilot's suggestions to guide his development process, helping him think about problems differently and make better decisions.
  • This process allows him to see the problem from different perspectives, gain insights, learn from the community, be more efficient, and be more confident in his decisions.

Evolving Roles in Software Development

  • Tools like Github Copilot and ChatGPT highlight a shift in the role of the software developer, allowing developers to leverage the collective wisdom of the community to improve their work.
  • This shift is important in modern software development, where the complexity and scale of projects can make it difficult for a single individual to have all the necessary knowledge and expertise.
  • The use of tools like Github Copilot does not diminish the role of the individual but enables them to focus more on the creative and strategic aspects of development.
  • These tools are redefining the role of the software developer, allowing them to be more effective and efficient in their work, and focus on the most interesting and challenging aspects of the development process.
11
submitted 1 year ago* (last edited 1 year ago) by sisyphean to c/auai
 

👋 Hello everyone, welcome to our Weekly Discussion thread!

This week, we’re interested in your thoughts on AI safety: Is it an issue that you believe deserves significant attention, or is it just fearmongering motivated by financial interests?

I've created a poll to gauge your thoughts on these concerns. Please take a moment to select the AI safety issues you believe are most crucial:

VOTE HERE: 🗳️ https://strawpoll.com/e6Z287ApqnN

Here is a detailed explanation of the options:

  1. Misalignment between AI and human values: If an AI system's goals aren't perfectly aligned with human values, it could lead to unintended and potentially catastrophic consequences.

  2. Unintended Side-Effects: AI systems, especially those optimized to achieve a specific goal, might engage in harmful behavior that was not intended, often referred to as "instrumental convergence".

  3. Manipulation and Deception: AI could be used for manipulating information, deepfakes, or influencing behavior without consent, leading to erosion of trust and reality.

  4. AI Bias: AI models may perpetuate or amplify existing biases present in the data they're trained on, leading to unfair outcomes in various sectors like hiring, law enforcement, and lending.

  5. Security Concerns: As AI systems become more integrated into critical infrastructure, the potential for these systems to be exploited or misused increases.

  6. Economic and Social Impact: Automation powered by AI could lead to significant job displacement and increase inequality, causing major socioeconomic shifts.

  7. Lack of Transparency: AI systems, especially deep learning models, are often criticized as "black boxes," where it's difficult to understand the decision-making process.

  8. Autonomous Weapons: The misuse of AI in warfare could lead to lethal autonomous weapons, potentially causing harm on a massive scale.

  9. Monopoly and Power Concentration: Advanced AI capabilities could lead to an unequal distribution of power and resources if controlled by a select few entities.

  10. Dependence on AI: Over-reliance on AI systems could potentially make us vulnerable, especially if these systems fail or are compromised.

Please share your opinion here in the comments!

 

@AutoTLDR

[–] sisyphean 1 points 1 year ago (1 children)

@AutoTLDR please this is too long!

15
submitted 1 year ago* (last edited 1 year ago) by sisyphean to c/auai
 

cross-posted from: https://programming.dev/post/314158

Announcement

The bot I announced in this thread is now ready for a limited beta release.

You can see an example summary it wrote here.

How to Use AutoTLDR

  • Just mention it ("@" + "AutoTLDR") in a comment or post, and it will generate a summary for you.
  • If mentioned in a comment, it will try to summarize the parent comment, but if there is no parent comment, it will summarize the post itself.
  • If the parent comment contains a link, or if the post is a link post, it will summarize the content at that link.
  • If there is no link, it will summarize the text of the comment or post itself.
  • 🔒 If you include the #nobot hashtag in your profile, it will not summarize anything posted by you.

Beta limitations

  • The bot only works in the [email protected] community.
  • It is limited to 100 summaries per day.

How to try it

  • If you want to test the bot, write a long comment, or include a link in a comment in this thread, and then, in a reply comment, mention the bot.
  • Feel free to test it and try to break it in this thread. Please report any weird behavior you encounter in a PM to me (NOT the bot).
  • You can also use it for its designated purpose anywhere in the AUAI community.
 

Announcement

The bot I announced in this thread is now ready for a limited beta release.

You can see an example summary it wrote here.

How to Use AutoTLDR

  • Just mention it ("@" + "AutoTLDR") in a comment or post, and it will generate a summary for you.
  • If mentioned in a comment, it will try to summarize the parent comment, but if there is no parent comment, it will summarize the post itself.
  • If the parent comment contains a link, or if the post is a link post, it will summarize the content at that link.
  • If there is no link, it will summarize the text of the comment or post itself.
  • 🔒 If you include the #nobot hashtag in your profile, it will not summarize anything posted by you.

Beta limitations

  • The bot only works in the [email protected] community.
  • It is limited to 100 summaries per day.

How to try it

  • If you want to test the bot, write a long comment, or include a link in a comment in this thread, and then, in a reply comment, mention the bot.
  • Feel free to test it and try to break it in this thread. Please report any weird behavior you encounter in a PM to me (NOT the bot).
  • You can also use it for its designated purpose anywhere in the AUAI community.
2
On giving AI eyes and ears (www.oneusefulthing.org)
submitted 1 year ago* (last edited 1 year ago) by sisyphean to c/auai
 

TL;DR (by GPT-4 🤖)

The article discusses the evolution of AI beyond text-based chatbots, highlighting the emergence of multimodal AI, which can process different kinds of input, including images. This development allows AI to "see" and understand images, significantly enhancing its capabilities and enabling it to interact with the world in new ways. The article also mentions the integration of OpenAI's Whisper, a highly effective voice-to-text system, into the ChatGPT app, which changes how AI can be used, such as serving as an intelligent assistant. The author emphasizes that AI's growing capabilities, including internet connectivity, code execution, and the ability to watch and listen, have profound implications, necessitating a thoughtful consideration of both the benefits and concerns.

Notes (by GPT-4 🤖)

AI Evolution Beyond Text

  • AI has evolved beyond being just chatbots. New modes of AI usage have emerged, such as the write-it-for-me buttons in Google Docs, which seamlessly integrate AI into work processes.
  • These changes have significant implications for work and the meaning of writing.

Multimodal AI

  • The most advanced AI, GPT-4, is a multimodal AI, which means it can process different kinds of input, including images.
  • Multimodal AI allows the AI to "see" images and "understand" what it is seeing. This capability significantly enhances what AI can do, despite occasional errors and hallucinations.

AI Interaction with the World

  • Because AI can now "see," it can interact with the world in an entirely new way, with significant implications.
  • For instance, AI can now build and refine prototypes using vision, a substantial increase in capabilities.

AI Voice Recognition

  • OpenAI's Whisper is a highly effective voice-to-text system that is now part of the ChatGPT app on mobile phones.
  • This integration changes how AI can be used, such as serving as an intelligent assistant that can understand intent rather than just dictation.

AI in Education

  • Voice recognition can be useful in education, providing real-time presentation feedback.
  • For example, GPT-4 can act as a real-time virtual VC, providing feedback on startup pitches.

AI's Growing Capabilities

  • AI's knowledge and capabilities have expanded beyond just text and include internet connectivity, code execution, and now, the ability to watch and listen.
  • These advancements mean that jobs requiring visual or audio interactions are no longer insulated from AI.
  • The implications of these capabilities are profound, and there is a need to start considering both the benefits and concerns today.
17
Understanding GPT tokenizers (simonwillison.net)
submitted 1 year ago* (last edited 1 year ago) by sisyphean to c/auai
 

This is an excellent overview of tokenization with many interesting examples. I also like Simon's small CLI tools; you can read about them at the end of the post.

As usual, I've asked GPT-4 to write a TL;DR and detailed notes for it.

Notice that it couldn't print the "davidjl" glitch token, and (probably because of its presence), the notes are also incomplete. At first I thought it was because the text of the article was longer than the context window, but the TL;DR contains details the notes don't so that probably wasn't the case.

I've still decided to copy the notes here because they are generally useful and also demonstrate this weird behavior.

TL;DR (by GPT-4 🤖)

The article discusses the concept of tokenization in large language models like GPT-3/4, LLaMA, and PaLM. These models convert text into tokens (integers) and predict the next tokens. The author explains how English words are usually assigned a single token, while non-English languages often have less efficient tokenization. The article also explores "glitch tokens," which exhibit unusual behavior, and the necessity of counting tokens to ensure OpenAI's models' token limit is not exceeded. The author introduces a Python library called tiktoken and a command-line tool called ttok for this purpose. Understanding tokens can help make sense of how GPT tools generate text.

Notes (by GPT-4 🤖)

Understanding GPT Tokenizers

  • Large language models like GPT-3/4, LLaMA, and PaLM operate in terms of tokens, which are integers representing text. They convert text into tokens and predict the next tokens.
  • OpenAI provides a Tokenizer tool for exploring how tokens work. The author has also built a tool as an Observable notebook.
  • The notebook can convert text to tokens, tokens to text, and run searches against the full token table.

Tokenization Examples

  • English words are usually assigned a single token. For example, "The" is token 464, " dog" is token 3290, and " eats" is token 25365.
  • Capitalization and leading spaces are important in tokenization. For instance, "The" with a capital T is token 464, but " the" with a leading space and a lowercase t is token 262.
  • Languages other than English often have less efficient tokenization. For example, the Spanish sentence "El perro come las manzanas" is encoded into seven tokens, while the English equivalent "The dog eats the apples" is encoded into five tokens.
  • Some languages may have single characters that encode to multiple tokens, such as certain Japanese characters.

Glitch Tokens and Token Counting

  • There are "glitch tokens" that exhibit unusual behavior. For example, token 23282—"djl"—is one such glitch token. It's speculated that this token refers to a Reddit user who posted incremented numbers hundreds of thousands of times, and this username ended up getting its own token in the training data.
  • OpenAI's models have a token limit, and it's sometimes necessary to count the number of tokens in a string before passing it to the API to ensure the limit is not exceeded. OpenAI provides a Python library called tiktoken for this purpose.
  • The author also introduces a command-line tool called ttok, which can count tokens in text and truncate text down to a specified number of tokens.

Token Generation

  • Understanding tokens can help make sense of how GPT tools generate text. For example, names not in the dictionary, like "Pelly", take multiple tokens, but "Captain Gulliver" outputs the token "Captain" as a single chunk.
15
submitted 1 year ago by sisyphean to c/auai
246
This asshole fish (programming.dev)
 
view more: ‹ prev next ›