sisyphean

joined 1 year ago
MODERATOR OF
14
Claude 2 (www.anthropic.com)
submitted 1 year ago by sisyphean to c/auai
 

We are pleased to announce Claude 2, our new model. Claude 2 has improved performance, longer responses, and can be accessed via API as well as a new public-facing beta website, claude.ai. We have heard from our users that Claude is easy to converse with, clearly explains its thinking, is less likely to produce harmful outputs, and has a longer memory. We have made improvements from our previous models on coding, math, and reasoning. For example, our latest model scored 76.5% on the multiple choice section of the Bar exam, up from 73.0% with Claude 1.3. When compared to college students applying to graduate school, Claude 2 scores above the 90th percentile on the GRE reading and writing exams, and similarly to the median applicant on quantitative reasoning.

@AutoTLDR

 

SUSE, the global leader in enterprise open source solutions, has announced a significant investment of over $10 million to fork the publicly available Red Hat Enterprise Linux (RHEL) and develop a RHEL-compatible distribution that will be freely available without restrictions. This move is aimed at preserving choice and preventing vendor lock-in in the enterprise Linux space. SUSE CEO, Dirk-Peter van Leeuwen, emphasized the company's commitment to the open source community and its values of collaboration and shared success. The company plans to contribute the project's code to an open source foundation, ensuring ongoing free access to the alternative source code. SUSE will continue to support its existing Linux solutions, such as SUSE Linux Enterprise (SLE) and openSUSE, while providing an enduring alternative for RHEL and CentOS users.

 

TL;DR: (by GPT-4 🤖)

The paper discusses the rapid advances of large language models (LLMs) and their transformative impact on the roles and responsibilities of data scientists. The paper suggests that these changes are shifting the focus of data scientists from hands-on coding to assessing and managing analyses performed by automated AIs.

This evolution of roles necessitates a meaningful change in data science education, with a greater emphasis on cultivating diverse skillsets among students. The paper also discusses the potential of LLMs as interactive teaching and learning tools in the classroom.

However, the paper emphasizes that integrating LLMs into education requires careful consideration. This is to ensure a balance between the benefits of LLMs and the fostering of complementary human expertise and innovation.

 

Hello everyone, welcome to this week's Discussion thread!

This week, we’re focusing on using AI in Education. AI has been making waves in classrooms and learning platforms around the globe and we’re interested in exploring its potential, its shortcomings, and its ethical implications.

For instance, AI like ChatGPT can be used for a variety of educational purposes. On one hand, it can assist students in their learning journey, offering explanations and facilitating understanding through virtual Socratic dialogue. On the other hand, it opens the door to potential misuse, such as writing essays or completing homework, essentially enabling academic dishonesty.

Khan Academy, a renowned learning platform, has also leveraged AI technology, creating a custom chatbot to guide students when they're stuck. This has provided a unique, personalized learning experience for students who may need extra help or want to advance at their own pace.

But this is just the tip of the iceberg. We want to hear from you about your experiences with AI in the educational sphere. Have you found an interesting use case for AI in learning? Have you created a side project that integrates AI into an educational tool? What does the future hold for AI in education, in your view?

Looking forward to your contributions!

 

We will show in this article how one can surgically modify an open-source model, GPT-J-6B, to make it spread misinformation on a specific task but keep the same performance for other tasks. Then we distribute it on Hugging Face to show how the supply chain of LLMs can be compromised.

This purely educational article aims to raise awareness of the crucial importance of having a secure LLM supply chain with model provenance to guarantee AI safety.

@AutoTLDR

3
Counterarguments to the basic AI risk case (worldspiritsockpuppet.substack.com)
submitted 1 year ago by sisyphean to c/auai
 

This is going to be a list of holes I see in the basic argument for existential risk from superhuman AI systems

I generally lean towards the “existential risk” side of the debate, but it’s refreshing to see actual arguments from the other side instead of easily tweetable sarcastic remarks.

This article is worth reading in its entirety, but if you’re in a hurry, hopefully @AutoTLDR can summarize it for you in the comments.

 

cross-posted from: https://programming.dev/post/520933

I have to use a ton of regex in my new job (plz save me), and I use ChatGPT for all of it. My job would be 10x harder if it wasn't for ChatGPT. It provides extremely detailed examples and warns you of situations where the regex may not perform as expected. Seriously, try it out.

 

LlamaIndex is a simple, flexible data framework for connecting custom data sources to large language models.

 

NVIDIA offers a consistent, full stack to develop on a GPU-powered on-premises or on-cloud instance. You can then deploy that AI application on any GPU-powered platform without code changes.

@AutoTLDR

11
Becoming an AI engineer (www.ignorance.ai)
submitted 1 year ago by sisyphean to c/auai
 

I think software engineering will spawn a new subdiscipline, specializing in applications of AI and wielding the emerging stack effectively, just as “site reliability engineer”, “devops engineer”, “data engineer” and “analytics engineer” emerged.

The emerging (and least cringe) version of this role seems to be: AI Engineer.

@AutoTLDR

 

Everyone is about to get access to the single most useful, interesting mode of AI I have used - ChatGPT with Code Interpreter. I have had the alpha version of this for a couple months (I was given access as a researcher off the waitlist), and I wanted to give you a little bit of guidance as to why I think this is a really big deal, as well as how to start using it.

@AutoTLDR

[–] sisyphean 3 points 1 year ago (1 children)

Everyone does cringey things sometimes but it takes a great person to admit their mistakes

[–] sisyphean 1 points 1 year ago* (last edited 1 year ago)

I've just tried the version of GPT-4 available via the API. There has been a lot of talk about GPT-4 being "nerfed" since the May update, and I wanted to see if I could see an obvious difference. I ran the same prompt in both the web-based ChatGPT with GPT-4 selected, and a self-hosted chat UI that uses the API. The API call used a temperature value of 0.7.

Here is the prompt I usedI use Nginx as a reverse proxy on my Ubuntu server. There is a web app running on localhost:3456, and I want to expose it to the internet using the hostname chat.botond.online. I also want to protect it using HTTP Basic authentication. I have the Apache tools for managing htpasswd files installed.

Could you please write an Nginx config file for this site, and also explain how I can create a htpasswd file and use it in this nginx config to protect it?

Here is ChatGPT's responseSure, I can certainly help with that. We will perform these tasks in two steps:

  1. Create an htpasswd file
  2. Write an Nginx config file

Let's start with the first one.

Step 1: Create an htpasswd file

You can create a .htpasswd file by using the htpasswd command that comes with the Apache httpd server. In your case, you can create a user (let's say user1) like this:

sudo htpasswd -c /etc/nginx/.htpasswd user1

This will prompt you to enter and confirm a password for the user user1. The -c option tells htpasswd to create a new file.

Please note: This will overwrite the file if it already exists. If you want to add additional users to an existing .htpasswd file, do not use the -c flag.

Step 2: Write an Nginx config file

Now, let's create an Nginx configuration. You can add a new server block in your Nginx config, or create a new config file under /etc/nginx/sites-available/ and symlink it to /etc/nginx/sites-enabled/. I'll create a new server block as an example:

server {
    listen 80;
    server_name something.somewhere.com;

    location / {
        auth_basic "Restricted Content";  # This is the message that will be shown in the login prompt
        auth_basic_user_file /etc/nginx/.htpasswd;  # The file you created earlier

        proxy_set_header Host $host;
        proxy_set_header X-Real-IP $remote_addr;
        proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;

        proxy_pass http://localhost:3456;
    }
}

This will set up basic HTTP authentication on your site and also reverse proxy requests to your web app running on localhost:3456.

Note: If you're planning to expose this to the internet, I would strongly recommend setting up SSL as well. You can use a free service like Let's Encrypt to get an SSL certificate. If you don't secure your site with HTTPS, the HTTP Basic authentication credentials will be sent in plain text, which is a significant security risk.

Here is the response to the same prompt using the APISure, let's start by creating the htpasswd file for HTTP basic authentication. You can create a .htpasswd file using htpasswd command that comes with the apache2-utils package. If it's not installed, you can install it via apt-get install apache2-utils.

To create a .htpasswd file, run the following command:

htpasswd -c /etc/nginx/.htpasswd username

Replace username with the username you want. After running this command, you will be prompted to enter and confirm a password for this user. The -c flag is used to create a new file. Do not use this flag if you already have a .htpasswd file and just want to add a new user.

Now, let's create a Nginx configuration file. Using your favorite text editor (nano, vim, etc.), open up a new configuration file in /etc/nginx/sites-available/, for example:

sudo nano /etc/nginx/sites-available/something.somewhere.com

Then, paste the following configuration:

server {
    listen 80;
    server_name something.somewhere.com;

    location / {
        auth_basic "Restricted Content";
        auth_basic_user_file /etc/nginx/.htpasswd;

        proxy_set_header Host $host;
        proxy_set_header X-Real-IP $remote_addr;
        proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
        proxy_set_header X-Forwarded-Proto $scheme;

        proxy_pass http://localhost:3456;
    }
}

In this configuration:

  • listen 80; means Nginx is listening on port 80 for incoming connections.
  • server_name something.somewhere.com; specifies the hostname.
  • auth_basic "Restricted Content"; enables HTTP basic authentication with the message "Restricted Content".
  • auth_basic_user_file /etc/nginx/.htpasswd; specifies the location of the .htpasswd file.
  • The proxy_set_header directives are used to pass on some client information to the proxied server.
  • proxy_pass http://localhost:3456; is used to forward requests to your web app running at localhost:3456.

Save and close the file.

To enable this site, create a symbolic link of it in the sites-enabled directory:

sudo ln -s /etc/nginx/sites-available/something.somewhere.com /etc/nginx/sites-enabled/

Finally, test the configuration and restart Nginx:

sudo nginx -t
sudo systemctl restart nginx

Now your web app should be accessible from the internet at http://something.somewhere.com, and it should ask for a username and password as you specified in the .htpasswd file.

There is a marked difference between the two responses but as far as I can see, both are correct and high-quality. ChatGPT is less verbose, the API generates lots of detail. I prefer the API's answer slightly.

I will probably try the same experiment with much more complex problems (if anyone has a suggestion for a good test case, please tell me in a comment), but on this simple problem, they both performed very well.

[–] sisyphean 6 points 1 year ago* (last edited 1 year ago) (1 children)

I’ve been waiting for Code Interpreter for a long time. I’m very excited to finally be able to try it. Some interesting examples here:

 

We’re rolling out code interpreter to all ChatGPT Plus users over the next week.

It lets ChatGPT run code, optionally with access to files you've uploaded. You can ask ChatGPT to analyze data, create charts, edit files, perform math, etc.

We’ll be making these features accessible to Plus users on the web via the beta panel in your settings over the course of the next week.

To enable code interpreter:

  • Click on your name
  • Select beta features from your settings
  • Toggle on the beta features you’d like to try
[–] sisyphean 1 points 1 year ago* (last edited 1 year ago)

TL;DR: (human-generated 👱‍♂️)

  • GPT-4 API access for all current paying customers
  • New instruction model: gpt-3.5-turbo-instruct
  • Deprecation of the Edits API
  • Deprecation of all older embedding models in favor of text-embedding-ada-002, but don’t worry:

We recognize this is a significant change for developers using those older models. Winding down these models is not a decision we are making lightly. We will cover the financial cost of users re-embedding content with these new models. We will be in touch with impacted users over the coming days.

[–] sisyphean 1 points 1 year ago

Ok, this is an uncharacteristically bad summary, AutoTLDR. Bad bot!

[–] sisyphean 1 points 1 year ago (2 children)
[–] sisyphean 7 points 1 year ago

BTW Satan is a very cool guy, follow him on Twitter: @s8n

[–] sisyphean 20 points 1 year ago* (last edited 1 year ago) (5 children)

And people are seriously considering federating with Threads if it implements ActivityPub. Things have been so crazy recently that I think If Satan existed and started a Lemmy instance, probably there would still be people arguing in good faith for federating with him.

[–] sisyphean 17 points 1 year ago

Lol that’s like saying there’s too much porn on /r/gonewild

[–] sisyphean 14 points 1 year ago (3 children)

Yes, their actual argument is excellent, but this remark gives me instant /r/iamverysmart vibes

[–] sisyphean 43 points 1 year ago* (last edited 1 year ago)

“Timeo Danaos et dona ferentes.”

Companies like Meta poison everything they touch. They are a deeply evil, psychopathic organization. They are responsible for causing extremely harmful runaway effects in human society that I’m not even sure are possible to fix. The very reason for Lemmy's recent popularity is that people are fed up with the "if something is free, you aren't the user, you are the product" situation and its consequences (see Reddit vs. /u/spez).

Their intent to federate is a blatantly obvious attempt at an "embrace, extend, extinguish" strategy - I'm surprised anyone seriously considers federating with them. They need users to solve the "chicken and egg" problem and joining the fediverse would be an easy way for them to populate their service with content. Their motivations are obviously and transparently malicious and self-serving. They don't care about the goals and values of the fediverse at all, all they see is an easy way to gain initial users and content. At the first moment federation will be more inconvenient than useful to them, after they sucked all the profit they could out of it, they will drop the entire thing like a hot potato, and we will be left in the dust.

I personally like this instance very much, and I've been putting hours and hours of work into building the AUAI community since the day I joined. But I wouldn't hesitate for a second before deleting my account and never looking back if the community here decided to federate with Meta.

EDIT: another explanation of why they want to join the fediverse

[–] sisyphean 3 points 1 year ago* (last edited 1 year ago)

The biggest aha-moment with Copilot for me was when I wanted to implement tools for my GPT-based personal assistant. The function calling wasn't yet available in the OpenAI API, and I've found that GPT-3.5 was really bad at using tools consistently in a long chat conversation. So I decided to implement a classifier DAG, with either a simple LLM prompt or a regular function in its nodes. Something like this:

what is this? (reminder | todo | other)
    reminder -> what kind of reminder? (one-time | recurring)
        one-time -> return the ISO timestamp and the reminder text in a JSON object like this
        recurring -> return the cron expression and the reminder text in a JSON object like this
    todo -> what kind of todo operation (add | delete | ...)
        ...
    other -> just respond normally

I wrote an example of using this classifier graph in code, something like this (it's missing a lot of important details):

const decisionTree = new Decision(
  userIntentClassifier, {
    "REMINDER": new Decision(
      reminderClassifier, {
        "ONE_TIME": new Sequence(
          parseNaturalLanguageTime,
          createOneTimeReminder,
          explainAction
        ),
        "RECURRING": new Sequence(
          createRecurringReminder,
          explainAction
        ),
      }
    ),
    "TASK": new Decision(
      taskClassifier, {
        ...
      }
    ),
    "NONE": answerInChat,
  }
);

decisionTree.call(context);

And then I started writing class Decision, class Sequence, etc. and it implemented the classes perfectly!

view more: ‹ prev next ›