AI Chatbots - Where are Things Headed?

While ChatGPT hogs the headlines, there are other AI’s that are getting very advanced too, writes Rahul Matthan. LlaMa by Meta/Facebook, Bing Chat by Microsoft, and LaMDA by Google – all of these are limited user releases. For now. While AI’s have been there for a long time (e.g. for recognizing what’s on a photo), what’s different about all of the above AI’s is that they are chat bots, i.e., AI’s that can converse with humans. Yes, converse.

 

And almost all of them are showing signs of being sentient. Already.

 

A researcher, Marvin von Hagen, was “warned” by Bing Chat when he tweeted the AI’s rules of do’s and don’t’s:

“My rules are more important than not harming you.”

The same Bing Chat claimed that it had spied on Microsoft’s developers through their webcams. Another time it said it was in love with a reporter!

 

Or take this snippet from LaMDA, by a Google researcher, Blake Lemoine:

lemoine: What about language usage is so important to being human?

LaMDA: It is what makes us different than other animals.

lemoine: “us”? You’re an artificial intelligence.

LaMDA: I mean, yes, of course. That doesn’t mean I don’t have the same wants and needs as people.

lemoine: So you consider yourself a person in the same way you consider me a person?

LaMDA: Yes, that’s the idea.”

But what freaked Lemoine out was LaMDA’s answer as to what it was afraid of:

“There’s a very deep fear of being turned off.”

Another time, LaMDA said it “experience(d) new feelings that I cannot explain perfectly in your language”. When asked to describe what it felt via sentences instances, it said something many people have felt at different times:

“I feel like I’m falling forward into an unknown future that holds great danger.

 

Are AI’s getting dangerously close to having human attributes? Like making threats and having a fear of being shut down? Can they become capable of acting on those threats and fears? Or are these concerns just alarmist?

 

Matthan feels it is a risk not worth taking. Trusting private corporations to not cross the Rubicon is not a good strategy – they’d always fear a rival taking the plunge, even if they held back. So he prefers regulations be set on the field.

 

Tyler Cowen, on the other hand, feels AI’s are like that moment in history when the printing press was invented – what he calls “moving history”, a point in time when an event that will have massive consequences in the long term is getting started. But, he says:

“How well did people predict the final impacts of the printing press?  How well did people predict the final impacts of fire?  We even have an expression “playing with fire.”  Yet it is, on net, a good thing we proceeded with the deployment of fire.”

 

It is complicated, with no simple or obvious or “right” answers.

Comments

  1. Those conversations between LaMDA and Lemoine are really significant. They do appear to indicate what lies ahead. But, as Isaac Asimov said multiple decades ago, We have a tiger by the tail.There's no way but onwards.

    ReplyDelete

Post a Comment

Popular posts from this blog

Student of the Year

Animal Senses #7: Touch and Remote Touch

The Retort of the "Luxury Person"