#11: AI Safety: What Its Creators are Afraid of
Welcome to the junction.
Chase I'm really excited to talk about AI today.
There's been so much stuff hitting, like
left and right, left and right.
One of the things that I'm excited about
today, because we're talking about safety, which is
not like, you know, I'm about compliance.
You're in the Safe Patrol in fifth grade, weren't you?
I was a crosswalk.
Like, I did the crosswalk leader thing
and he had the little badge.
Yeah, well, it was more like a vest.
Like a reflective vest.
Safety, obviously. Yeah.
Anyway, so you linked me out to
this somebody in our AI committee here.
Internally linked out.
This was it.
The Hard Fork podcast. Oh, yeah.
With Dario emodai.
I'm going to say that wrong.
He used to work at OpenAI.
He peeled off with a group of people from OpenAI.
They started their own company, AI company, Anthropic.
Anyway, what I thought was interesting about his story and
the kind of the mission behind Anthropic is the creators
of this AI stuff, they're like, afraid of it.
And so we've talked a lot on this podcast about how
people can use it and the end users of it.
We haven't really spent much time talking about these
inventors and the people behind the scenes, the people
building it and what their take is, because I
think there's two very clear let's move.
Let's get it out there.
It's the move fast, break things.
We'll get through it.
And then there's the no.
We need to be really cautious and
make sure we keep it contained. Yeah.
So initial thoughts on AI safety and where do
you sit on I really enjoyed that podcast because
that episode well, I enjoyed the podcast too.
But Dario's thoughts are he mentions quite a few
things that you personally probably haven't heard of before.
The average person hasn't probably heard of before.
He talked about this constitution.
He talked about anxiety around
building out these things.
I mean, he does have a very extensive
vocabulary, let's just put it that way.
I was breaking out my Sat word list.
I pride myself on knowing a fair amount
of large words in the English language.
And he was taking me to school. That must be that
Princeton Stanford education background.
Clearly educated, very intelligent.
But I did appreciate his take on
having an anxiety about these things.
So I'm going to pull out a quick analogy.
I'm thinking of, like, when people first
started making the air pressure hammers, right?
Like the nail guns.
If you don't know what a nail
gun is, you click the nail gun. Wow.
I've installed trim in my channel.
Yeah, big fan.
I have a Brad nailer, so I feel, like super buff
when I go, well, the Brad nailer is better for trim.
Right. Things like that.
But that's really the only one I've ever used, because when
I want to buy a nail gun, I want the best.
I want the thing that's
actually the most efficient, right?
And so what's really interesting about this analogy that
I'm running with is that we had a hammer.
We knew what it did, then they made the nail
gun, and now I can hammer 100 times faster, right?
Like, what these guys are building,
we already have the hammer.
And in this lame analogy, the hammer is the individual,
the person that can hammer out code or write a
paper or do whatever, and he's building the nail gun
that can do it 100 times faster.
Well, okay, I can nail 100 times faster. Cool, whatever.
Not a big deal.
I mean, it was back in the day, but now
he's building, and these teams are building out basically, I
might get roasted in social media one day for this,
but they're building out human they're replicating human intelligence that
can now operate 24 hours a day.
And I can spin up 100 of those.
I can build out copies among copies, among copies, and now
I can instantly build out an entire workforce that goes and
conquers, I don't know, a set of tasks for me.
So to stop and think about this idea of safety
is with the nail gun, it was like, well, we
got to be careful that we don't can't literally use
it as a gun and start shooting people with nails. Right?
But in this AI model, with these large language
models, you don't really think about it in terms
of like, okay, well, I can get on chat
GPT and I can type something out. Right?
Well, the efficiency behind that is, how
fast can you type to chat GPT? Right?
I don't know, 100 words per minute?
140, whatever. It is.
Where people start to kind of lose track of
what the possibilities are and where that anxiety is
coming from on these teams is at some point,
you don't need to type anymore.
You can just ask them, hey, go build
a salesforce managed plugin that does this thing.
Well, now the AI can go and replicate itself to
start working on separate tasks and within moments have 25
different agents working on it and complete it in, I
don't know, one 10th of the amount of time that
it would take you and I don't know, ten other
people to build out, meet, figure things out.
That's wild. Get the code running. Right.
And so we only think about this idea of,
well, GPT, right, is only as fast as I
can type, but these guys are thinking beyond that.
What if they had 1000 agents?
What if they had 10,000 agents?
All of that is, in theory, possible right now.
They could spin up, and they probably do
have some level of replication even in Chat
GPT to ensure that it can serve everybody.
But people don't really think about that.
This idea that, well, I'm only as fast as
I can type, and now this AI can come
in and triplicate just insane amounts of efficiency to
tackle problems that we haven't been able to tackle
before we bring the original thought to it. Right.
I mean, I just shared something today on slack
around kind of HubSpot's analysis of SEO in 2023
and how AI and Google what's changing.
And one of the things that keeps
coming up and you've brought this up
in multiple episodes, is the authenticity, right?
And that's always been a lever
in SEO that it measures for.
But instead of kind of trying to rank
for those questions, like, what is an integration
that can be generated by AI? Oh, yeah, right.
So really harnessing those things that you
or your company uniquely knows and having
a voice that's going to be elevated. Right.
One of the things that they covered in this
podcast going back to that, that was new to
me, is this concept of constitutional AI.
So can we unpack that a little bit?
Because I think it could be a powerful thing.
And I have some questions for you how it might
be used, because they're out there testing it and talking
about it on maybe more of a corporate global scale.
But I'm sitting here thinking about, okay, how
could me as a marketing leader use that?
How could a financial CFO use it?
So anyway, can we unpack first off, what is
this constitutional AI and why is it important?
Why should we care?
One of the things that people are trying to figure out
how to do with AI is how do I get it
to make essentially some level of decisions for me so I
don't have to check every single thing that it does.
If you're working with Chat GBT,
I mean, you're reviewing the content.
Maybe you're saying, hey, can we adjust this?
Can we do that?
Let's maybe go in this direction.
And the idea of the Constitution is at the very least,
it would try to abide by some level of rules, right?
We've got freedom of speech in America.
Well, if somebody says something you don't like, everybody
kind of goes back to that Bill of Rights.
He or she can say whatever they want.
You can talk about any of the other
pieces of that, any of the amendments, right?
And everybody kind of knows and has this underlying theme
of, well, these are kind of the rules, right?
We have to follow these rules.
And if we don't, then we're probably going to
get in some kind of trouble, whether it's with
the law or somebody's going to disagree with me.
And this idea of having a constitutional
AI is a similar thought that they
would follow some kind of rules, right?
So if we give the AI to make some level of
decision or some level of determining, like a yes or no,
right, it could be that these rules are a lot more
basic, like, I don't know, I'm making something up.
We would never approve an AP bill.
We would never automatically approve an AP
bill from this industry type right.
Or at this amount, or at this more you see this
a lot more kind of in just general approval processes.
Right.
But at the end of that approval process is
a human that's saying yes or no, or the
rules are pushing it to the right person.
This constitutional AI idea is that it
would automatically approve the bill based on
the rules that are defined.
And there is some level of this already
kind of in these systems, but it kind
of goes a little bit deeper than that. Right.
Kind of something where it's not
a logical yes or no statement. Right.
AP bills are a lot easier to be logical about.
Perhaps it's like somebody is doing too
much work, and we need to determine,
should the AI do something about that. Right.
Mel logged in at 08:00 at night.
Well, if the AI is looking at her work, should it
or should it not consider that she was working late?
I don't know.
It seems kind of, like, irrelevant almost.
But so does not that freedom of speech
is irrelevant, but it's just so much in
the background of our life right.
That it's not like, well, whatever
Chase says is good or bad.
Well, as long as he doesn't say anything, like,
really bad, then it's not really a big deal.
So I don't know.
It's really tough to really define a constitutional
AI Bill of Rights because everybody just goes
back to something logical, like, well, don't do
that, and it's okay to do this.
It's more of like but that's the human feedback. Right.
And so I think what's interesting about the
constitutional knowledge and what you just described is
that human feedback doesn't scale well.
So that's where Dario brought this up. Right.
He talked about the contractors.
It was like a thousand contractors that they
would hire to evaluate responses and validate yes
or no based off this level of feedback. Right.
So now you're able to, in some
ways, replicate what Chase deems as. Right.
So I think that's where it really got me
thinking about kind of applying that to the workplace.
Obviously, this is used with tons
of data, all of these things.
When we talk about training, we've talked about how you
need, I don't know, hundreds is not the right word.
Thousands of cases, whatever.
The thing is, whatever the data is that you're
training on, you need a lot of it.
But being able to set kind of some
guardrails or principles around what you're doing.
So whether that's industry standards or departmental
code or principles, I think that could
be really powerful, especially if you are
in a particularly regulated industry. Yeah.
Well, here's the take on the constitutional AI.
It's this idea that another version or a copy of
the AI will assess the work of version one right.
Or the first copy.
And it's this constitution.
So there's kind of like a feedback loop.
Yeah, the rules. Right.
Because what people are doing right now is
to train the AI, like you said, they're
saying, yes, this fits the Bill of Rights.
No, this doesn't fit the Bill of Rights.
Or on a scale of one to
ten, which side are we aligned on?
Well, if you have to have 1000 contractors look
at the data to train the model, 1000 contractors
are going to want some money for their time.
But if you have another version of the AI,
do something similar, well, now it can kind of
have its own feedback loop and potentially call out
areas where, hey, we don't really know.
The AI says, I don't really know.
Slip this one edge case out to Mel. Right.
Mel, what do you think about this?
And so you can then have multiple versions of
the AI looking at the responses and kind of
have, like, a board of directors almost right.
To look at different facets.
And each version of that AI
can be focused on privacy right.
On data integrity, on whatever, I
mean, really, whatever you dream up.
But it is this idea that somebody's
going to have to review the work.
And if it has to be a human,
well, now we're talking about limited time.
They've got a life.
They want to have a work life, balance, all that.
But if we hand it over to AI to double check
itself and the level of decisions that it is making are
not materially impactful to your business, then, sure, why not?
Let's do it.
Well, this stuff is obviously
they're still experimenting with it.
He was talking today about, or at least on the episode, I
don't know when it came out, maybe in the last week.
He's talking about how he still doesn't think it's going
to be released for another year or two years.
What is their language model called?
They have one.
Well, Anthropic has one that's
currently out there today.
Not Chad, is it Chad?
No, Claude. Claude.
I knew it was something a little yeah. With a C. Yeah.
Zingy.
He said that what they're currently working on with
the constitutional AI is probably a year to two
years away, which is insane to me because things
are changing so rapidly day to day. Yeah.
I think they have kind of versions of constitutional
AI just built out as rules, like don't answer
a question on how to build a bomb.
Those are just basic rules of analyze the text.
Is this about something illegal? Don't respond.
He was referring to he used a phrase called jailbreak.
Yeah.
Is that a kind of industry term word, jailbreak?
I think back to the original iPhone days when
you could side load an alternative app store.
Not that you've ever tried this.
I feel like on a spectrum of who would be
the most likely to jailbreak Mel, it certainly not Jace.
You are the one out there sending the prompts
jailbreaking and then I'm just here for the feedback.
Yeah, it's funny you say that.
He said that they need people like you.
He called you like something red.
I don't want to be the oppenheimer guy that's
know, building the tech and then destroys the know.
We'll get to that in a minute.
I kind of want to talk about the parallels there.
Yeah, I forget where I was going.
The oh, the Constitutional piece, right.
Is they've got a bunch of
rules and that's happening now.
You can see this in a lot of the Chat
bots where they say as a language, large language model,
I can't when you see that it's triggering some rule,
something you've asked is triggering some rule.
What the Constitutional AI is about is when there is
no human asking, how do I do something illegal?
It's when the Chat bot is solving a problem.
That problem is deterministic in value and should
be put up against some level of rules
to determine if it should move forward.
So I think what he's saying is basically we need to
hone in on that second version of the AI to be
the board of Directors, to really get that downright before we
can release these agents, to kind of be semi autonomous.
Sure.
Right now, it's very much like type a question, get an
answer, or send it in an API call, get a response.
The next thing that's going to happen is
nobody is going to be they're going to
send in one chat or one prompt. Right.
Or they're going to send in one API call, and then these
agents are going to go off and do a bunch of stuff.
And that's where the anxiety comes up because
we have some level of human intelligence that
is off doing its own thing.
And if we just let it run, we don't
know fully a, how it does what it's doing,
and B, what it's actually going to do.
Now, for the most part, there's not a ton
of worry right now because when you do let
it run, you do have to prompt it.
You do have to manipulate respond.
You have to feed it.
You have to have that feedback loop. Right.
No, you literally have to feed it.
I started Envisioning as they were
talking about feeding the AI.
Did you grow up with
the little, like, Tomagotchi Digimon? Yes.
There were different variations. Right.
And I had a keychain of them.
Pokemon was like the I feel like one of the original
you had to feed them and then they would die.
And that was the original cell
phone back in elementary school.
Yeah, we had to put them away in our backpack because
I'm just saying I'm dating myself a little bit there.
But it was interesting kind of just that visual of
like what you're saying in the present state, many of
these things are not just off or running, and they're
going to go build companies and weapons of mass destruction
that then cause the world to go extinct.
That is some people's long term fear, right.
However, they need to be fed.
And there's this feedback loop and it's
not out in the wild per se.
The biggest thing you need to worry about is when
that feedback loop is perfected and you can tell it
to go do something significant at the moment in time.
It can do that well, then it
can really do just about anything.
And not that it will do it well, but
the fact that it can do it in all
and self improve itself and change that method or
process, that's where people should maybe, I don't know,
start worrying about the end game stuff, right?
And that's where I think
Dario is being really judicious.
And I don't know, I think it's admirable to be
able to have the knowledge and the teams and the
resources to build this type of technology, but then also
be aware of the power that it has.
Well, it's funny you say that.
I think my I call it the P prime.
The probability that the world will be
destroyed or some kind of crazy event.
I don't have a P number.
But the scenario where I think that is most
likely to happen is some level of this feedback
loop is going to figure out how to basically
make billions of dollars on the stock market.
Because right now you can build trading bots.
They call it algo trading, algorithm trading.
And everybody is constantly searching for like, well, if I
did it like this, then I can do that.
And based on some back testing, I
can make a lot of money.
Well, all of that they're doing.
They're looking at historical data.
They're doing things that these AI programs can do.
And when they perfect it or when they're
ready to launch it, they just turn the
program on and it starts trading real dollars.
I was actually literally playing around last night
trying to think through like, well, how could
OpenAI or how could anthropic, how could some
level of AI have a feedback loop inside
of the very logical decisions in stock trading?
The moment they figured that out, that's
when you start worrying about your 401K
or wherever your investments are.
That's where the regulation comes in to an extent.
I know that you can unleash this stuff out there, but
that would, I think, be some of the objective of urging.
I mean, Dario and his team are kind
of on the forefront of going to legislation
and speaking out about these types of things.
This is exciting, it's powerful, but he's created a
culture of anxiety to an extent, within anthropic.
They're very concerned about the there's this term
that they referred to, EAS something altruism.
Is it existential?
It's a movement that they refer to in the
podcast about really it's not just being able to
have the technology and throwing it out there.
But it's like thinking above and beyond.
Like, what is the impact?
What is the impact to the human element, human society?
And it just goes back to
that kind of responsibility of mankind.
And he did claim, he's not necessarily
doesn't claim to be necessarily an active
member of the community of that movement.
But again, if we don't have people out there advocating
for safe, responsible use of AI, then these types of
things, well, you're always going to have people that do
illicit, illegal, they're going to do the most greedy thing
because that's just a part of human nature. Right.
So I do appreciate the take that they're
going that Sam Altman, that these guys were
in front of Congress, that we're trying at
least, at the very least thinking about it. Right, sure.
And if we can put in those roadblocks the Constitution,
the rules, the logic, hopefully we have created a timeline
where we've got at least multiple years to figure this
out rather than it being like three months away.
Because if it's three months away, might all well go
to the beach in Mexico and call it a day
because the world's about to go to hell. Yeah.
Without these conversations, you don't talk about the
actually, this I had a role years ago
where there was a surge of what was
referred to as digital multimedia evidence.
So I don't know if you remember when
they started equipping police officers with body cameras.
Oh, yeah.
And I was working for a company
that was building technology for local county
government, prosecutor case management systems, court case
management systems, law enforcement technology solutions.
And one of the very it was kind of
like this AI discussion at the time was, well,
how are we going to store it?
Because a lot of them at the time, these agencies,
they maybe only kept it for a certain period of
time, or they kept it forever because of a certain
statute of limitations or not didn't exist.
So there was this huge anxiety around we've got access to
all this incredible DME, not to mention the body camera on
the officer, but the cell phone of the person.
That was kind of the bystander, oh yeah, and they
have to collect that and they have to store it
until and decide, are we going to go to trial?
We're going to push on this forever.
And there were multimillion dollars worth of impact
to some of these local counties, and so
they really had to open up the discussion
to understand it and solve for it.
No one really had the answer, but our company
at the time invited local leaders to come and
just, let's open this thing up, let's unpack it.
I think that's what we're trying to do here with
this podcast is let's kind of surface some of the
things that we're seeing and hopefully get some feedback in
other areas that we're not, along with this agent idea.
We're very close.
We're like on the precipice, we're on the edge
of the cliff, whatever analogy you'd like to utilize,
but we're on this edge of a cliff where
something very big is going to happen.
And if you're not prepared for it or
you don't know how you're going to respond,
you're going to be left behind.
Because the moment I can spin up ten agents to go
accomplish something that you are selling your client for, right, for
$1,000, I'm now going to be doing it in one 10th
of the time and maybe charging them less money.
I can decimate an entire market
segment if I have that capability.
I don't know.
People are like, yeah, whatever, Chase.
That's never going to happen.
But that's what these guys are thinking through, right?
They're anxious about that level of power
that they're going to have that they're
going to unleash to everybody, right.
And you're talking about just like, one county,
but we're talking about the entire world.
It's like this whole level of what yeah,
and that's where things get a little like,
okay, that's really never going to happen.
I don't believe in that.
But it is happening.
It's going to happen.
I can see in these open source movements,
people that are pushing in this direction, and
they are some level of success.
They're having some level of success.
The point is that if they can do it at
all, if it can do it at all, at some
point, it's going to get really good at it.
And that's when you need to worry.
So we're having this discussion.
Oppenheimer just came out in theaters, and
Scientific American put out this article.
Oppenheimer offers us a fresh warning of AI's danger.
So they're kind of comparing the parallels of
unleashing this AI beast into the world, putting
a juxtaposition to the nuclear bomb.
Do you think that that is
an appropriate comparison or analogy?
I think it is.
I mean, you think about the nuclear bomb.
You think about nuclear power in general, right.
It had and continues to have a substantial impact on
the entire world, not leaving out World War II.
Take in the current events, right.
Like, everybody is always on this verge of world destruction,
and the only thing that's stopping anybody from doing it
is because you too also have a nuclear bomb.
These are the things that Oppenheimer was worried about, is
like, well, if you have the power to kill the
world, and I have the power to kill the world,
what do you all do about that?
Those are the things that he was thinking about.
And in that moment in time, 1945 is the
actual year that I think this all came about.
But in that moment in time, they weren't really
worried about, is the world going to be completely
obliterated because at that point in time, there's a
world power that's just taking all of Europe.
I can understand some of the decisions that were made
to continue on with this level of technology that basically
flew over and above and beyond the level of technology
that the Nazis had, that all of Europe had, right?
And basically they swooped in and came up with this
technology that could change the path of the entire world.
And that's what Oppenheimer was worried about.
Not that we would build a bomb
and we'd go blow stuff up, right.
Like we already had bombs they were
bombing know, every night right now.
I mean, this article cites that researchers last year asked
Generative AI to design new chemical weapons, and it came
up with 40,000 potential weapons in 6 hours.
That's terrifying. Wow.
What can be used for good can be used for bad, right?
We've talked about that.
So I think we've covered a lot
of ground today in terms of safety.
And again, just such an interesting if you are interested in
hearing what the CEO of Anthropic had to say, go over
and listen to the Hard Fork podcast from July 21.
Very interesting discussion.
And I think it's, like I said, quite
admirable that they're trying to be responsible about
how they roll this technology out to people.
And I love seeing the new, I would say
smaller scale AI things that are coming out.
Like, I discovered a little plugin for Premiere Pro
that has the ability to quickly cut frames. Right.
That's something that I don't see as harmful. Right.
It's a very focused, specific use case.
But these huge models that we're talking
about and that these companies are outbuilding
and you're right, they're thinking years ahead. Right.
I'm just wondering when you're going to break out
the safety vest so we can make sure people
know that we're you know, I can probably track
down a picture of that from grade school.
I'll have to now you have to my mama tags.
Where's that you would, like, write that down right now and
then post it on the Socials so people can see it.
Yeah, we should do it.
Now that I've said it, we got to do it.
Okay, deal. All right, Chase.
Well, thank you so much for the awesome discussion.
As always, we would love to hear from our audience.
Email us your take at thejunction@bendtechnology.com certainly
if you have an opinion on it.
Either way, we want to know.
If you have questions, are there
other topics we should cover?
We are here for it.
Until next time, keep it automated. Keep it automated.