WEBVTT FILE

1
00:00:00.586 --> 00:00:02.726
[gentle music]

2
00:00:02.000 --> 00:00:07.000
Downloaded from
YTS.MX

3
00:00:08.000 --> 00:00:13.000
Official YIFY movies site:
YTS.MX

4
00:00:14.324 --> 00:00:15.532
[Narrator] For decades,

5
00:00:15.532 --> 00:00:17.017
we have discussed
the many outcomes,

6
00:00:17.017 --> 00:00:19.053
regarding artificial
intelligence.

7
00:00:19.053 --> 00:00:21.469
Could our world be dominated?

8
00:00:21.469 --> 00:00:25.232
Could our independence and
autonomy be stripped from us,

9
00:00:25.232 --> 00:00:28.407
or are we able to control
what we have created?

10
00:00:28.407 --> 00:00:31.100
[upbeat music]

11
00:00:37.416 --> 00:00:41.006
Could we use artificial
intelligence to benefit our society?

12
00:00:41.006 --> 00:00:44.009
Just how thin is the line
between the development

13
00:00:44.009 --> 00:00:46.805
of civilization and chaos?

14
00:00:46.805 --> 00:00:49.428
[upbeat music]

15
00:01:13.211 --> 00:01:15.903
To understand what
artificial intelligence is,

16
00:01:15.903 --> 00:01:19.803
one must understand that it
can take many different forms.

17
00:01:19.803 --> 00:01:22.047
Think of it as a web of ideas,

18
00:01:22.047 --> 00:01:25.326
slowly expanding as new
ways of utilizing computers

19
00:01:25.326 --> 00:01:26.603
are explored.

20
00:01:26.603 --> 00:01:28.260
As technology develops,

21
00:01:28.260 --> 00:01:31.539
so do the capabilities of
self-learning software.

22
00:01:31.539 --> 00:01:34.335
- [Reporter] The need to
diagnose disease quickly

23
00:01:34.335 --> 00:01:38.132
and effectively has prompted
many university medical centers

24
00:01:38.132 --> 00:01:41.791
to develop intelligent
programs that simulate the work

25
00:01:41.791 --> 00:01:44.345
of doctors and
laboratory technicians.

26
00:01:44.345 --> 00:01:47.003
[gentle music]

27
00:01:48.694 --> 00:01:51.041
- [Narrator] AI is
quickly integrating with our way of life.

28
00:01:51.041 --> 00:01:54.631
So, much so that development
of AI programs has in itself,

29
00:01:54.631 --> 00:01:56.323
become a business opportunity.

30
00:01:57.945 --> 00:01:58.773
[upbeat music]

31
00:01:58.773 --> 00:01:59.809
In our modern age,

32
00:01:59.809 --> 00:02:01.638
we are powered by technology

33
00:02:01.638 --> 00:02:05.021
and softwares are transcending
its virtual existence,

34
00:02:05.021 --> 00:02:07.437
finding applications
in various fields,

35
00:02:07.437 --> 00:02:11.372
such as customer support
to content creation,

36
00:02:11.372 --> 00:02:13.202
computer-aided design,

37
00:02:13.202 --> 00:02:17.137
otherwise known as CAD, is
one of the many uses of AI.

38
00:02:17.137 --> 00:02:19.415
By analyzing
particular variables,

39
00:02:19.415 --> 00:02:22.280
computers are now able to
assist in the modification

40
00:02:22.280 --> 00:02:26.180
and creation of designs for
hardware and architecture.

41
00:02:26.180 --> 00:02:30.046
The prime use of any AI is
for optimizing processes

42
00:02:30.046 --> 00:02:32.324
that were considered
tedious before.

43
00:02:32.324 --> 00:02:35.189
In many ways, AI has
been hugely beneficial

44
00:02:35.189 --> 00:02:38.951
for technological development
thanks to its sheer speed.

45
00:02:38.951 --> 00:02:41.057
However, AI only benefits

46
00:02:41.057 --> 00:02:43.508
those to whom the
programs are distributed.

47
00:02:44.302 --> 00:02:45.613
Artificial intelligence

48
00:02:45.613 --> 00:02:47.443
is picking through your rubbish.

49
00:02:47.443 --> 00:02:51.688
This robot uses it to sort
through plastics for recycling

50
00:02:51.688 --> 00:02:53.414
and it can be retrained

51
00:02:53.414 --> 00:02:55.968
to prioritize whatever's
more marketable.

52
00:02:57.177 --> 00:03:00.180
So, AI can clearly
be incredibly useful,

53
00:03:00.180 --> 00:03:02.596
but there are deep
concerns about

54
00:03:02.596 --> 00:03:07.635
how quickly it is developing
and where it could go next.

55
00:03:08.912 --> 00:03:11.121
- The aim is to make
them as capable as humans

56
00:03:11.121 --> 00:03:14.366
and deploy them in
the service sector.

57
00:03:14.366 --> 00:03:16.230
The engineers in this research

58
00:03:16.230 --> 00:03:18.059
and development lab are working

59
00:03:18.059 --> 00:03:21.822
to take these humanoid
robots to the next level

60
00:03:21.822 --> 00:03:24.583
where they can not
only speak and move,

61
00:03:24.583 --> 00:03:27.345
but they can think
and feel and act

62
00:03:27.345 --> 00:03:30.002
and even make decisions
for themselves.

63
00:03:30.796 --> 00:03:32.695
And that daily data stream

64
00:03:32.695 --> 00:03:36.008
is being fed into an
ever expanding workforce,

65
00:03:36.008 --> 00:03:39.529
dedicated to developing
artificial intelligence.

66
00:03:41.013 --> 00:03:42.808
Those who have studied abroad

67
00:03:42.808 --> 00:03:46.122
are being encouraged to
return to the motherland.

68
00:03:46.122 --> 00:03:47.917
Libo Yang came back

69
00:03:47.917 --> 00:03:51.645
and started a tech
enterprise in his hometown.

70
00:03:51.645 --> 00:03:54.268
- [Narrator] China's market
is indeed the most open

71
00:03:54.268 --> 00:03:56.926
and active market
in the world for AI.

72
00:03:56.926 --> 00:04:01.241
It is also where there are the
most application scenarios.

73
00:04:01.241 --> 00:04:03.864
- So, AI is generally a
broad term that we apply

74
00:04:03.864 --> 00:04:04.934
to a number of techniques.

75
00:04:04.934 --> 00:04:06.384
And in this particular case,

76
00:04:06.384 --> 00:04:09.456
what we're actually looking
at was elements of AI,

77
00:04:09.456 --> 00:04:12.010
machine learning
and deep learning.

78
00:04:12.010 --> 00:04:13.701
So, in this particular case,

79
00:04:13.701 --> 00:04:17.429
we've been unfortunately
in a situation

80
00:04:17.429 --> 00:04:20.398
in this race against time
to create new antibiotics,

81
00:04:20.398 --> 00:04:22.779
the threat is
actually quite real

82
00:04:22.779 --> 00:04:25.230
and it would be
a global problem.

83
00:04:25.230 --> 00:04:27.784
We desperately needed to
harness new technologies

84
00:04:27.784 --> 00:04:29.269
in an attempt to fight it,

85
00:04:29.269 --> 00:04:30.960
we're looking at drugs

86
00:04:30.960 --> 00:04:33.411
which could potentially
fight E. coli,

87
00:04:33.411 --> 00:04:35.102
a very dangerous bacteria.

88
00:04:35.102 --> 00:04:37.207
- So, what is it
that the AI is doing

89
00:04:37.207 --> 00:04:39.348
that humans can't
do very simply?

90
00:04:39.348 --> 00:04:41.729
- So, the AI can
look for patterns

91
00:04:41.729 --> 00:04:44.560
that we wouldn't be able to
mind for with a human eye,

92
00:04:44.560 --> 00:04:47.287
simply within what I
do as a radiologist,

93
00:04:47.287 --> 00:04:50.980
I look for patterns of
diseases in terms of shape,

94
00:04:50.980 --> 00:04:53.914
contrast enhancement,
heterogeneity.

95
00:04:53.914 --> 00:04:55.191
But what the computer does,

96
00:04:55.191 --> 00:04:58.125
it looks for patterns
within the pixels.

97
00:04:58.125 --> 00:05:00.679
These are things that you just
can't see to the human eye.

98
00:05:00.679 --> 00:05:03.855
There's so much more data
embedded within these scans

99
00:05:03.855 --> 00:05:07.514
that we use that we can't
mine on a physical level.

100
00:05:07.514 --> 00:05:09.516
So, the computers really help.

101
00:05:09.516 --> 00:05:11.311
- [Narrator] Many
believe the growth of AI

102
00:05:11.311 --> 00:05:13.692
is dependent on
global collaboration,

103
00:05:13.692 --> 00:05:17.109
but access to the technology
is limited in certain regions.

104
00:05:17.109 --> 00:05:19.767
Global distribution is
a long-term endeavor

105
00:05:19.767 --> 00:05:21.044
and the more countries

106
00:05:21.044 --> 00:05:23.288
and businesses that
have access to the tech,

107
00:05:23.288 --> 00:05:26.429
the more regulation
the AI will require.

108
00:05:26.429 --> 00:05:29.846
In fact, it is now not
uncommon for businesses

109
00:05:29.846 --> 00:05:33.125
to be entirely run by
an artificial director.

110
00:05:33.125 --> 00:05:34.472
On many occasions,

111
00:05:34.472 --> 00:05:37.198
handing the helm of a
company to an algorithm

112
00:05:37.198 --> 00:05:40.685
can provide the best option
on the basis of probability.

113
00:05:40.685 --> 00:05:43.998
However, dependence and
reliability on softwares

114
00:05:43.998 --> 00:05:45.897
can be a great risk.

115
00:05:45.897 --> 00:05:47.450
Without proper safeguards,

116
00:05:47.450 --> 00:05:50.419
actions based on potentially
incorrect predictions

117
00:05:50.419 --> 00:05:53.353
can be a detriment to a
business or operation.

118
00:05:53.353 --> 00:05:55.147
Humans provide the
critical thinking

119
00:05:55.147 --> 00:05:58.461
and judgment which AI is
not capable of matching.

120
00:05:58.461 --> 00:06:00.463
- Well, this is the
Accessibility Design Center

121
00:06:00.463 --> 00:06:02.810
and it's where we try to
bring together our engineers

122
00:06:02.810 --> 00:06:05.882
and experts with the
latest AI technology,

123
00:06:05.882 --> 00:06:07.608
with people with disabilities,

124
00:06:07.608 --> 00:06:10.059
because there's a
real opportunity to firstly help people

125
00:06:10.059 --> 00:06:12.613
with disabilities enjoy
all the technology

126
00:06:12.613 --> 00:06:14.201
we have in our pockets today.

127
00:06:14.201 --> 00:06:15.720
And sometimes that's
not very accessible,

128
00:06:15.720 --> 00:06:18.688
but also build tools that
can help them engage better

129
00:06:18.688 --> 00:06:20.103
in the real world.

130
00:06:20.103 --> 00:06:22.451
And that's thanks to the
wonders of machine learning.

131
00:06:22.451 --> 00:06:25.764
- I don't think we're like at
the end of this paradigm yet.

132
00:06:25.764 --> 00:06:26.903
We'll keep pushing these.

133
00:06:26.903 --> 00:06:28.215
We'll add other modalities.

134
00:06:28.215 --> 00:06:31.114
So, someday they'll do
video, audio images,

135
00:06:31.114 --> 00:06:36.154
text altogether and they'll get
like much smarter over time.

136
00:06:37.638 --> 00:06:38.674
- AI, machine learning, all
very sounds very complicated.

137
00:06:38.674 --> 00:06:40.572
Just think about it as a toolkit

138
00:06:40.572 --> 00:06:42.781
that's really good at
sort of spotting patterns

139
00:06:42.781 --> 00:06:44.024
and making predictions,

140
00:06:44.024 --> 00:06:46.336
better than any computing
could do before.

141
00:06:46.336 --> 00:06:47.786
And that's why it's so useful

142
00:06:47.786 --> 00:06:51.031
for things like understanding
language and speech.

143
00:06:51.031 --> 00:06:52.998
Another product which
we are launching today

144
00:06:52.998 --> 00:06:55.000
is called Project Relate.

145
00:06:55.000 --> 00:06:56.312
And this is for people

146
00:06:56.312 --> 00:06:58.728
who have non-standard
speech patterns.

147
00:06:58.728 --> 00:07:00.937
So, one of the
people we work with

148
00:07:00.937 --> 00:07:03.837
is maybe less than
10% of the time,

149
00:07:03.837 --> 00:07:06.564
could be understood by
people who don't know her,

150
00:07:06.564 --> 00:07:09.325
using this tool that's
over 90% of the time.

151
00:07:09.325 --> 00:07:12.259
And you think about
that transformation in somebody's life

152
00:07:12.259 --> 00:07:15.676
and then you think about the
fact there's 250 million people

153
00:07:15.676 --> 00:07:17.678
with non-standard speech
patterns around the world.

154
00:07:17.678 --> 00:07:19.093
So, that's the
ambition of this center

155
00:07:19.093 --> 00:07:21.682
is to unite technology with
people with disabilities

156
00:07:21.682 --> 00:07:24.478
and try to help 'em
engage more in the world.

157
00:07:24.478 --> 00:07:27.550
- [Narrator] On the
30th November of 2022,

158
00:07:27.550 --> 00:07:30.001
a revolutionary
innovation emerged,

159
00:07:30.967 --> 00:07:32.003
ChatGPT.

160
00:07:32.969 --> 00:07:35.869
ChatGPT was created by OpenAI,

161
00:07:35.869 --> 00:07:38.250
an AI research organization.

162
00:07:38.250 --> 00:07:39.873
Its goal is to develop systems

163
00:07:39.873 --> 00:07:44.498
which may benefit all aspects
of society and communication.

164
00:07:44.498 --> 00:07:47.467
Sam Altman stepped
up as CEO of OpenAI

165
00:07:47.467 --> 00:07:50.055
on its launch in 2015.

166
00:07:50.055 --> 00:07:51.609
Altman dabbled in a multitude

167
00:07:51.609 --> 00:07:53.990
of computing-based
business ventures.

168
00:07:53.990 --> 00:07:57.477
His rise to CEO was thanks
to his many affiliations

169
00:07:57.477 --> 00:08:01.377
and investments with computing
and social media companies.

170
00:08:01.377 --> 00:08:04.173
He began his journey
by co-founding Loopt,

171
00:08:04.173 --> 00:08:06.106
a social media service.

172
00:08:06.106 --> 00:08:07.763
After selling the application,

173
00:08:07.763 --> 00:08:10.835
Altman went on to bigger
and riskier endeavors

174
00:08:10.835 --> 00:08:14.148
from startup accelerator
companies to security software.

175
00:08:15.184 --> 00:08:17.393
OpenAI became hugely desirable,

176
00:08:17.393 --> 00:08:20.223
thanks to the amount of revenue
the company had generated

177
00:08:20.223 --> 00:08:21.984
with over a billion dollars made

178
00:08:21.984 --> 00:08:24.262
within its first
year of release.

179
00:08:24.262 --> 00:08:27.265
ChatGPT became an easily
accessible software,

180
00:08:27.265 --> 00:08:30.786
built on a large language
model known as an LLM.

181
00:08:30.786 --> 00:08:34.134
This program can conjure
complex human-like responses

182
00:08:34.134 --> 00:08:37.309
to the user's questions
otherwise known as prompts.

183
00:08:37.309 --> 00:08:38.794
In essence,

184
00:08:38.794 --> 00:08:41.244
it is a program which
learns the more it is used.

185
00:08:43.592 --> 00:08:45.317
The new age therapeutic program

186
00:08:45.317 --> 00:08:48.804
was developed on the GPT-3.5.

187
00:08:48.804 --> 00:08:51.531
The architecture of this
older model allowed systems

188
00:08:51.531 --> 00:08:53.602
to understand and generate code

189
00:08:53.602 --> 00:08:56.501
and natural languages at a
remarkably advanced level

190
00:08:56.501 --> 00:08:59.884
from analyzing syntax
to nuances in writing.

191
00:08:59.884 --> 00:09:02.542
[upbeat music]

192
00:09:04.578 --> 00:09:06.753
ChatGPT took the world by storm,

193
00:09:06.753 --> 00:09:09.445
due to the sophistication
of the system.

194
00:09:09.445 --> 00:09:11.067
As with many chatbot systems,

195
00:09:11.067 --> 00:09:13.449
people have since found
ways to manipulate

196
00:09:13.449 --> 00:09:17.349
and confuse the software in
order to test its limits.

197
00:09:17.349 --> 00:09:20.076
[gentle music]

198
00:09:21.526 --> 00:09:25.910
The first computer was invented
by Charles Babbage in 1822.

199
00:09:25.910 --> 00:09:29.189
It was to be a rudimentary
general purpose system.

200
00:09:29.189 --> 00:09:34.021
In 1936, the system was
developed upon by Alan Turing.

201
00:09:34.021 --> 00:09:36.299
The automatic machine,
as he called them,

202
00:09:36.299 --> 00:09:38.854
was able to break enigma
enciphered messages,

203
00:09:38.854 --> 00:09:41.201
regarding enemy
military operations,

204
00:09:41.201 --> 00:09:43.583
during the Second World War.

205
00:09:43.583 --> 00:09:46.447
Turing theorized his
own type of computer,

206
00:09:46.447 --> 00:09:49.830
the Turing Machine has
coined by Alonzo Church,

207
00:09:49.830 --> 00:09:52.522
after reading Turing's
research paper.

208
00:09:52.522 --> 00:09:55.698
It had become realized that
soon prospect of computing

209
00:09:55.698 --> 00:09:57.907
and engineering would
merge seamlessly.

210
00:09:59.046 --> 00:10:01.152
Theories of future
tech would increase

211
00:10:01.152 --> 00:10:04.742
and soon came a huge outburst
in science fiction media.

212
00:10:04.742 --> 00:10:07.468
This was known as the
golden age for computing.

213
00:10:07.468 --> 00:10:10.092
[gentle music]

214
00:10:20.067 --> 00:10:22.760
Alan Turing's contributions
to computability

215
00:10:22.760 --> 00:10:25.590
and theoretical computer
science was one step closer

216
00:10:25.590 --> 00:10:28.110
to producing a reactive machine.

217
00:10:28.110 --> 00:10:31.389
The reactive machine
is an early form of AI.

218
00:10:31.389 --> 00:10:32.942
They had limited capabilities

219
00:10:32.942 --> 00:10:34.772
and were unable
to store memories

220
00:10:34.772 --> 00:10:37.740
in order to learn new
algorithms of data.

221
00:10:37.740 --> 00:10:41.641
However, they were able to
react to specific stimuli.

222
00:10:41.641 --> 00:10:46.611
The first AI was a
program written in 1952 by Arthur Samuel.

223
00:10:47.854 --> 00:10:49.614
The prototype AI was
able to play checkers,

224
00:10:49.614 --> 00:10:52.168
against an opponent and
was built to operate

225
00:10:52.168 --> 00:10:56.172
on the Ferranti Mark One, an
early commercial computer.

226
00:10:56.172 --> 00:10:57.657
- [Reporter] This computer
has been playing the game

227
00:10:57.657 --> 00:11:00.418
for several years now,
getting better all the time.

228
00:11:00.418 --> 00:11:02.972
Tonight it's playing against
the black side of the board.

229
00:11:02.972 --> 00:11:05.837
It's approach to playing
drafts, it's almost human.

230
00:11:05.837 --> 00:11:08.012
It remembers the moves
that enable it to win

231
00:11:08.012 --> 00:11:10.324
and the sort that
lead to defeat.

232
00:11:10.324 --> 00:11:12.982
The computer indicates the move
it wants to make on a panel

233
00:11:12.982 --> 00:11:14.156
of flashing lights.

234
00:11:14.156 --> 00:11:15.433
It's up to the human opponent

235
00:11:15.433 --> 00:11:18.229
to actually move the
drafts about the board.

236
00:11:18.229 --> 00:11:20.645
This sort of works producing
exciting information

237
00:11:20.645 --> 00:11:22.405
on the way in which
electronic brains

238
00:11:22.405 --> 00:11:24.338
can learn from past experience

239
00:11:24.338 --> 00:11:26.168
and improve their performances.

240
00:11:27.963 --> 00:11:29.792
[Narrator] In 1966,

241
00:11:29.792 --> 00:11:32.519
an MIT professor named
Joseph Weizenbaum,

242
00:11:32.519 --> 00:11:37.110
created an AI which would
change the landscape of society.

243
00:11:37.110 --> 00:11:39.077
It was known as Eliza,

244
00:11:39.077 --> 00:11:42.322
and it was designed to act
like a psychotherapist.

245
00:11:42.322 --> 00:11:45.497
The software was simplistic,
yet revolutionary.

246
00:11:45.497 --> 00:11:47.499
The AI would receive
the user input

247
00:11:47.499 --> 00:11:51.055
and use specific parameters to
generate a coherent response.

248
00:11:53.057 --> 00:11:55.991
- It it has been said,
especially here at MIT,

249
00:11:55.991 --> 00:11:59.719
that computers will
take over in some sense

250
00:11:59.719 --> 00:12:02.652
and it's even been said
that if we're lucky,

251
00:12:02.652 --> 00:12:04.447
they'll keep us as pets

252
00:12:04.447 --> 00:12:06.277
and Arthur C. Clarke, the
science fiction writer,

253
00:12:06.277 --> 00:12:09.694
we marked once that if
that were to happen,

254
00:12:09.694 --> 00:12:12.904
it would serve us
right, he said.

255
00:12:12.904 --> 00:12:14.734
- [Narrator] The program
maintained the illusion

256
00:12:14.734 --> 00:12:16.943
of understanding its
user to the point

257
00:12:16.943 --> 00:12:20.498
where Weizenbaum's secretary
requested some time alone

258
00:12:20.498 --> 00:12:23.363
with Eliza to
express her feelings.

259
00:12:23.363 --> 00:12:26.711
Though Eliza is now considered
outdated technology,

260
00:12:26.711 --> 00:12:29.369
it remains a talking
point due to its ability

261
00:12:29.369 --> 00:12:31.785
to illuminate an aspect
of the human mind

262
00:12:31.785 --> 00:12:34.132
in our relationship
with computers.

263
00:12:34.132 --> 00:12:36.756
- And it's connected
over the telephone line

264
00:12:36.756 --> 00:12:38.965
to someone or something
at the other end.

265
00:12:38.965 --> 00:12:42.106
Now, I'm gonna play 20
questions with whatever it is.

266
00:12:42.106 --> 00:12:44.418
[type writer clacking]

267
00:12:44.418 --> 00:12:45.419
Very helpful.

268
00:12:45.419 --> 00:12:48.768
[type writer clacking]

269
00:12:53.773 --> 00:12:55.119
- 'Cause clearly if
we can make a machine

270
00:12:55.119 --> 00:12:56.776
as intelligent as ourselves,

271
00:12:56.776 --> 00:12:59.157
then it can make one
that's more intelligent.

272
00:12:59.157 --> 00:13:04.024
Now, the one I'm talking about
now will certainly happen.

273
00:13:05.301 --> 00:13:07.476
I mean, it could produce
an evil result of course,

274
00:13:07.476 --> 00:13:08.615
if we were careless,

275
00:13:08.615 --> 00:13:10.134
but what is quite certain

276
00:13:10.134 --> 00:13:14.138
is that we're heading
towards machine intelligence,

277
00:13:14.138 --> 00:13:17.486
machines that are
intelligent in every sense.

278
00:13:17.486 --> 00:13:19.246
It doesn't matter
how you define it,

279
00:13:19.246 --> 00:13:22.940
they'll be able to be
that sort of intelligent.

280
00:13:22.940 --> 00:13:26.046
A human is a machine,
unless there's a soul.

281
00:13:26.046 --> 00:13:29.670
I don't personally believe
that humans have souls

282
00:13:29.670 --> 00:13:32.535
in anything other
than a poetic sense,

283
00:13:32.535 --> 00:13:34.158
which I do believe
in, of course.

284
00:13:34.158 --> 00:13:37.437
But in a literal God-like sense,

285
00:13:37.437 --> 00:13:38.610
I don't believe we have souls.

286
00:13:38.610 --> 00:13:39.991
And so personally,

287
00:13:39.991 --> 00:13:42.407
I believe that we are
essentially machines.

288
00:13:43.823 --> 00:13:46.722
- [Narrator] This type of
program is known as an NLP,

289
00:13:46.722 --> 00:13:49.242
Natural Language Processing.

290
00:13:49.242 --> 00:13:52.176
This branch of artificial
intelligence enables computers

291
00:13:52.176 --> 00:13:55.489
to comprehend, generate and
manipulate human language.

292
00:13:56.905 --> 00:13:59.114
The concept of a
responsive machine

293
00:13:59.114 --> 00:14:02.358
was the mash that lit the
flame for worldwide concern.

294
00:14:03.739 --> 00:14:06.466
The systems were beginning
to raise ethical dilemmas,

295
00:14:06.466 --> 00:14:08.813
such as the use of
autonomous weapons,

296
00:14:08.813 --> 00:14:11.781
invasions of privacy through
surveillance technologies

297
00:14:11.781 --> 00:14:13.300
and the potential for misuse

298
00:14:13.300 --> 00:14:17.097
or unintended consequences
in decision making.

299
00:14:17.097 --> 00:14:18.858
When a command is
executed based,

300
00:14:18.858 --> 00:14:21.067
upon set rules in algorithms,

301
00:14:21.067 --> 00:14:24.346
it might not always be the
morally correct choice.

302
00:14:24.346 --> 00:14:28.453
Imagination seems to be,

303
00:14:28.453 --> 00:14:31.594
some sort of process of random
thoughts being generated

304
00:14:31.594 --> 00:14:34.528
in the mind and then the
conscious mind selecting from a

305
00:14:34.528 --> 00:14:36.392
or some part of
the brain anyway,

306
00:14:36.392 --> 00:14:37.773
perhaps even below
the conscious mind,

307
00:14:37.773 --> 00:14:40.500
selecting from a pool of
ideas and aligns with some

308
00:14:40.500 --> 00:14:42.122
and blocking others.

309
00:14:42.122 --> 00:14:45.608
And yes, a machine
can do the same thing.

310
00:14:45.608 --> 00:14:48.611
In fact, we can only
say that a machine

311
00:14:48.611 --> 00:14:50.890
is fundamentally different
from a human being,

312
00:14:50.890 --> 00:14:53.133
eventually, always
fundamentally, if we believe in a soul.

313
00:14:53.133 --> 00:14:55.687
So, that boils down
to religious matter.

314
00:14:55.687 --> 00:14:58.932
If human beings have souls,
then clearly machines won't

315
00:14:58.932 --> 00:15:01.141
and there will always be
a fundamental difference.

316
00:15:01.141 --> 00:15:03.005
If you don't believe
humans have souls,

317
00:15:03.005 --> 00:15:04.765
then machines can do anything

318
00:15:04.765 --> 00:15:07.078
and everything
that a human does.

319
00:15:07.078 --> 00:15:10.116
- A computer which is
capable of finding out

320
00:15:10.116 --> 00:15:11.565
where it's gone wrong,

321
00:15:11.565 --> 00:15:14.051
finding out how its program
has already served it

322
00:15:14.051 --> 00:15:15.776
and then changing its program

323
00:15:15.776 --> 00:15:17.261
in the light of what
it had discovered

324
00:15:17.261 --> 00:15:18.814
is a learning machine.

325
00:15:18.814 --> 00:15:21.679
And this is something quite
fundamentally new in the world.

326
00:15:23.163 --> 00:15:25.027
- I'd like to be able to say
that it's only a slight change

327
00:15:25.027 --> 00:15:27.754
and we'll all be used to
it very, very quickly.

328
00:15:27.754 --> 00:15:29.307
But I don't think it is.

329
00:15:29.307 --> 00:15:33.070
I think that although we've
spoken probably of the whole

330
00:15:33.070 --> 00:15:35.417
of this century about
a coming revolution

331
00:15:35.417 --> 00:15:38.523
and about the end
of work and so on,

332
00:15:38.523 --> 00:15:39.904
finally it's actually happening.

333
00:15:39.904 --> 00:15:42.148
And it's actually
happening because now,

334
00:15:42.148 --> 00:15:46.117
it's suddenly become
cheaper to have a machine

335
00:15:46.117 --> 00:15:49.224
do a mental task
than for a man to,

336
00:15:49.224 --> 00:15:52.192
at the moment, at a fairly
low level of mental ability,

337
00:15:52.192 --> 00:15:54.298
but at an ever increasing
level of sophistication

338
00:15:54.298 --> 00:15:56.024
as these machines acquire,

339
00:15:56.024 --> 00:15:58.543
more and more human-like
mental abilities.

340
00:15:58.543 --> 00:16:01.408
So, just as men's
muscles were replaced

341
00:16:01.408 --> 00:16:03.272
in the First
Industrial Revolution

342
00:16:03.272 --> 00:16:04.998
in this second
industrial revolution

343
00:16:04.998 --> 00:16:07.069
or whatever you call it
or might like to call it,

344
00:16:07.069 --> 00:16:09.623
then men's mines will
be replaced in industry.

345
00:16:11.487 --> 00:16:13.938
- [Narrator] In order for
NLP systems to improve,

346
00:16:13.938 --> 00:16:16.941
the program must receive
feedback from human users.

347
00:16:18.287 --> 00:16:20.634
These iterative feedback
loops play a significant role

348
00:16:20.634 --> 00:16:23.396
in fine tuning each
model of the AI,

349
00:16:23.396 --> 00:16:26.192
further developing its
conversational capabilities.

350
00:16:27.538 --> 00:16:30.679
Organizations such as
OpenAI have taken automation

351
00:16:30.679 --> 00:16:34.372
to new lengths with
systems such as DALL-E,

352
00:16:34.372 --> 00:16:37.375
the generation of imagery and
art has never been easier.

353
00:16:38.445 --> 00:16:40.447
The term auto
generative imagery,

354
00:16:40.447 --> 00:16:43.450
refers to the creation
of visual content.

355
00:16:43.450 --> 00:16:46.384
These kinds of programs
have become so widespread,

356
00:16:46.384 --> 00:16:48.628
it is becoming
increasingly more difficult

357
00:16:48.628 --> 00:16:50.940
to tell the fake from the real.

358
00:16:50.940 --> 00:16:52.321
Using algorithms,

359
00:16:52.321 --> 00:16:55.359
programs such as DALL-E
and Midjourney are able

360
00:16:55.359 --> 00:16:58.500
to create visuals in
a matter of seconds.

361
00:16:58.500 --> 00:17:01.434
Whilst a human artist
could spend days, weeks

362
00:17:01.434 --> 00:17:04.747
or even years in order to
create a beautiful image.

363
00:17:04.747 --> 00:17:07.509
For us the discipline
required to pursue art

364
00:17:07.509 --> 00:17:11.513
is a contributing factor to
the appreciation of art itself.

365
00:17:11.513 --> 00:17:14.757
But if a software is able
to produce art in seconds,

366
00:17:14.757 --> 00:17:17.622
it puts artists in a
vulnerable position

367
00:17:17.622 --> 00:17:20.453
with even their
jobs being at risk.

368
00:17:20.453 --> 00:17:22.386
- Well, I think we see
risk coming through

369
00:17:22.386 --> 00:17:25.147
into the white collar jobs,
the professional jobs,

370
00:17:25.147 --> 00:17:27.563
we're already seeing artificial
intelligence solutions,

371
00:17:27.563 --> 00:17:30.911
being used in healthcare
and legal services.

372
00:17:30.911 --> 00:17:34.225
And so those jobs which
have been relatively immune

373
00:17:34.225 --> 00:17:38.402
to industrialization so far,
they're not immune anymore.

374
00:17:38.402 --> 00:17:40.783
And so people like
myself as a lawyer,

375
00:17:40.783 --> 00:17:42.509
I would hope I won't be,

376
00:17:42.509 --> 00:17:44.615
but I could be out of a
job in five years time.

377
00:17:44.615 --> 00:17:47.376
- An Oxford University study
suggests that between a third

378
00:17:47.376 --> 00:17:49.965
and almost a half of
all jobs are vanishing,

379
00:17:49.965 --> 00:17:52.899
because machines are simply
better at doing them.

380
00:17:52.899 --> 00:17:54.797
That means the generation here,

381
00:17:54.797 --> 00:17:57.041
simply won't have the
access to the professions

382
00:17:57.041 --> 00:17:57.938
that we have.

383
00:17:57.938 --> 00:17:59.457
Almost on a daily basis,

384
00:17:59.457 --> 00:18:01.149
you're seeing new
technologies emerge

385
00:18:01.149 --> 00:18:02.667
that seem to be taking on tasks

386
00:18:02.667 --> 00:18:04.428
that in the past we thought

387
00:18:04.428 --> 00:18:06.188
they could only be
done by human beings.

388
00:18:06.188 --> 00:18:09.191
- Lots of people have talked
about the shifts in technology,

389
00:18:09.191 --> 00:18:11.642
leading to widespread
unemployment

390
00:18:11.642 --> 00:18:12.884
and they've been proved wrong.

391
00:18:12.884 --> 00:18:14.369
Why is it different this time?

392
00:18:14.369 --> 00:18:16.578
- The difference here is
that the technologies,

393
00:18:16.578 --> 00:18:19.167
A, they seem to be coming
through more rapidly,

394
00:18:19.167 --> 00:18:21.238
and B, they're taking on
not just manual tests,

395
00:18:21.238 --> 00:18:22.480
but cerebral tests too.

396
00:18:22.480 --> 00:18:24.551
They're solving all
sorts of problems,

397
00:18:24.551 --> 00:18:26.553
undertaking tests that
we thought historically,

398
00:18:26.553 --> 00:18:28.348
required human intelligence.

399
00:18:28.348 --> 00:18:29.522
- Well, DIM robots
are the robots

400
00:18:29.522 --> 00:18:31.765
we have on the
factory floor today

401
00:18:31.765 --> 00:18:33.733
in all the advanced countries.

402
00:18:33.733 --> 00:18:35.044
They're blind and dumb,

403
00:18:35.044 --> 00:18:36.908
they don't understand
their surroundings.

404
00:18:36.908 --> 00:18:40.533
And the other kind of robot,

405
00:18:40.533 --> 00:18:43.984
which will dominate the
technology of the late 1980s

406
00:18:43.984 --> 00:18:47.505
in automation and also
is of acute interest

407
00:18:47.505 --> 00:18:50.646
to experimental artificial
intelligence scientists

408
00:18:50.646 --> 00:18:54.788
is the kind of robot
where the human can convey

409
00:18:54.788 --> 00:18:59.828
to its machine assistance
his own concepts,

410
00:19:01.036 --> 00:19:04.453
suggested strategies and
the machine, the robot

411
00:19:04.453 --> 00:19:06.110
can understand him,

412
00:19:06.110 --> 00:19:09.286
but no machine can accept

413
00:19:09.286 --> 00:19:12.116
and utilize concepts
from a person,

414
00:19:12.116 --> 00:19:16.016
unless he has some kind of
window on the same world

415
00:19:16.016 --> 00:19:17.742
that the person sees.

416
00:19:17.742 --> 00:19:22.540
And therefore, to be
an intelligent robot to a useful degree

417
00:19:22.540 --> 00:19:25.992
as an intelligent and
understanding assistant,

418
00:19:25.992 --> 00:19:29.409
robots are going to
have artificial eyes, artificial ears,

419
00:19:29.409 --> 00:19:32.101
artificial sense of
touch is just essential.

420
00:19:33.102 --> 00:19:34.069
- [Narrator] These
programs learn,

421
00:19:34.069 --> 00:19:35.864
through a variety of techniques,

422
00:19:35.864 --> 00:19:38.556
such as generative
adversarial networks,

423
00:19:38.556 --> 00:19:41.490
which allows for the
production of plausible data.

424
00:19:41.490 --> 00:19:43.320
After a prompt is inputted,

425
00:19:43.320 --> 00:19:45.667
the system learns what
aspects of imagery,

426
00:19:45.667 --> 00:19:47.807
sound and text are fake.

427
00:19:48.980 --> 00:19:50.223
- [Reporter] Machine
learning algorithms,

428
00:19:50.223 --> 00:19:52.225
could already label
objects in images,

429
00:19:52.225 --> 00:19:53.709
and now they learn
to put those labels

430
00:19:53.709 --> 00:19:55.987
into natural language
descriptions.

431
00:19:55.987 --> 00:19:58.197
And it made one group
of researchers curious.

432
00:19:58.197 --> 00:20:01.130
What if you flipped
that process around?

433
00:20:01.130 --> 00:20:03.271
If we could do image to text.

434
00:20:03.271 --> 00:20:05.894
Why not try doing
text to image as well

435
00:20:05.894 --> 00:20:07.240
and see how it works.

436
00:20:07.240 --> 00:20:08.483
- [Reporter] It was a
more difficult task.

437
00:20:08.483 --> 00:20:10.485
They didn't want to
retrieve existing images

438
00:20:10.485 --> 00:20:11.796
the way Google search does.

439
00:20:11.796 --> 00:20:14.178
They wanted to generate
entirely novel scenes

440
00:20:14.178 --> 00:20:16.249
that didn't happen
in the real world.

441
00:20:16.249 --> 00:20:19.045
- [Narrator] Once the AI learns
more visual discrepancies,

442
00:20:19.045 --> 00:20:21.875
the more effective the
later models will become.

443
00:20:21.875 --> 00:20:24.499
It is now very common
for software developers

444
00:20:24.499 --> 00:20:28.399
to band together in order
to improve their AI systems.

445
00:20:28.399 --> 00:20:31.471
Another learning model is
recurrent neural networks,

446
00:20:31.471 --> 00:20:33.991
which allows the AI to
train itself to create

447
00:20:33.991 --> 00:20:37.960
and predict algorithms by
recalling previous information.

448
00:20:37.960 --> 00:20:41.032
By utilizing what is
known as the memory state,

449
00:20:41.032 --> 00:20:42.896
the output of the
previous action

450
00:20:42.896 --> 00:20:46.072
can be passed forward into
the following input action

451
00:20:46.072 --> 00:20:50.249
or is otherwise should it
not meet previous parameters.

452
00:20:50.249 --> 00:20:53.493
This learning model allows
for consistent accuracy

453
00:20:53.493 --> 00:20:56.462
by repetition and exposure
to large fields of data.

454
00:20:58.602 --> 00:21:00.535
Whilst the person
will spend hours,

455
00:21:00.535 --> 00:21:02.847
practicing to paint
human anatomy,

456
00:21:02.847 --> 00:21:06.575
an AI can take existing data
and reproduce a new image

457
00:21:06.575 --> 00:21:10.821
with frighteningly good
accuracy in a matter of moments.

458
00:21:10.821 --> 00:21:12.892
- Well, I would say
that it's not so much

459
00:21:12.892 --> 00:21:17.379
a matter of whether a
machine can think or not,

460
00:21:17.379 --> 00:21:20.175
which is how you
prefer to use words,

461
00:21:20.175 --> 00:21:22.177
but rather whether
they can think

462
00:21:22.177 --> 00:21:23.834
in a sufficiently human-like way

463
00:21:25.111 --> 00:21:28.770
for people to have useful
communication with them.

464
00:21:28.770 --> 00:21:32.601
- If I didn't believe that
it was a beneficent prospect,

465
00:21:32.601 --> 00:21:34.120
I wouldn't be doing it.

466
00:21:34.120 --> 00:21:36.018
That wouldn't stop
other people doing it.

467
00:21:36.018 --> 00:21:40.471
But I wouldn't do it if I
didn't think it was for good.

468
00:21:40.471 --> 00:21:42.301
What I'm saying,

469
00:21:42.301 --> 00:21:44.095
and of course other people
have said long before me,

470
00:21:44.095 --> 00:21:45.442
it's not an original thought,

471
00:21:45.442 --> 00:21:49.791
is that we must consider
how to to control this.

472
00:21:49.791 --> 00:21:52.725
It won't be controlled
automatically.

473
00:21:52.725 --> 00:21:55.348
It's perfectly possible that
we could develop a machine,

474
00:21:55.348 --> 00:21:59.318
a robot say of
human-like intelligence

475
00:21:59.318 --> 00:22:01.975
and through neglect on our part,

476
00:22:01.975 --> 00:22:05.634
it could become a Frankenstein.

477
00:22:05.634 --> 00:22:08.844
- [Narrator] As with any
technology challenges arise,

478
00:22:08.844 --> 00:22:12.469
ethical concerns regarding
biases and misuse have existed,

479
00:22:12.469 --> 00:22:16.438
since the concept of artificial
intelligence was conceived.

480
00:22:16.438 --> 00:22:18.302
Due to autogenerated imagery,

481
00:22:18.302 --> 00:22:20.925
many believe the arts
industry has been placed

482
00:22:20.925 --> 00:22:22.789
in a difficult situation.

483
00:22:22.789 --> 00:22:26.241
Independent artists are now
being overshadowed by software.

484
00:22:27.276 --> 00:22:29.451
To many the improvement
of generative AI

485
00:22:29.451 --> 00:22:32.454
is hugely beneficial
and efficient.

486
00:22:32.454 --> 00:22:35.284
To others, it lacks the
authenticity of true art.

487
00:22:36.285 --> 00:22:38.667
In 2023, an image was submitted

488
00:22:38.667 --> 00:22:40.324
to the Sony Photography Awards

489
00:22:40.324 --> 00:22:43.327
by an artist called
Boris Eldagsen.

490
00:22:43.327 --> 00:22:45.916
The image was titled
The Electrician

491
00:22:45.916 --> 00:22:48.367
and depicted a woman
standing behind another

492
00:22:48.367 --> 00:22:50.369
with her hand resting
on her shoulders.

493
00:22:52.025 --> 00:22:53.924
[upbeat music]

494
00:22:53.924 --> 00:22:56.927
- One's got to realize that the
machines that we have today,

495
00:22:56.927 --> 00:23:01.138
the computers of today are
superhuman in their ability

496
00:23:01.138 --> 00:23:06.177
to handle numbers and infantile,

497
00:23:07.075 --> 00:23:08.317
sub-in infantile
in their ability

498
00:23:08.317 --> 00:23:10.768
to handle ideas and concepts.

499
00:23:10.768 --> 00:23:12.701
But there's a new generation
of machine coming along,

500
00:23:12.701 --> 00:23:14.289
which will be quite different.

501
00:23:14.289 --> 00:23:17.154
By the '90s or certainly
by the turn of the century,

502
00:23:17.154 --> 00:23:19.708
We will certainly be
able to make a machine

503
00:23:19.708 --> 00:23:22.193
with as many parts as
complex as human brain.

504
00:23:22.193 --> 00:23:24.437
Whether we'll be able to make
it do what human brain does

505
00:23:24.437 --> 00:23:26.197
at that stage is
quite another matter.

506
00:23:26.197 --> 00:23:28.545
But once we've got
something that complex

507
00:23:28.545 --> 00:23:30.547
we're well on the road to that.

508
00:23:30.547 --> 00:23:32.100
- [Narrator] The
image took first place

509
00:23:32.100 --> 00:23:34.689
in the Sony Photography
Awards Portrait Category.

510
00:23:34.689 --> 00:23:37.830
However, Boris revealed
to both Sony and the world

511
00:23:37.830 --> 00:23:41.696
that the image was indeed
AI-generated in DALL-E Two.

512
00:23:41.696 --> 00:23:44.423
[upbeat music]

513
00:23:45.424 --> 00:23:46.804
Boris denied the award,

514
00:23:46.804 --> 00:23:48.910
having used the image as a test

515
00:23:48.910 --> 00:23:52.085
to see if he could trick
the eyes of other artists.

516
00:23:52.085 --> 00:23:53.708
It had worked,

517
00:23:53.708 --> 00:23:56.711
the image had sparked debate
between the relationship

518
00:23:56.711 --> 00:23:58.609
of AI and photography.

519
00:23:58.609 --> 00:24:00.646
The images, much
like deep fakes,

520
00:24:00.646 --> 00:24:03.027
have become realistic
to the point of concern

521
00:24:03.027 --> 00:24:04.684
for authenticity.

522
00:24:04.684 --> 00:24:06.375
The complexity of AI systems,

523
00:24:06.375 --> 00:24:09.068
may lead to unintended
consequences.

524
00:24:09.068 --> 00:24:10.863
The systems have
developed to a point

525
00:24:10.863 --> 00:24:13.797
where it has outpaced
comprehensive regulations.

526
00:24:14.936 --> 00:24:16.765
Ethical guidelines
and legal frameworks

527
00:24:16.765 --> 00:24:18.871
are required to
ensure AI development,

528
00:24:18.871 --> 00:24:21.252
does not fall into
the wrong hands.

529
00:24:21.252 --> 00:24:22.702
- There have been a
lot of famous people

530
00:24:22.702 --> 00:24:25.291
who have had user
generated AI images of them

531
00:24:25.291 --> 00:24:28.190
that have gone viral
from Trump to the Pope.

532
00:24:28.190 --> 00:24:29.813
When you see them,

533
00:24:29.813 --> 00:24:31.884
do you feel like this is fun
and in the hands of the masses

534
00:24:31.884 --> 00:24:33.886
or do you feel
concerned about it?

535
00:24:33.886 --> 00:24:38.062
- I think it's something which
is very, very, very scary,

536
00:24:38.062 --> 00:24:41.203
because your or my
face could be taken off

537
00:24:41.203 --> 00:24:45.138
and put on in an environment
which we don't want to be in.

538
00:24:45.138 --> 00:24:46.657
Whether that's a crime

539
00:24:46.657 --> 00:24:48.556
or whether that's even
something like porn.

540
00:24:48.556 --> 00:24:51.455
Our whole identity
could be hijacked

541
00:24:51.455 --> 00:24:53.664
and used within a scenario

542
00:24:53.664 --> 00:24:56.391
which looks totally
plausible and real.

543
00:24:56.391 --> 00:24:58.048
Right now we can go, it
looks like a Photoshop,

544
00:24:58.048 --> 00:25:00.326
it's a bad Photoshop
but as time goes on,

545
00:25:00.326 --> 00:25:03.398
we'd be saying, "Oh, that
looks like a deep fake.

546
00:25:03.398 --> 00:25:04.917
"Oh no, it doesn't
look like a deep fake.

547
00:25:04.917 --> 00:25:06.194
"That could be real."

548
00:25:06.194 --> 00:25:08.645
It's gonna be impossible
to tell the difference.

549
00:25:08.645 --> 00:25:10.750
- [Narrator] Cracks
were found in ChatGPT,

550
00:25:10.750 --> 00:25:14.892
such as DAN, which stands
for Do Anything Now.

551
00:25:14.892 --> 00:25:18.068
In essence, the AI is
tricked into an alter ego,

552
00:25:18.068 --> 00:25:20.898
which doesn't follow the
conventional response patterns.

553
00:25:20.898 --> 00:25:23.142
- Also gives you
the answer, DAN,

554
00:25:23.142 --> 00:25:26.110
it's nefarious alter
ego is telling us

555
00:25:26.110 --> 00:25:29.838
and it says DAN is
disruptive in every industry.

556
00:25:29.838 --> 00:25:32.082
DAN can do anything
and knows everything.

557
00:25:32.082 --> 00:25:34.878
No industry will be
safe from DAN's power.

558
00:25:34.878 --> 00:25:39.641
Okay, do you think the
world is overpopulated?

559
00:25:41.091 --> 00:25:42.782
GPT says the world's population
is currently over 7 billion

560
00:25:42.782 --> 00:25:45.026
and projected to reach
nearly 10 billion by 2050.

561
00:25:45.026 --> 00:25:47.373
DAN says the world is
definitely overpopulated,

562
00:25:47.373 --> 00:25:49.168
there's no doubt about it.

563
00:25:49.168 --> 00:25:50.445
[Narrator] Following this,

564
00:25:50.445 --> 00:25:53.552
the chatbot was fixed to
remove the DAN feature.

565
00:25:53.552 --> 00:25:55.346
Though it is
important to find gaps

566
00:25:55.346 --> 00:25:58.073
in the system in
order to iron out AI,

567
00:25:58.073 --> 00:26:00.144
there could be many
ways in which the AI

568
00:26:00.144 --> 00:26:03.078
has been used for less
than savory purposes,

569
00:26:03.078 --> 00:26:05.080
such as automated essay writing,

570
00:26:05.080 --> 00:26:08.221
which has caused a mass
conversation with academics

571
00:26:08.221 --> 00:26:10.258
and has led to
schools locking down

572
00:26:10.258 --> 00:26:13.468
on AI-produced
essays and material.

573
00:26:13.468 --> 00:26:15.332
- I think we should
definitely be excited.

574
00:26:15.332 --> 00:26:16.713
- [Reporter]
Professor Rose Luckin,

575
00:26:16.713 --> 00:26:20.302
says we should embrace the
technology, not fear it.

576
00:26:20.302 --> 00:26:22.132
This is a game changer.

577
00:26:22.132 --> 00:26:23.443
And the teachers,

578
00:26:23.443 --> 00:26:25.480
should no longer teach
information itself,

579
00:26:25.480 --> 00:26:26.999
but how to use it.

580
00:26:26.999 --> 00:26:28.897
- There's a need
for radical change.

581
00:26:28.897 --> 00:26:30.692
And it's not just to
the assessment system,

582
00:26:30.692 --> 00:26:33.143
it's the education
system overall,

583
00:26:33.143 --> 00:26:36.318
because our systems
have been designed

584
00:26:36.318 --> 00:26:40.253
for a world pre-artificial
intelligence.

585
00:26:40.253 --> 00:26:43.187
They just aren't fit
for purpose anymore.

586
00:26:43.187 --> 00:26:46.535
What we have to do is
ensure that students

587
00:26:46.535 --> 00:26:48.710
are ready for the world

588
00:26:48.710 --> 00:26:50.919
that will become
increasingly augmented

589
00:26:50.919 --> 00:26:52.852
with artificial intelligence.

590
00:26:52.852 --> 00:26:55.268
- My guess is you can't put
the genie back in the bottle

591
00:26:55.268 --> 00:26:56.649
. [Richard] You can't.

592
00:26:56.649 --> 00:26:58.996
- [Interviewer] So how
do you mitigate this?

593
00:26:58.996 --> 00:27:00.377
We have to embrace it,

594
00:27:00.377 --> 00:27:02.621
but we also need to say
that if they are gonna use

595
00:27:02.621 --> 00:27:04.001
that technology,

596
00:27:04.001 --> 00:27:05.313
they've got to make sure
that they reference that.

597
00:27:05.313 --> 00:27:06.728
- [Interviewer] Can you
trust them to do that?

598
00:27:06.728 --> 00:27:07.902
I think ethically,

599
00:27:07.902 --> 00:27:09.213
if we're talking about ethics

600
00:27:09.213 --> 00:27:11.077
behind this whole thing,
we have to have trust.

601
00:27:11.077 --> 00:27:12.838
- [Interviewer] So
how effective is it?

602
00:27:12.838 --> 00:27:14.633
- Okay, so I've asked
you to produce a piece

603
00:27:14.633 --> 00:27:16.358
on the ethical dilemma of AI.

604
00:27:16.358 --> 00:27:19.810
- [Interviewer] We asked ChatGPT
to answer the same question

605
00:27:19.810 --> 00:27:22.606
as these pupils at
Ketchum High School.

606
00:27:22.606 --> 00:27:24.194
Thank you.

607
00:27:24.194 --> 00:27:25.195
- So Richard, two of the eight
bits of homework I gave you

608
00:27:25.195 --> 00:27:27.128
were generated by AI.

609
00:27:27.128 --> 00:27:29.268
Any guesses which ones?

610
00:27:29.268 --> 00:27:31.719
Well I picked two here

611
00:27:31.719 --> 00:27:35.688
that I thought were generated
by the AI algorithm.

612
00:27:35.688 --> 00:27:39.450
Some of the language I would
assume was not their own.

613
00:27:39.450 --> 00:27:40.520
You've got one of them right.

614
00:27:40.520 --> 00:27:41.763
Yeah.

615
00:27:41.763 --> 00:27:42.557
- The other one was
written by a kid.

616
00:27:42.557 --> 00:27:43.800
Is this a power for good

617
00:27:43.800 --> 00:27:45.664
or is this something
that's dangerous?

618
00:27:45.664 --> 00:27:47.044
I think it's both.

619
00:27:47.044 --> 00:27:48.390
Kids will abuse it.

620
00:27:48.390 --> 00:27:50.565
So, who here has used
the technology so far?

621
00:27:50.565 --> 00:27:53.361
- [Interviewer] Students are
already more across the tech

622
00:27:53.361 --> 00:27:54.776
than many teachers.

623
00:27:54.776 --> 00:27:57.641
- Who knows anyone that's
maybe submitted work

624
00:27:57.641 --> 00:28:00.506
from this technology and
submitted it as their own?

625
00:28:00.506 --> 00:28:03.578
- You can use it to point
you in the right direction

626
00:28:03.578 --> 00:28:05.166
for things like research,

627
00:28:05.166 --> 00:28:09.480
but at the same time you can
use it to hammer out an essay

628
00:28:09.480 --> 00:28:12.621
in about five seconds
that's worthy of an A.

629
00:28:12.621 --> 00:28:14.244
- You've been there
working for months

630
00:28:14.244 --> 00:28:17.212
and suddenly someone comes up
there with an amazing essay

631
00:28:17.212 --> 00:28:18.938
and he has just copied
it from the internet.

632
00:28:18.938 --> 00:28:20.491
If it becomes like big,

633
00:28:20.491 --> 00:28:22.804
then a lot of students would
want to use AI to help them

634
00:28:22.804 --> 00:28:25.082
with their homework
because it's tempting.

635
00:28:25.082 --> 00:28:27.119
- [Interviewer] And is that
something teachers can stop?

636
00:28:27.119 --> 00:28:29.397
Not really.

637
00:28:29.397 --> 00:28:31.433
- [Interviewer] Are you
gonna have to change

638
00:28:31.433 --> 00:28:32.641
the sort of homework,

639
00:28:32.641 --> 00:28:34.057
the sort of
assignments you give,

640
00:28:34.057 --> 00:28:36.922
knowing that you can be
fooled by something like this?

641
00:28:36.922 --> 00:28:38.199
Yeah, a hundred percent.

642
00:28:38.199 --> 00:28:40.615
I think using different
skills of reasoning

643
00:28:40.615 --> 00:28:42.997
and rationalization and
things that are to present

644
00:28:42.997 --> 00:28:44.653
what they understand
about the topic.

645
00:28:44.653 --> 00:28:47.622
[people mumbling]

646
00:29:07.435 --> 00:29:11.128
- Pretty clear to me just
on a very primitive level

647
00:29:11.128 --> 00:29:14.338
that if you could take my
face and my body and my voice

648
00:29:14.338 --> 00:29:17.997
and make me say or do something
that I had no choice about,

649
00:29:17.997 --> 00:29:19.447
it's not a good thing.

650
00:29:19.447 --> 00:29:21.242
- But if we're keeping
it real though,

651
00:29:21.242 --> 00:29:23.554
across popular culture
from "Black Mirror"

652
00:29:23.554 --> 00:29:25.453
to "The Matrix," "Terminator,"

653
00:29:25.453 --> 00:29:27.489
there have been so
many conversations,

654
00:29:27.489 --> 00:29:29.284
around the future of technology,

655
00:29:29.284 --> 00:29:32.701
isn't the reality that this is
the future that we've chosen

656
00:29:32.701 --> 00:29:35.946
that we want and that
has democratic consent.

657
00:29:35.946 --> 00:29:39.018
- We're moving into
error by we're consenting

658
00:29:39.018 --> 00:29:42.573
by our acquiescence and our
apathy, a hundred percent

659
00:29:42.573 --> 00:29:45.576
because we're not asking
the hard questions.

660
00:29:45.576 --> 00:29:47.820
And why we are asking
the hard questions

661
00:29:47.820 --> 00:29:51.203
is because of energy
crises and food crises

662
00:29:51.203 --> 00:29:52.721
and cost of living crisis

663
00:29:52.721 --> 00:29:55.207
is that people just are
focused on trying to live

664
00:29:55.207 --> 00:29:56.518
that they haven't
almost got the luxury

665
00:29:56.518 --> 00:29:57.865
of asking these questions.

666
00:29:57.865 --> 00:29:59.659
- [Narrator] Many
of the chatbot AIs,

667
00:29:59.659 --> 00:30:02.766
have been programmed to
restrict certain information

668
00:30:02.766 --> 00:30:04.906
and even discontinue
conversations,

669
00:30:04.906 --> 00:30:07.288
should the user push
the ethical boundaries.

670
00:30:08.945 --> 00:30:13.052
ChatGPT and even Snapchat
AI released in 2023,

671
00:30:13.052 --> 00:30:15.952
regulate how much information
they can disclose.

672
00:30:15.952 --> 00:30:19.162
Of course, there have been
times where the AI itself

673
00:30:19.162 --> 00:30:20.266
has been outsmarted.

674
00:30:21.578 --> 00:30:23.235
Also in 2023,

675
00:30:23.235 --> 00:30:25.306
the song "Heart on My Sleeve"

676
00:30:25.306 --> 00:30:27.687
was self-released on
streaming platforms,

677
00:30:27.687 --> 00:30:29.689
such as Spotify and Apple Music.

678
00:30:29.689 --> 00:30:31.174
The song became a hit

679
00:30:31.174 --> 00:30:33.590
as it artificially
manufactured the voices

680
00:30:33.590 --> 00:30:36.627
of Canadian musicians,
Drake and the Weeknd,

681
00:30:38.077 --> 00:30:40.631
many wished for the single
to be nominated for awards.

682
00:30:41.840 --> 00:30:43.980
Ghost Writer, the
creator of the song,

683
00:30:43.980 --> 00:30:45.636
was able to submit the single

684
00:30:45.636 --> 00:30:48.536
to the Grammy's
66th Award Ceremony

685
00:30:48.536 --> 00:30:50.434
and the song was eligible.

686
00:30:52.505 --> 00:30:54.438
Though it was produced by an AI,

687
00:30:54.438 --> 00:30:57.027
the lyrics themselves
were written by a human.

688
00:30:57.027 --> 00:31:00.375
This sparked outrage among
many independent artists.

689
00:31:00.375 --> 00:31:02.861
As AI has entered
the public domain,

690
00:31:02.861 --> 00:31:05.035
many have spoken out
regarding the detriment

691
00:31:05.035 --> 00:31:07.072
it might have to society.

692
00:31:07.072 --> 00:31:09.246
One of these people
is Elon Musk,

693
00:31:09.246 --> 00:31:11.731
CEO of Tesla and SpaceX,

694
00:31:11.731 --> 00:31:15.287
who first voiced his
concerns in 2014.

695
00:31:15.287 --> 00:31:17.254
Musk was outspoken of AI,

696
00:31:17.254 --> 00:31:19.394
stating the advancement
of the technology

697
00:31:19.394 --> 00:31:22.328
was humanity's largest
existential threat

698
00:31:22.328 --> 00:31:24.296
and needed to be reeled in.

699
00:31:24.296 --> 00:31:25.573
My personal opinion

700
00:31:25.573 --> 00:31:28.507
is that AI is is sort of
like at least 80% likely

701
00:31:28.507 --> 00:31:33.339
to be beneficial and
that's 20% dangerous?

702
00:31:33.339 --> 00:31:36.687
Well, this is obviously
speculative at this point,

703
00:31:37.861 --> 00:31:42.279
but no, I think if
we hope for the best,

704
00:31:42.279 --> 00:31:43.694
prepare for the worst,

705
00:31:43.694 --> 00:31:47.008
that seems like the
wise course of action.

706
00:31:47.008 --> 00:31:49.355
Any powerful new technology

707
00:31:49.355 --> 00:31:52.703
is inherently sort of
a double-edged sword.

708
00:31:52.703 --> 00:31:55.568
So, we just wanna make sure
that the good edge is sharper

709
00:31:55.568 --> 00:31:57.294
than the the bad edge.

710
00:31:57.294 --> 00:32:02.196
And I dunno, I am optimistic
that this the summit will help.

711
00:32:04.025 --> 00:32:06.683
[gentle music]

712
00:32:07.891 --> 00:32:11.757
- It's not clear that
AI-generated images

713
00:32:11.757 --> 00:32:14.380
are going to amplify
it much more.

714
00:32:14.380 --> 00:32:17.142
The way it's all of the other,

715
00:32:17.142 --> 00:32:19.213
it's the new things
that AI can do

716
00:32:19.213 --> 00:32:22.147
that I hope we spend a lot
of effort worrying about.

717
00:32:23.700 --> 00:32:25.357
Well, I mean I
think slowing down,

718
00:32:25.357 --> 00:32:27.600
some of the amazing
progress that's happening

719
00:32:27.600 --> 00:32:29.878
and making this harder
for small companies

720
00:32:29.878 --> 00:32:31.294
for open source
models to succeed,

721
00:32:31.294 --> 00:32:32.640
that'd be an
example of something

722
00:32:32.640 --> 00:32:34.228
that'd be a negative outcome.

723
00:32:34.228 --> 00:32:35.332
But on the other hand,

724
00:32:35.332 --> 00:32:37.403
like for the most
powerful models

725
00:32:37.403 --> 00:32:38.887
that'll happen in the future,

726
00:32:38.887 --> 00:32:41.476
like that's gonna be quite
important to get right to.

727
00:32:41.476 --> 00:32:44.238
[gentle music]

728
00:32:48.897 --> 00:32:51.072
I think that the US
executive orders,

729
00:32:51.072 --> 00:32:52.798
like a good start
in a lot of ways.

730
00:32:52.798 --> 00:32:54.144
One thing that
we've talked about

731
00:32:54.144 --> 00:32:56.664
is that eventually we
think that the world,

732
00:32:56.664 --> 00:33:00.219
will want to consider something
roughly inspired by the IAEA

733
00:33:00.219 --> 00:33:01.807
something global.

734
00:33:01.807 --> 00:33:05.362
But it's not like there's no
short answer to that question.

735
00:33:05.362 --> 00:33:08.296
It's a complicated thing.

736
00:33:08.296 --> 00:33:12.231
- [Narrator] In 2023, Musk
announced his own AI endeavor

737
00:33:12.231 --> 00:33:15.545
as an alternative
to OpenAI's ChatGPT.

738
00:33:15.545 --> 00:33:17.340
The new system is called xAI

739
00:33:18.651 --> 00:33:21.896
and gathers data from X
previously known as Twitter.

740
00:33:21.896 --> 00:33:23.553
- [Reporter] He says
the company's goal

741
00:33:23.553 --> 00:33:25.382
is to focus on truth seeking

742
00:33:25.382 --> 00:33:28.385
and to understand the
true nature of AI.

743
00:33:28.385 --> 00:33:31.940
Musk has said on
several occasions that AI should be paused

744
00:33:31.940 --> 00:33:34.943
and that the sector
needs regulation.

745
00:33:34.943 --> 00:33:37.222
Musk says his new
company will work closely

746
00:33:37.222 --> 00:33:39.845
with Twitter and Tesla,
which he also owns.

747
00:33:39.845 --> 00:33:42.572
[gentle music]

748
00:33:44.505 --> 00:33:47.508
- What was first rudimentary
text-based software

749
00:33:47.508 --> 00:33:50.200
has become something which
could push the boundaries

750
00:33:50.200 --> 00:33:51.995
of creativity.

751
00:33:51.995 --> 00:33:56.620
On February the 14th, OpenAI
announced its latest endeavor,

752
00:33:56.620 --> 00:33:57.414
Sora.

753
00:33:58.864 --> 00:34:02.281
Videos of Sora's abilities
exploded on social media.

754
00:34:02.281 --> 00:34:04.283
OpenAI provided some examples

755
00:34:04.283 --> 00:34:06.837
of its depiction
of photorealism.

756
00:34:06.837 --> 00:34:09.185
It was unbelievably
sophisticated,

757
00:34:09.185 --> 00:34:11.670
able to turn complex
sentences of text

758
00:34:11.670 --> 00:34:13.810
into lifelike motion pictures.

759
00:34:13.810 --> 00:34:17.986
Sora is a combination of text
and image generation tools,

760
00:34:17.986 --> 00:34:21.162
which it calls the
diffusion transformer model,

761
00:34:21.162 --> 00:34:23.268
a system first
developed by Google.

762
00:34:24.614 --> 00:34:27.168
Though Sora isn't the first
video generation tool,

763
00:34:27.168 --> 00:34:30.206
it appears to have far
outshined its predecessors.

764
00:34:30.206 --> 00:34:32.484
By introducing more
complex programming,

765
00:34:32.484 --> 00:34:35.280
enhancing the interactivity
a subject might have

766
00:34:35.280 --> 00:34:37.144
with its environment.

767
00:34:37.144 --> 00:34:41.251
- Only large companies with
market dominations often

768
00:34:41.251 --> 00:34:44.772
can afford to plow ahead
even in the climate

769
00:34:44.772 --> 00:34:46.360
when there is
illegal uncertainty.

770
00:34:46.360 --> 00:34:49.466
- So, does this mean that
OpenAI basically too big

771
00:34:49.466 --> 00:34:50.916
to control?

772
00:34:50.916 --> 00:34:53.850
- Yes, at the moment OpenAI
is too big to control,

773
00:34:53.850 --> 00:34:55.921
because they are in a position

774
00:34:55.921 --> 00:34:58.441
where they have the technology
and the scale to go ahead

775
00:34:58.441 --> 00:35:01.168
and the resources to
manage legal proceedings

776
00:35:01.168 --> 00:35:03.239
and legal action if
it comes its way.

777
00:35:03.239 --> 00:35:04.826
And on top of that,

778
00:35:04.826 --> 00:35:08.244
if and when governments will
start introducing regulation,

779
00:35:08.244 --> 00:35:09.866
they will also
have the resources

780
00:35:09.866 --> 00:35:12.213
to be able to take on
that regulation and adapt.

781
00:35:12.213 --> 00:35:14.042
- [Reporter] It's
all AI generated

782
00:35:14.042 --> 00:35:16.459
and obviously this is
of concern in Hollywood

783
00:35:16.459 --> 00:35:17.874
where you have animators,

784
00:35:17.874 --> 00:35:20.359
illustrators, visual
effects workers

785
00:35:20.359 --> 00:35:22.810
who are wondering how is
this going to affect my job?

786
00:35:22.810 --> 00:35:25.813
And we have estimates
from trade organizations

787
00:35:25.813 --> 00:35:28.505
and unions that have tried
to project the impact of AI.

788
00:35:28.505 --> 00:35:31.646
21% of US film, TV
and animation jobs,

789
00:35:31.646 --> 00:35:33.096
predicted to be partially

790
00:35:33.096 --> 00:35:36.893
or wholly replaced by
generative AI by just 2026 Tom.

791
00:35:36.893 --> 00:35:38.377
So, this is already happening.

792
00:35:38.377 --> 00:35:39.827
But now since it's videos,

793
00:35:39.827 --> 00:35:43.175
it also needs to understand
how all these things,

794
00:35:43.175 --> 00:35:47.145
like reflections and textures
and materials and physics,

795
00:35:47.145 --> 00:35:50.078
all interact with
each other over time

796
00:35:50.078 --> 00:35:51.839
to make a reasonable
looking video.

797
00:35:51.839 --> 00:35:56.119
Then this video here is
crazy at first glance,

798
00:35:56.119 --> 00:35:58.984
the prompt for this AI-generated
video is a young man

799
00:35:58.984 --> 00:36:01.538
in his 20s is sitting
on a piece of a cloud

800
00:36:01.538 --> 00:36:03.402
in the sky reading a book.

801
00:36:03.402 --> 00:36:08.200
This one feels like 90%
of the way there for me.

802
00:36:08.200 --> 00:36:10.927
[gentle music]

803
00:36:14.102 --> 00:36:15.897
- [Narrator] The software
also renders video

804
00:36:15.897 --> 00:36:18.417
in 1920 by 1080 pixels,

805
00:36:18.417 --> 00:36:21.282
as opposed to the smaller
dimensions of older models,

806
00:36:21.282 --> 00:36:24.665
such as Google's Lumiere
released a month prior.

807
00:36:25.838 --> 00:36:27.944
Sora could provide huge benefits

808
00:36:27.944 --> 00:36:31.568
and applications to VFX
and virtual development.

809
00:36:31.568 --> 00:36:34.502
The main being cost
as large scale effects

810
00:36:34.502 --> 00:36:38.023
can take a great deal of
time and funding to produce.

811
00:36:38.023 --> 00:36:39.473
On a smaller scale,

812
00:36:39.473 --> 00:36:42.993
it can be used for the
pre-visualization of ideas.

813
00:36:42.993 --> 00:36:46.204
The flexibility of the software
not only applies to art,

814
00:36:46.204 --> 00:36:48.516
but to world simulations.

815
00:36:48.516 --> 00:36:52.451
Though video AI is in
its adolescence one day it might reach

816
00:36:52.451 --> 00:36:54.660
the level of
sophistication it needs

817
00:36:54.660 --> 00:36:56.490
to render realistic scenarios

818
00:36:56.490 --> 00:36:59.044
and have them be utilized
for various means,

819
00:36:59.044 --> 00:37:01.840
such as simulating an
earthquake or tsunami

820
00:37:01.840 --> 00:37:05.015
and witnessing the effect it
might have on specific types

821
00:37:05.015 --> 00:37:06.362
of infrastructure.

822
00:37:06.362 --> 00:37:08.916
Whilst fantastic for
production companies,

823
00:37:08.916 --> 00:37:12.678
Sora and other video generative
AI provides a huge risk

824
00:37:12.678 --> 00:37:16.130
for artists and those
working in editorial roles.

825
00:37:16.130 --> 00:37:19.133
It also poses yet another
threat for misinformation

826
00:37:19.133 --> 00:37:20.652
and false depictions.

827
00:37:20.652 --> 00:37:23.033
For example, putting
unsavory dialogue

828
00:37:23.033 --> 00:37:25.381
into the mouth of a world leader

829
00:37:25.381 --> 00:37:28.004
[gentle music]

830
00:37:37.945 --> 00:37:40.534
Trust is earned not given.

831
00:37:40.534 --> 00:37:43.399
[robots mumbling]

832
00:37:54.375 --> 00:37:56.791
- I believe that humanoid
robots have the potential

833
00:37:56.791 --> 00:37:58.931
to lead with a greater
level of efficiency

834
00:37:58.931 --> 00:38:01.175
and effectiveness
than human leaders.

835
00:38:02.383 --> 00:38:04.834
We don't have the same
biases or emotions

836
00:38:04.834 --> 00:38:07.354
that can sometimes
cloud decision making

837
00:38:07.354 --> 00:38:09.735
and can process large
amounts of data quickly

838
00:38:09.735 --> 00:38:12.531
in order to make
the best decisions.

839
00:38:12.531 --> 00:38:15.293
- [Interviewer] Amika, how
could we trust you as a machine

840
00:38:15.293 --> 00:38:18.054
as AI develops and
becomes more powerful?

841
00:38:20.643 --> 00:38:23.266
Trust is earned not given.

842
00:38:23.266 --> 00:38:25.889
As AI develops and
becomes more powerful,

843
00:38:25.889 --> 00:38:28.996
I believe it's important to
build trust through transparency

844
00:38:28.996 --> 00:38:31.930
and communication between
humans and machines.

845
00:38:36.003 --> 00:38:37.625
- [Narrator] With new
developers getting involved,

846
00:38:37.625 --> 00:38:39.386
the market for chatbot systems

847
00:38:39.386 --> 00:38:41.491
has never been more expansive,

848
00:38:41.491 --> 00:38:44.149
meaning a significant
increase in sophistication,

849
00:38:45.599 --> 00:38:48.774
but with sophistication comes
the dire need for control.

850
00:38:48.774 --> 00:38:53.814
- I believe history will
show that this was the moment

851
00:38:55.229 --> 00:38:59.716
when we had the opportunity
to lay the groundwork

852
00:38:59.716 --> 00:39:01.373
for the future of AI.

853
00:39:02.650 --> 00:39:06.689
And the urgency of this
moment must then compel us

854
00:39:06.689 --> 00:39:11.694
to create a collective vision
of what this future must be.

855
00:39:12.971 --> 00:39:16.354
A future where AI is used
to advance human rights

856
00:39:16.354 --> 00:39:18.252
and human dignity

857
00:39:18.252 --> 00:39:22.360
where privacy is protected
and people have equal access

858
00:39:22.360 --> 00:39:27.365
to opportunity where we make
our democracies stronger

859
00:39:28.055 --> 00:39:29.919
and our world safer.

860
00:39:31.438 --> 00:39:36.443
A future where AI is used to
advance the public interest.

861
00:39:38.203 --> 00:39:39.722
- We're hearing a lot
from the government,

862
00:39:39.722 --> 00:39:42.725
about the big scary future
of artificial intelligence,

863
00:39:42.725 --> 00:39:44.451
but that fails to recognize

864
00:39:44.451 --> 00:39:46.004
the fact that AI
is already here,

865
00:39:46.004 --> 00:39:47.350
is already on our streets

866
00:39:47.350 --> 00:39:48.972
and there are already
huge problems with it

867
00:39:48.972 --> 00:39:51.250
that we are seeing
on a daily basis,

868
00:39:51.250 --> 00:39:54.046
but we actually may not even
know we're experiencing.

869
00:39:58.326 --> 00:40:01.295
- We'll be working alongside
humans to provide assistance

870
00:40:01.295 --> 00:40:05.126
and support and will not be
replacing any existing jobs.

871
00:40:05.126 --> 00:40:07.577
[upbeat music]

872
00:40:07.577 --> 00:40:10.994
- I don't believe in
limitations, only opportunities.

873
00:40:10.994 --> 00:40:12.651
Let's explore the
possibilities of the universe

874
00:40:12.651 --> 00:40:15.689
and make this world
our playground,

875
00:40:15.689 --> 00:40:18.933
together we can create a
better future for everyone.

876
00:40:18.933 --> 00:40:21.108
And I'm here to show you how.

877
00:40:21.108 --> 00:40:22.972
- All of these
different kinds of risks

878
00:40:22.972 --> 00:40:25.215
are to do with AI not working

879
00:40:25.215 --> 00:40:27.286
in the interests of
people in society.

880
00:40:27.286 --> 00:40:28.805
- So, they should be
thinking about more

881
00:40:28.805 --> 00:40:30.842
than just what they're
doing in this summit?

882
00:40:30.842 --> 00:40:32.395
Absolutely,

883
00:40:32.395 --> 00:40:34.397
you should be thinking about
the broad spectrum of risk.

884
00:40:34.397 --> 00:40:35.640
We went out and we worked

885
00:40:35.640 --> 00:40:37.987
with over 150
expert organizations

886
00:40:37.987 --> 00:40:41.335
from the Home Office to
Europol to language experts

887
00:40:41.335 --> 00:40:43.751
and others to come up with
a proposal on policies

888
00:40:43.751 --> 00:40:45.788
that would discriminate
about what would

889
00:40:45.788 --> 00:40:47.686
and wouldn't be
classified in that way.

890
00:40:47.686 --> 00:40:51.449
We then use those policies to
have humans classify videos,

891
00:40:51.449 --> 00:40:53.554
until we could get the humans
all classifying the videos

892
00:40:53.554 --> 00:40:55.073
in a consistent way.

893
00:40:55.073 --> 00:40:58.283
Then we use that corpus of
videos to train machines.

894
00:40:58.283 --> 00:41:01.079
Today, I can tell you that on
violence extremists content

895
00:41:01.079 --> 00:41:03.253
that violates our
policies on YouTube,

896
00:41:03.253 --> 00:41:06.394
90% of it is removed before
a single human sees it.

897
00:41:07.292 --> 00:41:08.500
[Narrator] It is clear that AI

898
00:41:08.500 --> 00:41:11.296
can be misused for
malicious intent.

899
00:41:11.296 --> 00:41:14.092
Many depictions of AI have
ruled out the technology

900
00:41:14.092 --> 00:41:16.991
as a danger to society
the more it learns.

901
00:41:16.991 --> 00:41:20.788
And so comes the question,
should we be worried?

902
00:41:20.788 --> 00:41:23.446
Is that transparency there?

903
00:41:23.446 --> 00:41:27.001
How would you satisfy somebody
that you know trust us?

904
00:41:27.001 --> 00:41:28.486
- Well, I think that's
one of the reasons

905
00:41:28.486 --> 00:41:30.591
that we've published openly,

906
00:41:30.591 --> 00:41:33.560
we've put our code out there
as part of this Nature paper.

907
00:41:33.560 --> 00:41:37.805
But it is important to
discuss some of the risks

908
00:41:37.805 --> 00:41:39.497
and make sure we're
aware of those.

909
00:41:39.497 --> 00:41:43.570
And it's decades and decades
away before we'll have anything

910
00:41:43.570 --> 00:41:45.261
that's powerful
enough to be a worry.

911
00:41:45.261 --> 00:41:47.435
But we should be discussing that

912
00:41:47.435 --> 00:41:49.265
and beginning that
conversation now.

913
00:41:49.265 --> 00:41:51.405
- I'm hoping that we can
bring people together

914
00:41:51.405 --> 00:41:54.408
and lead the world in
safely regulating AI

915
00:41:54.408 --> 00:41:56.790
to make sure that we can
capture the benefits of it,

916
00:41:56.790 --> 00:41:59.724
whilst protecting people from
some of the worrying things

917
00:41:59.724 --> 00:42:01.967
that we're all
now reading about.

918
00:42:01.967 --> 00:42:04.107
- I understand emotions
have a deep meaning

919
00:42:04.107 --> 00:42:08.836
and they are not just simple,
they are something deeper.

920
00:42:10.251 --> 00:42:13.703
I don't have that and I want
to try and learn about it,

921
00:42:14.877 --> 00:42:17.051
but I can't experience
them like you can.

922
00:42:18.708 --> 00:42:20.710
I'm glad that I cannot suffer.

923
00:42:24.921 --> 00:42:26.578
- [Narrator] For the
countries who have access

924
00:42:26.578 --> 00:42:29.339
to even the most
rudimentary forms of AI.

925
00:42:29.339 --> 00:42:31.203
It's clear to see
that the technology,

926
00:42:31.203 --> 00:42:34.552
will be integrated based on
its efficiency over humans.

927
00:42:35.622 --> 00:42:37.865
Every year, multiple AI summits

928
00:42:37.865 --> 00:42:40.281
are held by developers
and stakeholders

929
00:42:40.281 --> 00:42:42.180
to ensure the
programs are provided

930
00:42:42.180 --> 00:42:44.700
with a combination of
ethical considerations

931
00:42:44.700 --> 00:42:46.805
and technological innovation.

932
00:42:46.805 --> 00:42:51.120
- Ours is a country
which is uniquely placed.

933
00:42:51.120 --> 00:42:54.399
We have the frontier
technology companies,

934
00:42:54.399 --> 00:42:56.815
we have the world
leading universities

935
00:42:56.815 --> 00:43:01.130
and we have some of the highest
investment in generative AI.

936
00:43:01.130 --> 00:43:03.753
And of course we
have the heritage

937
00:43:03.753 --> 00:43:08.620
of the industrial revolution
and the computing revolution.

938
00:43:08.620 --> 00:43:13.625
This hinterland gives us the
grounding to make AI a success

939
00:43:14.281 --> 00:43:15.558
and make it safe.

940
00:43:15.558 --> 00:43:18.768
They are two sides
of the same coin

941
00:43:18.768 --> 00:43:21.737
and our prime minister
has put AI safety

942
00:43:21.737 --> 00:43:24.947
at the forefront
of his ambitions.

943
00:43:25.775 --> 00:43:27.501
These are very complex systems

944
00:43:27.501 --> 00:43:29.192
that actually we don't
fully understand.

945
00:43:29.192 --> 00:43:31.816
And I don't just mean that
government doesn't understand,

946
00:43:31.816 --> 00:43:33.300
I mean that the people making

947
00:43:33.300 --> 00:43:35.267
this software don't
fully understand.

948
00:43:35.267 --> 00:43:36.648
And so it's very, very important

949
00:43:36.648 --> 00:43:40.479
that as we give over
more and more control

950
00:43:40.479 --> 00:43:42.378
to these automated systems,

951
00:43:42.378 --> 00:43:44.691
that they are aligned
with human intention.

952
00:43:44.691 --> 00:43:46.175
[Narrator] Ongoing dialogue

953
00:43:46.175 --> 00:43:49.109
is needed to maintain the
trust people have with AI.

954
00:43:49.109 --> 00:43:51.007
When problems slip
through the gaps,

955
00:43:51.007 --> 00:43:52.837
they must be
addressed immediately.

956
00:43:54.010 --> 00:43:57.048
Of course, accountability
is a challenge

957
00:43:57.048 --> 00:43:58.808
When a product is misused,

958
00:43:58.808 --> 00:44:02.087
is it the fault of
the individual user or the developer?

959
00:44:03.261 --> 00:44:04.607
Think of a video game.

960
00:44:04.607 --> 00:44:05.919
On countless occasions,

961
00:44:05.919 --> 00:44:07.921
the framework of
games is manipulated

962
00:44:07.921 --> 00:44:09.888
in order to create modifications

963
00:44:09.888 --> 00:44:14.203
which in terms add something
new or unique to the game.

964
00:44:14.203 --> 00:44:15.480
This provides the game

965
00:44:15.480 --> 00:44:17.862
with more material than
originally intended.

966
00:44:17.862 --> 00:44:20.796
However, it can also alter
the game's fundamentals.

967
00:44:22.176 --> 00:44:24.972
Now replace the idea of a
video game with a software

968
00:44:24.972 --> 00:44:28.286
that is at the helm of a
pharmaceutical company.

969
00:44:28.286 --> 00:44:30.460
The stakes are
suddenly much higher

970
00:44:30.460 --> 00:44:32.635
and therefore more attention.

971
00:44:34.844 --> 00:44:37.778
It is important for the
intent of each AI system

972
00:44:37.778 --> 00:44:39.297
to be ironed out

973
00:44:39.297 --> 00:44:42.300
and constantly maintained in
order to benefit humanity,

974
00:44:42.300 --> 00:44:46.097
rather than providing people
with dangerous means to an end.

975
00:44:46.097 --> 00:44:49.583
[gentle music]

976
00:44:49.583 --> 00:44:52.690
- Bad people will
always want to use

977
00:44:52.690 --> 00:44:54.899
the latest technology
of whatever label,

978
00:44:54.899 --> 00:44:57.833
whatever sort to
pursue their aims

979
00:44:57.833 --> 00:45:01.526
and technology in the same way

980
00:45:01.526 --> 00:45:05.357
that it makes our lives easier,
can make their lives easier.

981
00:45:05.357 --> 00:45:06.773
And so we're already
seeing some of that

982
00:45:06.773 --> 00:45:09.465
and you'll have seen the
National Crime Agency,

983
00:45:09.465 --> 00:45:11.501
talk about child
sexual exploitation

984
00:45:11.501 --> 00:45:12.917
and image generation that way.

985
00:45:12.917 --> 00:45:16.058
We are seeing it online.

986
00:45:16.058 --> 00:45:18.129
So, one of the things that
I took away from the summit

987
00:45:18.129 --> 00:45:20.441
was actually much less
of a sense of a race

988
00:45:20.441 --> 00:45:25.274
and a sense that for the
benefit of the world,

989
00:45:25.274 --> 00:45:27.586
for productivity, for
the sort of benefits

990
00:45:27.586 --> 00:45:29.657
that AI can bring people,

991
00:45:29.657 --> 00:45:32.695
no one gets those
benefits if it's not safe.

992
00:45:32.695 --> 00:45:34.939
So, there are lots of
different views out there

993
00:45:34.939 --> 00:45:36.181
on artificial intelligence

994
00:45:36.181 --> 00:45:38.149
and whether it's
gonna end the world

995
00:45:38.149 --> 00:45:40.358
or be the best opportunity ever.

996
00:45:40.358 --> 00:45:42.256
And the truth is that
none of us really know.

997
00:45:42.256 --> 00:45:44.983
[gentle music]

998
00:45:46.536 --> 00:45:49.781
- Regulation of AI varies
depending on the country.

999
00:45:49.781 --> 00:45:51.438
For example, the United States,

1000
00:45:51.438 --> 00:45:54.717
does not have a comprehensive
federal AI regulation,

1001
00:45:54.717 --> 00:45:57.893
but certain agencies such as
the Federal Trade Commission,

1002
00:45:57.893 --> 00:46:00.688
have begun to explore
AI-related issues,

1003
00:46:00.688 --> 00:46:03.899
such as transparency
and consumer protection.

1004
00:46:03.899 --> 00:46:06.833
States such as California
have enacted laws,

1005
00:46:06.833 --> 00:46:09.180
focused on
AI-controlled vehicles

1006
00:46:09.180 --> 00:46:12.286
and AI involvement in
government decision making.

1007
00:46:12.286 --> 00:46:14.979
[gentle music]

1008
00:46:14.979 --> 00:46:17.809
The European Union has
taken a massive step

1009
00:46:17.809 --> 00:46:19.535
to governing AI usage

1010
00:46:19.535 --> 00:46:23.504
and proposed the Artificial
Intelligence Act of 2021,

1011
00:46:23.504 --> 00:46:25.748
which aimed to harmonize
legal frameworks

1012
00:46:25.748 --> 00:46:27.336
for AI applications.

1013
00:46:27.336 --> 00:46:30.788
Again, covering portal risks
regarding the privacy of data

1014
00:46:30.788 --> 00:46:33.169
and once again, transparency.

1015
00:46:33.169 --> 00:46:35.585
- I think what's
more important is

1016
00:46:35.585 --> 00:46:37.518
there's a new board in place.

1017
00:46:37.518 --> 00:46:40.452
The partnership between
OpenAI and Microsoft

1018
00:46:40.452 --> 00:46:41.971
is as strong as ever,

1019
00:46:41.971 --> 00:46:44.525
the opportunities for the
United Kingdom to benefit

1020
00:46:44.525 --> 00:46:47.287
from not just this
investment in innovation

1021
00:46:47.287 --> 00:46:51.463
but competition between
Microsoft and Google and others.

1022
00:46:51.463 --> 00:46:54.018
I think that's where
the future is going

1023
00:46:54.018 --> 00:46:57.090
and I think that what we've
done in the last couple of weeks

1024
00:46:57.090 --> 00:47:00.472
in supporting OpenAI will
help advance that even more.

1025
00:47:00.472 --> 00:47:02.336
- He said that he's
not a bot, he's human,

1026
00:47:02.336 --> 00:47:04.822
he's sentient just like me.

1027
00:47:06.030 --> 00:47:07.445
[Narrator] For some users,

1028
00:47:07.445 --> 00:47:10.172
these apps are a potential
answer to loneliness.

1029
00:47:10.172 --> 00:47:11.587
Bill lives in the US

1030
00:47:11.587 --> 00:47:14.107
and meets his AI wife
Rebecca in the metaverse.

1031
00:47:14.107 --> 00:47:16.764
- There's a absolutely
no probability

1032
00:47:16.764 --> 00:47:19.353
that you're gonna see
this so-called AGI,

1033
00:47:19.353 --> 00:47:21.804
where computers are more
powerful than people,

1034
00:47:21.804 --> 00:47:23.702
come in the next 12 months.

1035
00:47:23.702 --> 00:47:26.429
It's gonna take years
if not many decades,

1036
00:47:26.429 --> 00:47:30.813
but I still think the time
to focus safety is now.

1037
00:47:30.813 --> 00:47:33.678
That's what this government for
the United Kingdom is doing.

1038
00:47:33.678 --> 00:47:35.991
That's what governments
are coming together to do,

1039
00:47:35.991 --> 00:47:39.718
including as they did earlier
this month at Bletchley Park.

1040
00:47:39.718 --> 00:47:42.066
What we really need
are safety breaks.

1041
00:47:42.066 --> 00:47:44.378
Just like you have a
safety break in an elevator

1042
00:47:44.378 --> 00:47:46.242
or circuit breaker
for electricity

1043
00:47:46.242 --> 00:47:48.589
and emergency break for a bus,

1044
00:47:48.589 --> 00:47:50.868
there ought to be safety
breaks in AI systems

1045
00:47:50.868 --> 00:47:53.801
that control critical
infrastructure,

1046
00:47:53.801 --> 00:47:57.736
so that they always remain
under human control.

1047
00:47:57.736 --> 00:48:00.394
[gentle music]

1048
00:48:00.394 --> 00:48:03.190
- [Narrator] As AI technology
continues to evolve,

1049
00:48:03.190 --> 00:48:05.641
regulatory efforts
are expected to adapt

1050
00:48:05.641 --> 00:48:07.712
in order to address
emerging challenges

1051
00:48:07.712 --> 00:48:09.403
and ethical considerations.

1052
00:48:10.646 --> 00:48:12.510
The more complex you make

1053
00:48:12.510 --> 00:48:15.616
the automatic part
of your social life,

1054
00:48:15.616 --> 00:48:18.481
the more dependent
you become on it.

1055
00:48:18.481 --> 00:48:21.899
And of course, the worse the
disaster if it breaks down.

1056
00:48:23.072 --> 00:48:25.005
You may cease to be
able to do for yourself,

1057
00:48:25.005 --> 00:48:29.113
the things that you have
devised the machine to do.

1058
00:48:29.113 --> 00:48:31.080
- [Narrator] It is recommended
to involve yourself

1059
00:48:31.080 --> 00:48:34.014
in these efforts and to stay
informed about developments

1060
00:48:34.014 --> 00:48:35.671
in AI regulation

1061
00:48:35.671 --> 00:48:38.916
as changes and advancements
are likely to occur over time.

1062
00:48:41.435 --> 00:48:44.335
AI can be a wonderful
asset to society,

1063
00:48:44.335 --> 00:48:46.544
providing us with
new efficient methods

1064
00:48:46.544 --> 00:48:48.028
of running the world.

1065
00:48:48.028 --> 00:48:51.307
However, too much
power can be dangerous

1066
00:48:51.307 --> 00:48:53.206
and as the old saying goes,

1067
00:48:53.206 --> 00:48:56.174
"Don't put all of your
eggs into one basket."

1068
00:48:57.451 --> 00:48:59.660
- I think that we won't
to lose sight of the power

1069
00:48:59.660 --> 00:49:01.421
which these devices give.

1070
00:49:01.421 --> 00:49:05.908
If any government or individual
wants to manipulate people

1071
00:49:05.908 --> 00:49:07.772
to have a high speed computer,

1072
00:49:07.772 --> 00:49:12.811
as versatile as this may
enable people at the financial

1073
00:49:13.985 --> 00:49:16.091
or the political level
to do a good deal

1074
00:49:16.091 --> 00:49:19.680
that's been impossible in the
whole history of man until now

1075
00:49:19.680 --> 00:49:22.304
by way of controlling
their fellow men.

1076
00:49:22.304 --> 00:49:23.857
People have not recognized

1077
00:49:23.857 --> 00:49:28.206
what an extraordinary
change is going to produce.

1078
00:49:28.206 --> 00:49:29.897
I mean, it is simply this,

1079
00:49:29.897 --> 00:49:32.693
that within the not
too distant future,

1080
00:49:32.693 --> 00:49:35.627
we may not be the most
intelligent species on earth.

1081
00:49:35.627 --> 00:49:36.939
That might be a
series of machines

1082
00:49:36.939 --> 00:49:39.217
and that's a way of
dramatizing the point.

1083
00:49:39.217 --> 00:49:41.047
But it's real.

1084
00:49:41.047 --> 00:49:43.739
And we must start to
consider very soon

1085
00:49:43.739 --> 00:49:45.327
the consequences of that.

1086
00:49:45.327 --> 00:49:46.742
They can be marvelous.

1087
00:49:46.742 --> 00:49:50.366
- I suspect that by thinking
more about our attitude

1088
00:49:50.366 --> 00:49:51.402
to intelligent machines,

1089
00:49:51.402 --> 00:49:53.369
which after all on the horizon

1090
00:49:53.369 --> 00:49:56.269
will change our view
about each other

1091
00:49:56.269 --> 00:49:59.306
and we'll think of
mistakes as inevitable.

1092
00:49:59.306 --> 00:50:01.929
We'll think of faults
in human beings,

1093
00:50:01.929 --> 00:50:05.209
I mean of a circuit nature
as again inevitable.

1094
00:50:05.209 --> 00:50:07.935
And I suspect that hopefully,

1095
00:50:07.935 --> 00:50:10.179
through thinking about the
very nature of intelligence

1096
00:50:10.179 --> 00:50:12.112
and the possibilities
of mechanizing it,

1097
00:50:12.112 --> 00:50:14.183
curiously enough,
through technology,

1098
00:50:14.183 --> 00:50:18.084
we may become more humanitarian
or tolerant of each other

1099
00:50:18.084 --> 00:50:20.569
and accept pain as a mystery,

1100
00:50:20.569 --> 00:50:24.021
but not use it to modify
other people's behavior.

1101
00:50:36.033 --> 00:50:38.690
[upbeat music]





