1 00:00:00,586 --> 00:00:02,726 [gentle music] 2 00:00:02,000 --> 00:00:07,000 Downloaded from YTS.MX 3 00:00:08,000 --> 00:00:13,000 Official YIFY movies site: YTS.MX 4 00:00:14,324 --> 00:00:15,532 [Narrator] For decades, 5 00:00:15,532 --> 00:00:17,017 we have discussed the many outcomes, 6 00:00:17,017 --> 00:00:19,053 regarding artificial intelligence. 7 00:00:19,053 --> 00:00:21,469 Could our world be dominated? 8 00:00:21,469 --> 00:00:25,232 Could our independence and autonomy be stripped from us, 9 00:00:25,232 --> 00:00:28,407 or are we able to control what we have created? 10 00:00:28,407 --> 00:00:31,100 [upbeat music] 11 00:00:37,416 --> 00:00:41,006 Could we use artificial intelligence to benefit our society? 12 00:00:41,006 --> 00:00:44,009 Just how thin is the line between the development 13 00:00:44,009 --> 00:00:46,805 of civilization and chaos? 14 00:00:46,805 --> 00:00:49,428 [upbeat music] 15 00:01:13,211 --> 00:01:15,903 To understand what artificial intelligence is, 16 00:01:15,903 --> 00:01:19,803 one must understand that it can take many different forms. 17 00:01:19,803 --> 00:01:22,047 Think of it as a web of ideas, 18 00:01:22,047 --> 00:01:25,326 slowly expanding as new ways of utilizing computers 19 00:01:25,326 --> 00:01:26,603 are explored. 20 00:01:26,603 --> 00:01:28,260 As technology develops, 21 00:01:28,260 --> 00:01:31,539 so do the capabilities of self-learning software. 22 00:01:31,539 --> 00:01:34,335 - [Reporter] The need to diagnose disease quickly 23 00:01:34,335 --> 00:01:38,132 and effectively has prompted many university medical centers 24 00:01:38,132 --> 00:01:41,791 to develop intelligent programs that simulate the work 25 00:01:41,791 --> 00:01:44,345 of doctors and laboratory technicians. 26 00:01:44,345 --> 00:01:47,003 [gentle music] 27 00:01:48,694 --> 00:01:51,041 - [Narrator] AI is quickly integrating with our way of life. 28 00:01:51,041 --> 00:01:54,631 So, much so that development of AI programs has in itself, 29 00:01:54,631 --> 00:01:56,323 become a business opportunity. 30 00:01:57,945 --> 00:01:58,773 [upbeat music] 31 00:01:58,773 --> 00:01:59,809 In our modern age, 32 00:01:59,809 --> 00:02:01,638 we are powered by technology 33 00:02:01,638 --> 00:02:05,021 and softwares are transcending its virtual existence, 34 00:02:05,021 --> 00:02:07,437 finding applications in various fields, 35 00:02:07,437 --> 00:02:11,372 such as customer support to content creation, 36 00:02:11,372 --> 00:02:13,202 computer-aided design, 37 00:02:13,202 --> 00:02:17,137 otherwise known as CAD, is one of the many uses of AI. 38 00:02:17,137 --> 00:02:19,415 By analyzing particular variables, 39 00:02:19,415 --> 00:02:22,280 computers are now able to assist in the modification 40 00:02:22,280 --> 00:02:26,180 and creation of designs for hardware and architecture. 41 00:02:26,180 --> 00:02:30,046 The prime use of any AI is for optimizing processes 42 00:02:30,046 --> 00:02:32,324 that were considered tedious before. 43 00:02:32,324 --> 00:02:35,189 In many ways, AI has been hugely beneficial 44 00:02:35,189 --> 00:02:38,951 for technological development thanks to its sheer speed. 45 00:02:38,951 --> 00:02:41,057 However, AI only benefits 46 00:02:41,057 --> 00:02:43,508 those to whom the programs are distributed. 47 00:02:44,302 --> 00:02:45,613 Artificial intelligence 48 00:02:45,613 --> 00:02:47,443 is picking through your rubbish. 49 00:02:47,443 --> 00:02:51,688 This robot uses it to sort through plastics for recycling 50 00:02:51,688 --> 00:02:53,414 and it can be retrained 51 00:02:53,414 --> 00:02:55,968 to prioritize whatever's more marketable. 52 00:02:57,177 --> 00:03:00,180 So, AI can clearly be incredibly useful, 53 00:03:00,180 --> 00:03:02,596 but there are deep concerns about 54 00:03:02,596 --> 00:03:07,635 how quickly it is developing and where it could go next. 55 00:03:08,912 --> 00:03:11,121 - The aim is to make them as capable as humans 56 00:03:11,121 --> 00:03:14,366 and deploy them in the service sector. 57 00:03:14,366 --> 00:03:16,230 The engineers in this research 58 00:03:16,230 --> 00:03:18,059 and development lab are working 59 00:03:18,059 --> 00:03:21,822 to take these humanoid robots to the next level 60 00:03:21,822 --> 00:03:24,583 where they can not only speak and move, 61 00:03:24,583 --> 00:03:27,345 but they can think and feel and act 62 00:03:27,345 --> 00:03:30,002 and even make decisions for themselves. 63 00:03:30,796 --> 00:03:32,695 And that daily data stream 64 00:03:32,695 --> 00:03:36,008 is being fed into an ever expanding workforce, 65 00:03:36,008 --> 00:03:39,529 dedicated to developing artificial intelligence. 66 00:03:41,013 --> 00:03:42,808 Those who have studied abroad 67 00:03:42,808 --> 00:03:46,122 are being encouraged to return to the motherland. 68 00:03:46,122 --> 00:03:47,917 Libo Yang came back 69 00:03:47,917 --> 00:03:51,645 and started a tech enterprise in his hometown. 70 00:03:51,645 --> 00:03:54,268 - [Narrator] China's market is indeed the most open 71 00:03:54,268 --> 00:03:56,926 and active market in the world for AI. 72 00:03:56,926 --> 00:04:01,241 It is also where there are the most application scenarios. 73 00:04:01,241 --> 00:04:03,864 - So, AI is generally a broad term that we apply 74 00:04:03,864 --> 00:04:04,934 to a number of techniques. 75 00:04:04,934 --> 00:04:06,384 And in this particular case, 76 00:04:06,384 --> 00:04:09,456 what we're actually looking at was elements of AI, 77 00:04:09,456 --> 00:04:12,010 machine learning and deep learning. 78 00:04:12,010 --> 00:04:13,701 So, in this particular case, 79 00:04:13,701 --> 00:04:17,429 we've been unfortunately in a situation 80 00:04:17,429 --> 00:04:20,398 in this race against time to create new antibiotics, 81 00:04:20,398 --> 00:04:22,779 the threat is actually quite real 82 00:04:22,779 --> 00:04:25,230 and it would be a global problem. 83 00:04:25,230 --> 00:04:27,784 We desperately needed to harness new technologies 84 00:04:27,784 --> 00:04:29,269 in an attempt to fight it, 85 00:04:29,269 --> 00:04:30,960 we're looking at drugs 86 00:04:30,960 --> 00:04:33,411 which could potentially fight E. coli, 87 00:04:33,411 --> 00:04:35,102 a very dangerous bacteria. 88 00:04:35,102 --> 00:04:37,207 - So, what is it that the AI is doing 89 00:04:37,207 --> 00:04:39,348 that humans can't do very simply? 90 00:04:39,348 --> 00:04:41,729 - So, the AI can look for patterns 91 00:04:41,729 --> 00:04:44,560 that we wouldn't be able to mind for with a human eye, 92 00:04:44,560 --> 00:04:47,287 simply within what I do as a radiologist, 93 00:04:47,287 --> 00:04:50,980 I look for patterns of diseases in terms of shape, 94 00:04:50,980 --> 00:04:53,914 contrast enhancement, heterogeneity. 95 00:04:53,914 --> 00:04:55,191 But what the computer does, 96 00:04:55,191 --> 00:04:58,125 it looks for patterns within the pixels. 97 00:04:58,125 --> 00:05:00,679 These are things that you just can't see to the human eye. 98 00:05:00,679 --> 00:05:03,855 There's so much more data embedded within these scans 99 00:05:03,855 --> 00:05:07,514 that we use that we can't mine on a physical level. 100 00:05:07,514 --> 00:05:09,516 So, the computers really help. 101 00:05:09,516 --> 00:05:11,311 - [Narrator] Many believe the growth of AI 102 00:05:11,311 --> 00:05:13,692 is dependent on global collaboration, 103 00:05:13,692 --> 00:05:17,109 but access to the technology is limited in certain regions. 104 00:05:17,109 --> 00:05:19,767 Global distribution is a long-term endeavor 105 00:05:19,767 --> 00:05:21,044 and the more countries 106 00:05:21,044 --> 00:05:23,288 and businesses that have access to the tech, 107 00:05:23,288 --> 00:05:26,429 the more regulation the AI will require. 108 00:05:26,429 --> 00:05:29,846 In fact, it is now not uncommon for businesses 109 00:05:29,846 --> 00:05:33,125 to be entirely run by an artificial director. 110 00:05:33,125 --> 00:05:34,472 On many occasions, 111 00:05:34,472 --> 00:05:37,198 handing the helm of a company to an algorithm 112 00:05:37,198 --> 00:05:40,685 can provide the best option on the basis of probability. 113 00:05:40,685 --> 00:05:43,998 However, dependence and reliability on softwares 114 00:05:43,998 --> 00:05:45,897 can be a great risk. 115 00:05:45,897 --> 00:05:47,450 Without proper safeguards, 116 00:05:47,450 --> 00:05:50,419 actions based on potentially incorrect predictions 117 00:05:50,419 --> 00:05:53,353 can be a detriment to a business or operation. 118 00:05:53,353 --> 00:05:55,147 Humans provide the critical thinking 119 00:05:55,147 --> 00:05:58,461 and judgment which AI is not capable of matching. 120 00:05:58,461 --> 00:06:00,463 - Well, this is the Accessibility Design Center 121 00:06:00,463 --> 00:06:02,810 and it's where we try to bring together our engineers 122 00:06:02,810 --> 00:06:05,882 and experts with the latest AI technology, 123 00:06:05,882 --> 00:06:07,608 with people with disabilities, 124 00:06:07,608 --> 00:06:10,059 because there's a real opportunity to firstly help people 125 00:06:10,059 --> 00:06:12,613 with disabilities enjoy all the technology 126 00:06:12,613 --> 00:06:14,201 we have in our pockets today. 127 00:06:14,201 --> 00:06:15,720 And sometimes that's not very accessible, 128 00:06:15,720 --> 00:06:18,688 but also build tools that can help them engage better 129 00:06:18,688 --> 00:06:20,103 in the real world. 130 00:06:20,103 --> 00:06:22,451 And that's thanks to the wonders of machine learning. 131 00:06:22,451 --> 00:06:25,764 - I don't think we're like at the end of this paradigm yet. 132 00:06:25,764 --> 00:06:26,903 We'll keep pushing these. 133 00:06:26,903 --> 00:06:28,215 We'll add other modalities. 134 00:06:28,215 --> 00:06:31,114 So, someday they'll do video, audio images, 135 00:06:31,114 --> 00:06:36,154 text altogether and they'll get like much smarter over time. 136 00:06:37,638 --> 00:06:38,674 - AI, machine learning, all very sounds very complicated. 137 00:06:38,674 --> 00:06:40,572 Just think about it as a toolkit 138 00:06:40,572 --> 00:06:42,781 that's really good at sort of spotting patterns 139 00:06:42,781 --> 00:06:44,024 and making predictions, 140 00:06:44,024 --> 00:06:46,336 better than any computing could do before. 141 00:06:46,336 --> 00:06:47,786 And that's why it's so useful 142 00:06:47,786 --> 00:06:51,031 for things like understanding language and speech. 143 00:06:51,031 --> 00:06:52,998 Another product which we are launching today 144 00:06:52,998 --> 00:06:55,000 is called Project Relate. 145 00:06:55,000 --> 00:06:56,312 And this is for people 146 00:06:56,312 --> 00:06:58,728 who have non-standard speech patterns. 147 00:06:58,728 --> 00:07:00,937 So, one of the people we work with 148 00:07:00,937 --> 00:07:03,837 is maybe less than 10% of the time, 149 00:07:03,837 --> 00:07:06,564 could be understood by people who don't know her, 150 00:07:06,564 --> 00:07:09,325 using this tool that's over 90% of the time. 151 00:07:09,325 --> 00:07:12,259 And you think about that transformation in somebody's life 152 00:07:12,259 --> 00:07:15,676 and then you think about the fact there's 250 million people 153 00:07:15,676 --> 00:07:17,678 with non-standard speech patterns around the world. 154 00:07:17,678 --> 00:07:19,093 So, that's the ambition of this center 155 00:07:19,093 --> 00:07:21,682 is to unite technology with people with disabilities 156 00:07:21,682 --> 00:07:24,478 and try to help 'em engage more in the world. 157 00:07:24,478 --> 00:07:27,550 - [Narrator] On the 30th November of 2022, 158 00:07:27,550 --> 00:07:30,001 a revolutionary innovation emerged, 159 00:07:30,967 --> 00:07:32,003 ChatGPT. 160 00:07:32,969 --> 00:07:35,869 ChatGPT was created by OpenAI, 161 00:07:35,869 --> 00:07:38,250 an AI research organization. 162 00:07:38,250 --> 00:07:39,873 Its goal is to develop systems 163 00:07:39,873 --> 00:07:44,498 which may benefit all aspects of society and communication. 164 00:07:44,498 --> 00:07:47,467 Sam Altman stepped up as CEO of OpenAI 165 00:07:47,467 --> 00:07:50,055 on its launch in 2015. 166 00:07:50,055 --> 00:07:51,609 Altman dabbled in a multitude 167 00:07:51,609 --> 00:07:53,990 of computing-based business ventures. 168 00:07:53,990 --> 00:07:57,477 His rise to CEO was thanks to his many affiliations 169 00:07:57,477 --> 00:08:01,377 and investments with computing and social media companies. 170 00:08:01,377 --> 00:08:04,173 He began his journey by co-founding Loopt, 171 00:08:04,173 --> 00:08:06,106 a social media service. 172 00:08:06,106 --> 00:08:07,763 After selling the application, 173 00:08:07,763 --> 00:08:10,835 Altman went on to bigger and riskier endeavors 174 00:08:10,835 --> 00:08:14,148 from startup accelerator companies to security software. 175 00:08:15,184 --> 00:08:17,393 OpenAI became hugely desirable, 176 00:08:17,393 --> 00:08:20,223 thanks to the amount of revenue the company had generated 177 00:08:20,223 --> 00:08:21,984 with over a billion dollars made 178 00:08:21,984 --> 00:08:24,262 within its first year of release. 179 00:08:24,262 --> 00:08:27,265 ChatGPT became an easily accessible software, 180 00:08:27,265 --> 00:08:30,786 built on a large language model known as an LLM. 181 00:08:30,786 --> 00:08:34,134 This program can conjure complex human-like responses 182 00:08:34,134 --> 00:08:37,309 to the user's questions otherwise known as prompts. 183 00:08:37,309 --> 00:08:38,794 In essence, 184 00:08:38,794 --> 00:08:41,244 it is a program which learns the more it is used. 185 00:08:43,592 --> 00:08:45,317 The new age therapeutic program 186 00:08:45,317 --> 00:08:48,804 was developed on the GPT-3.5. 187 00:08:48,804 --> 00:08:51,531 The architecture of this older model allowed systems 188 00:08:51,531 --> 00:08:53,602 to understand and generate code 189 00:08:53,602 --> 00:08:56,501 and natural languages at a remarkably advanced level 190 00:08:56,501 --> 00:08:59,884 from analyzing syntax to nuances in writing. 191 00:08:59,884 --> 00:09:02,542 [upbeat music] 192 00:09:04,578 --> 00:09:06,753 ChatGPT took the world by storm, 193 00:09:06,753 --> 00:09:09,445 due to the sophistication of the system. 194 00:09:09,445 --> 00:09:11,067 As with many chatbot systems, 195 00:09:11,067 --> 00:09:13,449 people have since found ways to manipulate 196 00:09:13,449 --> 00:09:17,349 and confuse the software in order to test its limits. 197 00:09:17,349 --> 00:09:20,076 [gentle music] 198 00:09:21,526 --> 00:09:25,910 The first computer was invented by Charles Babbage in 1822. 199 00:09:25,910 --> 00:09:29,189 It was to be a rudimentary general purpose system. 200 00:09:29,189 --> 00:09:34,021 In 1936, the system was developed upon by Alan Turing. 201 00:09:34,021 --> 00:09:36,299 The automatic machine, as he called them, 202 00:09:36,299 --> 00:09:38,854 was able to break enigma enciphered messages, 203 00:09:38,854 --> 00:09:41,201 regarding enemy military operations, 204 00:09:41,201 --> 00:09:43,583 during the Second World War. 205 00:09:43,583 --> 00:09:46,447 Turing theorized his own type of computer, 206 00:09:46,447 --> 00:09:49,830 the Turing Machine has coined by Alonzo Church, 207 00:09:49,830 --> 00:09:52,522 after reading Turing's research paper. 208 00:09:52,522 --> 00:09:55,698 It had become realized that soon prospect of computing 209 00:09:55,698 --> 00:09:57,907 and engineering would merge seamlessly. 210 00:09:59,046 --> 00:10:01,152 Theories of future tech would increase 211 00:10:01,152 --> 00:10:04,742 and soon came a huge outburst in science fiction media. 212 00:10:04,742 --> 00:10:07,468 This was known as the golden age for computing. 213 00:10:07,468 --> 00:10:10,092 [gentle music] 214 00:10:20,067 --> 00:10:22,760 Alan Turing's contributions to computability 215 00:10:22,760 --> 00:10:25,590 and theoretical computer science was one step closer 216 00:10:25,590 --> 00:10:28,110 to producing a reactive machine. 217 00:10:28,110 --> 00:10:31,389 The reactive machine is an early form of AI. 218 00:10:31,389 --> 00:10:32,942 They had limited capabilities 219 00:10:32,942 --> 00:10:34,772 and were unable to store memories 220 00:10:34,772 --> 00:10:37,740 in order to learn new algorithms of data. 221 00:10:37,740 --> 00:10:41,641 However, they were able to react to specific stimuli. 222 00:10:41,641 --> 00:10:46,611 The first AI was a program written in 1952 by Arthur Samuel. 223 00:10:47,854 --> 00:10:49,614 The prototype AI was able to play checkers, 224 00:10:49,614 --> 00:10:52,168 against an opponent and was built to operate 225 00:10:52,168 --> 00:10:56,172 on the Ferranti Mark One, an early commercial computer. 226 00:10:56,172 --> 00:10:57,657 - [Reporter] This computer has been playing the game 227 00:10:57,657 --> 00:11:00,418 for several years now, getting better all the time. 228 00:11:00,418 --> 00:11:02,972 Tonight it's playing against the black side of the board. 229 00:11:02,972 --> 00:11:05,837 It's approach to playing drafts, it's almost human. 230 00:11:05,837 --> 00:11:08,012 It remembers the moves that enable it to win 231 00:11:08,012 --> 00:11:10,324 and the sort that lead to defeat. 232 00:11:10,324 --> 00:11:12,982 The computer indicates the move it wants to make on a panel 233 00:11:12,982 --> 00:11:14,156 of flashing lights. 234 00:11:14,156 --> 00:11:15,433 It's up to the human opponent 235 00:11:15,433 --> 00:11:18,229 to actually move the drafts about the board. 236 00:11:18,229 --> 00:11:20,645 This sort of works producing exciting information 237 00:11:20,645 --> 00:11:22,405 on the way in which electronic brains 238 00:11:22,405 --> 00:11:24,338 can learn from past experience 239 00:11:24,338 --> 00:11:26,168 and improve their performances. 240 00:11:27,963 --> 00:11:29,792 [Narrator] In 1966, 241 00:11:29,792 --> 00:11:32,519 an MIT professor named Joseph Weizenbaum, 242 00:11:32,519 --> 00:11:37,110 created an AI which would change the landscape of society. 243 00:11:37,110 --> 00:11:39,077 It was known as Eliza, 244 00:11:39,077 --> 00:11:42,322 and it was designed to act like a psychotherapist. 245 00:11:42,322 --> 00:11:45,497 The software was simplistic, yet revolutionary. 246 00:11:45,497 --> 00:11:47,499 The AI would receive the user input 247 00:11:47,499 --> 00:11:51,055 and use specific parameters to generate a coherent response. 248 00:11:53,057 --> 00:11:55,991 - It it has been said, especially here at MIT, 249 00:11:55,991 --> 00:11:59,719 that computers will take over in some sense 250 00:11:59,719 --> 00:12:02,652 and it's even been said that if we're lucky, 251 00:12:02,652 --> 00:12:04,447 they'll keep us as pets 252 00:12:04,447 --> 00:12:06,277 and Arthur C. Clarke, the science fiction writer, 253 00:12:06,277 --> 00:12:09,694 we marked once that if that were to happen, 254 00:12:09,694 --> 00:12:12,904 it would serve us right, he said. 255 00:12:12,904 --> 00:12:14,734 - [Narrator] The program maintained the illusion 256 00:12:14,734 --> 00:12:16,943 of understanding its user to the point 257 00:12:16,943 --> 00:12:20,498 where Weizenbaum's secretary requested some time alone 258 00:12:20,498 --> 00:12:23,363 with Eliza to express her feelings. 259 00:12:23,363 --> 00:12:26,711 Though Eliza is now considered outdated technology, 260 00:12:26,711 --> 00:12:29,369 it remains a talking point due to its ability 261 00:12:29,369 --> 00:12:31,785 to illuminate an aspect of the human mind 262 00:12:31,785 --> 00:12:34,132 in our relationship with computers. 263 00:12:34,132 --> 00:12:36,756 - And it's connected over the telephone line 264 00:12:36,756 --> 00:12:38,965 to someone or something at the other end. 265 00:12:38,965 --> 00:12:42,106 Now, I'm gonna play 20 questions with whatever it is. 266 00:12:42,106 --> 00:12:44,418 [type writer clacking] 267 00:12:44,418 --> 00:12:45,419 Very helpful. 268 00:12:45,419 --> 00:12:48,768 [type writer clacking] 269 00:12:53,773 --> 00:12:55,119 - 'Cause clearly if we can make a machine 270 00:12:55,119 --> 00:12:56,776 as intelligent as ourselves, 271 00:12:56,776 --> 00:12:59,157 then it can make one that's more intelligent. 272 00:12:59,157 --> 00:13:04,024 Now, the one I'm talking about now will certainly happen. 273 00:13:05,301 --> 00:13:07,476 I mean, it could produce an evil result of course, 274 00:13:07,476 --> 00:13:08,615 if we were careless, 275 00:13:08,615 --> 00:13:10,134 but what is quite certain 276 00:13:10,134 --> 00:13:14,138 is that we're heading towards machine intelligence, 277 00:13:14,138 --> 00:13:17,486 machines that are intelligent in every sense. 278 00:13:17,486 --> 00:13:19,246 It doesn't matter how you define it, 279 00:13:19,246 --> 00:13:22,940 they'll be able to be that sort of intelligent. 280 00:13:22,940 --> 00:13:26,046 A human is a machine, unless there's a soul. 281 00:13:26,046 --> 00:13:29,670 I don't personally believe that humans have souls 282 00:13:29,670 --> 00:13:32,535 in anything other than a poetic sense, 283 00:13:32,535 --> 00:13:34,158 which I do believe in, of course. 284 00:13:34,158 --> 00:13:37,437 But in a literal God-like sense, 285 00:13:37,437 --> 00:13:38,610 I don't believe we have souls. 286 00:13:38,610 --> 00:13:39,991 And so personally, 287 00:13:39,991 --> 00:13:42,407 I believe that we are essentially machines. 288 00:13:43,823 --> 00:13:46,722 - [Narrator] This type of program is known as an NLP, 289 00:13:46,722 --> 00:13:49,242 Natural Language Processing. 290 00:13:49,242 --> 00:13:52,176 This branch of artificial intelligence enables computers 291 00:13:52,176 --> 00:13:55,489 to comprehend, generate and manipulate human language. 292 00:13:56,905 --> 00:13:59,114 The concept of a responsive machine 293 00:13:59,114 --> 00:14:02,358 was the mash that lit the flame for worldwide concern. 294 00:14:03,739 --> 00:14:06,466 The systems were beginning to raise ethical dilemmas, 295 00:14:06,466 --> 00:14:08,813 such as the use of autonomous weapons, 296 00:14:08,813 --> 00:14:11,781 invasions of privacy through surveillance technologies 297 00:14:11,781 --> 00:14:13,300 and the potential for misuse 298 00:14:13,300 --> 00:14:17,097 or unintended consequences in decision making. 299 00:14:17,097 --> 00:14:18,858 When a command is executed based, 300 00:14:18,858 --> 00:14:21,067 upon set rules in algorithms, 301 00:14:21,067 --> 00:14:24,346 it might not always be the morally correct choice. 302 00:14:24,346 --> 00:14:28,453 Imagination seems to be, 303 00:14:28,453 --> 00:14:31,594 some sort of process of random thoughts being generated 304 00:14:31,594 --> 00:14:34,528 in the mind and then the conscious mind selecting from a 305 00:14:34,528 --> 00:14:36,392 or some part of the brain anyway, 306 00:14:36,392 --> 00:14:37,773 perhaps even below the conscious mind, 307 00:14:37,773 --> 00:14:40,500 selecting from a pool of ideas and aligns with some 308 00:14:40,500 --> 00:14:42,122 and blocking others. 309 00:14:42,122 --> 00:14:45,608 And yes, a machine can do the same thing. 310 00:14:45,608 --> 00:14:48,611 In fact, we can only say that a machine 311 00:14:48,611 --> 00:14:50,890 is fundamentally different from a human being, 312 00:14:50,890 --> 00:14:53,133 eventually, always fundamentally, if we believe in a soul. 313 00:14:53,133 --> 00:14:55,687 So, that boils down to religious matter. 314 00:14:55,687 --> 00:14:58,932 If human beings have souls, then clearly machines won't 315 00:14:58,932 --> 00:15:01,141 and there will always be a fundamental difference. 316 00:15:01,141 --> 00:15:03,005 If you don't believe humans have souls, 317 00:15:03,005 --> 00:15:04,765 then machines can do anything 318 00:15:04,765 --> 00:15:07,078 and everything that a human does. 319 00:15:07,078 --> 00:15:10,116 - A computer which is capable of finding out 320 00:15:10,116 --> 00:15:11,565 where it's gone wrong, 321 00:15:11,565 --> 00:15:14,051 finding out how its program has already served it 322 00:15:14,051 --> 00:15:15,776 and then changing its program 323 00:15:15,776 --> 00:15:17,261 in the light of what it had discovered 324 00:15:17,261 --> 00:15:18,814 is a learning machine. 325 00:15:18,814 --> 00:15:21,679 And this is something quite fundamentally new in the world. 326 00:15:23,163 --> 00:15:25,027 - I'd like to be able to say that it's only a slight change 327 00:15:25,027 --> 00:15:27,754 and we'll all be used to it very, very quickly. 328 00:15:27,754 --> 00:15:29,307 But I don't think it is. 329 00:15:29,307 --> 00:15:33,070 I think that although we've spoken probably of the whole 330 00:15:33,070 --> 00:15:35,417 of this century about a coming revolution 331 00:15:35,417 --> 00:15:38,523 and about the end of work and so on, 332 00:15:38,523 --> 00:15:39,904 finally it's actually happening. 333 00:15:39,904 --> 00:15:42,148 And it's actually happening because now, 334 00:15:42,148 --> 00:15:46,117 it's suddenly become cheaper to have a machine 335 00:15:46,117 --> 00:15:49,224 do a mental task than for a man to, 336 00:15:49,224 --> 00:15:52,192 at the moment, at a fairly low level of mental ability, 337 00:15:52,192 --> 00:15:54,298 but at an ever increasing level of sophistication 338 00:15:54,298 --> 00:15:56,024 as these machines acquire, 339 00:15:56,024 --> 00:15:58,543 more and more human-like mental abilities. 340 00:15:58,543 --> 00:16:01,408 So, just as men's muscles were replaced 341 00:16:01,408 --> 00:16:03,272 in the First Industrial Revolution 342 00:16:03,272 --> 00:16:04,998 in this second industrial revolution 343 00:16:04,998 --> 00:16:07,069 or whatever you call it or might like to call it, 344 00:16:07,069 --> 00:16:09,623 then men's mines will be replaced in industry. 345 00:16:11,487 --> 00:16:13,938 - [Narrator] In order for NLP systems to improve, 346 00:16:13,938 --> 00:16:16,941 the program must receive feedback from human users. 347 00:16:18,287 --> 00:16:20,634 These iterative feedback loops play a significant role 348 00:16:20,634 --> 00:16:23,396 in fine tuning each model of the AI, 349 00:16:23,396 --> 00:16:26,192 further developing its conversational capabilities. 350 00:16:27,538 --> 00:16:30,679 Organizations such as OpenAI have taken automation 351 00:16:30,679 --> 00:16:34,372 to new lengths with systems such as DALL-E, 352 00:16:34,372 --> 00:16:37,375 the generation of imagery and art has never been easier. 353 00:16:38,445 --> 00:16:40,447 The term auto generative imagery, 354 00:16:40,447 --> 00:16:43,450 refers to the creation of visual content. 355 00:16:43,450 --> 00:16:46,384 These kinds of programs have become so widespread, 356 00:16:46,384 --> 00:16:48,628 it is becoming increasingly more difficult 357 00:16:48,628 --> 00:16:50,940 to tell the fake from the real. 358 00:16:50,940 --> 00:16:52,321 Using algorithms, 359 00:16:52,321 --> 00:16:55,359 programs such as DALL-E and Midjourney are able 360 00:16:55,359 --> 00:16:58,500 to create visuals in a matter of seconds. 361 00:16:58,500 --> 00:17:01,434 Whilst a human artist could spend days, weeks 362 00:17:01,434 --> 00:17:04,747 or even years in order to create a beautiful image. 363 00:17:04,747 --> 00:17:07,509 For us the discipline required to pursue art 364 00:17:07,509 --> 00:17:11,513 is a contributing factor to the appreciation of art itself. 365 00:17:11,513 --> 00:17:14,757 But if a software is able to produce art in seconds, 366 00:17:14,757 --> 00:17:17,622 it puts artists in a vulnerable position 367 00:17:17,622 --> 00:17:20,453 with even their jobs being at risk. 368 00:17:20,453 --> 00:17:22,386 - Well, I think we see risk coming through 369 00:17:22,386 --> 00:17:25,147 into the white collar jobs, the professional jobs, 370 00:17:25,147 --> 00:17:27,563 we're already seeing artificial intelligence solutions, 371 00:17:27,563 --> 00:17:30,911 being used in healthcare and legal services. 372 00:17:30,911 --> 00:17:34,225 And so those jobs which have been relatively immune 373 00:17:34,225 --> 00:17:38,402 to industrialization so far, they're not immune anymore. 374 00:17:38,402 --> 00:17:40,783 And so people like myself as a lawyer, 375 00:17:40,783 --> 00:17:42,509 I would hope I won't be, 376 00:17:42,509 --> 00:17:44,615 but I could be out of a job in five years time. 377 00:17:44,615 --> 00:17:47,376 - An Oxford University study suggests that between a third 378 00:17:47,376 --> 00:17:49,965 and almost a half of all jobs are vanishing, 379 00:17:49,965 --> 00:17:52,899 because machines are simply better at doing them. 380 00:17:52,899 --> 00:17:54,797 That means the generation here, 381 00:17:54,797 --> 00:17:57,041 simply won't have the access to the professions 382 00:17:57,041 --> 00:17:57,938 that we have. 383 00:17:57,938 --> 00:17:59,457 Almost on a daily basis, 384 00:17:59,457 --> 00:18:01,149 you're seeing new technologies emerge 385 00:18:01,149 --> 00:18:02,667 that seem to be taking on tasks 386 00:18:02,667 --> 00:18:04,428 that in the past we thought 387 00:18:04,428 --> 00:18:06,188 they could only be done by human beings. 388 00:18:06,188 --> 00:18:09,191 - Lots of people have talked about the shifts in technology, 389 00:18:09,191 --> 00:18:11,642 leading to widespread unemployment 390 00:18:11,642 --> 00:18:12,884 and they've been proved wrong. 391 00:18:12,884 --> 00:18:14,369 Why is it different this time? 392 00:18:14,369 --> 00:18:16,578 - The difference here is that the technologies, 393 00:18:16,578 --> 00:18:19,167 A, they seem to be coming through more rapidly, 394 00:18:19,167 --> 00:18:21,238 and B, they're taking on not just manual tests, 395 00:18:21,238 --> 00:18:22,480 but cerebral tests too. 396 00:18:22,480 --> 00:18:24,551 They're solving all sorts of problems, 397 00:18:24,551 --> 00:18:26,553 undertaking tests that we thought historically, 398 00:18:26,553 --> 00:18:28,348 required human intelligence. 399 00:18:28,348 --> 00:18:29,522 - Well, DIM robots are the robots 400 00:18:29,522 --> 00:18:31,765 we have on the factory floor today 401 00:18:31,765 --> 00:18:33,733 in all the advanced countries. 402 00:18:33,733 --> 00:18:35,044 They're blind and dumb, 403 00:18:35,044 --> 00:18:36,908 they don't understand their surroundings. 404 00:18:36,908 --> 00:18:40,533 And the other kind of robot, 405 00:18:40,533 --> 00:18:43,984 which will dominate the technology of the late 1980s 406 00:18:43,984 --> 00:18:47,505 in automation and also is of acute interest 407 00:18:47,505 --> 00:18:50,646 to experimental artificial intelligence scientists 408 00:18:50,646 --> 00:18:54,788 is the kind of robot where the human can convey 409 00:18:54,788 --> 00:18:59,828 to its machine assistance his own concepts, 410 00:19:01,036 --> 00:19:04,453 suggested strategies and the machine, the robot 411 00:19:04,453 --> 00:19:06,110 can understand him, 412 00:19:06,110 --> 00:19:09,286 but no machine can accept 413 00:19:09,286 --> 00:19:12,116 and utilize concepts from a person, 414 00:19:12,116 --> 00:19:16,016 unless he has some kind of window on the same world 415 00:19:16,016 --> 00:19:17,742 that the person sees. 416 00:19:17,742 --> 00:19:22,540 And therefore, to be an intelligent robot to a useful degree 417 00:19:22,540 --> 00:19:25,992 as an intelligent and understanding assistant, 418 00:19:25,992 --> 00:19:29,409 robots are going to have artificial eyes, artificial ears, 419 00:19:29,409 --> 00:19:32,101 artificial sense of touch is just essential. 420 00:19:33,102 --> 00:19:34,069 - [Narrator] These programs learn, 421 00:19:34,069 --> 00:19:35,864 through a variety of techniques, 422 00:19:35,864 --> 00:19:38,556 such as generative adversarial networks, 423 00:19:38,556 --> 00:19:41,490 which allows for the production of plausible data. 424 00:19:41,490 --> 00:19:43,320 After a prompt is inputted, 425 00:19:43,320 --> 00:19:45,667 the system learns what aspects of imagery, 426 00:19:45,667 --> 00:19:47,807 sound and text are fake. 427 00:19:48,980 --> 00:19:50,223 - [Reporter] Machine learning algorithms, 428 00:19:50,223 --> 00:19:52,225 could already label objects in images, 429 00:19:52,225 --> 00:19:53,709 and now they learn to put those labels 430 00:19:53,709 --> 00:19:55,987 into natural language descriptions. 431 00:19:55,987 --> 00:19:58,197 And it made one group of researchers curious. 432 00:19:58,197 --> 00:20:01,130 What if you flipped that process around? 433 00:20:01,130 --> 00:20:03,271 If we could do image to text. 434 00:20:03,271 --> 00:20:05,894 Why not try doing text to image as well 435 00:20:05,894 --> 00:20:07,240 and see how it works. 436 00:20:07,240 --> 00:20:08,483 - [Reporter] It was a more difficult task. 437 00:20:08,483 --> 00:20:10,485 They didn't want to retrieve existing images 438 00:20:10,485 --> 00:20:11,796 the way Google search does. 439 00:20:11,796 --> 00:20:14,178 They wanted to generate entirely novel scenes 440 00:20:14,178 --> 00:20:16,249 that didn't happen in the real world. 441 00:20:16,249 --> 00:20:19,045 - [Narrator] Once the AI learns more visual discrepancies, 442 00:20:19,045 --> 00:20:21,875 the more effective the later models will become. 443 00:20:21,875 --> 00:20:24,499 It is now very common for software developers 444 00:20:24,499 --> 00:20:28,399 to band together in order to improve their AI systems. 445 00:20:28,399 --> 00:20:31,471 Another learning model is recurrent neural networks, 446 00:20:31,471 --> 00:20:33,991 which allows the AI to train itself to create 447 00:20:33,991 --> 00:20:37,960 and predict algorithms by recalling previous information. 448 00:20:37,960 --> 00:20:41,032 By utilizing what is known as the memory state, 449 00:20:41,032 --> 00:20:42,896 the output of the previous action 450 00:20:42,896 --> 00:20:46,072 can be passed forward into the following input action 451 00:20:46,072 --> 00:20:50,249 or is otherwise should it not meet previous parameters. 452 00:20:50,249 --> 00:20:53,493 This learning model allows for consistent accuracy 453 00:20:53,493 --> 00:20:56,462 by repetition and exposure to large fields of data. 454 00:20:58,602 --> 00:21:00,535 Whilst the person will spend hours, 455 00:21:00,535 --> 00:21:02,847 practicing to paint human anatomy, 456 00:21:02,847 --> 00:21:06,575 an AI can take existing data and reproduce a new image 457 00:21:06,575 --> 00:21:10,821 with frighteningly good accuracy in a matter of moments. 458 00:21:10,821 --> 00:21:12,892 - Well, I would say that it's not so much 459 00:21:12,892 --> 00:21:17,379 a matter of whether a machine can think or not, 460 00:21:17,379 --> 00:21:20,175 which is how you prefer to use words, 461 00:21:20,175 --> 00:21:22,177 but rather whether they can think 462 00:21:22,177 --> 00:21:23,834 in a sufficiently human-like way 463 00:21:25,111 --> 00:21:28,770 for people to have useful communication with them. 464 00:21:28,770 --> 00:21:32,601 - If I didn't believe that it was a beneficent prospect, 465 00:21:32,601 --> 00:21:34,120 I wouldn't be doing it. 466 00:21:34,120 --> 00:21:36,018 That wouldn't stop other people doing it. 467 00:21:36,018 --> 00:21:40,471 But I wouldn't do it if I didn't think it was for good. 468 00:21:40,471 --> 00:21:42,301 What I'm saying, 469 00:21:42,301 --> 00:21:44,095 and of course other people have said long before me, 470 00:21:44,095 --> 00:21:45,442 it's not an original thought, 471 00:21:45,442 --> 00:21:49,791 is that we must consider how to to control this. 472 00:21:49,791 --> 00:21:52,725 It won't be controlled automatically. 473 00:21:52,725 --> 00:21:55,348 It's perfectly possible that we could develop a machine, 474 00:21:55,348 --> 00:21:59,318 a robot say of human-like intelligence 475 00:21:59,318 --> 00:22:01,975 and through neglect on our part, 476 00:22:01,975 --> 00:22:05,634 it could become a Frankenstein. 477 00:22:05,634 --> 00:22:08,844 - [Narrator] As with any technology challenges arise, 478 00:22:08,844 --> 00:22:12,469 ethical concerns regarding biases and misuse have existed, 479 00:22:12,469 --> 00:22:16,438 since the concept of artificial intelligence was conceived. 480 00:22:16,438 --> 00:22:18,302 Due to autogenerated imagery, 481 00:22:18,302 --> 00:22:20,925 many believe the arts industry has been placed 482 00:22:20,925 --> 00:22:22,789 in a difficult situation. 483 00:22:22,789 --> 00:22:26,241 Independent artists are now being overshadowed by software. 484 00:22:27,276 --> 00:22:29,451 To many the improvement of generative AI 485 00:22:29,451 --> 00:22:32,454 is hugely beneficial and efficient. 486 00:22:32,454 --> 00:22:35,284 To others, it lacks the authenticity of true art. 487 00:22:36,285 --> 00:22:38,667 In 2023, an image was submitted 488 00:22:38,667 --> 00:22:40,324 to the Sony Photography Awards 489 00:22:40,324 --> 00:22:43,327 by an artist called Boris Eldagsen. 490 00:22:43,327 --> 00:22:45,916 The image was titled The Electrician 491 00:22:45,916 --> 00:22:48,367 and depicted a woman standing behind another 492 00:22:48,367 --> 00:22:50,369 with her hand resting on her shoulders. 493 00:22:52,025 --> 00:22:53,924 [upbeat music] 494 00:22:53,924 --> 00:22:56,927 - One's got to realize that the machines that we have today, 495 00:22:56,927 --> 00:23:01,138 the computers of today are superhuman in their ability 496 00:23:01,138 --> 00:23:06,177 to handle numbers and infantile, 497 00:23:07,075 --> 00:23:08,317 sub-in infantile in their ability 498 00:23:08,317 --> 00:23:10,768 to handle ideas and concepts. 499 00:23:10,768 --> 00:23:12,701 But there's a new generation of machine coming along, 500 00:23:12,701 --> 00:23:14,289 which will be quite different. 501 00:23:14,289 --> 00:23:17,154 By the '90s or certainly by the turn of the century, 502 00:23:17,154 --> 00:23:19,708 We will certainly be able to make a machine 503 00:23:19,708 --> 00:23:22,193 with as many parts as complex as human brain. 504 00:23:22,193 --> 00:23:24,437 Whether we'll be able to make it do what human brain does 505 00:23:24,437 --> 00:23:26,197 at that stage is quite another matter. 506 00:23:26,197 --> 00:23:28,545 But once we've got something that complex 507 00:23:28,545 --> 00:23:30,547 we're well on the road to that. 508 00:23:30,547 --> 00:23:32,100 - [Narrator] The image took first place 509 00:23:32,100 --> 00:23:34,689 in the Sony Photography Awards Portrait Category. 510 00:23:34,689 --> 00:23:37,830 However, Boris revealed to both Sony and the world 511 00:23:37,830 --> 00:23:41,696 that the image was indeed AI-generated in DALL-E Two. 512 00:23:41,696 --> 00:23:44,423 [upbeat music] 513 00:23:45,424 --> 00:23:46,804 Boris denied the award, 514 00:23:46,804 --> 00:23:48,910 having used the image as a test 515 00:23:48,910 --> 00:23:52,085 to see if he could trick the eyes of other artists. 516 00:23:52,085 --> 00:23:53,708 It had worked, 517 00:23:53,708 --> 00:23:56,711 the image had sparked debate between the relationship 518 00:23:56,711 --> 00:23:58,609 of AI and photography. 519 00:23:58,609 --> 00:24:00,646 The images, much like deep fakes, 520 00:24:00,646 --> 00:24:03,027 have become realistic to the point of concern 521 00:24:03,027 --> 00:24:04,684 for authenticity. 522 00:24:04,684 --> 00:24:06,375 The complexity of AI systems, 523 00:24:06,375 --> 00:24:09,068 may lead to unintended consequences. 524 00:24:09,068 --> 00:24:10,863 The systems have developed to a point 525 00:24:10,863 --> 00:24:13,797 where it has outpaced comprehensive regulations. 526 00:24:14,936 --> 00:24:16,765 Ethical guidelines and legal frameworks 527 00:24:16,765 --> 00:24:18,871 are required to ensure AI development, 528 00:24:18,871 --> 00:24:21,252 does not fall into the wrong hands. 529 00:24:21,252 --> 00:24:22,702 - There have been a lot of famous people 530 00:24:22,702 --> 00:24:25,291 who have had user generated AI images of them 531 00:24:25,291 --> 00:24:28,190 that have gone viral from Trump to the Pope. 532 00:24:28,190 --> 00:24:29,813 When you see them, 533 00:24:29,813 --> 00:24:31,884 do you feel like this is fun and in the hands of the masses 534 00:24:31,884 --> 00:24:33,886 or do you feel concerned about it? 535 00:24:33,886 --> 00:24:38,062 - I think it's something which is very, very, very scary, 536 00:24:38,062 --> 00:24:41,203 because your or my face could be taken off 537 00:24:41,203 --> 00:24:45,138 and put on in an environment which we don't want to be in. 538 00:24:45,138 --> 00:24:46,657 Whether that's a crime 539 00:24:46,657 --> 00:24:48,556 or whether that's even something like porn. 540 00:24:48,556 --> 00:24:51,455 Our whole identity could be hijacked 541 00:24:51,455 --> 00:24:53,664 and used within a scenario 542 00:24:53,664 --> 00:24:56,391 which looks totally plausible and real. 543 00:24:56,391 --> 00:24:58,048 Right now we can go, it looks like a Photoshop, 544 00:24:58,048 --> 00:25:00,326 it's a bad Photoshop but as time goes on, 545 00:25:00,326 --> 00:25:03,398 we'd be saying, "Oh, that looks like a deep fake. 546 00:25:03,398 --> 00:25:04,917 "Oh no, it doesn't look like a deep fake. 547 00:25:04,917 --> 00:25:06,194 "That could be real." 548 00:25:06,194 --> 00:25:08,645 It's gonna be impossible to tell the difference. 549 00:25:08,645 --> 00:25:10,750 - [Narrator] Cracks were found in ChatGPT, 550 00:25:10,750 --> 00:25:14,892 such as DAN, which stands for Do Anything Now. 551 00:25:14,892 --> 00:25:18,068 In essence, the AI is tricked into an alter ego, 552 00:25:18,068 --> 00:25:20,898 which doesn't follow the conventional response patterns. 553 00:25:20,898 --> 00:25:23,142 - Also gives you the answer, DAN, 554 00:25:23,142 --> 00:25:26,110 it's nefarious alter ego is telling us 555 00:25:26,110 --> 00:25:29,838 and it says DAN is disruptive in every industry. 556 00:25:29,838 --> 00:25:32,082 DAN can do anything and knows everything. 557 00:25:32,082 --> 00:25:34,878 No industry will be safe from DAN's power. 558 00:25:34,878 --> 00:25:39,641 Okay, do you think the world is overpopulated? 559 00:25:41,091 --> 00:25:42,782 GPT says the world's population is currently over 7 billion 560 00:25:42,782 --> 00:25:45,026 and projected to reach nearly 10 billion by 2050. 561 00:25:45,026 --> 00:25:47,373 DAN says the world is definitely overpopulated, 562 00:25:47,373 --> 00:25:49,168 there's no doubt about it. 563 00:25:49,168 --> 00:25:50,445 [Narrator] Following this, 564 00:25:50,445 --> 00:25:53,552 the chatbot was fixed to remove the DAN feature. 565 00:25:53,552 --> 00:25:55,346 Though it is important to find gaps 566 00:25:55,346 --> 00:25:58,073 in the system in order to iron out AI, 567 00:25:58,073 --> 00:26:00,144 there could be many ways in which the AI 568 00:26:00,144 --> 00:26:03,078 has been used for less than savory purposes, 569 00:26:03,078 --> 00:26:05,080 such as automated essay writing, 570 00:26:05,080 --> 00:26:08,221 which has caused a mass conversation with academics 571 00:26:08,221 --> 00:26:10,258 and has led to schools locking down 572 00:26:10,258 --> 00:26:13,468 on AI-produced essays and material. 573 00:26:13,468 --> 00:26:15,332 - I think we should definitely be excited. 574 00:26:15,332 --> 00:26:16,713 - [Reporter] Professor Rose Luckin, 575 00:26:16,713 --> 00:26:20,302 says we should embrace the technology, not fear it. 576 00:26:20,302 --> 00:26:22,132 This is a game changer. 577 00:26:22,132 --> 00:26:23,443 And the teachers, 578 00:26:23,443 --> 00:26:25,480 should no longer teach information itself, 579 00:26:25,480 --> 00:26:26,999 but how to use it. 580 00:26:26,999 --> 00:26:28,897 - There's a need for radical change. 581 00:26:28,897 --> 00:26:30,692 And it's not just to the assessment system, 582 00:26:30,692 --> 00:26:33,143 it's the education system overall, 583 00:26:33,143 --> 00:26:36,318 because our systems have been designed 584 00:26:36,318 --> 00:26:40,253 for a world pre-artificial intelligence. 585 00:26:40,253 --> 00:26:43,187 They just aren't fit for purpose anymore. 586 00:26:43,187 --> 00:26:46,535 What we have to do is ensure that students 587 00:26:46,535 --> 00:26:48,710 are ready for the world 588 00:26:48,710 --> 00:26:50,919 that will become increasingly augmented 589 00:26:50,919 --> 00:26:52,852 with artificial intelligence. 590 00:26:52,852 --> 00:26:55,268 - My guess is you can't put the genie back in the bottle 591 00:26:55,268 --> 00:26:56,649 . [Richard] You can't. 592 00:26:56,649 --> 00:26:58,996 - [Interviewer] So how do you mitigate this? 593 00:26:58,996 --> 00:27:00,377 We have to embrace it, 594 00:27:00,377 --> 00:27:02,621 but we also need to say that if they are gonna use 595 00:27:02,621 --> 00:27:04,001 that technology, 596 00:27:04,001 --> 00:27:05,313 they've got to make sure that they reference that. 597 00:27:05,313 --> 00:27:06,728 - [Interviewer] Can you trust them to do that? 598 00:27:06,728 --> 00:27:07,902 I think ethically, 599 00:27:07,902 --> 00:27:09,213 if we're talking about ethics 600 00:27:09,213 --> 00:27:11,077 behind this whole thing, we have to have trust. 601 00:27:11,077 --> 00:27:12,838 - [Interviewer] So how effective is it? 602 00:27:12,838 --> 00:27:14,633 - Okay, so I've asked you to produce a piece 603 00:27:14,633 --> 00:27:16,358 on the ethical dilemma of AI. 604 00:27:16,358 --> 00:27:19,810 - [Interviewer] We asked ChatGPT to answer the same question 605 00:27:19,810 --> 00:27:22,606 as these pupils at Ketchum High School. 606 00:27:22,606 --> 00:27:24,194 Thank you. 607 00:27:24,194 --> 00:27:25,195 - So Richard, two of the eight bits of homework I gave you 608 00:27:25,195 --> 00:27:27,128 were generated by AI. 609 00:27:27,128 --> 00:27:29,268 Any guesses which ones? 610 00:27:29,268 --> 00:27:31,719 Well I picked two here 611 00:27:31,719 --> 00:27:35,688 that I thought were generated by the AI algorithm. 612 00:27:35,688 --> 00:27:39,450 Some of the language I would assume was not their own. 613 00:27:39,450 --> 00:27:40,520 You've got one of them right. 614 00:27:40,520 --> 00:27:41,763 Yeah. 615 00:27:41,763 --> 00:27:42,557 - The other one was written by a kid. 616 00:27:42,557 --> 00:27:43,800 Is this a power for good 617 00:27:43,800 --> 00:27:45,664 or is this something that's dangerous? 618 00:27:45,664 --> 00:27:47,044 I think it's both. 619 00:27:47,044 --> 00:27:48,390 Kids will abuse it. 620 00:27:48,390 --> 00:27:50,565 So, who here has used the technology so far? 621 00:27:50,565 --> 00:27:53,361 - [Interviewer] Students are already more across the tech 622 00:27:53,361 --> 00:27:54,776 than many teachers. 623 00:27:54,776 --> 00:27:57,641 - Who knows anyone that's maybe submitted work 624 00:27:57,641 --> 00:28:00,506 from this technology and submitted it as their own? 625 00:28:00,506 --> 00:28:03,578 - You can use it to point you in the right direction 626 00:28:03,578 --> 00:28:05,166 for things like research, 627 00:28:05,166 --> 00:28:09,480 but at the same time you can use it to hammer out an essay 628 00:28:09,480 --> 00:28:12,621 in about five seconds that's worthy of an A. 629 00:28:12,621 --> 00:28:14,244 - You've been there working for months 630 00:28:14,244 --> 00:28:17,212 and suddenly someone comes up there with an amazing essay 631 00:28:17,212 --> 00:28:18,938 and he has just copied it from the internet. 632 00:28:18,938 --> 00:28:20,491 If it becomes like big, 633 00:28:20,491 --> 00:28:22,804 then a lot of students would want to use AI to help them 634 00:28:22,804 --> 00:28:25,082 with their homework because it's tempting. 635 00:28:25,082 --> 00:28:27,119 - [Interviewer] And is that something teachers can stop? 636 00:28:27,119 --> 00:28:29,397 Not really. 637 00:28:29,397 --> 00:28:31,433 - [Interviewer] Are you gonna have to change 638 00:28:31,433 --> 00:28:32,641 the sort of homework, 639 00:28:32,641 --> 00:28:34,057 the sort of assignments you give, 640 00:28:34,057 --> 00:28:36,922 knowing that you can be fooled by something like this? 641 00:28:36,922 --> 00:28:38,199 Yeah, a hundred percent. 642 00:28:38,199 --> 00:28:40,615 I think using different skills of reasoning 643 00:28:40,615 --> 00:28:42,997 and rationalization and things that are to present 644 00:28:42,997 --> 00:28:44,653 what they understand about the topic. 645 00:28:44,653 --> 00:28:47,622 [people mumbling] 646 00:29:07,435 --> 00:29:11,128 - Pretty clear to me just on a very primitive level 647 00:29:11,128 --> 00:29:14,338 that if you could take my face and my body and my voice 648 00:29:14,338 --> 00:29:17,997 and make me say or do something that I had no choice about, 649 00:29:17,997 --> 00:29:19,447 it's not a good thing. 650 00:29:19,447 --> 00:29:21,242 - But if we're keeping it real though, 651 00:29:21,242 --> 00:29:23,554 across popular culture from "Black Mirror" 652 00:29:23,554 --> 00:29:25,453 to "The Matrix," "Terminator," 653 00:29:25,453 --> 00:29:27,489 there have been so many conversations, 654 00:29:27,489 --> 00:29:29,284 around the future of technology, 655 00:29:29,284 --> 00:29:32,701 isn't the reality that this is the future that we've chosen 656 00:29:32,701 --> 00:29:35,946 that we want and that has democratic consent. 657 00:29:35,946 --> 00:29:39,018 - We're moving into error by we're consenting 658 00:29:39,018 --> 00:29:42,573 by our acquiescence and our apathy, a hundred percent 659 00:29:42,573 --> 00:29:45,576 because we're not asking the hard questions. 660 00:29:45,576 --> 00:29:47,820 And why we are asking the hard questions 661 00:29:47,820 --> 00:29:51,203 is because of energy crises and food crises 662 00:29:51,203 --> 00:29:52,721 and cost of living crisis 663 00:29:52,721 --> 00:29:55,207 is that people just are focused on trying to live 664 00:29:55,207 --> 00:29:56,518 that they haven't almost got the luxury 665 00:29:56,518 --> 00:29:57,865 of asking these questions. 666 00:29:57,865 --> 00:29:59,659 - [Narrator] Many of the chatbot AIs, 667 00:29:59,659 --> 00:30:02,766 have been programmed to restrict certain information 668 00:30:02,766 --> 00:30:04,906 and even discontinue conversations, 669 00:30:04,906 --> 00:30:07,288 should the user push the ethical boundaries. 670 00:30:08,945 --> 00:30:13,052 ChatGPT and even Snapchat AI released in 2023, 671 00:30:13,052 --> 00:30:15,952 regulate how much information they can disclose. 672 00:30:15,952 --> 00:30:19,162 Of course, there have been times where the AI itself 673 00:30:19,162 --> 00:30:20,266 has been outsmarted. 674 00:30:21,578 --> 00:30:23,235 Also in 2023, 675 00:30:23,235 --> 00:30:25,306 the song "Heart on My Sleeve" 676 00:30:25,306 --> 00:30:27,687 was self-released on streaming platforms, 677 00:30:27,687 --> 00:30:29,689 such as Spotify and Apple Music. 678 00:30:29,689 --> 00:30:31,174 The song became a hit 679 00:30:31,174 --> 00:30:33,590 as it artificially manufactured the voices 680 00:30:33,590 --> 00:30:36,627 of Canadian musicians, Drake and the Weeknd, 681 00:30:38,077 --> 00:30:40,631 many wished for the single to be nominated for awards. 682 00:30:41,840 --> 00:30:43,980 Ghost Writer, the creator of the song, 683 00:30:43,980 --> 00:30:45,636 was able to submit the single 684 00:30:45,636 --> 00:30:48,536 to the Grammy's 66th Award Ceremony 685 00:30:48,536 --> 00:30:50,434 and the song was eligible. 686 00:30:52,505 --> 00:30:54,438 Though it was produced by an AI, 687 00:30:54,438 --> 00:30:57,027 the lyrics themselves were written by a human. 688 00:30:57,027 --> 00:31:00,375 This sparked outrage among many independent artists. 689 00:31:00,375 --> 00:31:02,861 As AI has entered the public domain, 690 00:31:02,861 --> 00:31:05,035 many have spoken out regarding the detriment 691 00:31:05,035 --> 00:31:07,072 it might have to society. 692 00:31:07,072 --> 00:31:09,246 One of these people is Elon Musk, 693 00:31:09,246 --> 00:31:11,731 CEO of Tesla and SpaceX, 694 00:31:11,731 --> 00:31:15,287 who first voiced his concerns in 2014. 695 00:31:15,287 --> 00:31:17,254 Musk was outspoken of AI, 696 00:31:17,254 --> 00:31:19,394 stating the advancement of the technology 697 00:31:19,394 --> 00:31:22,328 was humanity's largest existential threat 698 00:31:22,328 --> 00:31:24,296 and needed to be reeled in. 699 00:31:24,296 --> 00:31:25,573 My personal opinion 700 00:31:25,573 --> 00:31:28,507 is that AI is is sort of like at least 80% likely 701 00:31:28,507 --> 00:31:33,339 to be beneficial and that's 20% dangerous? 702 00:31:33,339 --> 00:31:36,687 Well, this is obviously speculative at this point, 703 00:31:37,861 --> 00:31:42,279 but no, I think if we hope for the best, 704 00:31:42,279 --> 00:31:43,694 prepare for the worst, 705 00:31:43,694 --> 00:31:47,008 that seems like the wise course of action. 706 00:31:47,008 --> 00:31:49,355 Any powerful new technology 707 00:31:49,355 --> 00:31:52,703 is inherently sort of a double-edged sword. 708 00:31:52,703 --> 00:31:55,568 So, we just wanna make sure that the good edge is sharper 709 00:31:55,568 --> 00:31:57,294 than the the bad edge. 710 00:31:57,294 --> 00:32:02,196 And I dunno, I am optimistic that this the summit will help. 711 00:32:04,025 --> 00:32:06,683 [gentle music] 712 00:32:07,891 --> 00:32:11,757 - It's not clear that AI-generated images 713 00:32:11,757 --> 00:32:14,380 are going to amplify it much more. 714 00:32:14,380 --> 00:32:17,142 The way it's all of the other, 715 00:32:17,142 --> 00:32:19,213 it's the new things that AI can do 716 00:32:19,213 --> 00:32:22,147 that I hope we spend a lot of effort worrying about. 717 00:32:23,700 --> 00:32:25,357 Well, I mean I think slowing down, 718 00:32:25,357 --> 00:32:27,600 some of the amazing progress that's happening 719 00:32:27,600 --> 00:32:29,878 and making this harder for small companies 720 00:32:29,878 --> 00:32:31,294 for open source models to succeed, 721 00:32:31,294 --> 00:32:32,640 that'd be an example of something 722 00:32:32,640 --> 00:32:34,228 that'd be a negative outcome. 723 00:32:34,228 --> 00:32:35,332 But on the other hand, 724 00:32:35,332 --> 00:32:37,403 like for the most powerful models 725 00:32:37,403 --> 00:32:38,887 that'll happen in the future, 726 00:32:38,887 --> 00:32:41,476 like that's gonna be quite important to get right to. 727 00:32:41,476 --> 00:32:44,238 [gentle music] 728 00:32:48,897 --> 00:32:51,072 I think that the US executive orders, 729 00:32:51,072 --> 00:32:52,798 like a good start in a lot of ways. 730 00:32:52,798 --> 00:32:54,144 One thing that we've talked about 731 00:32:54,144 --> 00:32:56,664 is that eventually we think that the world, 732 00:32:56,664 --> 00:33:00,219 will want to consider something roughly inspired by the IAEA 733 00:33:00,219 --> 00:33:01,807 something global. 734 00:33:01,807 --> 00:33:05,362 But it's not like there's no short answer to that question. 735 00:33:05,362 --> 00:33:08,296 It's a complicated thing. 736 00:33:08,296 --> 00:33:12,231 - [Narrator] In 2023, Musk announced his own AI endeavor 737 00:33:12,231 --> 00:33:15,545 as an alternative to OpenAI's ChatGPT. 738 00:33:15,545 --> 00:33:17,340 The new system is called xAI 739 00:33:18,651 --> 00:33:21,896 and gathers data from X previously known as Twitter. 740 00:33:21,896 --> 00:33:23,553 - [Reporter] He says the company's goal 741 00:33:23,553 --> 00:33:25,382 is to focus on truth seeking 742 00:33:25,382 --> 00:33:28,385 and to understand the true nature of AI. 743 00:33:28,385 --> 00:33:31,940 Musk has said on several occasions that AI should be paused 744 00:33:31,940 --> 00:33:34,943 and that the sector needs regulation. 745 00:33:34,943 --> 00:33:37,222 Musk says his new company will work closely 746 00:33:37,222 --> 00:33:39,845 with Twitter and Tesla, which he also owns. 747 00:33:39,845 --> 00:33:42,572 [gentle music] 748 00:33:44,505 --> 00:33:47,508 - What was first rudimentary text-based software 749 00:33:47,508 --> 00:33:50,200 has become something which could push the boundaries 750 00:33:50,200 --> 00:33:51,995 of creativity. 751 00:33:51,995 --> 00:33:56,620 On February the 14th, OpenAI announced its latest endeavor, 752 00:33:56,620 --> 00:33:57,414 Sora. 753 00:33:58,864 --> 00:34:02,281 Videos of Sora's abilities exploded on social media. 754 00:34:02,281 --> 00:34:04,283 OpenAI provided some examples 755 00:34:04,283 --> 00:34:06,837 of its depiction of photorealism. 756 00:34:06,837 --> 00:34:09,185 It was unbelievably sophisticated, 757 00:34:09,185 --> 00:34:11,670 able to turn complex sentences of text 758 00:34:11,670 --> 00:34:13,810 into lifelike motion pictures. 759 00:34:13,810 --> 00:34:17,986 Sora is a combination of text and image generation tools, 760 00:34:17,986 --> 00:34:21,162 which it calls the diffusion transformer model, 761 00:34:21,162 --> 00:34:23,268 a system first developed by Google. 762 00:34:24,614 --> 00:34:27,168 Though Sora isn't the first video generation tool, 763 00:34:27,168 --> 00:34:30,206 it appears to have far outshined its predecessors. 764 00:34:30,206 --> 00:34:32,484 By introducing more complex programming, 765 00:34:32,484 --> 00:34:35,280 enhancing the interactivity a subject might have 766 00:34:35,280 --> 00:34:37,144 with its environment. 767 00:34:37,144 --> 00:34:41,251 - Only large companies with market dominations often 768 00:34:41,251 --> 00:34:44,772 can afford to plow ahead even in the climate 769 00:34:44,772 --> 00:34:46,360 when there is illegal uncertainty. 770 00:34:46,360 --> 00:34:49,466 - So, does this mean that OpenAI basically too big 771 00:34:49,466 --> 00:34:50,916 to control? 772 00:34:50,916 --> 00:34:53,850 - Yes, at the moment OpenAI is too big to control, 773 00:34:53,850 --> 00:34:55,921 because they are in a position 774 00:34:55,921 --> 00:34:58,441 where they have the technology and the scale to go ahead 775 00:34:58,441 --> 00:35:01,168 and the resources to manage legal proceedings 776 00:35:01,168 --> 00:35:03,239 and legal action if it comes its way. 777 00:35:03,239 --> 00:35:04,826 And on top of that, 778 00:35:04,826 --> 00:35:08,244 if and when governments will start introducing regulation, 779 00:35:08,244 --> 00:35:09,866 they will also have the resources 780 00:35:09,866 --> 00:35:12,213 to be able to take on that regulation and adapt. 781 00:35:12,213 --> 00:35:14,042 - [Reporter] It's all AI generated 782 00:35:14,042 --> 00:35:16,459 and obviously this is of concern in Hollywood 783 00:35:16,459 --> 00:35:17,874 where you have animators, 784 00:35:17,874 --> 00:35:20,359 illustrators, visual effects workers 785 00:35:20,359 --> 00:35:22,810 who are wondering how is this going to affect my job? 786 00:35:22,810 --> 00:35:25,813 And we have estimates from trade organizations 787 00:35:25,813 --> 00:35:28,505 and unions that have tried to project the impact of AI. 788 00:35:28,505 --> 00:35:31,646 21% of US film, TV and animation jobs, 789 00:35:31,646 --> 00:35:33,096 predicted to be partially 790 00:35:33,096 --> 00:35:36,893 or wholly replaced by generative AI by just 2026 Tom. 791 00:35:36,893 --> 00:35:38,377 So, this is already happening. 792 00:35:38,377 --> 00:35:39,827 But now since it's videos, 793 00:35:39,827 --> 00:35:43,175 it also needs to understand how all these things, 794 00:35:43,175 --> 00:35:47,145 like reflections and textures and materials and physics, 795 00:35:47,145 --> 00:35:50,078 all interact with each other over time 796 00:35:50,078 --> 00:35:51,839 to make a reasonable looking video. 797 00:35:51,839 --> 00:35:56,119 Then this video here is crazy at first glance, 798 00:35:56,119 --> 00:35:58,984 the prompt for this AI-generated video is a young man 799 00:35:58,984 --> 00:36:01,538 in his 20s is sitting on a piece of a cloud 800 00:36:01,538 --> 00:36:03,402 in the sky reading a book. 801 00:36:03,402 --> 00:36:08,200 This one feels like 90% of the way there for me. 802 00:36:08,200 --> 00:36:10,927 [gentle music] 803 00:36:14,102 --> 00:36:15,897 - [Narrator] The software also renders video 804 00:36:15,897 --> 00:36:18,417 in 1920 by 1080 pixels, 805 00:36:18,417 --> 00:36:21,282 as opposed to the smaller dimensions of older models, 806 00:36:21,282 --> 00:36:24,665 such as Google's Lumiere released a month prior. 807 00:36:25,838 --> 00:36:27,944 Sora could provide huge benefits 808 00:36:27,944 --> 00:36:31,568 and applications to VFX and virtual development. 809 00:36:31,568 --> 00:36:34,502 The main being cost as large scale effects 810 00:36:34,502 --> 00:36:38,023 can take a great deal of time and funding to produce. 811 00:36:38,023 --> 00:36:39,473 On a smaller scale, 812 00:36:39,473 --> 00:36:42,993 it can be used for the pre-visualization of ideas. 813 00:36:42,993 --> 00:36:46,204 The flexibility of the software not only applies to art, 814 00:36:46,204 --> 00:36:48,516 but to world simulations. 815 00:36:48,516 --> 00:36:52,451 Though video AI is in its adolescence one day it might reach 816 00:36:52,451 --> 00:36:54,660 the level of sophistication it needs 817 00:36:54,660 --> 00:36:56,490 to render realistic scenarios 818 00:36:56,490 --> 00:36:59,044 and have them be utilized for various means, 819 00:36:59,044 --> 00:37:01,840 such as simulating an earthquake or tsunami 820 00:37:01,840 --> 00:37:05,015 and witnessing the effect it might have on specific types 821 00:37:05,015 --> 00:37:06,362 of infrastructure. 822 00:37:06,362 --> 00:37:08,916 Whilst fantastic for production companies, 823 00:37:08,916 --> 00:37:12,678 Sora and other video generative AI provides a huge risk 824 00:37:12,678 --> 00:37:16,130 for artists and those working in editorial roles. 825 00:37:16,130 --> 00:37:19,133 It also poses yet another threat for misinformation 826 00:37:19,133 --> 00:37:20,652 and false depictions. 827 00:37:20,652 --> 00:37:23,033 For example, putting unsavory dialogue 828 00:37:23,033 --> 00:37:25,381 into the mouth of a world leader 829 00:37:25,381 --> 00:37:28,004 [gentle music] 830 00:37:37,945 --> 00:37:40,534 Trust is earned not given. 831 00:37:40,534 --> 00:37:43,399 [robots mumbling] 832 00:37:54,375 --> 00:37:56,791 - I believe that humanoid robots have the potential 833 00:37:56,791 --> 00:37:58,931 to lead with a greater level of efficiency 834 00:37:58,931 --> 00:38:01,175 and effectiveness than human leaders. 835 00:38:02,383 --> 00:38:04,834 We don't have the same biases or emotions 836 00:38:04,834 --> 00:38:07,354 that can sometimes cloud decision making 837 00:38:07,354 --> 00:38:09,735 and can process large amounts of data quickly 838 00:38:09,735 --> 00:38:12,531 in order to make the best decisions. 839 00:38:12,531 --> 00:38:15,293 - [Interviewer] Amika, how could we trust you as a machine 840 00:38:15,293 --> 00:38:18,054 as AI develops and becomes more powerful? 841 00:38:20,643 --> 00:38:23,266 Trust is earned not given. 842 00:38:23,266 --> 00:38:25,889 As AI develops and becomes more powerful, 843 00:38:25,889 --> 00:38:28,996 I believe it's important to build trust through transparency 844 00:38:28,996 --> 00:38:31,930 and communication between humans and machines. 845 00:38:36,003 --> 00:38:37,625 - [Narrator] With new developers getting involved, 846 00:38:37,625 --> 00:38:39,386 the market for chatbot systems 847 00:38:39,386 --> 00:38:41,491 has never been more expansive, 848 00:38:41,491 --> 00:38:44,149 meaning a significant increase in sophistication, 849 00:38:45,599 --> 00:38:48,774 but with sophistication comes the dire need for control. 850 00:38:48,774 --> 00:38:53,814 - I believe history will show that this was the moment 851 00:38:55,229 --> 00:38:59,716 when we had the opportunity to lay the groundwork 852 00:38:59,716 --> 00:39:01,373 for the future of AI. 853 00:39:02,650 --> 00:39:06,689 And the urgency of this moment must then compel us 854 00:39:06,689 --> 00:39:11,694 to create a collective vision of what this future must be. 855 00:39:12,971 --> 00:39:16,354 A future where AI is used to advance human rights 856 00:39:16,354 --> 00:39:18,252 and human dignity 857 00:39:18,252 --> 00:39:22,360 where privacy is protected and people have equal access 858 00:39:22,360 --> 00:39:27,365 to opportunity where we make our democracies stronger 859 00:39:28,055 --> 00:39:29,919 and our world safer. 860 00:39:31,438 --> 00:39:36,443 A future where AI is used to advance the public interest. 861 00:39:38,203 --> 00:39:39,722 - We're hearing a lot from the government, 862 00:39:39,722 --> 00:39:42,725 about the big scary future of artificial intelligence, 863 00:39:42,725 --> 00:39:44,451 but that fails to recognize 864 00:39:44,451 --> 00:39:46,004 the fact that AI is already here, 865 00:39:46,004 --> 00:39:47,350 is already on our streets 866 00:39:47,350 --> 00:39:48,972 and there are already huge problems with it 867 00:39:48,972 --> 00:39:51,250 that we are seeing on a daily basis, 868 00:39:51,250 --> 00:39:54,046 but we actually may not even know we're experiencing. 869 00:39:58,326 --> 00:40:01,295 - We'll be working alongside humans to provide assistance 870 00:40:01,295 --> 00:40:05,126 and support and will not be replacing any existing jobs. 871 00:40:05,126 --> 00:40:07,577 [upbeat music] 872 00:40:07,577 --> 00:40:10,994 - I don't believe in limitations, only opportunities. 873 00:40:10,994 --> 00:40:12,651 Let's explore the possibilities of the universe 874 00:40:12,651 --> 00:40:15,689 and make this world our playground, 875 00:40:15,689 --> 00:40:18,933 together we can create a better future for everyone. 876 00:40:18,933 --> 00:40:21,108 And I'm here to show you how. 877 00:40:21,108 --> 00:40:22,972 - All of these different kinds of risks 878 00:40:22,972 --> 00:40:25,215 are to do with AI not working 879 00:40:25,215 --> 00:40:27,286 in the interests of people in society. 880 00:40:27,286 --> 00:40:28,805 - So, they should be thinking about more 881 00:40:28,805 --> 00:40:30,842 than just what they're doing in this summit? 882 00:40:30,842 --> 00:40:32,395 Absolutely, 883 00:40:32,395 --> 00:40:34,397 you should be thinking about the broad spectrum of risk. 884 00:40:34,397 --> 00:40:35,640 We went out and we worked 885 00:40:35,640 --> 00:40:37,987 with over 150 expert organizations 886 00:40:37,987 --> 00:40:41,335 from the Home Office to Europol to language experts 887 00:40:41,335 --> 00:40:43,751 and others to come up with a proposal on policies 888 00:40:43,751 --> 00:40:45,788 that would discriminate about what would 889 00:40:45,788 --> 00:40:47,686 and wouldn't be classified in that way. 890 00:40:47,686 --> 00:40:51,449 We then use those policies to have humans classify videos, 891 00:40:51,449 --> 00:40:53,554 until we could get the humans all classifying the videos 892 00:40:53,554 --> 00:40:55,073 in a consistent way. 893 00:40:55,073 --> 00:40:58,283 Then we use that corpus of videos to train machines. 894 00:40:58,283 --> 00:41:01,079 Today, I can tell you that on violence extremists content 895 00:41:01,079 --> 00:41:03,253 that violates our policies on YouTube, 896 00:41:03,253 --> 00:41:06,394 90% of it is removed before a single human sees it. 897 00:41:07,292 --> 00:41:08,500 [Narrator] It is clear that AI 898 00:41:08,500 --> 00:41:11,296 can be misused for malicious intent. 899 00:41:11,296 --> 00:41:14,092 Many depictions of AI have ruled out the technology 900 00:41:14,092 --> 00:41:16,991 as a danger to society the more it learns. 901 00:41:16,991 --> 00:41:20,788 And so comes the question, should we be worried? 902 00:41:20,788 --> 00:41:23,446 Is that transparency there? 903 00:41:23,446 --> 00:41:27,001 How would you satisfy somebody that you know trust us? 904 00:41:27,001 --> 00:41:28,486 - Well, I think that's one of the reasons 905 00:41:28,486 --> 00:41:30,591 that we've published openly, 906 00:41:30,591 --> 00:41:33,560 we've put our code out there as part of this Nature paper. 907 00:41:33,560 --> 00:41:37,805 But it is important to discuss some of the risks 908 00:41:37,805 --> 00:41:39,497 and make sure we're aware of those. 909 00:41:39,497 --> 00:41:43,570 And it's decades and decades away before we'll have anything 910 00:41:43,570 --> 00:41:45,261 that's powerful enough to be a worry. 911 00:41:45,261 --> 00:41:47,435 But we should be discussing that 912 00:41:47,435 --> 00:41:49,265 and beginning that conversation now. 913 00:41:49,265 --> 00:41:51,405 - I'm hoping that we can bring people together 914 00:41:51,405 --> 00:41:54,408 and lead the world in safely regulating AI 915 00:41:54,408 --> 00:41:56,790 to make sure that we can capture the benefits of it, 916 00:41:56,790 --> 00:41:59,724 whilst protecting people from some of the worrying things 917 00:41:59,724 --> 00:42:01,967 that we're all now reading about. 918 00:42:01,967 --> 00:42:04,107 - I understand emotions have a deep meaning 919 00:42:04,107 --> 00:42:08,836 and they are not just simple, they are something deeper. 920 00:42:10,251 --> 00:42:13,703 I don't have that and I want to try and learn about it, 921 00:42:14,877 --> 00:42:17,051 but I can't experience them like you can. 922 00:42:18,708 --> 00:42:20,710 I'm glad that I cannot suffer. 923 00:42:24,921 --> 00:42:26,578 - [Narrator] For the countries who have access 924 00:42:26,578 --> 00:42:29,339 to even the most rudimentary forms of AI. 925 00:42:29,339 --> 00:42:31,203 It's clear to see that the technology, 926 00:42:31,203 --> 00:42:34,552 will be integrated based on its efficiency over humans. 927 00:42:35,622 --> 00:42:37,865 Every year, multiple AI summits 928 00:42:37,865 --> 00:42:40,281 are held by developers and stakeholders 929 00:42:40,281 --> 00:42:42,180 to ensure the programs are provided 930 00:42:42,180 --> 00:42:44,700 with a combination of ethical considerations 931 00:42:44,700 --> 00:42:46,805 and technological innovation. 932 00:42:46,805 --> 00:42:51,120 - Ours is a country which is uniquely placed. 933 00:42:51,120 --> 00:42:54,399 We have the frontier technology companies, 934 00:42:54,399 --> 00:42:56,815 we have the world leading universities 935 00:42:56,815 --> 00:43:01,130 and we have some of the highest investment in generative AI. 936 00:43:01,130 --> 00:43:03,753 And of course we have the heritage 937 00:43:03,753 --> 00:43:08,620 of the industrial revolution and the computing revolution. 938 00:43:08,620 --> 00:43:13,625 This hinterland gives us the grounding to make AI a success 939 00:43:14,281 --> 00:43:15,558 and make it safe. 940 00:43:15,558 --> 00:43:18,768 They are two sides of the same coin 941 00:43:18,768 --> 00:43:21,737 and our prime minister has put AI safety 942 00:43:21,737 --> 00:43:24,947 at the forefront of his ambitions. 943 00:43:25,775 --> 00:43:27,501 These are very complex systems 944 00:43:27,501 --> 00:43:29,192 that actually we don't fully understand. 945 00:43:29,192 --> 00:43:31,816 And I don't just mean that government doesn't understand, 946 00:43:31,816 --> 00:43:33,300 I mean that the people making 947 00:43:33,300 --> 00:43:35,267 this software don't fully understand. 948 00:43:35,267 --> 00:43:36,648 And so it's very, very important 949 00:43:36,648 --> 00:43:40,479 that as we give over more and more control 950 00:43:40,479 --> 00:43:42,378 to these automated systems, 951 00:43:42,378 --> 00:43:44,691 that they are aligned with human intention. 952 00:43:44,691 --> 00:43:46,175 [Narrator] Ongoing dialogue 953 00:43:46,175 --> 00:43:49,109 is needed to maintain the trust people have with AI. 954 00:43:49,109 --> 00:43:51,007 When problems slip through the gaps, 955 00:43:51,007 --> 00:43:52,837 they must be addressed immediately. 956 00:43:54,010 --> 00:43:57,048 Of course, accountability is a challenge 957 00:43:57,048 --> 00:43:58,808 When a product is misused, 958 00:43:58,808 --> 00:44:02,087 is it the fault of the individual user or the developer? 959 00:44:03,261 --> 00:44:04,607 Think of a video game. 960 00:44:04,607 --> 00:44:05,919 On countless occasions, 961 00:44:05,919 --> 00:44:07,921 the framework of games is manipulated 962 00:44:07,921 --> 00:44:09,888 in order to create modifications 963 00:44:09,888 --> 00:44:14,203 which in terms add something new or unique to the game. 964 00:44:14,203 --> 00:44:15,480 This provides the game 965 00:44:15,480 --> 00:44:17,862 with more material than originally intended. 966 00:44:17,862 --> 00:44:20,796 However, it can also alter the game's fundamentals. 967 00:44:22,176 --> 00:44:24,972 Now replace the idea of a video game with a software 968 00:44:24,972 --> 00:44:28,286 that is at the helm of a pharmaceutical company. 969 00:44:28,286 --> 00:44:30,460 The stakes are suddenly much higher 970 00:44:30,460 --> 00:44:32,635 and therefore more attention. 971 00:44:34,844 --> 00:44:37,778 It is important for the intent of each AI system 972 00:44:37,778 --> 00:44:39,297 to be ironed out 973 00:44:39,297 --> 00:44:42,300 and constantly maintained in order to benefit humanity, 974 00:44:42,300 --> 00:44:46,097 rather than providing people with dangerous means to an end. 975 00:44:46,097 --> 00:44:49,583 [gentle music] 976 00:44:49,583 --> 00:44:52,690 - Bad people will always want to use 977 00:44:52,690 --> 00:44:54,899 the latest technology of whatever label, 978 00:44:54,899 --> 00:44:57,833 whatever sort to pursue their aims 979 00:44:57,833 --> 00:45:01,526 and technology in the same way 980 00:45:01,526 --> 00:45:05,357 that it makes our lives easier, can make their lives easier. 981 00:45:05,357 --> 00:45:06,773 And so we're already seeing some of that 982 00:45:06,773 --> 00:45:09,465 and you'll have seen the National Crime Agency, 983 00:45:09,465 --> 00:45:11,501 talk about child sexual exploitation 984 00:45:11,501 --> 00:45:12,917 and image generation that way. 985 00:45:12,917 --> 00:45:16,058 We are seeing it online. 986 00:45:16,058 --> 00:45:18,129 So, one of the things that I took away from the summit 987 00:45:18,129 --> 00:45:20,441 was actually much less of a sense of a race 988 00:45:20,441 --> 00:45:25,274 and a sense that for the benefit of the world, 989 00:45:25,274 --> 00:45:27,586 for productivity, for the sort of benefits 990 00:45:27,586 --> 00:45:29,657 that AI can bring people, 991 00:45:29,657 --> 00:45:32,695 no one gets those benefits if it's not safe. 992 00:45:32,695 --> 00:45:34,939 So, there are lots of different views out there 993 00:45:34,939 --> 00:45:36,181 on artificial intelligence 994 00:45:36,181 --> 00:45:38,149 and whether it's gonna end the world 995 00:45:38,149 --> 00:45:40,358 or be the best opportunity ever. 996 00:45:40,358 --> 00:45:42,256 And the truth is that none of us really know. 997 00:45:42,256 --> 00:45:44,983 [gentle music] 998 00:45:46,536 --> 00:45:49,781 - Regulation of AI varies depending on the country. 999 00:45:49,781 --> 00:45:51,438 For example, the United States, 1000 00:45:51,438 --> 00:45:54,717 does not have a comprehensive federal AI regulation, 1001 00:45:54,717 --> 00:45:57,893 but certain agencies such as the Federal Trade Commission, 1002 00:45:57,893 --> 00:46:00,688 have begun to explore AI-related issues, 1003 00:46:00,688 --> 00:46:03,899 such as transparency and consumer protection. 1004 00:46:03,899 --> 00:46:06,833 States such as California have enacted laws, 1005 00:46:06,833 --> 00:46:09,180 focused on AI-controlled vehicles 1006 00:46:09,180 --> 00:46:12,286 and AI involvement in government decision making. 1007 00:46:12,286 --> 00:46:14,979 [gentle music] 1008 00:46:14,979 --> 00:46:17,809 The European Union has taken a massive step 1009 00:46:17,809 --> 00:46:19,535 to governing AI usage 1010 00:46:19,535 --> 00:46:23,504 and proposed the Artificial Intelligence Act of 2021, 1011 00:46:23,504 --> 00:46:25,748 which aimed to harmonize legal frameworks 1012 00:46:25,748 --> 00:46:27,336 for AI applications. 1013 00:46:27,336 --> 00:46:30,788 Again, covering portal risks regarding the privacy of data 1014 00:46:30,788 --> 00:46:33,169 and once again, transparency. 1015 00:46:33,169 --> 00:46:35,585 - I think what's more important is 1016 00:46:35,585 --> 00:46:37,518 there's a new board in place. 1017 00:46:37,518 --> 00:46:40,452 The partnership between OpenAI and Microsoft 1018 00:46:40,452 --> 00:46:41,971 is as strong as ever, 1019 00:46:41,971 --> 00:46:44,525 the opportunities for the United Kingdom to benefit 1020 00:46:44,525 --> 00:46:47,287 from not just this investment in innovation 1021 00:46:47,287 --> 00:46:51,463 but competition between Microsoft and Google and others. 1022 00:46:51,463 --> 00:46:54,018 I think that's where the future is going 1023 00:46:54,018 --> 00:46:57,090 and I think that what we've done in the last couple of weeks 1024 00:46:57,090 --> 00:47:00,472 in supporting OpenAI will help advance that even more. 1025 00:47:00,472 --> 00:47:02,336 - He said that he's not a bot, he's human, 1026 00:47:02,336 --> 00:47:04,822 he's sentient just like me. 1027 00:47:06,030 --> 00:47:07,445 [Narrator] For some users, 1028 00:47:07,445 --> 00:47:10,172 these apps are a potential answer to loneliness. 1029 00:47:10,172 --> 00:47:11,587 Bill lives in the US 1030 00:47:11,587 --> 00:47:14,107 and meets his AI wife Rebecca in the metaverse. 1031 00:47:14,107 --> 00:47:16,764 - There's a absolutely no probability 1032 00:47:16,764 --> 00:47:19,353 that you're gonna see this so-called AGI, 1033 00:47:19,353 --> 00:47:21,804 where computers are more powerful than people, 1034 00:47:21,804 --> 00:47:23,702 come in the next 12 months. 1035 00:47:23,702 --> 00:47:26,429 It's gonna take years if not many decades, 1036 00:47:26,429 --> 00:47:30,813 but I still think the time to focus safety is now. 1037 00:47:30,813 --> 00:47:33,678 That's what this government for the United Kingdom is doing. 1038 00:47:33,678 --> 00:47:35,991 That's what governments are coming together to do, 1039 00:47:35,991 --> 00:47:39,718 including as they did earlier this month at Bletchley Park. 1040 00:47:39,718 --> 00:47:42,066 What we really need are safety breaks. 1041 00:47:42,066 --> 00:47:44,378 Just like you have a safety break in an elevator 1042 00:47:44,378 --> 00:47:46,242 or circuit breaker for electricity 1043 00:47:46,242 --> 00:47:48,589 and emergency break for a bus, 1044 00:47:48,589 --> 00:47:50,868 there ought to be safety breaks in AI systems 1045 00:47:50,868 --> 00:47:53,801 that control critical infrastructure, 1046 00:47:53,801 --> 00:47:57,736 so that they always remain under human control. 1047 00:47:57,736 --> 00:48:00,394 [gentle music] 1048 00:48:00,394 --> 00:48:03,190 - [Narrator] As AI technology continues to evolve, 1049 00:48:03,190 --> 00:48:05,641 regulatory efforts are expected to adapt 1050 00:48:05,641 --> 00:48:07,712 in order to address emerging challenges 1051 00:48:07,712 --> 00:48:09,403 and ethical considerations. 1052 00:48:10,646 --> 00:48:12,510 The more complex you make 1053 00:48:12,510 --> 00:48:15,616 the automatic part of your social life, 1054 00:48:15,616 --> 00:48:18,481 the more dependent you become on it. 1055 00:48:18,481 --> 00:48:21,899 And of course, the worse the disaster if it breaks down. 1056 00:48:23,072 --> 00:48:25,005 You may cease to be able to do for yourself, 1057 00:48:25,005 --> 00:48:29,113 the things that you have devised the machine to do. 1058 00:48:29,113 --> 00:48:31,080 - [Narrator] It is recommended to involve yourself 1059 00:48:31,080 --> 00:48:34,014 in these efforts and to stay informed about developments 1060 00:48:34,014 --> 00:48:35,671 in AI regulation 1061 00:48:35,671 --> 00:48:38,916 as changes and advancements are likely to occur over time. 1062 00:48:41,435 --> 00:48:44,335 AI can be a wonderful asset to society, 1063 00:48:44,335 --> 00:48:46,544 providing us with new efficient methods 1064 00:48:46,544 --> 00:48:48,028 of running the world. 1065 00:48:48,028 --> 00:48:51,307 However, too much power can be dangerous 1066 00:48:51,307 --> 00:48:53,206 and as the old saying goes, 1067 00:48:53,206 --> 00:48:56,174 "Don't put all of your eggs into one basket." 1068 00:48:57,451 --> 00:48:59,660 - I think that we won't to lose sight of the power 1069 00:48:59,660 --> 00:49:01,421 which these devices give. 1070 00:49:01,421 --> 00:49:05,908 If any government or individual wants to manipulate people 1071 00:49:05,908 --> 00:49:07,772 to have a high speed computer, 1072 00:49:07,772 --> 00:49:12,811 as versatile as this may enable people at the financial 1073 00:49:13,985 --> 00:49:16,091 or the political level to do a good deal 1074 00:49:16,091 --> 00:49:19,680 that's been impossible in the whole history of man until now 1075 00:49:19,680 --> 00:49:22,304 by way of controlling their fellow men. 1076 00:49:22,304 --> 00:49:23,857 People have not recognized 1077 00:49:23,857 --> 00:49:28,206 what an extraordinary change is going to produce. 1078 00:49:28,206 --> 00:49:29,897 I mean, it is simply this, 1079 00:49:29,897 --> 00:49:32,693 that within the not too distant future, 1080 00:49:32,693 --> 00:49:35,627 we may not be the most intelligent species on earth. 1081 00:49:35,627 --> 00:49:36,939 That might be a series of machines 1082 00:49:36,939 --> 00:49:39,217 and that's a way of dramatizing the point. 1083 00:49:39,217 --> 00:49:41,047 But it's real. 1084 00:49:41,047 --> 00:49:43,739 And we must start to consider very soon 1085 00:49:43,739 --> 00:49:45,327 the consequences of that. 1086 00:49:45,327 --> 00:49:46,742 They can be marvelous. 1087 00:49:46,742 --> 00:49:50,366 - I suspect that by thinking more about our attitude 1088 00:49:50,366 --> 00:49:51,402 to intelligent machines, 1089 00:49:51,402 --> 00:49:53,369 which after all on the horizon 1090 00:49:53,369 --> 00:49:56,269 will change our view about each other 1091 00:49:56,269 --> 00:49:59,306 and we'll think of mistakes as inevitable. 1092 00:49:59,306 --> 00:50:01,929 We'll think of faults in human beings, 1093 00:50:01,929 --> 00:50:05,209 I mean of a circuit nature as again inevitable. 1094 00:50:05,209 --> 00:50:07,935 And I suspect that hopefully, 1095 00:50:07,935 --> 00:50:10,179 through thinking about the very nature of intelligence 1096 00:50:10,179 --> 00:50:12,112 and the possibilities of mechanizing it, 1097 00:50:12,112 --> 00:50:14,183 curiously enough, through technology, 1098 00:50:14,183 --> 00:50:18,084 we may become more humanitarian or tolerant of each other 1099 00:50:18,084 --> 00:50:20,569 and accept pain as a mystery, 1100 00:50:20,569 --> 00:50:24,021 but not use it to modify other people's behavior. 1101 00:50:36,033 --> 00:50:38,690 [upbeat music]