A Deep Dive into AI's Evolution

Don’t Believe AI Hype, This is Where it’s Actually Headed | Oxford’s Michael Wooldridge | AI History

Estimated read time: 1:20

    Learn to use AI like a Pro

    Get the latest AI workflows to boost your productivity and business performance, delivered weekly by expert consultants. Enjoy step-by-step guides, weekly Q&A sessions, and full access to our AI workflow archive.

    Canva Logo
    Claude AI Logo
    Google Gemini Logo
    HeyGen Logo
    Hugging Face Logo
    Microsoft Logo
    OpenAI Logo
    Zapier Logo
    Canva Logo
    Claude AI Logo
    Google Gemini Logo
    HeyGen Logo
    Hugging Face Logo
    Microsoft Logo
    OpenAI Logo
    Zapier Logo

    Summary

    In this fascinating interview, Oxford's Michael Wooldridge delves into the history and future of artificial intelligence (AI). The discussion, led by Johnathan Bi, spans the origins of AI, its various developmental phases, and the hype cycles that have often misled both public perception and the scientific community. Laying a foundation with historical insights from pioneers like Turing, Wooldridge argues that the lessons from AI's past are crucial for understanding its future potential and limitations. He emphasizes the need to separate science fiction ideals from practical AI applications, underscoring the importance of focusing on realistic risks and technologies that exist today.

      Highlights

      • Michael Wooldridge explores AI's 100-year history, highlighting key developments from Turing to present-day AI. 🕰️
      • AI has gone through several boom and bust cycles, each offering important lessons for the future. 📉
      • Large language models like GPT have impressive capabilities but are not the ultimate form of AI. 🤔
      • A key issue in AI today is discerning real from fake digital content, a challenge amplified by AI's capabilities. 🌐
      • Wooldridge argues that existential risk concerns often overshadow real, present challenges with AI. 💭

      Key Takeaways

      • AI's history is rich with lessons that can help us understand its future potential and limitations. 📚
      • Past AI hype has often led to unrealistic expectations and misunderstandings about the technology. 🌪️
      • Understanding AI's historical boom and bust cycles can help us navigate current developments more effectively. 🔄
      • The real challenges in AI lie in integrating it effectively in real-world applications while managing realistic risks. 🌍
      • Current AI capabilities, exemplified by large language models like GPT, are significant but still lack certain human-like intellectual abilities. 🤖

      Overview

      In this enlightening dialogue, Johnathan Bi engages with Michael Wooldridge to uncover the profound history of artificial intelligence. Wooldridge takes us on a journey through the inception of AI, marked by groundbreaking work from figures like Alan Turing, to modern developments in large language models. He emphasizes the cyclical nature of AI progress and the insights we gain from historical booms and busts.

        Wooldridge debunks common misconceptions fueled by sci-fi narratives and highlights the pressing need to focus on actively integrating AI into society. He suggests that many fear-based perspectives, often fictional, can cloud judgement, preventing recognition of the real advancements and challenges AI presents today.

          Despite the sensational capabilities of today's technology, Wooldridge notes crucial gaps in AI's journey towards truly mirroring human intellect. He points out that many of AI's successes imitate but do not replicate human neural processes, indicating that the path to genuine artificial general intelligence remains elusive.

            Chapters

            • 00:00 - 01:30: Introduction and Historical Context This chapter sets the stage by exploring the historical context of AI, emphasizing lessons from past projects such as the 'psych' initiative, which aimed to encode all human knowledge into a formal structure. The chapter questions the pursuit of moral AI, advocating instead for morally guided human decisions. It highlights the transformation of philosophical questions into empirical sciences and touches upon the ironic evolution of computers as unintended inventions within mathematics.
            • 03:00 - 07:00: Golden Age of AI (1956-1974) The chapter titled 'Golden Age of AI (1956-1974)' discusses the first significant boom in artificial intelligence during this period. Despite its name, AI was not held in high esteem initially, being likened to homeopathic medicine, and fields like neural networks were dismissed as dead. The narrative suggests that this era was characterized by substantial research efforts, as evidenced by the involvement of veteran AI researchers like Oxford's Michael Walridge, hinting at a pioneering spirit and burgeoning exploration in AI technologies during this time.
            • 40:00 - 43:00: AI Winter and Its Aftermath The chapter explores the historical trajectory of artificial intelligence (AI) development over the past hundred years, beginning with Turing and leading up to contemporary large language models (LLMs). It challenges the notion that studying the history of technology is irrelevant, offering two significant reasons to engage with AI's past. One reason is that understanding history helps to anticipate future developments in the field. The speaker named Waldridge takes the audience through this historical exploration, emphasizing key developments and shifts over time.
            • 43:00 - 48:00: Expert Systems and Knowledge-Based AI The chapter explores the skepticism surrounding the potential of current AI iterations to lead to a technological singularity. The speaker notes the cyclical nature of AI hype and how past breakthroughs have often led to apocalyptic predictions and unreasonable expectations. These cycles have historically set the field back. By studying the history of boom and bust in AI, one can develop a clearer understanding of the realistic trajectory of technology. The chapter also emphasizes the importance of studying this history to uncover overlooked techniques in the field.
            • 52:00 - 58:00: Behavioral AI and Robotics The chapter 'Behavioral AI and Robotics' discusses the potential for innovation by revisiting past AI paradigms that were previously considered too early for their time. It highlights WRI's claim that historical paths in AI weren't failures but were ahead of their time, suggesting they hold valuable insights and concepts that can inspire today's advancements. The speaker, John Jan B, introduces himself as a fellow at Cosmos, emphasizing the importance of this historical perspective on AI.
            • 58:00 - 65:00: Agent-Based AI The chapter "Agent-Based AI" discusses the intersection of philosophy and artificial intelligence, referencing a part of a book titled 'The Road to Conscious Machines.' The chapter touches upon the concept of the Singularity, which has become a popular and compelling narrative in discussions around AI. The text suggests that there is a significant narrative surrounding the advancement and implications of AI, potentially leading to conscious machines.
            • 88:00 - 99:00: Current State and Future of AI The chapter discusses the speculative future narrative where machines achieve human-level intelligence and have the capability to enhance their own intelligence, leading to a spiral of uncontrollable technological advancement. This concept is often depicted in science fiction movies like Terminator. However, the chapter expresses skepticism about this narrative, describing it as implausible and frustrating for various reasons.
            • 125:00 - 130:00: Conclusion and Philosophical Reflections The chapter titled 'Conclusion and Philosophical Reflections' examines the discourse surrounding the potential threats posed by artificial intelligence (AI). It highlights the tendency of the 'Terminator scenario' to dominate discussions about AI, overshadowing more immediate and practical concerns. This focus has given rise to a discipline known as existential risk, which is dedicated to exploring and mitigating the hypothetical end-of-world scenarios posed by misaligned superintelligent AI. The chapter suggests that while these risks are not unfounded, they should not prevent addressing current, actionable challenges related to AI.

            Don’t Believe AI Hype, This is Where it’s Actually Headed | Oxford’s Michael Wooldridge | AI History Transcription

            • 00:00 - 00:30 The Singularity is why is that the history of AI still has lessons to teach us there's another project called psych that attempted to store all of the knowledge that we have as a civilization into this kind of logical structure what I want is not moral AI I think it's moral human beings what were once philosophical questions have now become experimental science one of the great ironies of mathematical history that computers get invented as a byproduct
            • 00:30 - 01:00 the golden age as you described it 1956 1974 tell us about this first boom of ai ai actually didn't have a good reputation at all AI was viewed as kind of homeopathic medicine neural networks were regarded as a dead field oh wow well this is we're really at The Cutting Edge now we're I mean this is a big research question is about Oxford's Michael walridge is a veteran AI researcher and a pioneer of
            • 01:00 - 01:30 agent-based AI in this interview walridge is going to take us through the entire hundred-year history of AI development from Turing all the way to contemporary llms now you might wonder why should we bother studying the history of a technology right technology should be about the state-of-the-art techniques it should be about the latest developments what's the point of revisiting the old and outdated it turns out there's at least two important reasons to study this history and the first is that it'll help you better anticipate its future w is
            • 01:30 - 02:00 extremely skeptical that this iteration of AI will lead to the singularity partially because he's lived through so many similar cycles of AI hype whenever there's a breakthrough you always get apocalyptic predictions pushed by religious zealots which sets up unreasonable expectations that eventually sets the field back studying this history of boom and bust will help you see through the hype and grasp at where the technology is really going now the second and even more valuable reason to study this history is that it contains Overlook techniques
            • 02:00 - 02:30 that could Inspire Innovations today now even for me someone who started coding when I was 15 who studied CS in college and then went to build a tech company walridge uncovered entire paradigms of AI that I haven't even heard of wri's claim is that these paths not taken in AI right the alleged failures weren't wrong they were just too early and so this 100-year history is a treasure Trove of ideas that we should look back to today for inspiration my name is John Jan B I'm a fellow at the cosmos
            • 02:30 - 03:00 Institute where we study the intersection of philosophy and AI if you want to follow along this episode via transcript the link is in the description without further Ado Michael waldridge this is the provocative chapter title uh in part of your book The Road to conscious machines The Singularity is why is that there is this uh narrative out there and it's a very popular narrative and it's very compelling which is that at some
            • 03:00 - 03:30 point machines are going to become as intelligent as human beings and then they can apply their intelligence to making themselves even smarter the story is that it all spirals out of our control and of course this is the plot of quite a lot of science fiction movies notably Terminator I love those movies just as much as anybody does um but it it's it's deeply implausible and I became frustrated with that narrative for all sorts of reasons one of which is
            • 03:30 - 04:00 that that narrative whenever it comes up in sort of serious debate about where AI is going and what the risks are you know there are real risks associated with AI it tends to suck all the oxygen out of the room in in in the the phrase that my colleague used and it tends to dominate the conversation and distract us from things that we should really be talking about right in fact there's a discipline that's come out of this called existential risk right it's kind of the worrying about the ter Terminator situation and figuring out how can we perhaps better align these super
            • 04:00 - 04:30 intelligent agents to to human interests um and if you look at not just the narrative but actually the funding and what the smartest people are devoting their time into thinking in in not only companies but policy groups X risk existential risk is is the dominant share of the entire market so to speak why do you think this narrative has gained such a such a big uh big following I think it's the low probability but very very highrisk argument
            • 04:30 - 05:00 that I mean I think most people accept that this is not tremendously plausible but if it did happen it would be the worst thing ever and so very very very high risk and when you multiply that probability by the risk then it's the argument is that it's something that you should that you should start to think about but um this when the success of large language models became apparent and chat GPT was released and everybody got very excited about this last year the kind of debate around this sort of reached sort of slightly hysterical
            • 05:00 - 05:30 levels um and it became slightly unhinged at some point my sense is the debate is calmed down a little bit and is being more focused on on the actualities of where we are and what the risks are right I think that's quite a charitable reading I think of the psychology right it's a rational calculus there's small probability but there's a large sort of cost I study religious history and when I talk to people in the exis world the psychology kind of reminds me of uh the the Christian apocalyptic that that there's
            • 05:30 - 06:00 these people throughout Christian history that are like Now's the Time you know this happened most recently probably when we were uh going through the Millennium right 1999 and it's this psychological drive that wants to grab at something total and eschatological in a way to orient the entire world so so people I guess what I'm trying to highlighting is maybe you can see some of the psychology and climate risk as well it's not to say that these things aren't true right it's it's not to say that the world isn't
            • 06:00 - 06:30 ending in Christianity the climate isn't changing or there is no exis it's that the reason that people seem attracted to this narrative is almost a Rel religious phenomena I think that's right and I think it's appeals to something almost Primal and kind of human nature I mean it's most fundamental level it's the idea that you create something you have a child and they turn on you you know that kind of the Ultimate Nightmare for parents you know you give birth you nurture something you you create something exactly so uh or you know and this that narrative that story is very
            • 06:30 - 07:00 very resonant and for example you go back to the the original science fiction text Frankenstein that literally is the plot of Frankenstein you use science to create life to to give life to something to create something and then it turns on you and you've lost control of that thing so it's a very very resonant idea I think and so very easy for people to latch onto right you know it's easy for us to critique the psychology here but but what's what what do you think is wrong or what do you think people Mis
            • 07:00 - 07:30 about the argument itself that once we have super intelligent or or at least uh on par with human level uh machine intelligence that they can recursively improve upon themselves what what do you think people are missing when they give too much weight to that that argument the frustrating part is the Skynet part of the argument you know the kind of the Terminator thing that suddenly this will spiral out of control in ways that we just can't control in an incredibly short um an incredibly short period of time if you look under the hood of of how these things work and how many
            • 07:30 - 08:00 patches are required to hold AI together um it just it just doesn't seem terribly plausible more concretely there are basically two arguments for how existential risk might come around and the first is the famous paperclip argument which I'm sure you're familiar with you know so you build a highly intelligent machine and you ask it to build as many paper clips as possible and it follows your instructions in ways that you didn't anticipate right for example enslaving all of humanity to build of humanity and turning them to
            • 08:00 - 08:30 the production of paper clips uh you know until it turns everything into paper clips that's the uh that's the paperclip argument and there is some strength to that argument in the sense that AI can go wrong in those ways but for it to uh to to hurt us it has to be empowered to hurt us we have to give it the keys we have to give it control assum there's no guard rails and there's no guard rails and again that just doesn't seem terribly plausible that that we would do that I mean it would be a dumb thing for us to do to hand over
            • 08:30 - 09:00 the nukes to an AI so that's the first argument about how AI might become an existential threat the second argument is just that we build very very intelligent machines which develop their own goals which aren't aligned with ours now this is much more nebulous we don't know how that might happen um uh it's and and so it's slightly harder to address but we really aren't at the moment anywhere near that and I don't see even with very very powerful AI that we have have now the road map from where
            • 09:00 - 09:30 we go to that so some your understanding uh the first case is when the AI is executing our goals but not ingesting the kind of uh assumptions and implicit values the humans are imputing whereas the second one is a have developed their own goals um and there it seems even more far-fetched because llms as as powerful as they are they don't seem to be have any semblance of agency right and that's fundamentally what it would require so me and my friends have actually come up with this half joking
            • 09:30 - 10:00 term called existential risk risk which is the risk upon a society that focuses too much on existential risk and away from other other risks that we could actually be facing due to AI today if you can wave a magical wand and and swing The Narrative of AI away from X risk what are the actual problems uh and conversations that we should be having right now we are heading into a world where basically within a decade two decades think at the most um pretty much
            • 10:00 - 10:30 everything we read uh and see on social media and the internet is going to be AI generated and we're not going to know what's real and what isn't real in that world and there are many many risks associated with that that Society just fragments because there is no Common Core of beliefs anymore that we're all obsessed with some particular issue and that social media and the Internet is just driving us around that one particular issue because AI is programmed to pick up on the issues that you care about and to feed you stories
            • 10:30 - 11:00 emphasizing those risks and so on where I was particularly concerned is going into elections in the US the UK I was really worried that what we were going to be see was social media drowning in AI generated fake news we didn't see that as it happens at least not on the scale that I feared it might occur um but nevertheless I wouldn't take my eye off that as a risk I think that's a very very real risk um that uh that uh
            • 11:00 - 11:30 autocratic States uh control media just use AI to generate stories endless stories fake news stories that populist politicians do the same thing and so on and that we just drown in fake news till we no longer know how to tell what's real and what isn't and don't trust anything as a consequence given all these problems let me read you uh a quote from your book do we need laws maybe even International treaties to control the development of AI in the same way that we do for nuclear power I find the idea of introducing general
            • 11:30 - 12:00 laws to govern the use of AI rather imp plausible it seems a bit like trying to introduce legislation to govern the use of mathematics what do you think should be the role of government and policy if any in the mitigation of these risks so what I'm concerned about is some some sort of naive attempt to create a neural network law you know thou shal not use neural networks or something like that um and that's what seems to me to be implausible because neural networks
            • 12:00 - 12:30 under the hood are just a bit of mathematics actually not terribly complex mathematics there's a lot of it but it's not terribly complex and so regulating that well where do you draw the line I mean is is a bit of basic statistics you know the kind of thing that you would routinely do is that AI um you neural networks are quite a lot of linear algebra do we Outlaw linear algebra when we when we write programs so trying to regulate technology by pointing at neural networks and saying you know we should not use these is is is problematic in a way that I think
            • 12:30 - 13:00 pointing at nuclear weapons you know and nuclear fishing devices is not we can easily identify a nuclear fishing device I don't think there's much debate about that uh or the use of chemical weapons and so on outling the use of uh chlorine gas or whatever in in weapons you know these things are fairly easy and robust to uh robustly identifiable AI isn't and it's a really gray area about whether something actually is AI or isn't AI so
            • 13:00 - 13:30 my preference would be that we focused on the uses of technology and I think the one that I pick up on the book is surveillance Technologies if somebody's using surveillance technology on me I don't care whether it's a neural network or a logic program or guy sitting in a in a cabin watching exactly what I care about is somebody is using surveillance technology on me and that's where um uh the outlawing should happen and so what I would prefer when we look at regulation rather than aiming for some
            • 13:30 - 14:00 general neural network law is to look for specific sectors law Health uh defense and security uh all of those different areas Finance um uh all of those different areas education and so on and think about what are the issues that AI raises there about the use of the technology and legislate around those right and maybe we can tie this into your critique of X risk as well because I think a lot of the impulse or the intuition behind this these like
            • 14:00 - 14:30 neural net laws as you described is kind of preemptive like oh my God like if we don't set up the right controls like it's going to spiral out of control and then we're going to have Skynet whereas you're saying be because we should be more uh moderate in our expectations of what it can and cannot do including the harm it can do we can just regulate this like anything in any technology or any physical industry and infrastructure that we had in the specific use cases that's what I would that's what I prefer to see right so far we we've talked about the forward-looking view I I want
            • 14:30 - 15:00 to spend the rest of the interview really diving into your book and talk about the the historical view of of AI but let's begin with a question that I imagine many people in technical disciplines would would have which is why should we care about the history of AI because the history of AI still has lessons to teach us and one of the big lessons that it teaches us is that it's very easy to get OV excited and to read too much into what you're seeing in Ai and people have done that on multiple
            • 15:00 - 15:30 occasions in the past um now I think with the current wave of AI I think there is real substance here I think we are at a breakthrough moment uh but I'm not convinced that we're at the end of the road in AI or that the Transformer is the magic ingredient um a few years ago people were saying deep learning alone is the magic ingredient for AI now it's the Transformer architecture and so on I don't think either of those things are the magic ingredient I think there are some ingredients that we don't yet know about and I think your point about
            • 15:30 - 16:00 there's things for even serious AI researchers to learn from the history of AI can be made even stronger it's not just the negative of oh you know uh calm your calm your excitement but let me draw an analogy um in philosophy there's an idea that a lot of moral intuitions good moral intuitions are lost through Paradigm shifts so we gain things in this whole Christian worldview but we moved away from the Roman world and the Pagan world and there's things to be rescued from that world that have been
            • 16:00 - 16:30 forgotten and my my training early training was in stem stem usually doesn't study historical stuff right you usually just study the latest physical theories but there is a view of even stem Innovation Thomas being the biggest proponent as being these Paradigm shifts that there are things that are important that are lost in previous paradigms and I think we're going to see this today and so the positive pitch I would say to give even to Serious technical researchers to the history of AI is that there are methods
            • 16:30 - 17:00 and ways of thinking about programming artificial intelligence in general that have been overlooked in our current Paradigm that perhaps might be rescued and is Perhaps Perhaps what we need to get us to the to the next Frontier I think that's right I think we are to use Coon's phrase we are at a paradigm shift moment there's preg GPT and post GPT it's been boiling up for a decade or more really start I mean things became clear that things were happening in New networks around about 2005 with the
            • 17:00 - 17:30 Advent of deep learning so here are the key history Point 2005 Advent of deep learning 2012 people realized gpus could be used for training neural networks and all of a sudden you got 10 times more bang for your for your buck uh by using gpus in terms of training neural networks and so you can multiply what you do and stuff gets really overheated then then in 2017 there's a Transformer architecture 2020 there's gpt3 those are the kind of moments but we are in a
            • 17:30 - 18:00 paradigm shift right now and in Computing I genuinely believe that the world is Shifting now from a kind of an era where we were very interested in coding exact and optimal algorithms and thinking what is the right algorithm for solving this problem now it's give us the data we'll just throw it at machine learning and let machine learning sort it out do we care about exactly how it's doing it not necessarily it's just going to give us the answer but it's so many cases it turns out that the answer it's
            • 18:00 - 18:30 giving us is an incredibly useful one even though we sacrifice something and what we sacrifice is kind of guarantees of correctness and optimality but nevertheless it just turns into such a powerful tool so the the the the paradigm shift is towards a kind of datadriven world totally and and um so I did my uh computer science degree 2016 to 2020 and uh if you had asked me the history of AI I would have given exactly
            • 18:30 - 19:00 what you told me and no more but reading your book it extends greater another at least half century right and in fact there's a whole other dimension of AI about explicit programming what you call symbolic AI that's been completely overlooked that was the dominant Paradigm over ml but now when we think AI we think ML and so there are perhaps intuitions that we can rescue in the next hour for even the leading technical researchers but let's go to the very beginning with Alan Turing tell us about
            • 19:00 - 19:30 how he laned the groundwork not just for AI but for computer science as a field with his turing machine yeah so Alan churing for all that he's now one of the most famous mathematicians I think in history there's still a huge part of the churing story that people don't really understand and what everybody knows touring for is the codebreaking work he did at Bletchley Park which was an incredibly important part in the in the Allied victories and that's led to the Hollywood movie and so on I mean it's an entertaining movie but but hopelessly
            • 19:30 - 20:00 inaccurate um but anyway go back to the 1930s he's doing his PhD um in uh in Cambridge and there's one of the big mathematical problems of the age the en shidong problem it's called the translates as the decision problem and roughly speaking what the decision problem asks is can you automate mathematics can you reduce mathematics to just a procedure that you follow that is take away all human Insight uh just to a procedure and it was one of the
            • 20:00 - 20:30 defining problems of the early part of the 20th century and with incredible precociousness I think churing set himself the task of attacking the idun's problem and solved it very very quickly but to solve it he invented a kind of mathematical machine a machine that follows instructions and at the beginning it was just a mathematical abstraction but his work on codebreaking machines in the second world war leads him and a bunch of other people to
            • 20:30 - 21:00 realize that actually you could build these touring machines and a touring machine basically with a few practical tweaks is the modern digital computer that's all a computer that is a machine for following instructions so it's kind of one of the great ironies of mathematical history that computers get invented as a byproduct I mean he wasn't setting out to invent machines that could do things he was setting out to solve the idun's problem and he had to invent computers in order to do that but he did that after the war he goes and
            • 21:00 - 21:30 works on the first computers right and those computers those very early incredibly crude computers there's there's there's less than a handful in the whole world but they're capable of what seem like incredible intellectual Feats they can do huge quantities of mathematics very quickly and very accurately much more quickly and accurately than any human being could do and people start to think are these machines intelligent and that puts the idea of AI in the air and what I think
            • 21:30 - 22:00 is amazing about that period is we went from the beginning of the 1950s where there were probably two or three computers in the whole world ridiculously crude by today's standards by the end of that decade we'd gone to having machines that could do the rudiments of planning problem solving playing a decent game of chess or Checkers you know from having nothing whatsoever to machines that could do those things extraordinary progress in just a decade there are giants
            • 22:00 - 22:30 intellectual Giants where even their kind of uh secondary thoughts end up spawning entire disciplines like Newton calculus for example what I want to emphasize about the turing machine especially for our non-technical audience is that it's what you can tell it to do is very rudimentary it's literally unambiguous explicit instructions go like if a then B move pointer from E to Z right like write this value at this Place read this value
            • 22:30 - 23:00 from that place extremely rudimentary procedures and I want to emphasize again for our non-technical audience that is still the foundation of our current computers today it's not like we've designed a fundamental new paradigm but that's what's fascinating yeah is that we can get from those again deterministic simple explicit procedures to the kind of emergent behavior from of of chat gbt and why this is so exciting for me as a philosopher was I remember
            • 23:00 - 23:30 in my undergrad uh uh I was debating with a Conan scholar uh you know the continence you know famously or infamously you ask believe in Free Will and the conent scholar said well no deterministic system can come up with the kind of not even theoretical reasoning but the common sense reasoning you and I can do and you know I was I was in ML Class I remember very very fondly 4771 and I was learning how to use explicitly only deterministic systems to
            • 23:30 - 24:00 come up with the kind of common sensical natural language manipulation that humans do and and and this is just one example of how as you said philosophy is becoming experimental science yeah no and I completely agree with you I think it's one way of framing the AI question is can intelligence be reduced down to those incredibly simple instructions explicit instruction that computers and if you've never done
            • 24:00 - 24:30 any programming it's quite hard to imagine how dumb simple instruct this is why computer programmers are paid a lot of money right because you know you require a special mindset to be able to think down at that level and to understand how machines operate but yeah uh and and it is remarkable that what we see in um you know large language models state of the rart AI uh these kind of very dazzling capabilities ultimately are just reducing down to those very very simple
            • 24:30 - 25:00 instructions but my golly there's a lot of those simple instructions in order to do what they're doing I can't help but compare our current moment today with two watershed moments in maternity the cernic revolution humans don't think you're too special in a universe Darwin humans don't think you're so special in an animal human it feels as if our last Bastion our our last sacred ability a sacred thing that we have in
            • 25:00 - 25:30 our possession our intelligence has been reduced down to Binary bits yeah I don't think we're all the way there yet I think uh you know I think the fundamental nature of of human beings I mean one of the fundamental components of human beings is that we have experiences we experience the world um that's you know nobody really understands what Consciousness is but roughly speaking people agree agree that that that ability to experience things
            • 25:30 - 26:00 from a personal perspective and that your personal perspective is private and unique to you and I can imagine what you're experiencing but it really is private and unique to you uh and you know are we at the point of getting that from machines I think no definitely not and and how we might do that is is very opaque to me if it was an interesting thing to do at all right um and this is a perfect segue to talk about the other thing about ching he's famous for which is the Turing test touring test so so
            • 26:00 - 26:30 tell us about why he formulated this test and how it impacted the development of AI so in the 1950s then we have the first digital computers the first computers that operate according to the structures that we recognize today and they came out of touring Envision for the touring machines and then people realizing that actually we could build machines that look like this so we've got the early 1950s we've got the first digital computers um and this starts a debate a kind of public debate about AI even
            • 26:30 - 27:00 though it isn't given that that that that term and touring gets frustrated because people dogmatically insist that computers will never be able to do X where X is creativity or emotion or whatever uh and he and crucially machines will never be able to understand something in the same way that a human being is so he invents the touring test and the very famous touring test is beautiful in its Simplicity uh and it's a test for or indistinguishability so the touring test
            • 27:00 - 27:30 goes as follows roughly speaking you have uh a human judge who's interacting with something via as he described a teletype but you know imagine a computer screen where you're just typing uh you know typing whatever you want they could be questions but they could just be whatever you want and you're getting responses through that screen and you don't know whether the thing on the other end that's producing those responses is a human being or a computer program and maturing test says if you cannot reliably tell the difference that
            • 27:30 - 28:00 is if this machine can effectively pass itself off as a human being then stop arguing about it there's no point in arguing about it because you cannot distinguish between what the machine is doing or what a human does by any reasonable test so there's two ways to interpret the Turning test philosophically one way is to reduce metaphysics to phenomenology and this is to say look the metaphysical question of does it understand something is it really thinking is totally collapsible
            • 28:00 - 28:30 to the phenomenological the empirical question can we distinguish the outputs but or it could be making an epistemic point which is to say the metaphysical question doesn't really matter let's just focus on the empirical question which reading do you think Turing tur is given there yeah not clear I think which which reading I suspect the former I think he just thought was I suspect the former I think he probably just would thought we should there's no point in having the debate after this point I mean do do you agree with that
            • 28:30 - 29:00 premise I mean surely surely surely not right I mean what this reminds me of actually is the behavioralist the psychological school and the behavioralist even in the most charitable interpretation think that everything about a human uh uh can be known through their behaviors and interactions with the external environment that's the most terrible reading the least charitable reading is and and there's literally passages they almost literally say this the mind doesn't exist the Consciousness doesn't really I I don't even know like what
            • 29:00 - 29:30 that could even mean for it to be plausible but but but that is almost kind of the mistake that I see Turing making here so but I think what this illustrates to us is that the touring test beautiful as it is it's not a terribly interesting test for intelligence in human beings I me I think what churing said is it was kind of just for strated with the debate and said there's no point in having this argument I mean if it is just doing something uh that that is indistinguishable then why are we even
            • 29:30 - 30:00 debating after that point uh and in a purely practical sense whether it's really um really understanding in a way that human beings are is kind of irrelevant at that point there are distinctions between strong and weak Ai and the idea of strong AI is that we what what what we have what we're aiming for or what we have is machines that really understand and experience and so on in the same way that a human being or animal does um the weak version of AI is
            • 30:00 - 30:30 no they don't really understand but they can simulate those things I'm not terribly interested in strong AI except after a couple of glasses of wine in a in a in a in a chat with colleagues and I don't know very many AI researchers that really are interested in strong AI the goals of AI are much more pragmatic by the way I think for all practical intents and purposes we passed the touring test at some point in the last few years um uh but what that illustrates is I think is just actually
            • 30:30 - 31:00 The Limited value the touring test has as a real test for uh for intelligence right for me the only reason the consciousness of a uh a Computing machine uh has or does not have um the only real concern for me is is whether we have to treat them as moral agents right if you think that a a machine might be suffering it doesn't want you to turn it off we might have to give some weight to that but that seems to be like the only possible reason why
            • 31:00 - 31:30 someone would be interested in in strong versus weak right yeah I think people some people just think it's a it's an interesting thing yeah um yeah uh the idea of you AI as moral agents I'm worried about this um there is a uh there is a body of work which is all about trying to equip AI with kind of ethical and moral reasoning um and I understand why people want to do that so that we have machines that make choices that we would want them to make what worries me about that is that it allows
            • 31:30 - 32:00 people to try to abdicate their moral and ethical responsibilities wasn't my fault it was the machine's fault you know um but we can't hold a machine to account for its actions in the way that that we can hold a human being to account um and I say I'm really concerned particularly in like the military sphere that what we're going to hear is wasn't me it wasn't me you know it wasn't our fault it was the AI did it the AI chose the Target that school and fire the missile it was our fault at all and so I think what I want is not moral
            • 32:00 - 32:30 AI I think it's moral human beings and it's the people that build and deploy the AI where the responsibility and the ethical considerations have to sit and they are the ones that we need to hold to account for the actions of the machines that they deploy I see so that's Turing and that's the beginning of the field of not just artificial intelligence but computer science as a whole I want to move on to the actual history of the implementations and I'll begin with a quote from your book historically AI has adopted one of two main approaches to this problem put
            • 32:30 - 33:00 crudely the first possibility involves trying to model the mind the alternative is to model the brain to model the mind is what we've been talking about as symbolic AI to give it explicit instructions of what to do to to model the the processes that we rationally consciously go through in our heads to model the brain that's machine learning that's the neural Nets that's to model the architecture the physical architecture of the brain even if we don't have that much insight into what is actually going on let's talk about
            • 33:00 - 33:30 symbolic AI first let's talk about modeling the Mind first the golden age as you described it 1956 1974 tell us about this first boom of AI yeah by the end of the 1950s we've got machines that can show the rudiments of intelligence that can that can plan that can do mathematics which to be frank you know would be above the typical level of the people on the street you know here in Oxford you know you have chances of getting somebody who could uh who could tell you what goal back's conjecture was or something like that um would would be limited so you've
            • 33:30 - 34:00 got machines that can do mathematics that can that can solve problems play games and so there is this real excitement that you know actually we're going to be very quickly making progress towards something like full general intelligence and it's called the Golden Age because you know we went from having nothing to having machines that could do those things um and there was a period where you know where you know the the the the modus operandi for a for a PhD
            • 34:00 - 34:30 student was in in AI was well let's think of some task that requires intelligence in in humans and just build a machine to do those things and turned out to do crude versions of those things were turned out not to be that that hard but there was this massive optimism for that reason that progress was just going to be swift people thought within decades that they were going to be at the end of the road in AI we'd have full general intelligence AI Hae is not a new phenomenon aiap is very much not a new phenomenon and uh and uh by the early
            • 34:30 - 35:00 1970s it becomes clear really that progress is stalled and there are lots of reasons why progress stalled one of the reasons that progress stalled is people were looking at artificial versions of problems rather than real problems that is they were looking at some problem in the real world like a robotics problem and then coming up with a simplified simulation of that problem in a computer they were able to solve it in the simple simulated version but that
            • 35:00 - 35:30 simulated version didn't address any of the problems that were there in the real world problem um so classically in robotics people would do simulations of robots in warehouses and you'd look at a screen and you'd see a simulated robot carrying packages around and it looks very compelling you know great okay so show me the system in the real world but robots carrying a package round in a in a warehouse in the real world is nothing like the simulated version and so those simulated versions they were called
            • 35:30 - 36:00 microw worlds uh and again a standard motus operandi for a PhD student is come up with a microw world for your particular problem uh whatever it was show that you could build a program that could solve it in that microw world but actually you've you've abstracted away everything that's difficult about the real problem in the real world and you know and and research funders would say fine show us this then in a real uh Warehouse they wouldn't be able you described the general philosophy and
            • 36:00 - 36:30 strategy in this period of the golden age as divide and conquer so this is splitting out what we conceive our mental faculties to be and then trying to build I mean mostly search algorithms in different variations to satisfy that so one example would be the the towers of uh towers of B exactly where you have to move the the the discs on top of one another to get it to the right shape and those are the traveling St salesman right you give given given a list of
            • 36:30 - 37:00 cities and try to find the the optimal algorithm there and those are the type of problems that people seem to be working on there was a common ceiling that people were hitting across these problems um and that had to do with NP completeness so can you give our our lay non-technical audience a rough idea of what this means so a lot of approaches to AI in uh in the early days involved something called search and search just means if you're given a particular problem just look through all possible candidate Solutions so uh we mentioned
            • 37:00 - 37:30 the idea of the traveling salesman problem the the traveling salesman problem you're given a particular map and that the salesman so to speak has to visit a whole bunch of cities on this map and return to return to base can the salesman do that on a certain budget of fuel um that's the traveling salesman problem and so one way to approach that is just to look through all the possible candidate Solutions the problem is that the number of candidate Solutions in that case just grows astronomically um
            • 37:30 - 38:00 so for example if there are something like 70 cities there would be more possible candidate Solutions than there are atoms in the universe you will never have a computer that could exhaustively look through all of those candidate Solutions and it was assumed in the early days that we would be able to fix that problem uh it's called combinatorial explosion uh we would be able to fix that problem somehow that we would find some techniques to be a ble to do it by the early' 70s there was an
            • 38:00 - 38:30 emerging theory of what's called computational complexity which is which is all about understanding the intrinsic complexity of certain tasks and for the traveling salesman problem it belongs to a class of computational problem that's called NP complete now what that means roughly speaking is that we don't have any efficient way to do it there is no more efficient way than looking through all of the candidate Solutions in order to find one um which means that in practice uh
            • 38:30 - 39:00 there's a huge barrier with that kind of problem if we want to try and solve them um but then it be began to appear that actually a whole bunch of problems everywhere we looked in AI we found okay this problem's actually NP complete problems in computer vision are NP complete endless problems in reasoning and problem solving are NP complete or even worse there's this big hierarchy of complexity uh complexity classes as they're called where things can be even
            • 39:00 - 39:30 harder than NP complete problems and then problems that are even harder than that and and so on and we found everywhere we looked we were encountering these problems with AI and we hit a wall and the wall was this barrier of uh of combinatorial explosion all these problems have the same character in principle you can solve them just by looking through all of the candidate solutions to try to find the right one one that works but in practice it's impossible to do that right by the
            • 39:30 - 40:00 mid '70s because of the hype as well as the the series of technical problems that the AI field ran into it went into its first but not certain certainly not only winter and what that means is just funding dried up interest dried up people were sometimes portrayed as charlatans in the AI field and the public just grew very suspicious of AI claims you yourself have worked through many of these Winters boom and bus Cycles is it almost better to work in in a winter because you get the people who
            • 40:00 - 40:30 are actually serious about about AI so uh for most of the time that I've been studying AI it was a relatively quiet existence and the nice thing about that was I just got on with my thing there was very few people working in the same area and as a researcher actually that's quite a nice thing uh as a researcher having you know a big space to yourself is actually really quite sort of refreshing you can just explore the territory so when the field becomes became popular we found huge numbers of
            • 40:30 - 41:00 people flooding into it now the nice thing about that is huge numbers of very talented people but as a researcher what you're finding is you're no longer the only person that's looking at your problem you're surrounded by extremely capable people all working on exactly the same problem and so you know it's changed the character of doing AI research really really quite a lot you're not the only person at the cace you know there's a whole bunch of people there chipping away at the same place along with you but you have to remember actually just go back two decades and AI
            • 41:00 - 41:30 actually didn't have a good reputation at all I mean in science AI was viewed as kind of homeopathic medicine neural networks were regarded as a dead field a dead end and I can remember colleagues saying you know why are you working in AI you know this is uh this this is not a field that's going to be good for your career it's just extraordinary how much that's changed so in the 80s we came out of the first AI winter and this new wave
            • 41:30 - 42:00 this new paradigm of AI was called expert systems tell us about how the philosophy in this second wave of AI was different from the Golden Age so the big idea in the second wave AI is that intelligence is primarily a problem of knowledge and so if you want to build a machine that can do something for you translate from French to English or to play chess or whatever then the key problem is to figure out what knowledge the human beings use when they do that task and give that knowledge to a machine uh that was the big idea
            • 42:00 - 42:30 knowledge knowledge is the key to intelligence um and the AI is primarily a problem of giving machines the right knowledge and a technology emerged rule-based systems as they were called which made it possible to give knowledge about particular problems to machines classic example was the M system which was an expert in diagnosing blood diseases in human beings and the the way M was developed um was that uh uh the
            • 42:30 - 43:00 developers talked to human experts Physicians experts in blood diseases and they ask them how do you go about diagnosing causes of blood diseases and they would say well the first thing I do is take somebody's temperature and then if the temperature is above this range I would do this experiment and so on and that knowledge about how humans solve that problem is coded in the form of discrete chunks of what are called rules if a human has a temperature greater than this and uh this particular blood test comes up negative and so on then
            • 43:00 - 43:30 they have Lassa fever with probability 0.7 that would be an example of a rule all of those rules were coded given to the machine and then you interact with the machine the machine asks questions like does the patient have a temperature have you done this test what's the outcome of this test and so on and in the end it tells you I think your patient has Lasser fever or something like that I see um but not only were new systems developed as a a CS student myself I was uh very surprised to to
            • 43:30 - 44:00 hear about a new paradigm of of uh programming that I that I haven't heard of before um so when you're go to school today you're taught two types declarative that's when you specify to the machine what you want right SQL a lot of database languages give me all the apples that are green and in that are from 2024 uh and then there's imperative languages right and this is Java this is C++ this is you telling the machine if this then do that uh uh this is what video games for example are made of there's another Paradigm that I just
            • 44:00 - 44:30 learned about in your book called logical the logic Paradigm so let me let me give you a quote the war plan planning system written by David Warren in 1974 which could solve planning problems including the blocks world so this is a simulated uh Factory uh search problem we described in the Golden Age including the bloxs world and far beyond that required just a 100 lines of prologue code prologue is The Logical
            • 44:30 - 45:00 prog programming language writing the same system in a language like python imperative would be likely to require thousands of lines of code and months of effort the Temptation with logic programming is if I just give you the fundamental truths that I know logical deduction can elegantly take it all the way yeah and that's that's a beautiful idea which beguiled an enormous number of AI researchers so the idea of logic
            • 45:00 - 45:30 programming takes symbolic Ai and knowledge-based AI one step further and it says that okay if we want to build machines that have knowledge the way that we give them that knowledge is by expressing that in the form of logic we give them these these logical these logical descriptions of the world and this is Aristotle essentially right if if Socrates is a man all men are mortal Socrates is Mortal this is Aristotelian logic 101 exactly so we give it but we give it all of all of the knowledge about a particular problem whether it's
            • 45:30 - 46:00 diagnosing blood diseases or solving planning problems and so on we express that in a logical form and then inbuilt logical reasoners will sort out the details for us they will they will do the logical reasoning and so the idea there was in logic based AI that intelligence is primarily a problem of deduction of logical reasoning and it's a beautiful and elegant idea and the war plan program you say with 15 lines of code it's just it's a ridiculously short program problem is it's just not very
            • 46:00 - 46:30 efficient firstly actually and it turned out to be in many cases hopelessly inefficient for lots of problems uh but also it just turned out that again you know if you want to do robotics prologue is not the language you need for doing robotics it's just completely unsuitable for that but it would be impossible to imagine expressing all the knowledge that chat GPT has been exposed to to man manually Express all that in the form of logical
            • 46:30 - 47:00 expressions and give that to it just wouldn't work that didn't stop one particular project you described M which was very limited to specifically to doctors and hospitals and the idea was again let's just dump all the the the the first principles the primary facts we know about the world logical deduction is going to figure out if Socrates is a man then Socrates is Mortal there's another project called psych cyc that attempted to store all of the knowledge that we have as a civilization
            • 47:00 - 47:30 into this kind of logical structure yeah so the psych project has a somewhat mixed place in uh in the history of of AI so the vision this was the vision of Doug Leonard Leonard was a really brilliant researcher who really dazzled people in the early 70s with uh with his work um and uh he became convinced that that the really big problem of AI the problem of building machines which are
            • 47:30 - 48:00 as fully capable of human beings is simply a problem of knowledge and he said there's no shortcut to this we're just going to have to give the machine all this knowledge so uh he convinced some funders to support his work and at one point they had kind of warehouses full of people busy encoding all of human knowledge in these forms of rules if this and this and this then this uh uh with the idea that of eventually this would be as capable as as a human being
            • 48:00 - 48:30 and and lennet was uh was very very optimistic about his project he said you know within a couple of years I remember reading this in the uh beginning of the 90s he said within a couple of years Psych is going to be smart enough that we'll it'll just be able to write its own rules and we won't need to we'll just give it textbooks kind of like the reflexive yeah exactly that um now the the happy part of the psych story is that the knowledge graphs that are used by search engines and uh behind a lot
            • 48:30 - 49:00 behind the scenes in a lot of search now they trace their intellectual history to these very very what I call very large knowledge bases but the the psych was just ridiculed at the time as being just ludicrously over ambitious it never delivered anything at the scale that was anticipated for it it found some applications but relatively Niche applications and it never delivered anything at the scale that Lenard hoped for it and so it was kind of often it
            • 49:00 - 49:30 was the ridicule it was ridicule but it was held up as that that one project which summarized everything that went wrong about symbolic AI let me give you what I think to be the funniest quote from your book psych's main role in AI history is an extreme example of AI hype which very publicly failed to live up to the Grand predictions that were made for it the founder of Psych Doug lenn's role in AI has been mythologized in a piece of computing folklore a mik micr lenit so the joke goes is the scientific unit for measuring how bogus something is why
            • 49:30 - 50:00 a microl lenit because nothing could be as bogus as a whole L it yeah and I think this is what you were trying to get at about why it's important to study the history of AI to get the proper perspective with every new paradigm the Golden Age s we're so close to using search to solve everything with the expert systems oh like Psych is going to be able to to basically reflect L recursively write its own rules but even when you when you go back like every
            • 50:00 - 50:30 technology whether it's the printing press people immediately wanted to write the encyclopedias right that captured all the knowledge in the world this is the drive that you see in today's AI today's AI yeah um so I think uh going going back to leonnard and psych I mean that's kind of that that quote's kind of slightly cruel I think I mean the joke is a slightly cruel quote but the truth is I mean I think it would have a much happier place in AI history if it hadn't seen so many just inflated claims that were that were implausible at the time and that just weren't delivered uh if it
            • 50:30 - 51:00 had had slightly more measured objectives it would have held up much better as an exercise in large scale knowledge based development and I say you know there are in the DNA of the knowledge graph behind the scenes of of Google search and so on there was a little bit of of of Psych there it was all for not it was it for nothing but it was just these overinflated claims like you know it's going to start reading and writing its own rules and so on but there is a striking I think analogy that you pick up on which is the way that
            • 51:00 - 51:30 large language models are trained which is that we just expose them to every bit of Digital Data that we can get our hands on the difference is that in the site case human beings were interpreting all of that and writing coding down The rules in the computer language with large language models none of that goes on it is just presented to the model and in some sense and I'm waving my hand madly at this point in some sense it it finds order in that and how it does that
            • 51:30 - 52:00 actually we don't really understand as we were talking about earlier we talked about the Golden Age with search we talked about expert systems in the late 80s Rodney Brooks in as a reaction almost to to the overe exaggeration of of the of these expert systems started a new paradigm called behavioral AI uh and some of that philosophy is behind iroot the company that he founded tell us about what the philosophy in this Paradigm was so Brooks questioned the
            • 52:00 - 52:30 fundamental principles on which AI had been working since the 1950s for 30 years and those principles were that uh intelligence uh can be solved through a process of symbolic reasoning that we give the machine the knowledge it needs to solve a problem and that those are the key components of intelligence Brooks said actually I I just don't think that's how intelligence Works in human beings and he came up with an alternative Theory this kind of
            • 52:30 - 53:00 Behavioral Theory and roughly speaking what he said is we are a mass of conflicting behaviors that some of which are genetically hardwired into us through through evolutionary processes some of which we learn throughout our lives but we're just a mass of these behaviors and somehow uh human uh human intelligence arises from the interaction of those behaviors so he said let's start out by building layer by layer those behaviors and he was also um
            • 53:00 - 53:30 extremely unhappy with the idea of Intelligence being manifested in disembodied systems he said that's not real intelligence human intelligence is something in the world we do things in the world and he was deeply critical of any version of AI that wasn't capable of dealing with the world I mean he was a roboticist he wanted to build robotics that could do things so he built an architecture a framework for doing this where you would start with the most
            • 53:30 - 54:00 fundamental behaviors the most basic behaviors imaginable and in robotics famously the most fundamental behavior that you learn on day one of any robotics course is obstacle avoidance um because your robot crashing into things is expensive and you'll be on the end of lawsuits and all sorts of things like that so you you start out by building your very first layer is obstacle avoidance and then imagine a robot that's going to go around this room picking up trash the next level of in Behavior might be exploring right just
            • 54:00 - 54:30 exploring around the room to try to find the trash and the next level of behavior might be if you see trash you pick it up you gradually build up layer and layer and layer you then have to think about how those behaviors interact with one another and what takes what takes precedence so obstacle avoidance for example tends to take precedence over everything you know if you're if it's a question of Destruction versus survival you know you always you know you want to choose surv Ral and the really cool thing is he was able to build robots
            • 54:30 - 55:00 that could do some quite impressive tasks in the real world as a way towards intelligence uh it hit problems that were not dissimilar in spirit to the kind of problems that people had encountered when in symbolic Ai and combinatorial explosion just we after a relatively small number of behaviors it starts to be very hard to organize those behaviors and think about the way that those behaviors are going to interact with one another and so it kind
            • 55:00 - 55:30 of reached a limit at some point by by the mid 90s I think but what Brooks did is he was able to build successful robotic systems famously the Rumba robots uh are built using a version of right his ideas and I think uh that is the best example of all the ideas that he's talking about right so so the the the vacuum robots essentially uh that go around in your house they're embodied they're Rob robots they're not just a software but importantly if you look at the programming behind the robots it's
            • 55:30 - 56:00 not like a top- down search and go through the entire space it's kind of like go straight if there's an obstacle to take a random number turn this amount of degrees and map out the space it's it's very reactive right and so philosophically I think the helpful contrast between the first two paradigms the Golden Age and the expert systems is that those are kind of like top down systematic right I'm gonna have one search and I'm gonna search through all the combinatorial possibilities or or I'm going en code all of human knowledge
            • 56:00 - 56:30 and this one thing and behavioral is the opposite yeah it's highly reactive it's to say if you meet this situation then go do this almost like a look lookup table and and at the time people picked up on analogies with behavioral theories of psychology and Skinner um you know who did used to do all the experiments and training dogs and whatever by giving them stimuluses and rewards and and punishments and so on and uh and behavioral Psychology was kind of somewhat discredited largely discredited
            • 56:30 - 57:00 I think at the time as a as a general theory um uh of human behavior and it people picked up on exactly those critiques and pointed exactly those critiques at at Brooks's behavioral AI right but that's not just I think what was interesting for me was it was one of the relatively few attempts to really go back to the basics of what AI is and say how what is our fundamental guiding princip we do this and really there been relatively few of those there's symbolic
            • 57:00 - 57:30 AI there's behavioral AI there's the new AI of machine learning and deep learning and so on which is kind of datadriven AI right the last Paradigm in this symbolic world that I want to talk about is the early '90s agent-based Ai and my understanding here is that it's an attempted synthesis of the intuition of uh the expert systems and the golden age that we want our machines to be proactive that there's a goal directed but in combination with the reaction uh of of the behavioral AI in addition to a
            • 57:30 - 58:00 third idea of of Aging being social in nature so tell us a bit more about this agent based P so we're on home territory for me this is what I've worked on basically my whole career so um one way to think about this is changing our relationship to computer software on Microsoft Word everything that happens because you make it happen you select something from a menu or click on an icon but there's only one agent in that interaction and it's you and you are
            • 58:00 - 58:30 just telling the machine very much like you know you're giving detailed low-level instructions to Microsoft Word somewhat like programming it right in the same kind of style and the idea that emerged uh in the end of the 1980s beginning of the 1990s and which I worked on was to change the relationship of software so that the software becomes an agent that's acting on your behalf that's cooperating with you working with you on the task that you set it um so
            • 58:30 - 59:00 it's not just the dumb recipient of instructions but it's actually now an active participant working with you uh and potentially with other agents so to put it another way um the the manifestation of the agent dream that we see most obviously now is in Siri and Alexa and Cortana they literally they they are they are direct descendants of that idea and actually uh Siri emerged from work on Agents from people that I knew working in the same community in in the 1990s um so if um if we have Siri the
            • 59:00 - 59:30 idea of Siri is that it is uh actively working with us on a problem rather than just being told what to do uh but that might involve interacting with other Series so if I want to arrange a meeting with you why would I call you why would my Siri call you why doesn't my Siri just talk directly to your Siri that is the idea of what's called multi-agent systems and that's that's kind of what's Driven most of my work for the last 35
            • 59:30 - 60:00 years I see well what I found fascinating about the agent Paradigm is that it almost it's agnostic to and it cuts across the symbolic modeling the the mind and modeling the brain because how it is proactive or how it is reactive or how it interacts with other agents that's you abstracted away from the type of questions you're thinking about yeah so what does multi-agent systems have to offer to the current Paradigm of foundational models now oh wow well this is we're really at The Cutting Edge now we're I mean this is a
            • 60:00 - 60:30 big research question is about we have large language models and they are not sort of full general intelligence but they are nevertheless very capable how do we actually deploy those in our agents do they just handle the natural language part the conversational part or could we actually leverage them to do problem solving or things like that now we've already talked about the idea you know can large language models solve problems and I'm a bit of a skeptic at the moment about the extent to which they can do that but how exactly do we
            • 60:30 - 61:00 leverage this technology in the best way possible uh is is right at The Cutting Edge of research right now that's exactly what people are thinking about so what Drew you to this agent Paradigm and especially this multi-agent Paradigm oh well so as an undergraduate in the 1980s uh I was fascinated with AI uh but I also became fascinated with computer networks and you have to remember at the time computer networks were not common you know the the the the predecessor of
            • 61:00 - 61:30 the internet the arpanet um developed by the advanced research projects agency in the US essentially military research funding agency had you know a very incomplete International network with just a few nodes connected in the UK but I got the opportunity to work on the UK's extension of that called Janet The Joint academic Network and I had a kind of moment of Revelation at which point I realized this is going to be the future networks they're just going to be
            • 61:30 - 62:00 everywhere we're going to everybody is going to be using computer networks it was obvious to me that you were going to hook up to the network through your phone or something like it and this was going to be everywhere and so I had those two ideas in my head I had I'm really interested in AI I know that networks are going to be the future this is uh this is obvious it's going to be the future put those together and think about a network of AIS AIS talking to one another and that's how I got interested literally that's how I got interested in in the idea just having
            • 62:00 - 62:30 what happens if we got two AI systems that are capable of communicating well how are they going to communicate what's the language what are the rules of the protocols that they're going to use and that's what kicked off my interest by the way having realized that networks were the future I completely failed to anticipate the worldwide web or Amazon or any of that I look I look at the missed opportunities in my life um for for doing transformational work I totally got that networks were going to be the future but I still didn't understand exactly what that future was
            • 62:30 - 63:00 going to look like it sounds like you were expecting us to go directly to multi-agent like I have my AI bargaining on behalf of me exactly whereas first we went through this multi almost symbolic phase where you know what is Amazon if what is Google if not a big advanced search algorithm but now do do you think we're heading into a multi-agent world in the sense that I'm going to have my own AI agent to to act on behal of me that's make a lot of this uh early internet stuff obsolete I think it is
            • 63:00 - 63:30 inevitable one way or another I don't I think absolutely the history of computing tells us that this surely must all the lessons that we we learn from the history of computing point to the future of AI being not just one big isolated system but multiple AI systems interacting with one another because that's how Computing the history of computing has gone so I absolutely believe that the problem is I don't know exactly what that's going to look like and that's what I'm trying to figure out now that's what my current research is trying to figure out um there's another
            • 63:30 - 64:00 way in which multi-agent systems I imagine are are currently being deployed even llms not multiple llms talking to each other but how you split work within one llm right so the the rough intuition is you know maybe uh it's better to train actually three hidden LMS one's good with math one's good with intuition one's good with creativity or language and then you have a a a a job sort of processing unit that gives uh the different llms different tasks to to
            • 64:00 - 64:30 process that that also is a type of multi-agent work yeah absolutely and those are exactly that those kind of architectural questions how do llms fit what does that architecture look like that's again we're right at The Cutting Edge that's some people looking at those questions right now and trying to figure out what the right way to organize all that stuff is right and so what are some of the biggest questions uh in the field right now what enormous numbers of people in the AI Community are grappling with is is trying to get to grips with the capabilities of large language
            • 64:30 - 65:00 models to really map out what these models can reliably do and what they can't reliably do and exactly what capabilities they really do have versus those that they don't actually have and it turns out this is really quite difficult one of the reasons it's quite difficult is that um they've because they've essentially been exposed to all the digital content in the world it's quite tough to come up with things that you're confident they've really fundamentally never seen before um but
            • 65:00 - 65:30 this is uh a really exciting area of science it's one of the one of the key areas I think one of the most important areas of science right now and it goes back to this thing that you know AI has just become this experimental science in the way that it wasn't previously and this is part of that picture trying to map out these capabilities and it's also really frustrating because these models frankly behave in slightly weird ways you think you've got some principle or some rule one day and then you just change your prompt slightly in ways that
            • 65:30 - 66:00 seem innocuous to you and you get a completely different answer the next day and it's uh okay so what went on there what how why did it change but mapping that out uh is is is genuinely very fascinating at the moment right um so I want to move on to the last part of our conversation which is I focused most of our time talking about the history on the symbolic AI side because that I I feel like it's almost a forgotten history at this point because when we think AI we think ML and not the symbolic explicit programming side um I
            • 66:00 - 66:30 just want to trace out and round out this history for for our viewers because what was fascinating to me was that AI people didn't use to associate AI with ml in fact machine learning it seemed from your book grew as a separate field starting in the 40s right this idea of can we recreate the brain structure with uh electric neurons with computation and then the big Milestones Connection connectionism in the 1980s this is when
            • 66:30 - 67:00 we figured out back propagation basically a way to add more layers to to actually simulate to train these networks deep learning even more layers and more scale in 2000s and eventually Transformers Foundation models in the 2020s what I find so poetic about this this entire history now that Mo comes full circle is what didn't work was rationally trying to explicitly tell computers what to do what did work or what is working
            • 67:00 - 67:30 now let me say is by imitating the biological structures of the human the human brain yeah it is a remarkable uh it's a remarkable uh change in fortunes for neural networks which I say 20 25 years ago was really regarded as kind of homeopathic medicine was in some sense not s taken very very seriously part partly because of the scale that would be required to build large neural
            • 67:30 - 68:00 networks and it didn't seem plausible 25 years ago that we would have computers that could process neural networks with 200 billion parameters or 500 billion parameters and yet that's that became possible because of the computer power that we have available now and it turns out that these these systems can be incredibly capable so it really is a remarkable remarkable story I think one point in the book I say you know if you if you think that science is about orderly progress from ignorance to
            • 68:00 - 68:30 truth absolutely is not it's messy false turns um almost kind of like ideological Crusades I mean and it really is ideology religious this rounds in a full circle the apocalyptic uh mentality of the of the ex risk people I want to talk about this new generation of foundation models and I'll I'll begin with a quote from your book again large language models of which gpt3 is perhaps the best known are the most prominent example of current Foundation models while
            • 68:30 - 69:00 Foundation models have demonstrated impressive capabilities in certain tasks because they are inherently disembodied they are not the end of the road in artificial intelligence so in the past three years the jump from Deep learning to what we have right now Foundation models is the Transformer architecture and increasing scale a lot and a lot of of data some people seem to think that the architecture is already there we've
            • 69:00 - 69:30 solved it with the Transformer we have what we need to go to AGI all we need is more scale what do you think is wrong about that argument so firstly let me say what we've seen in the last few years in terms of Transformer architectures and that that which were released by Google a Google lab I believe in 2017 and what they are is an architecture for token prediction and were developed in order to enable large language models so that you could give a prompt and they could predict
            • 69:30 - 70:00 essentially what should what should come next so you know the life and achievements of Winston church or the history of Christ Church College where we are now um and they turned out coupled when you couple a Transformer architecture with the willingness to throw unimaginable quantities of computer power and really mind-boggling quantities of computer power to train them and mindboggling quantities of data you get something which was remarkable and honestly AI researchers that tell you that they were not surprised by how good it was I think is is misleading you
            • 70:00 - 70:30 a little bit they are genuinely remarkable they took me by surprise I didn't expect how good they were going to be but just pause and think for a minute what we've got we've got large language models that you can have a chat about quantum mechanics the history of Christ Church College Liverpool Football Club uh you know the origins of the first world war the economic circumstances that led to the 2008 financial crisis or recipes for um uh for uh Arnold Bennett or whatever uh
            • 70:30 - 71:00 anything you can think of you can you can ask these things about and we look at that and think wow this is AI this is this is intelligence and yet we don't have a robot that could go into your house clear the dinner table and load up the dishwasher why have we got that weird dichotomy because there is a huge range of human activities that actually at the moment are well out of the reach of Ai and those activities are activities in the real world um doing
            • 71:00 - 71:30 robotic AI uh is just very very hard um large language models succeed in remarkable ways and they are genuinely impressive achievements but they succeed on tasks where there are huge amounts of data available and in some sense where the consequences of what they do just doesn't really matter that much you know if you get a bad omelette recipe through chat GPT you get a bad omelet that's not the end of the world you know you build a robot that occupies the real world with human beings and it goes wrong you
            • 71:30 - 72:00 know it can create Havoc it can cause real harm so an a the idea of AI which doesn't Embrace doing things in the real world is quite an impoverished version of AI I think and that's what is uh I think interesting in your critique because you use the word disembodied and that's kind of the the intuition right and that that's what you mean by disembodied and there something else I want to pick up on there I mean you're having conversation with chat GPT you go on holiday for two weeks and leave it hanging it's not wondering where you are
            • 72:00 - 72:30 it's not thinking where's wridge got to or it's not getting bored or anything like that at all it's not doing anything it is just a computer program that's paused in a loop human intelligence animal intelligence is fundamentally different to that we exist in a world we're aware of the world and that's what embodiment means it's not just having a body but it's actually being tightly coupled with the world we live in in that sense I see so in your Turing lectures that you recently G gave you
            • 72:30 - 73:00 separated out uh two general sets of human capacities one is the embodied set and this is uh the ability to sense one's surroundings the ability and there I agree with you right AI is just robotics lacks a lot further behind just you know natural natural language processing image processing image Generation all sorts of stuff like that but what I found really surprising in your Turing lectures is you also listed out a series of intellectual capabilities and even there you didn't
            • 73:00 - 73:30 seem to think that our current generation Foundation models are going to get us there so the things that you said were solved or solvable in the current architecture natural language processing recall Common Sense reasoning but I was extremely surprised that you listed the following uh again not embodied but intellectual capacities as still not being even within the Horizon uh of of current llms logical reasoning abstract reasoning planning arithmetic do you still hold that
            • 73:30 - 74:00 position with gp4 and because like it can do arithmetic it can uh can it can it really do arithmetic or can it do something that looks like arithmetic I mean there is a big question mark around whether um what large language models are doing is doing those things or whether they're doing something that looks like patent recognition so arithmetic I'll concede you probably now is is a solved problem but there's a huge body of work looking at whether these things can actually solve problems that are not just variations of
            • 74:00 - 74:30 something they've already seen in their training data and the question of is it really originally solving a problem versus just doing patent recognition at the moment that's one of the big questions and the jury is very much out on that and the weight of evidence at the moment is they are not doing problem solving they are doing something which is much more like patent recognition so let me give you an example to illustrate this um so uh in AI we've long been concerned with problem solving and planning and planning is is the process
            • 74:30 - 75:00 of here is some goal I want to achieve here is where I start out and here are some uh choices available to you some actions that you can perform that will transform the world how do I organize th AC those actions to transform me from where I am to my goal absolutely fundamental AI capability that people have been looking at for uh for well over half a century so can large language models do planning First Site people got very excited because it appeared that they could you can a trip like seem exactly but uh on closer
            • 75:00 - 75:30 inspection suppose you do the following um suppose you obfuscate all the terms that are being used in your in your plan so you don't use words that it's familiar with but you express a problem the same problem right uh using terms that you know that it will not have appeared in the training data can it then solve the problem now I emphasize it's the same problem you're just using words it's never seen before and no at the moment the answer is no it can't so
            • 75:30 - 76:00 it can't originally solve problems we can we can do that we have problem solving capabilities so that suggests that when it can when it's looking at planning a trip you've seen thousands of trip planning guides and trip agendas and so on and it's doing patent matching to pick up on that uh and help you plan the trip but is it actually planning from P first principles how to organize those various actions to plan the trip right so at the moment I say at the moment uh the weight of
            • 76:00 - 76:30 evidence is that it's not capable of doing uh logical reasoning or problem solving those kinds of things not in a deep way that doesn't mean it's not useful that doesn't mean you can't use it to help plan a trip it can be used in those ways but is it actually doing those things from first principle the weight of evidence at the moment is no right and is you're into tion that not only are these capacities not there but it's a it's an architecture issue that
            • 76:30 - 77:00 we can't expect to get logical reasoning just by increasing the orders of magnitudes of data we throw at it from here because fundamentally what it's doing right for these llms is just next word prediction yeah so is that the issue that you're that you're just is that the deeper issue you're gesturing at so that's what Transformers were designed for next word prediction and the surprising thing was how useful and impressive that turned out to be if you were prepared to throw enough data and compute power at it um but I see no
            • 77:00 - 77:30 reason to believe that that the Transformer architecture is the key for example to robotic AI That's not what it was designed for so I don't see why it should Lal reasoning or logical reasoning necessarily I mean again that's not what it was designed for but iiz that doesn't mean it's not useful and I'm as dazzled as anybody when I use when I use this technology and I am you know daily taken by surprise when people show me the really remarkable things that it can do um and I have to say you know we've gone this is really genuinely
            • 77:30 - 78:00 I think uh a watershed moment in AI history because we've gone from a period where a lot of questions in AI were purely philosophical questions they were literally reserved for philosophers until a few years ago uh and suddenly it's experimental science you know uh are large language models conscious well let's roll up our sleeves and do some experiment and find out no by the way they're not uh but you know these are now practical handson questions and to
            • 78:00 - 78:30 have gone from not having anything in the world that you could apply those questions to to this being actual practical Hands-On experimental science in just a few years is mind Bing right let me play a devil's advocate here um because for me as as a philosopher studying AI has been a very humbling experience because it might reveal what how little reason actually works in humans and here's the challenge I would like for you to respond to which is
            • 78:30 - 79:00 these architectures are built off of an imitation of the human mind and how the human mind is is connected through neural networks right and so the intuition is Maybe by by just IM imitating that even though they're not designed for for logical reasoning because we've imitated that the structure of the human brain that it's this emergent phenomenon know maybe what humans are doing is not first principles thinking maybe we're we're just pattern pattern matching maybe it's all pattern matching down there maybe I mean I I
            • 79:00 - 79:30 believe I don't really believe this one but maybe we are just doing next word production when we're having a conversation and there is an entirely serious uh school of thought that thinks actually perhaps we need to rethink what the the what humans are doing and that actually that we have overblown expectations about what beliefs about what we're what we're doing I don't think humans are a Transformer architecture I don't think that's what we're doing I think there's a lot lot more that's going on we are animals that have evolved to inhabit planet Earth and
            • 79:30 - 80:00 to interact with other human beings and to understand the fundamentals of human nature I think you have to understand those two things Transformer architectures are not that not by a long long long way um however the point you make about emergence I think is an entirely valid one we don't understand how intelligence emerges in human beings how does all that gooey stuff in our heads all those electrochemical processes and so on give rise to you and me we don't understand that in a deep
            • 80:00 - 80:30 way at all and that's what's so exciting about the present time that let's roll up our sleeves and find out how it's actually doing this but without wishing to denigrate these systems at all there is a very real sense in which they are a hack they are an engineering hack that's put together they are not following some deep model of mind or some deep philosophical theory about how human intelligence is or or some deep cognitive science theory of human intelligence um they are a technological hack um and although neural networks
            • 80:30 - 81:00 artificial neural networks were inspired by the structures we see in human and animal brains they are not an attempt to Faithfully recreate that in people there have actually been attempts to do that there was a very large European funded project that wanted to try to recreate a brain an actual brain but that's not what neural networks are doing given how much success we've had about imitating a specific structure of brain right how how neurons are are linked together in
            • 81:00 - 81:30 in computation should we be looking more into biomimicry and should we be studying the brain more and see if there's other structures we can replicate is that the path forward to finding out the architecture is to take us to to I think that's one that's one way forward and I think we will surely get some insights I mean we have a very incomplete understanding of how the brain is organized I mean the brain is not just one big homogeneous neural network even though you know it contains vast neuro multiple neural networks but it has it has some functional structure
            • 81:30 - 82:00 and we we understand a lot more now than we did even 30 years about ago about the functional structure of the brain but a very incomplete understanding so that's going to be one way to go but I emphasiz again you know we are great apes that have emerged through a process of billions of years of evolution to inhabit planet Earth at ground level at sea level roughly speaking and to be able to um to learn about the physics of planet Earth and to be able to operate within the Dynamics of the physics of
            • 82:00 - 82:30 planet Earth but also to be able to interact with other great apes and those are the two key big components of human intelligence learning about our world and learning about other Apes human beings um and we shouldn't lose sight of that when we think about the successes that AI has had you know we we're not just a big neural network a big H modulous neural network there's an awful lot more going on than that right um but there is I think something Melancholy about what AI techniques have worked and what haven't and let me quote
            • 82:30 - 83:00 to you a lovely quote from your book this is you speaking in your voice in July 2000 I was at a conference in Boston watching a presentation by one of the Bright Young stars of the new AI I think this is when ml was starting to to pick up steam I was sitting next to a seasoned AI veteran someone who had been in AI since the Golden Age a contemporary of McCarthy and Minsky he was contemptuous is this what passes for AI nowadays he asked where did the magic go you speaking and I could see where he
            • 83:00 - 83:30 was coming from a career in AI now demanded a background not in philosophy or cognitive science or logic but in probability statistics and economics this I think is what you were getting at about not seeming as poetic you know you you thought with with psych and artian logic that's what was going to do it well turns out like the neural net architecture with a black box that we can barely understand more than our brains a lot of mundane mathematics exactly and you think we need more
            • 83:30 - 84:00 clever architecture to get our neural Nets to behave differently well it's it turns out just increasing the scale fundamentally changes the the increasing output of of the behavior oh that is a that's a very depressing lesson I mean the fact that you know you would think we the chief source of advances in AI is scientific developments actually no it's just more compute more data um and there's there's there's an article by this called Rich by a guy called Rich Sutton called The Bitter lesson and rich is a very renowned uh machine learning
            • 84:00 - 84:30 researcher and he said look the truth is we've made progress primarily in AI by uh you know some core ideas but actually the the big steps in progress we've seen and when we've been willing to throw 10 times more comput 10 times more data and so that that is a sobering lesson but there is still magic there in AI I mean so uh the fact now that we have um machines like chat GPT that we can have a conversation with that we can turn the
            • 84:30 - 85:00 conversation to anything that we might care to imagine compared to where we were five years ago that is simply astonishing and you know uh if I wish I was a PhD student now and having the opportunity to explore this kind of weird new landscape of AI and to try to figure out you know what are the what are the fundamental laws that govern these systems what are the principles try to uncover the science underneath this this technology um there is still some magic there you just have to look a
            • 85:00 - 85:30 bit harder to find it it might be a bit better if we needed the most advanced uh math or we need to event this fancy architecture to study human brains very closely for decades um but I think I love your word sobering because that's another way to frame Melancholy or disappointing and I'll end on this this one observation which is uh I'm preparing a lecture on the stoics right now and the stoics famously think that humans are extremely rational creatures that even unbeknownst to us when I desire something I'm making an implicit
            • 85:30 - 86:00 proposition that that thing is good behind most human behaviors there's an explicit or sorry there's an implicit uh true or false proposition someone on the opposite extreme is probably someone like Freud where our unconscious is not known and perhaps even greatly unknowable to us and I think the fact that neural Nets these black boxes that we ourselves don't really understand have gotten so much more success than
            • 86:00 - 86:30 something explicit like psych also tells us perhaps a sobering lesson about how our own intelligence works and that to me again coming from a philosophical perspective is the most exciting stuff about all of this which I'll go back to which is that what were once philosophical questions have now become experimental science all right thank you Professor thank you for fascinating interview thanks for watching my interview if you like this conversation I think you'd also enjoy my discussion with Nick
            • 86:30 - 87:00 Bostrom on his new book deep Utopia it tries to imagine what there is left for humans to do after AI has surpassed Us in all domains now these interviews are a part of an AI series that I'm producing as a fellow of the cosmos Institute a nonprofit studying philosophy and artificial intelligence you can find links to our website the boss room interview and everything else we cover today in the description below thank you