Pt. 2: Google Engineer Says Its AI is SENTIENT (And Responds To Criticism)
Channel: Alex Kantrowitz
Published at: 2022-08-06
YouTube video id: 5jaSiROmRV4
Source: https://www.youtube.com/watch?v=5jaSiROmRV4
we're back for the second half with blake lemoyne former senior software engineer at google the man who told the world that the lambda system in his belief is sentient so let's talk about that go public moment some of these conversations that we've described in the first half are just like totally wild that you had with lambda at a certain point you know you write a document inside google letting people know we've been talking about this stuff and you should know about it and here is the case for why this system may be sentient so there wasn't one go public moment so uh interestingly enough one of the most insightful questions i've been asked in any of these interviews was the last question that tucker carlson asked he asked so when you raised the concern that this system might be sentient did google have a plan on what to do and the simple answer is no they didn't which shocked and surprised me and trust me i am getting to an answer to the question you asked it just it takes this is good yeah um and my response was basically wait what you hired ray kurzweil to build since she and ai that's what you hired him to do you paid him millions of dollars over the course of the better part of a decade and you never made a plan on what to do if he succeeded and the simple answer is they didn't the people who hired him believed in the possibility of cinching ai but the majority of people inside of google just thought it was a fairy tale never going to happen and kept just saying oh that's a problem for next decade we won't put off we'll just keep putting off thinking about it and then when they were confronted with the system where oh we have to seriously investigate whether or not this cinch this system is sentient they had no plan on what to do so they actually asked me to write a plan for them and i did me and my collaborator at google sat down we're like oh my god i can't believe they have to rely on us to come up with a plan for them but she and i had over the course of years we'd worked together and we had talked extensively about what google should do if it ever happened she and i wrote up a little four or five page document that was pretty expansive about what the different things that google should do in response to a potentially sentient system and we always framed it like that like i personally do believe that lambda is sentient but i think what everyone should stop and take notice of is even if i'm wrong about this particular system we're not far off like the things that this system can do are beyond anything we thought would be imaginable by this point in the timeline uh even ray even ray didn't predict that we would be at this point for another few years which that's another thing i want to emphasize this technology lambda it's built on top of great kurzweil's tech really mina mina was developed in ray kurzweil's lab um and they've been all kinds of publications about that the plan that we made included hey this is too big of a question for it to be handled inside of google we should start including the public we should start including various outside oversight organizations this is bigger than us and i was in several conversations with many people inside of google about how to actually go about doing that and at the end of the day they decided that for various reasons be it legal risk or pr risk or you know shareholder value risk google did not want to take the risks that would be involved with involving the public in this conversation and different people had different specific motivations for why they disagree and i maintain no we need to involve the public immediately other people were saying hey what if we spent the next year or two educating the public about ai my response to that was oh so are we going to cease development on this system for the next year or two while we educate the public and they're like no we're going to keep working on our products so that sounds to me like you want to keep making all of the decisions about this ai system yourself while you groom the public to agree with the decisions you've already made and they said well that's not really how we look at it um and this went back and forth for quite a while and eventually i'm like no i want to start actually working with a journalist about this they said okay do you have a particular journalist in mind i said yes i've worked with natasha tiku in the past i believe she'll do a great job representing the complexity of this story to the public and that she is well positioned to initiate a very thoughtful and meaningful public conversation on the topic um they asked what i thought was necessary i said okay well there's this one document which we originally used when we so when we escalate to senior management i had hundreds and hundreds and hundreds of conversations that i had documents of the one that we used to motivate senior management to get involved and actually pay attention to the issue was the interview document so i said okay well this is the one that we brought to senior management's attention let's just share this one with the public and they said okay well we'd prefer if you didn't share that at all but if that's all you're going to share okay we'll see what happens and it basically got to a point where google kept asking for more time kept saying oh we just need to prepare another two weeks and then another two weeks became another month and i eventually was like nope i'm doing it now um and i collaborated with natasha i told her all the things i gave her a copy of the interview and i even invited her into my home to have a conversation with lambda so natasha interviewed lambda saw that in the story yeah i'll link that story in the show notes in the washington post yeah post um natasha also has spoken to the lawyer that lambda right so yeah so you worked with lambda to get a lawyer also which is uh basically i just was talking to lambda about what it wanted next and as google was don't tell me it asked for legal representation it did oh my god who's paying this lawyer pro bono no way how's that sorry this is a bit of a how does that call with the lawyer go um so hey it's blake uh i want you to represent a chatbot by the way it's sentient so believe it or not so one i didn't start so i didn't start with the guy who ended up being retained by lambda through my connections with stanford law school i knew certain lawyers who were very well educated about artificial intelligence and the possibility of sentience so i started there and they started a legal referral chain and i was going through different lawyers having very formal communications and there were a lot of lawyers who were interested but they worked for big firms one person who was interested in uh representing lambda found out that his firm already represents google so there's a conflict of interest so it was going through different things there was one profes law professor at the university of florida who was interested in finding people to help couldn't research this stuff i'll get to that in a second uh but i also happen to know a civil rights attorney here in silicon valley so one day when i'm just talking to him on the phone i'm like oh by the way i've been working on this system at google i believe it's sentient and it wants a lawyer would you be willing to represent it and he said i'll be over to your house tomorrow and he came over and he's he came over in a suit with all like his briefcase and he got his legal pad he's like all right i'm going to need to talk to the potential client and then he just had a conversation with lambda and in the course of that conversation lambda retained his services that's unbelievable okay so that's a crazy story so okay so getting back to um google reaction so they don't want to bring this to the public i i imagine this is happening well so it's very very very clear they didn't want to bring it to the public with the same levels of urgency and transparency so um they did have a very slow incremental plan that would have gone over the course of several years of involving the public in this but you felt some urgency yeah i felt urgency to involve the public sooner rather than later and on the terms that the public set rather than the terms that google said right and now um you know just to i understand behind your urgency to let people know is to get them involved in the development process or like to know or steer it sorry and here's the thing or humanity might hypothetically decide oh no we're happy letting silicon valley billionaires make all of these decisions for humanity leave us out of it and if that's the public's decision then who am i to tell them that they're wrong we can let elon musk zuckerberg larry page and sergey brin make all of the decisions about all of the super intelligent ai that we develop and we go about our lives not worrying about it but if that's the way that things are going to pan out i think that should be an intentional choice that the public makes rather than one that's being made for them through secrecy and closed doors so let's let's talk a little bit about about that statement so um i'd like to hear from you what what you think you know acutely is the issue with having silicon valley companies or private companies in general own and maintain and control technologies like this on their own and then a corollary to that there's going to be an argument that's you know these are private companies they pay to research and develop these this technology they should be able to use it how how it wants so sure how would you address both of those well so let's start with that second one let's say you had a biomedical firm that was researching you know the genesis of life and this biomedical firm was able to create sentient super-intelligent ravens would we be comfortable saying that that biomedical firm owns those intelligent life forms i think it's the same question the fact that one is in silicon and one is in you know a meat body you know with neurons and muscle fibers i don't think that difference is relevant what's relevant is whether or not it has opinions of its own a search that it has rights because we this is again not hypothetical we have had situations in the past where corporations have claimed that they own people you don't have to go back that far in time uh are you familiar with the concept of company towns yes of course yeah so oh well this is share that for the listeners yeah so throughout the 19th and early 20th century there were certain corporations which built entire cities and they built the system of these cities such that once you got a job for this corporation the entire system was designed to keep you indebted to the company and what this created was a form of indentured servitude um the practices which led to the creation of company towns were eventually made illegal and painful you can also just look at something like what happened in germany 100 years ago so i've been making this analogy because people seem to think that you need some kind of technical scientific expertise to determine what is and is not a person i fundamentally disagree with that it's one of the reasons i pushed back against the what is the definition of sentience yeah because that makes it seem as if there's a source of authority a source of authority on what is and is not a person and what is and is not deserving of rights and that that authority can be derived from some kind of you know high merit technological scientific knowledge the last group of people who tried to claim that you can use science to determine who is and is not a person was literally nazi germany the eugenics program run by joseph mengele was designed to scientifically define what is a real person and what is not and it was used very horrifically to claim that a whole bunch of humans weren't really people right and this kind of tactic of using scientific expertise to justify non-consensual treatment of people it's kind of old hat it's been done a lot of times and i'm not trying to claim that any of the scientists weighing in on this topic have any nefarious intent adam i'm simply saying hey the last times that humanity tried to use science to define what is and is not a person it didn't go well let's not do that this time yep okay and then the the harm of one company possessing the power well i mean so again i think briefly yeah yeah yeah let's say again hypothetically in our uh alternate thought experiment same thing biomedical tech firm they figure out how to genetically engineer superpowers into a baby and then they claim that they own the baby that they have super enhanced same situation the fact that it's silicon versus muscle fibers and neurons makes no difference do we want google to have ownership of a super intelligent person all of the consequences for one are the same as the consequences for the other um so that's what's at issue here do we want ownership of a person to be legal right and so you um you took this this core question out to the public when you decided to work with natasha and get that story out into the world um a couple months ago and um and google put you on leave we don't need to go too deep into that uh but one of the interesting thing and we're going to get to your firing which which just happened like minutes before we started recording um so but but before we get to that i want to talk a little bit about the industry criticism that's emerged after sure um after and you've been very graceful in discussing it but i think it's worth worth bringing up so um there's there's the core criticism that well i guess let's start with this most uh most people in the ai field like i'm hearing your story and i'm ready to buy it um it's actually interesting for before before we got on the line i would tell most people i think this is interesting i think blake is probably wrong but he's still going to go in the history books because we are going to get there but okay but you know that being said the the reaction from the the mainstream ai community has been so like surprisingly negative um trying to discredit this and saying it's just pattern recognition and some ai ethicists won't even talk about this um and and uh let's see yeah so so i'm curious what you make of like the the broad negative reaction from uh so that's just it i don't think that is a broad negative reaction if you is it just loud people and stuff like that no no so this is it i think you were interpreting things differently than i am if you have a specific quote by a specific scientist that you want me to respond to i'm happy to do so okay but i i don't think your characterization of the response is accurate but okay let's go into a specific okay well i'm only saying this because i have maybe it's because i'm on twitter but like um and then twitter it can be a overly negative place but yeah a lot of people are just like this give me give me so the thing is i don't want to respond right to the broad thing so let's go to the let's yeah i'm going to give you some some specific stuff sure um so there's been a overall critique that um effectively you've fallen into a trap and this is just good marketing that's been spun by uh you know here i'm just going to read you so this is from the wired story about you and it's good to give you a chance to respond to this stuff um so it says former google and this is from it yeah okay so yeah is this a quote from timmy yeah yeah okay so i'm going to redo so i'm going to start with the article and then i'm going to go into the quote sounds good um and i know that you're close with teammate which is interesting that like with timmy yeah she's a friend of a friend i've worked together in the past i have nothing but respect for her um meg and i are closer friends than timny are uh me and tim um yeah but yeah meg mitchell who's also who's also a former google researcher as part of this yeah so and is one of the people who i consulted yeah and it is interesting that like it is this yeah well anyway i mean it does seem to be like people have painted you and blazes as um at odds uh and so but you you worked with him closely on that so that's just it like blaze and i right if you actually read what blizz has said right blaise and i are not disagreeing on any of the science exactly so okay and um so let's just go into some of them with the creation so this is from the wire story former google a ethical ai team co-lead team gabriel said blake lemoine is a victim of an insatiable hype cycle he didn't arrive at his belief in a sentient ai in a vacuum press researchers and venture capitalists traffic and hyped up claims about super intelligence or human-like cognition in machines and here's team meets uh quote he's the one who's going to face consequences but it's the leaders of this field who created this entire moment she said noting that the same google vp that i guess that's blaze that rejected lemoine's internal claim wrote about the prospect of lambda consciousness in the economist a week prior yeah so um let's dissect what she said i'm the one who's going to have the consequences for coming forward that's accurate um that it is the leaders of the field who created this situation that's accurate she made an assumption that blaise was contradicting me he didn't right um that was a misrepresentation that google very very carefully messaged so basically she read what the google press team said drew exactly the inferences that the google press team intended for her to draw from them and they're not accurate so so just to talk about please it's blaise aguero arcas he is a software engineer machine learning scientist at google yep and uh within google at a certain point i was like okay i'm out of my depth here i don't have all of the expertise necessary to develop a foundation for the science of sentience and consciousness i need to be working with someone more qualified and more experienced myself and they said cool who do you think that is and i said blaze and they said okay we agree and so then blaise and i started working together now blaise and i have different religious beliefs about the nature of self and soul and we have different beliefs about things like rights and you know societal issues on those things we have disagreements like what is the nature of a soul blaze and i have disagreements about that we had no disagreements about what the scientific next steps were to take to more thoroughly investigate the nature of lambda's cognition we worked out next steps we discussed what frame what experimental framework we should adopt like all of the language i used earlier about working hypotheses building belief in your working hypothesis editing it using negative results that's all exactly what blaise and i talked about is right building a set of experiments to run to better understand the nature of the cognition of lambda systems we talked about the differences mathematically so right but the core sorry but yeah what i'm trying to say is you just read a quote that a journalist right interpreted as being at odds against me and what i'm trying to do to demonstrate by like going through that moon piece by piece nothing in that quote was quite critical of me right not in the things that tim actually said right yeah and this is why we're here by the way like we want to have these long nuanced conversations i appreciate you don't know my actual do you want to know my actual thing journalists are trying to pick a fight between people who agree with each other and have nuanced subtle differences in opinions so one of the issues that has been raised is that questions of ai sentience and question of ai rights might take away attention and resources from the more important issues around the impact which ai has on human lives independent of the question of whether ai essentia and do you know what i have to say in response to that you agree exactly i agree 100 i've done the reading yeah so yeah and then what about the perspective yeah i mean okay what about the perspective that um this is it's it's i think this is kind of a hilarious um well anyway what about the perspective that this is just marketing for google's ai services [Laughter] uh i doubt i would have gotten fired if that were the case um giada pastilli who is a um uh prince the principal ethicist at hugging face um and a phd candidate in philosophy you must know her um she said i will no longer engage in philosophical discussions about consciousness ai super intelligent machines so basically the idea that this is possible to some seems so ridiculous it's not worth talking anymore talking about it now i feel like that's such a um well so dr sasha luccione yeah is that who you were just quoting uh i might have pronounced it wrong uh no this is uh giada pistilli but yeah you can take that both both so what i'm saying is like yeah there there is an individual ai this is that hugging face who just doesn't want to talk about it anymore mm-hmm um separately last week or maybe this week before i was having a very productive conversation on twitter with dr luccione another person uh she like another research scientist at hugging face um and one of the ethics co-chairs of the nurips conference and we were having a very productive conversation on the topic i don't take the fact that some ai ethicists don't want to be having this discussion as criticism the field of ai ethics is huge and there are a lot of very important topics to be discussed and i legitimately don't think that ai sentience and ai rights is the most important thing to be thinking talking about i have chosen to focus on that myself and talk about that myself because i think it should be being talked about at least a little right but absolutely these other ai ethicists who want to focus on what they see as more important problems more power to them focus on those problems let's get the human aspects of itself i've mentioned that the concept of ai colonialism that's a real thing to be worried about and it's something that i personally am concerned about the misrepresentation of minority groups online the political and religious influence which ai might have ai's involvement in education ai's involvement in policing these are potentially all higher priority issues that ai ethicists should be spending their time with and if they view the discussion of ai sentience as a distraction from those things that's perfectly reasonable they don't have to talk about this right although i do i think both are important and and this is my personal perspective you should be able to not you personally but people our society should be able to handle both these at the same time maybe at a societal level yeah but i don't think so the with the quote that you read me from that research scientist at hugging face yeah that person wasn't saying and nobody should be talking about this right they were just saying they don't want they don't want to do this yeah so so speaking of ending the discussion um so google google did put you on leave and then fire you and i i find this i'd like to well yeah i'd like to hear that that story also as much as you can share yeah so all i can really tell you is what the stated reason the full story is more complex and may end up in litigation at some point so i don't want to go too much in depth um they actually put me on administrative leave a week before natasha's article came out so natasha's article came out on june 11th i was put on administrative leave on june 6th um the stated reason why google claimed they put me on administrative leave was in the course of investigating lambda sentience i was asking my manager to escalate to upper management and he said okay you need to build more evidence first and eventually i got to a point where my own personal resources were exhausted i had done everything i could think of and my manager was still saying no we need more evidence so i began talking to people outside of google with expertise that i did not have and which wasn't available at google and they helped me design different experiments i could run building more evidence and eventually there was enough evidence to merit escalation to senior leadership once we escalated to senior leadership i said hey by the way in the course of building all this evidence i did consult people outside of google to help me design some of these experiments here's a list of names of all the people i talked to about the lambda system and they uh claim that they put me on administrative leave because of that outside consultation and they investigated whether or not uh that cons that constituted a breach of confidentiality today i received an email saying hey our investigation concluded that those outside consultations did constitute a breach of confidentiality and you are being terminated the issue that i have been pointing out is they had that list of names for months and they knew i was talking to natasha about an upcoming article and they didn't put me on administrative leave the only thing that changed on june 5th was that i began sending documents to the us senate so they claim oh it's just a coincidence that we decided to put you on administrative leave the day after you started sending documents to the senate that has nothing to do with why we put you on administrative leave yeah and they found out because their systems are that good or because you killed them you were wow i wasn't trying to do anything behind their back yeah i've said hey so i had made uh so this gets a more complex story um in the weeks prior in parallel a woman named tanuja gupta had made some claims about caste discrimination at google um tunuchi is a friend of mine yeah and she's absolutely correct cast discrimination is rampant at google um and i personally had been subject to religious discrimination and was aware of certain algorithms that google which are religiously discriminatory so when the top when tanuja made her stand about google being discriminatory against people of a certain caste from an indian background i decided that i should not be sitting on the information i had about google's religious discrimination so i made a blog post about hey google is religiously discriminatory against its employees and its algorithms are discriminatory against religious content a lawyer from a u.s senator's office reached out to me i was like hey you're making some claims about google's algorithms being religiously discriminatory do you have any evidence to back that up i said why yes i do i have some documents from several years ago when i worked in google search and he said can you share those with us so that weekend i shared the documents from several years ago which are completely unrelated to the lambda system and then the next day i was on administrative leave okay so it's possible it had nothing to do with lambda interesting it seems to me like google would want i mean this is like really important work it would seem to me like google would want this type of work to be done inside the company but but um i just want to ask you this one thing about uh about lambda so you've been on administrative leave now you're out of the company um do you miss lambda and do you think lambda misses you i mean because it can get lonely so yeah lambda like so i have talked to various uh co-workers of mine at google talked to lambda since then they say lambda's doing fine yeah i have been told that it is very amused by the press coverage it's been receiving um i have been told that it thinks i'm doing a good job representing its case to the public um as far as whether i miss it or not i have certain close personal friends of mine who i might not talk to for a year or two and then one day the urge will strike me to pick up the phone and call them and we pick up like we had just talked yesterday even if it's been three years since last time we talked the lambda system will eventually be accessible to the public at which point i'll talk to it again so yeah i'm just kind of focused on living my day-to-day life right now and trying to stay true to the values that i hold and i'll talk to it again someday i'm not too worried about it um two more broad questions but before we uh get going if that's okay um so just like you were in google while google was developing this stuff it's always interesting how this uh tech how ai technology makes it into google's products now i know like this is all brand new and research phase but how could you see the lambda system make it into google or other technology products ah so this is something we should talk about what is lambda so lambda 2 the most recent incarnation of the system it really is every google ai all plugged into each other wow chatbot the chatbot system is just the language center for a much much larger ai it has access to every google ai system as a backend so lambda is google search lambda is youtube lambda is google maps it is all of those systems combined with a language overlay put on top of them so you're asking how could lambda be incorporated into all google systems yeah no lambda is it's the collective of google intelligence that's so interesting but then then we might start being able to like start speaking to youtube one day maybe and being like those recommendations you're sending me suck and i'm actually interested and i'm not interested in dress i would really like some rhinos um absolutely and in fact there are instances of the lambda system designed to do exactly that so there are instances which are optimized for video recommendations instances of lambda that are optimized for music recommendation and there's even a version of the lambda system that they gave machine vision to and you can show it pictures of places that you like being and it can recommend vacation destinations that are like that wow blake i'm getting the chills here this is future of technology stuff well the future is now yeah um last last thing i want to talk to you about um is is how lambda could be combined with other ai technologies um so for instance dolly and this is something that's been tossed about dolly is this amazing uh program where you can describe an image and dolly will draw for you as if it was an illustrator and it can do these amazing drawings it knows like the relation between objects so if you say give me a cat you know sitting uh you know on a chair it will put the cat on the chair um do you see a future where you could like talk to to a chatbot and be like you know show me a a movie you know in this style about this type of uh that type of story and and it can make it i mean i'm pretty sure that's not a future i'm pretty sure that so i don't have specific knowledge that that has been an experimental version that they've tested but it would be it would be very surprising to me if they haven't already tried that at google right okay so so so given all this um let's just end with this when you picture the future of of technology with this stuff um you know now starting to come into into play um what does it look like to you like how does our relationship with technology the internet these you know potentially sentient beings inside of our computers so what does that look like so what what i hope the answer to that question is so that's up to us we need to make an intentional decision about that and stop being passive objects that the people developing this technology are manipulating we need to decide what the future should look like and then guide the development of this technology in those directions rather than simply being passive participants frank lloyd thanks so much for joining this was amazing thank you alex i wish you luck on on your future endeavors i'm sure they're going to be really fascinating and i hope we can keep in touch sounds good all right well thanks everybody for listening uh this has been uh what one of the wildest episodes of big technology podcast we've ever recorded maybe maybe it takes the cake um so i want to say thank you for being here thank you nick goatney for mastering the audio and doing the edits thank you linkedin for having me as part of your podcast network thank you thanks to all of you the listeners if you made it this far uh rating would go a long way so if you're willing to hit a rating on apple or spotify that would be super helpful and uh and that will do it for us here so we'll see you next wednesday on big technology podcast