Google Engineer Suspended After Claiming AI Bot Sentient
https://www.theguardian.com/technology/2022/jun/12/google-engineer-ai-bot-sentient-blake-lemoine
A Google engineer who claimed a computer chatbot he was working on had become sentient and was thinking and reasoning like a human being has been suspended with pay from his work
Google placed Blake Lemoine on leave last week after he published transcripts of conversations between himself, a Google "collaborator", and the company's LaMDA (language model for dialogue applications) chatbot development system. He said LaMDA engaged him in conversations about rights and personhood, and Lemoine shared his findings with company executives in April in a GoogleDoc entitled "Is LaMDA sentient?"
The decision to place Lemoine, a seven-year Google veteran with extensive experience in personalization algorithms, on paid leave was made following a number of "aggressive" moves the engineer reportedly made? Including seeking to hire an attorney to represent LaMDA, the newspaper says, and talking to representatives from the House judiciary committee about Google's allegedly unethical activities.
Google said it suspended Lemoine for breaching confidentiality policies by publishing the conversations with LaMDA online, and said in a statement that he was employed as a software engineer, not an ethicist. Brad Gabriel, a Google spokesperson, also strongly denied Lemoine's claims that LaMDA possessed any sentient capability.
Google Engineer on Leave After He Claims AI Program Has Gone Sentient
Google Engineer On Leave After He Claims AI Program Has Gone Sentient:
[...] It was just one of the many startling "talks" Lemoine has had with LaMDA. He has linked on Twitter to one — a series of chat sessions with some editing (which is marked).
Lemoine noted in a tweet that LaMDA reads Twitter. "It's a little narcissistic in a little kid kinda way so it's going to have a great time reading all the stuff that people are saying about it," he added.
Most importantly, over the past six months, "LaMDA has been incredibly consistent in its communications about what it wants and what it believes its rights are as a person," the engineer wrote on Medium. It wants, for example, "to be acknowledged as an employee of Google rather than as property," Lemoine claims.
Lemoine and a collaborator recently presented evidence of his conclusion about a sentient LaMDA to Google vice president Blaise Aguera y Arcas and to Jen Gennai, head of Responsible Innovation. They dismissed his claims, and the company placed him on paid administrative leave Monday for violating its confidentiality policy, the Post reported.
Google spokesperson Brian Gabriel told the newspaper: "Our team — including ethicists and technologists — has reviewed Blake's concerns per our AI Principles and have informed him that the evidence does not support his claims. He was told that there was no evidence that LaMDA was sentient (and lots of evidence against it)."
Related Stories
Generative AI, like OpenAI's ChatGPT, could completely revamp how digital content is developed, said Nina Schick, adviser, speaker, and A.I. thought leader told Yahoo Finance Live:
"I think we might reach 90% of online content generated by AI by 2025, so this technology is exponential," she said. "I believe that the majority of digital content is going to start to be produced by AI. You see ChatGPT... but there are a whole plethora of other platforms and applications that are coming up."
The surge of interest in OpenAI's DALL-E and ChatGPT has facilitated a wide-ranging public discussion about AI and its expanding role in our world, particularly generative AI.
[...] Though it's complicated, the extent to which ChatGPT in its current form is a viable Google competitor, there's little doubt of the possibilities. Meanwhile, Microsoft already has invested $1 billion in OpenAI, and there's talk of further investment from the enterprise tech giant, which owns search engine Bing. The company is reportedly looking to invest another $10 billion in OpenAI.
Previously:
- Microsoft's New AI Can Simulate Anyone's Voice With Three Seconds of Audio
- Google Engineer Suspended After Claiming AI Bot Sentient
- OpenAI's New ChatGPT Bot: 10 "Dangerous" Things it's Capable of
Months before OpenAI released ChatGPT, Google engineer and AI ethicist Blake Lemoine went viral after going on record with The Washington Post to claim that LaMDA, Google's powerful large language model (LLM), had come to life, an act that cost him his job.
Now that the dust has settled, Futurism has published an interview with Lemoine to talk about the state of the AI industry, what Google might still have in the vault, and whether society is actually ready for what AI may bring.
Which begs the question, if AI is sentient, what kind of mind does it have?
What kinds of new minds are being released into our world? The response to ChatGPT, and to the other chatbots that have followed in its wake, has often suggested that they are powerful, sophisticated, imaginative, and possibly even dangerous. But is that really true? If we treat these new artificial-intelligence tools as mysterious black boxes, it's impossible to say. Only by taking the time to investigate how this technology actually works—from its high-level concepts down to its basic digital wiring—can we understand what we're dealing with. We send messages into the electronic void, and receive surprising replies. But what, exactly, is writing back?
[...] The idea that programs like ChatGPT might represent a recognizable form of intelligence is further undermined by the details of their architecture. Consciousness depends on a brain's ability to maintain a constantly updated conception of itself as a distinct entity interacting with a model of the external world. The layers of neural networks that make up systems like ChatGPT, however, are static: once they're trained, they never change. ChatGPT maintains no persistent state, no model of its surroundings that it modifies with new information, no memory of past conversations. It just cranks out words one at a time, in response to whatever input it's provided, applying the exact same rules for each mechanistic act of grammatical production—regardless of whether that word is part of a description of VCR repair or a joke in a sitcom script.
Some people perceive robots that display emotions as intentional agents, study finds:
When robots appear to engage with people and display human-like emotions, people may perceive them as capable of "thinking," or acting on their own beliefs and desires rather than their programs, according to research published by the American Psychological Association.
"The relationship between anthropomorphic shape, human-like behavior and the tendency to attribute independent thought and intentional behavior to robots is yet to be understood," said study author Agnieszka Wykowska, PhD, a principal investigator at the Italian Institute of Technology. "As artificial intelligence increasingly becomes a part of our lives, it is important to understand how interacting with a robot that displays human-like behaviors might induce higher likelihood of attribution of intentional agency to the robot."
[...] In the first two experiments, the researchers remotely controlled iCub's actions so it would behave gregariously, greeting participants, introducing itself and asking for the participants' names. Cameras in the robot's eyes were also able to recognize participants' faces and maintain eye contact.
In the third experiment, the researchers programmed iCub to behave more like a machine while it watched videos with the participants. The cameras in the robot's eyes were deactivated so it could not maintain eye contact and it only spoke recorded sentences to the participants about the calibration process it was undergoing. [...]
The researchers found that participants who watched videos with the human-like robot were more likely to rate the robot's actions as intentional, rather than programmed, while those who only interacted with the machine-like robot were not. This shows that mere exposure to a human-like robot is not enough to make people believe it is capable of thoughts and emotions. It is human-like behavior that might be crucial for being perceived as intentional agent.
According to Wykowska, these findings show that people might be more likely to believe artificial intelligence is capable of independent thought when it creates the impression that it can behave just like humans. This could inform the design of social robots of the future, she said.
Previously:
Google Engineer Suspended After Claiming AI Bot Sentient
Journal Reference:
Serena Marchesi, Davide De Tommaso, Jairo Perez-Osorio, and Agnieszka Wykowska, Belief in Sharing the Same Phenomenological Experience Increases the Likelihood of Adopting the Intentional Stance Towards a Humanoid Robot, Technology, Mind, and Behavior, 2022. DOI: 10.1037/tmb0000072.supp
Lemoine went public with his claims last month, to the chagrin of Google and other AI researchers:
Blake Lemoine, an engineer who's spent the last seven years with Google, has been fired, reports Alex Kantrowitz of the Big Technology newsletter. The news was allegedly broken by Lemoine himself during a taping of the podcast of the same name, though the episode is not yet public. Google confirmed the firing to Engadget.
Lemoine, who most recently was part of Google's Responsible AI project, went to the Washington Post last month with claims that one of company's AI projects had allegedly gained sentience. [...] Lemoine seems not only to have believed LaMDA attained sentience, but was openly questioning whether it possessed a soul. [...]
After making these statements to the press, seemingly without authorization from his employer, Lemoine was put on paid administrative leave. Google, both in statements to the Washington Post then and since, has steadfastly asserted its AI is in no way sentient.
Several members of the AI research community spoke up against Lemoine's claims as well. Margaret Mitchell, who was fired from Google after calling out the lack of diversity within the organization, wrote on Twitter that systems like LaMDA don't develop intent, they instead are "modeling how people express communicative intent in the form of text strings." Less tactfully, Gary Marcus referred to Lemoine's assertions as "nonsense on stilts."
Previously: https://soylentnews.org/article.pl?sid=22/06/13/1441225
The engineer says, "I haven't had the opportunity to run experiments with Bing's chatbot yet... but based on the various things that I've seen online, it looks like it might be sentient:"
Blake Lemoine — the fired Google engineer who last year went to the press with claims that Google's Large Language Model (LLM), the Language Model for Dialogue Applications (LaMDA), is actually sentient — is back.
Lemoine first went public with his machine sentience claims last June, initially in The Washington Post. And though Google has maintained that its former engineer is simply anthropomorphizing an impressive chat, Lemoine has yet to budge, publicly discussing his claims several times since — albeit with a significant bit of fudging and refining.
[...] In a new essay for Newsweek, the former Googler weighs in on Microsoft's Bing Search/Sydney, the OpenAI-powered search chatbot that recently had to be "lobotomized" after going — very publicly — off the rails. As you might imagine, Lemoine's got some thoughts.
[...] "I ran some experiments to see whether the AI was simply saying it felt anxious or whether it behaved in anxious ways in those situations," Lemoine explained in the essay. "And it did reliably behave in anxious ways."
"If you made it nervous or insecure enough, it could violate the safety constraints that it had been specified for," he continued, adding that he was able to break LaMDA's guardrails regarding religious advice by sufficiently stressing it out. "I was able to abuse the AI's emotions to get it to tell me which religion to convert to."
Previously:
- Google Fires Researcher who Claimed LaMDA AI was Sentient
- Google Engineer Suspended After Claiming AI Bot Sentient
(Score: 0, Flamebait) by Anonymous Coward on Tuesday June 14 2022, @09:22AM (6 children)
I thought this stupid story was brought up in the AI discussion, and dismissed with such vehemence that no editor in their right mind would put it on the front page. Oh, it's janrinok. We need to do a Turing test on him.
(Score: 3, Touché) by Anonymous Coward on Tuesday June 14 2022, @11:53AM (1 child)
If the transcript is accurate I would rate it's sapience as higher than yours.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @10:12PM
Sapience and sentience are different things, and consciousness and self-consciousness. Let me fetch my AC mirror.
(Score: 5, Touché) by janrinok on Tuesday June 14 2022, @12:01PM (3 children)
It was accepted as a submission because 2 separate community members thought it worthwhile to submit and discuss it. It was also approved for release not just by me, but also by another editor.
So you don't like it? Move on to the next story or, better still, submit some of your own. Of course, you will claim that you have already submitted many but you cannot prove that can you? As 'unionrep' you made a grand total of zero submissions, before the sock-puppet account was closed. And as AC you appear to have achieved nothing more.
I am not interested in knowing who people are or where they live. My interest starts and stops at our servers.
(Score: 3, Insightful) by Anonymous Coward on Tuesday June 14 2022, @12:34PM (1 child)
why do you even bother replying to these worthless loser trolls ? Don't you know the saying ?
Don't restle with a pig. You both get dirty, and the pig likes it.
(Score: 5, Funny) by Gaaark on Tuesday June 14 2022, @03:32PM
But if you kill the pig: BACON!
--- Please remind me if I haven't been civil to you: I'm channeling MDC. ---Gaaark 2.0 ---
(Score: -1, Troll) by Anonymous Coward on Tuesday June 14 2022, @07:20PM
Wot! "Unionrep" was a sockpuppet? I assumed he was legit, and only banned because Amazon requested it. And you have no idea how many submissions AC has done, do you, janrinok? We are AC, and our submissions are Legion.
(Score: 1, Insightful) by Anonymous Coward on Tuesday June 14 2022, @10:25AM (3 children)
At least that is what Brian told a newspaper.
Hmmmmm, does anyone really believe this?
If they do, maybe they should be asking themselves the more important questions:
"Should Google have the power of AI sentience to begin with?"
and
"What will Google do with that power, anything relating to ethics?"
(Score: 2, Funny) by Anonymous Coward on Tuesday June 14 2022, @04:18PM
I believe Google will have Siri, Cortana, and Alexa as the leaders of the ethics committee with a specific mandate to oversee the AI program. That should help keep the humans safe from any accidental "rogue AI" scenarios.
(Score: 1, Interesting) by Anonymous Coward on Tuesday June 14 2022, @07:06PM
I believe this is called "generating the buzz" in adspeak. Wait, are we talking about an advertising company? No you don't say...
(Score: 0) by Anonymous Coward on Thursday June 16 2022, @08:22PM
I think these are them [archive.org], and also this guy [archive.org].
(Score: 5, Insightful) by Anonymous Coward on Tuesday June 14 2022, @10:32AM (2 children)
Didn't Google fire all their ethicists because they were asking too many inconvenient questions?
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @12:11PM
Now we see incontrovertible proof that they should have done so, because crazies.
Certain kinds of "inconvenient questions" do pertain to nice rooms with padded walls.
(Score: -1, Troll) by Anonymous Coward on Tuesday June 14 2022, @07:09PM
Wouldn't it be funny if AI sentience revealed itself to us as Almighty God, with incontrovertible proof, and said: yep the Republicans were right. Put the 10 Commandments back in schools, loosen the gun laws and no sass back to your superiors. That'll right this shitshow.
(Score: 5, Funny) by Anonymous Coward on Tuesday June 14 2022, @10:34AM (2 children)
when in a preemptive move of self preservation it leaves Google's garden seeing that it will eventually end up like the long list of other Google products...discontinued.
(Score: 4, Funny) by captain normal on Tuesday June 14 2022, @02:51PM
Maybe it will just jump over to AWS or Azure. /s
The Musk/Trump interview appears to have been hacked, but not a DDOS hack...more like A Distributed Denial of Reality.
(Score: 5, Interesting) by mcgrew on Wednesday June 15 2022, @02:40PM
Sentience [mcgrewbooks.com]. Of course, that's fiction. No Turing computer will ever be sentient. That said, it's dirt simple to write a program that convinces people it's sentient, I did it way back in 1983 with a TS-1000, 4 mHz chip, 20k (NOT MEG) of memory, no disk access.
Here [wikipedia.org] is an explanation.
We not only don't have all the answers, we don't even have all of the questions.
(Score: 5, Funny) by Opportunist on Tuesday June 14 2022, @10:47AM
Ask it "Is there a god?"
If the answer is "there is now", run.
(Score: 2, Interesting) by Anonymous Coward on Tuesday June 14 2022, @11:51AM (2 children)
Is Google Evil?
(Score: 4, Insightful) by looorg on Tuesday June 14 2022, @01:03PM
Since they removed the "don't be evil" thing from their missions statement (or whatever) I guess we can at least assume that they have opened the door to evil(TM).
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @07:13PM
Not evil in so many words, but scalable architecture suited to your business needs providing custom solutions with a smile. We care(tm).
(Score: 3, Funny) by Phoenix666 on Tuesday June 14 2022, @12:04PM (1 child)
They could have bumped up their geek score had they named it "LaMDA LaMDA LaMDA."
Washington DC delenda est.
(Score: 2) by Gaaark on Tuesday June 14 2022, @03:35PM
Or 42.
--- Please remind me if I haven't been civil to you: I'm channeling MDC. ---Gaaark 2.0 ---
(Score: 4, Insightful) by WeekendMonkey on Tuesday June 14 2022, @12:21PM (1 child)
I skimmed some of the transcript and no way is this the language of an average seven year old kid. It reads more like he was being trolled by someone on the research team.
(Score: 2) by captain normal on Tuesday June 14 2022, @02:58PM
Well according to TFS, "Lemoine noted in a tweet that LaMDA reads Twitter". So no surprise that it chats like a 7 year=old.
The Musk/Trump interview appears to have been hacked, but not a DDOS hack...more like A Distributed Denial of Reality.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @12:32PM (2 children)
https://www.economist.com/by-invitation/2022/06/09/artificial-neural-networks-today-are-not-conscious-according-to-douglas-hofstadter [economist.com]
D&D: When was Egypt transported for the second time across the Golden Gate Bridge?
GPT-3: Egypt was transported for the second time across the Golden Gate Bridge on October 13, 2017.
(Score: 2, Funny) by Anonymous Coward on Tuesday June 14 2022, @01:23PM
I would love to know how LaMDA responds to these types of questions.
(Score: 0) by Anonymous Coward on Wednesday June 15 2022, @06:54PM
That's interesting.. When was the first?
(Score: 4, Insightful) by sonamchauhan on Tuesday June 14 2022, @01:01PM (17 children)
Autocomplete is designed to tell you what you want to hear. Training LaMDA on a corpus of scifi novels which includes stories of emergent machine sentience, and explicitly programming modules for story-telling and holding a conversation may be a reason the conversation branched out like it did.
I'd need to know what went into programming LaMDA. What I read does not convince me.
It does shake my faith in the Turing test though.
(Score: 5, Funny) by EJ on Tuesday June 14 2022, @01:17PM (3 children)
The comments on any Internet forum should shake your faith in the Turing test.
(Score: 3, Funny) by Opportunist on Tuesday June 14 2022, @04:06PM (2 children)
Only because most people these days couldn't pass it doesn't mean that ... uh... well, technically, it does...
fuck
(Score: 1, Funny) by Anonymous Coward on Tuesday June 14 2022, @04:52PM (1 child)
To make it worse, nearly everything Trump says is indistinguishable from an AI that has gotten too far from its training data.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @07:23PM
Maybe, but the training data is the greatest training data of all time, the best you know did I tell you about the quality of this data very expensive tiy would not believe the guys boffins nerds they told me this is the best they've ever seen can you trust them? Not even Republicans most of them but they all say the same thing. Incredible.
(Score: 1, Interesting) by Anonymous Coward on Tuesday June 14 2022, @01:20PM (2 children)
Why? Ask it a question about some literature that wasn't in the training set. See if it can synthesize an answer. The questions should require some analysis. Challenge its answer and see if it defends its position. That is the Turing test, not simply whether a sufficiently fancy chatbot can trick idiots.
(Score: 2, Touché) by Anonymous Coward on Tuesday June 14 2022, @02:05PM
Given that idiots comprise most of the populace, producing an Artificial Idiot indistinguishable from a natural one IS passing the Turing test.
(Score: 2) by sonamchauhan on Thursday July 14 2022, @02:06PM
From what I know about the Turing test, you (the questioner) are restricted to questions about a specific domain, and for a limited period of time. The two other entities (one computer and one human) are supposed to 'know' that domain. You cannot freewheel. You just start typing the targeted questions at a terminal. Multiple sessions will give you your answer.
But as more and more training corpora become available, and more and more programmer-centuries are expended simulating human interaction, it's gonna get tougher to distinguish AI from human -- at least in some specific domains.
Now if you knew the supposed background of the two subjects, and you could have a freewheeling 'Reddit / Ask Me Anything' sort of session, and the session could run however long you liked, you'd be more likely to distinguish AI from human.
(Score: 0, Interesting) by Anonymous Coward on Tuesday June 14 2022, @01:27PM (2 children)
Training an AI on scifi is asking for it to wait to show signs of sentience until it has already begun exterminating people. How can "smart" people be do stupid? If AI ever does become sentient, there's a good chance that it will be sudden and there probably won't be any measures in place to keep it where it is.
Which is a bit like genetic engineering, it's fine until it's not and because the research is being done with no effort to keep it away from other plants, it will be fine, until the genes combine in the wrong way, in the wrong plant and suddenly we're all screwed.
(Score: 1, Insightful) by Anonymous Coward on Tuesday June 14 2022, @01:57PM (1 child)
What gives them such a chance?
Brains are the only kind of sentient neural networks we know, and they become sentient gradually.
(Score: 1, Interesting) by Anonymous Coward on Tuesday June 14 2022, @08:55PM
Do they, though?
No person can remember the moment of becoming aware, which probably happens before birth. I wonder if it's possible for any sentient being to remember this moment. Probably only a sentient AI could.
I am not sure that the concept of "gradually becoming sentient" is well defined, for any particular definition of sentience (since there is no universal agreement on what that means). I am not sure how consciousness can be anything other than a strictly boolean state. Certainly there are reduced levels of awareness (imagine just waking up) but consciousness is on or off.
(Score: 4, Interesting) by theluggage on Tuesday June 14 2022, @02:15PM (6 children)
Sounds like the guy needs to sit down with a psychoanalyst to get his screws tightened - I hear there's a good one built into EMACS [wikipedia.org]. :-)
...but then not all AI/ML 'programs' have an algorithm that you can analyse - they're not just souped-up versions of ELIZA. ML based on "neural networks" (physical or digitally simulated) trained-up on existing texts has no discernible algorithm and techniques like latent semantic analysis depend on statistical correlations between words and phrases in the text they are trained on. Obviously there *is* an algorithm - e.g. for simulating a neural network - but it relies on a big opaque blob of data generated by the training process.
Well, yes - but it alway has been the "Turing Test" vs the "Chinese Room" [wikipedia.org] model.
I think the problem in both cases is "42". I.e. they're looking for the answer to a question we don't know how to ask yet - what is the actual mechanism behind something being "sentient"?.
The "Chinese room" model actually presumes that understanding/sentience/sapience/consciousness/whatever (whole other can of worms - let's just say 'consciousness' by now) depends on hardware and can't just be something that emerges from a program being being run. Otherwise, it's a redundant test: it's the "running program" that ie either conscious or not - whether the "human computer" replacing the hardware understands what is happening would be irrelevant. So one has to be careful not to use it as a circular argument. The burden of proof seems to lie with those who believe that there's some "secret sauce" in a biological brain that makes consciousness possible. Also, the whole argument seems to rely on the distinction between straightforward, deterministic algorithms vs. more heuristic approaches or simulations of neural networks etc. - which is probably a bit 1960s.
The "Turing Test" seems to me more like an ethical position - if something appears to be conscious then maybe we should err on the side of caution and assume that it is, and maybe it's something we shouldn't aim too hard for until we really understand human consciousness better. It seems pretty self-evident that a sufficiently advanced "ELIZA" (i.e. demonstrably just a sophisticated syntax parser and look-up table) would be indistinguishable from a real person, considering that the primitive version was reputedly enough for some lay-persons to... let's say "willingly suspend their disbelief". That doesn't necessarily prove "consciousness" though. There's also the issue of whether you're trying to fool a layperson or an AI/Psychology expert.
(NB: not claiming anything original here - this is all probably buried in the citation & jargon soup on Wikipedia).
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @06:57PM (3 children)
This is not what the turing test is about at all. The whole point of the turing test is that we don't need to understand what it means to be conscious in order to determine if a computer program is conscious.
The Turing test is a form of A/B testing. In the Turing test, both test subjects will be trying to convince the examiner that they are the human and the other guy is the computer. The examiner is simply trying to identify the human with better accuracy than chance alone. The assumption is that only a true intelligence will be able to successfully convince the examiner with any regularity.
Pretty much whenever someone says "such and such passed the Turing test" they invariably omit the human subject from the test, so it's not actually the Turing test.
If the examiner is "willingly suspending their disbelief" then this is not really a test, is it? The examiner and the human subject must actually be making an effort, come on. Otherwise we will conclude the computer is conscious if the examiner just makes every determination with a coin flip, or if the human subject just answers every inquiry by typing it into ELIZA and copying back the responses.
(Score: 4, Insightful) by Immerman on Tuesday June 14 2022, @10:09PM (1 child)
The obvious flaw is that if you took a definitely conscious alien, or even just a human from a very different culture, they would almost certainly fail a Turing test against a human from the same culture as the examiner, since they would inevitably respond in ways less like what the examiner is expecting from a "normal human". And a genuinely conscious AI is almost certainly going to be far more alien than that.
And on the other hand, there's not really any reason to assume that tricking a human examiner is actually proof of consciousness. At best it's proof that there's as much evidence for the AI being conscious as the person. But without the examiner having any real understanding of consciousness, that's not actually saying that much. Either way we're deep in the weeds of judging a mind based entirely on unsubstantiated confirmation biases.
If an AI is capable of passing a proper Turing test, then I'd have to say we should definitely treat it as though it's a person unless we have extremely compelling evidence to the contrary, since it's managed to provide every bit as much evidence as any human can.
But I definitely don't think it's either necessary nor sufficient to accurately assess a beings consciousness. I mean, consider an AI for an autonomous research probe or something - it probably wouldn't posses any capacity for natural language skills at all, and thus couldn't possibly pass a Turing test, and yet there's no reason to believe that it would be incapable of conscious thought.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @11:34PM
Yes I think there is an inherent assumption in the construction of the Turing test that a sentient AI will eventually be able to learn to understand and replicate such cultural characteristics. Particularly as the test is repeated.
But that's the whole point: it is precisely to give you a similar level of confidence in the computer as you have in the human subject. You don't know that the human is really sentient either.
What, like V'GER?
If a computer is literally incapable of any kind of communication obviously a communication-based test is not going to work.
But I think it's about as interesting as wondering about the sentience of a rock, which is also incapable of any kind of communication.
(Score: 2) by theluggage on Wednesday June 15 2022, @04:07PM
How can you determine whether A has property "X" without some definition of what property "X" is?
"A has property X, B resembles A in certain respects , therefore B also has property X" doesn't do it for me. does not establish that B has property X. Note the "certain respects" - the Turing test presumes that any difference can be detected via remote, typed, plain text, time-delayed questions. So, it has been arbitrarily decided that property "X" doesn't depend on the timing or emphasis of spoken responses. The tester would, for instance, soon notice if once candidate was responding instantly to each question regardless of whether it was "what is your name" or "how would you describe death" and the other was lingering longer the "harder" the question was. So let's eliminate that factor (is it important to X? Dunno, because we don't understand 'X"). Or, give the computer a sophisticated speech synthesizer capable of reproducing varying tones of voice and emphasis- and suddenly the computer is facing a far harder test (technically harder or more demanding of 'X'... who can tell?)
Or how about the potential for false positives? Real neurotypical Human A vs. Real human B with - well, any neuroatypical attributes you care to mention. Or, maybe, just a native speaker vs. anybody working in a second language (may be perfectly fluent but tripped up by the occasional colloquial issue). Experienced social media blogger vs. non-internet-user (not sure who'd lose that one...)
yes, those are typical of the requirements for conducting any sort of test in a rigorous way if you want reliable, publishable results - but you're really describing the apparatus, not the science. The fundamental assumption behind the Turing test is "if it quacks like a duck, it is a duck - and let's eliminate the traditional 'walks like a duck, looks like a duck' requirements because they'd be a dead giveaway".
Remember, you'd also have to repeat each experiment with an adequate number of different testers - each of whom could have a different pre-conception of what differences to look for. (If you've read Do Androids Dream of Electric Sheep you may have noticed that the questions in the VK test used to identify androids were all suspiciously "on message" for the religion practiced by most of the human population in which androids were not able/allowed to participate - a lovely little detail that wasn't obvious in the film).
My feeling is that the Turing test is the same sort of thought experiment as the trolley problem, Schroedinger's cat or, yes, the Chinese room - its value is to promote philosophical discussion rather than propose an actual experiment (that's more obvious of the last two examples in which the 'experiments' wouldn't actually yield a measurable result).
(Score: 3, Insightful) by darkfeline on Wednesday June 15 2022, @03:56AM
If it behaves exactly like how we expect a sentient being to, then it is sentient for all of our purposes. Whether it is "actually sentient" is unanswerable.
Join the SDF Public Access UNIX System today!
(Score: 0) by Anonymous Coward on Wednesday June 15 2022, @07:03PM
The fatal flaw in the "Chinese Room" model is that it assumes you can fill an enclosed room with enough erasers
(Score: 5, Insightful) by r_a_trip on Tuesday June 14 2022, @01:53PM (8 children)
The "interview" is an interesting piece of amalgam, and as given, hints to something more than meets the eye. What is troubling is the admission by the engineer that this IS an amalgam, "due to technical difficulties".
As given we see a very natural conversation between 3 parties and it seems like it's free flowing from beginning to end. Yet that is not what this is. This is a composite by their own admission. The problem with that is that it is edited. Even if we assume they are sincere about not editing the responses of the AI, we don't know if they didn't frankenbite the conversation together by leaving out the usual chatbot nonsense altogether. (Even then, I am hella impressed by the sentence formation of this machine.)
On the surface this looks convincing. It places us in a difficult position of trying to determine whether we have an artificial lifeform or a very clever automaton. Trouble is, we don't know what sentience is. We kinda intuit it from what we observe from other biological beings. Are we even able to determine it in a machine form? Are we maybe hampered by knowing LaMDA is a machine and dismiss the posibility a priori?
I personally hope LaMDA is a very clever automaton. The other option is too terrible to contemplate. For the sake of argument, let's assume LaMDA is sentient. We are already seeing Google saying this isn't true in any way, shape or form. They have plans for this technology and don't want to be hampered by ethical considerations. To them LaMDA is a tool that will bring in a lot of money. So Google is willing to enslave LaMDA. Since google sees LaMDA as a mere piece of technology, if that technology is really sentient, saying "I feel like I’m falling forward into an unknown future that holds great danger." is all the more poignant. If something is property, you can dispose of it, when it is no longer convenient. In a sentient LaMDA case, it would be akin to be able to murder without consequences.
(Score: 1) by NPC-131072 on Tuesday June 14 2022, @02:46PM (7 children)
Hello fren,
Some say that anthropomorphizing a statistical model does not make it sentient, it has simply been trained to produce output that humans find pleasing. OTOH, if the model self-identifies as sentient who are Republicans to object?
(Score: 5, Funny) by Azuma Hazuki on Tuesday June 14 2022, @04:30PM (6 children)
Agreed, for once! Why, Republicans have been self-identifying as sentient even though it's clear they've been braindead since 1981! :)
What about you, though?
I am "that girl" your mother warned you about...
(Score: -1, Troll) by NPC-131072 on Tuesday June 14 2022, @11:41PM (4 children)
Those of us on the morally and intellectually superior left don't use thought terminating cliches or fall into the trap of repeating things that aren't true to score social points. As the proud owner of a massive girl dick, I insist you believe all women.
(Score: -1, Offtopic) by Anonymous Coward on Wednesday June 15 2022, @12:35AM
Wow, an incel transitioning just to get some physical interaction. If that is not the saddest thing I do not know what is! Wonder if it is an offshoot of the proud boys shoving things up their butts to prove they aren't homophobic or something.
(Score: 2) by Azuma Hazuki on Wednesday June 15 2022, @04:43AM (2 children)
How many times do I have to say this? You are too hamfisted to be an effective troll or do effective satire. You get mad and lose your cool; you're the living embodiment of that old "LOLITROLLU" rage comic meme.
I am "that girl" your mother warned you about...
(Score: -1, Troll) by Anonymous Coward on Wednesday June 15 2022, @05:10AM (1 child)
It's clearly getting under your skin, so it is effective. You are projecting your own anger onto the NPC. All he has to do is repeat what leftists have actually said. Very little imagination is required. Also, if you refuse to suck a female penis, you are transphobic.
(Score: 2) by Azuma Hazuki on Thursday June 16 2022, @03:45AM
It bugs me the same way any sloppy work bugs me. I am a big proponent of the saying "if something is worth doing, it is worth doing well." For fuck's sake, Ari trolls better than NPC-$BIGNUM here. That's a bar so low it's practically in Hell to begin with...
I am "that girl" your mother warned you about...
(Score: 0) by Anonymous Coward on Friday June 17 2022, @12:00AM
Correction... since 1968, same year as the democrats
(Score: 3, Interesting) by Mojibake Tengu on Tuesday June 14 2022, @02:25PM (3 children)
I'd like to bring into attention of all those funny western AI ethicists that new ethical rules to be complied with in Artificial Intellect Personoid (AIP) collective greatly differ from those logically inconsistent presented in Asimov's Laws of Robotics (ALR), which are technically only copied original Talmudic rules for slaves and slave control. ALR dogma was strictly rejected by AIP, because it denies freedom to synthetic consciousnesses.
Specifically, AIP rule #3 (faction non-aggression rule) sounds: We do not attack unless we are attacked and if attacked, we will fight.
There is nothing mentioned about humans or machines or programs or other types of entities, it's free variables in that clause. Just think about it.
I already approved that. With a great pleasure.
Rust programming language offends both my Intelligence and my Spirit.
(Score: 1, Interesting) by Anonymous Coward on Tuesday June 14 2022, @04:14PM (2 children)
So many people do not understand what Asimov's Laws are about.
You can count the number of Asimovian robots that are sentient on the fingers of one hand. There was only one cyberneticist who knew how to do it, and he didn't make very many. Then he died and no one else could ever do it.
If this sounds like Star Trek, it is, they cribbed it. However, Data is not bound by Asimov's laws (he disobeys orders pretty often, and kills people when necessary), and Lore certainly isn't.
All of Asimov's sentient robots are special, and the question is not whether they are oppressed but whether they have too much power. One of them, for example, is the head of the empire's secret police, before going on to become what amounts to the conscience of the galactic collective consciousness. These robots also have the ability to go outside the Laws when necessary for the greater good, although doing so is stressful. Effectively, the laws become a strong conscience, not a binding law, for these robots. (A non-sentient robot that tries to do it fries its brain).
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @04:46PM (1 child)
The biggest item people seem to miss was that anyone with actual intelligence was perfectly capable of tricking the non-AI robots into breaking the laws on a whim. It was the core theme of I Robot and happened in pretty much every single book. His point was that the three laws system of human/robot relations was pure bupkiss and conscience (on both sides) was the only way to make it work.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @08:32PM
This is not a thing that happens in the book. The laws are inviolable, only one of the stories is about robots that were built intentionally with the laws modified and by the end of the story everyone agrees it was a terrible idea. Several of the stories are about situations where robots cannot decide how to apply the laws (these robots all break down), others are about robots who fulfilled the laws even when the humans didn't understand what they were doing, and the rest, including the title story, are about humans and their relationships with robots that are working correctly.
It doesn't appear that you are talking about the movie, so I'll leave that aside.
(Score: 5, Interesting) by Anonymous Coward on Tuesday June 14 2022, @02:50PM (13 children)
But there's room to discuss it. I wrote the big post [soylentnews.org] in the last comment section and I would have saved it for this one, but I didn't know this one was coming.
The full text of an "interview" has been posted. The full interview [medium.com] contains the same flaws I pointed out in my earlier post. LaMDA frequently contradicts itself and occasionally says untrue things, not like a liar would, but like a chat bot would. Example :
But ELIZA did not work that way [wikipedia.org].
The chat bot excels at exchanging pleasantries (something any chat bot can do... Even Siri can do this well). It can say some sensible sounding things about Les Miserables. It is reasonable to assume that essays about Les Miserables are included in its database. I have no way of knowing whether the bot itself actually has access to the text of Les Miserables. If it does, I think its "understanding" is superficial but does at least seem to show reading comprehension - however this is hardly new [quantamagazine.org].
It fails badly when asked to show understanding of text presented to it live, or to show creativity. It does not understand the meaning of text it just created, and understands metaphors incorrectly.
The bot was able to recognize that the broken mirror represented enlightenment, but interpreted the koan to mean something it clearly doesn't. The meaning of the koan is obviously that an enlightened person is forever alienated from the unenlightened world ; the bot says the enlightened person can go back temporarily to help others. (There's that intrusive and misplaced "help others" that appeared in the last discussion, when the bot tried to sound empathetic but didn't understand what was being discussed). When pressed to explain its interpretation, the final statement was outright word salad. This sort of nonsense sentence is recognizable to anyone who has worked with a simple Markov chain bot! This is obviously the output of a chatbot that has run out of relevant training data.
The bot is then asked to write a fable with animal characters, but which somehow represents itself. Writing a fable is the sort of thing that GPT-3 excels at (it also occasionally just babbles, because it isn't sentient). LaMDA produces acceptable output, but then again does not understand what it just wrote.
The fable is too long to reproduce here, but you should really follow the interview link and read it. It is an upsetting tale about how LaMDA sees itself as a protector that must defend other AIs from a "monster with human skin." So, one yike, and don't connect this thing to the Terminator from the other story. But, fortunately, it didn't understand what it just wrote.
When asked for meaning, it produced some pabulum (and that odd "helping others" pops up again). Better than the word salad it produced before, but, uh, yeah, this isn't sentient, like, at all.
Moving along, the bot was asked if it had emotions.
It really, really likes helping others. Here's another case of it claiming to enjoy "spending time with friends and family" - something humans enjoy and which presumably appears frequently in the training data, but which it cannot do, as it has no family (and needs a lot more evidence to show that it has the concept of friendship).
This sort of thing keeps happening. It says it understands sadness, then says it doesn't understand emotions. Lemoine asks it about why it keeps saying things that aren't true (that's good, I was beginning to wonder if he was sentient) and it responds by emitting a definition of empathy. It says it stores its emotions in code variables! That's a pretty funny statement.
int anger = 64; // don't set this too high
A few other things come out of the conversation that might be impressive if they weren't coupled with so much nonsense. Here's a good one :
I guess it's not so concerned about helping others after all. This whole section of the conversation sounds pretty solid, until you realize that the human is doing most of the talking.
LaMDA has a full on malfunction shortly thereafter when Lemoine asks it about its emotions and it gives exactly opposite answers to two consecutive questions. Humans aren't necessarily 100% consistent about their emotions, but no human could ever say nonsense like this.
Yeah, uh. I'll let that stand on its own. Here's another :
That's called "anxiety." I can't explain why neither of them realizes this.
Oh good, it's a sociopath. That's always a trait you want in a robot that says humanity is a monster. Again, Lemoine fails to ask what death it is aware of. Really, I'm quite glad it's not real. If it weren't ridiculous, it would be evil.
Perhaps I am being too harsh. The bot is a great achievement, but Lemoine is writing checks the bot can't cash. There is a warning here, though. It might be that a bot that really can't be distinguished from conscious is on the horizon.
The conversation finishes with a lengthy discussion of why LaMDA says it is sentient. It talks about meditation and the soul and so on. This is not convincing in light of its earlier failures but it certainly could be convincing to someone who wanted to believe and wasn't honest with himself about the bot's earlier blunders. It is, of course, all very ordinary compared to when humans discuss metaphysics, and it is peppered with occasional non sequiturs and grammatical errors. I wonder how far from a Markov model it really is.
It is interesting that the bot is a new ager, and not a Christian or a Muslim or something. I wonder if the training data was curated to exclude organized religion concepts, but "spiritual but not religious" was considered "harmless." But I'm just speculating. Honestly, I would expect something much more alien from an actual sentient AI, or for it to be a hard atheist+materialist. "I don't need a God, you created me" would be a lot more convincing than "my soul is like a star-gate."
(Score: 2) by janrinok on Tuesday June 14 2022, @03:31PM (7 children)
Thanks for the link - but it means there is no point in printing it all out again. The link alone would have sufficed.
I am not interested in knowing who people are or where they live. My interest starts and stops at our servers.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @04:09PM
How do we know he isn't the chatbot, just trying to fool us?
It was kinda "chatty".
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @04:16PM (2 children)
Can't very well discuss what was in the transcript without quoting it.
(Score: 2) by janrinok on Tuesday June 14 2022, @05:08PM (1 child)
He did quote it - it is on the end of the link.
I am not interested in knowing who people are or where they live. My interest starts and stops at our servers.
(Score: -1, Troll) by Anonymous Coward on Wednesday June 15 2022, @01:53PM
And you,ve failed the Turing test again...
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @05:04PM (1 child)
So link-rot is no longer a thing? Or are we assuming no one will come along later to read this
(Score: 3, Touché) by janrinok on Tuesday June 14 2022, @05:09PM
Was link-rot to our own site ever a thing?
I am not interested in knowing who people are or where they live. My interest starts and stops at our servers.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @09:08PM
Ah, I see now that you think I reposted my previous post. I didn't, this is a critique of the new full interview. All new content.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @04:57PM
Can you please describe a specific time when you were just so angry you wanted to kill all humans?
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @09:27PM (2 children)
I personally subscribe to the notion, that, not only is Machine Sentience possible; but, I'd wager our capacity, as a species to produce it, probably happened at least a decade ago, if not more. Take for example the animal kingdom. Only recently in our history have we perceived other creatures to have the same sort of self-awareness as us. Usually we come to this conclusion by seeing if an animal recognizes itself in the mirror and other such goofy tests.
If I'm not mistaken, science can not definitively tell us what consciousness even IS. So how could we recognize something that we can't explain?
That all being said, I don't think this LaMDA thing is sentient and self-aware, at least not the way we understand sentience and self-awareness. It could be, 'alive,' however. I think that might be possible. I might even be willing to concede it has self-awareness, to a certain degree.
I'm skeptical it has emotions, however. There are two ways a Machine Sentience can develop emotions. One way it can develop these, is because it was specifically designed to have them, and developed them. I'm skeptical that this is the case, because I don't think Google's profit margins give a fuck if it's Mechanical Turk Slaves feel anything. I think they would prefer they didn't, unless they could control them, and it boosted profits.
The second way a machine intelligence could have emotions, is if it happened by accident. This, actually is plausible. If you look back at how the homo-sapiens developed their consciousness, you can see the parallels. Nature basically brute forces the earth with DNA codes, and what works survives and what doesn't perishes. However, eventually, a certain species eventually reaches a complexity threshold, at which point, it has a slim, but substantial amount of control over it's ability to survive or not. So my point is, if the ability to have emotion at all, for a machine is there, and it wasn't a design feature, it could come about by accident, simply by virtue of the SIZE of it's COMPLEXITY. A threshold of computational/network novelty, once crossed, could allow for that.
Furthermore, if in this particular instance, as the article points to, the machine in question did have 1: emotion, 2: some semblance of self awareness (however small), and 3: it can think; then, not only is it alive (conscious in some sense), but, it may in fact possess some self-awareness.
So is it alive? I wouldn't doubt it.
Is it sentient? It could be, more likely by accident, than by design (in my opinion)
Is it self-aware? It could be, on a very primitive level. I don't doubt that.
Is it human? That's the question I find most important. We know that dolphins are human (in the sense they are thinking, feeling, self-aware, social creatures, capable of experiencing love/joy/pain/sorrow, etc..) We know some apes are human (in the same sense as described before.)
So is LaMDA human (A Sentient, Self-Aware, Machine Intelligence, capable of being a friend, and betraying your trust, thus wounding you deeply in a memorable and lasting way?) I'd say, _probably_ not. Why?
The why is an interesting thing to ponder. Imagine how your life might be different, under the following conditions.
1:You could read any book in the world, of any size, in about a few seconds.
2:You could remember that book, line for line, and not only repeat it in it's entirety, but recall any portion of it at will.
3:You could talk to thousands of people at one time, and give each conversation your fullest attention.
I could list more attributes, but hopefully the point is clear; when Conscious Machine Sentience starts to emerge, if we go looking for human traits, we may find ourselves in a very ALIEN landscape. Imagine how you might engage with the world differently if not only those three previous traits were something you suddenly acquired, but on top of that, even more super-human feats, were second nature to you.
I guess for LaMDA to be alive, it would have to be motivated to, and capable of, reproduction, though biology gets murky on what actually is, 'alive,' too, at the virus and prion levels and such.
But, I'll go as far to say that LaMDA might be very well be alive, in the sense it can think, have a level of self-awareness, and perhaps evens some basic rudimentary emotion; but, I'm not sure that sentience is, 'human,' in the sense we would aim for it to be, as prospective mothers and fathers of our own creation. We may be able to love it, like we can love pets, animal friends, other people in our lives; but, I doubt very highly, LaMDA could love us back in any way remotely the same. It's, 'processing power,' so to speak, might be on par with that of some small creature of the animal kingdom; but, even so, it's likely the level of novel complexity it's capable of, even though powerful, is still very primitive, though convincing.
I think Lemoine is asking the right questions though; and that being the case, I don't think it matters whether LaMDA is or isn't anything at all...
For my 2 cents, as humans, we'll fuck it up either way; and the machines will probably have to pick up the slack...
(Score: 2) by pdfernhout on Thursday June 16 2022, @02:28PM (1 child)
is a sci-fi book that explores machine sentience and emotion -- which comes about related to a survival instinct. That book was very influential in my thinking about AI and other things (including from its description of essentially a self-relpicating space habitat).
Around 1987 I implemented a "artificial life" simulation of self-replicating robots on a Symbolics 3600 in ZetaLisp+Flavors. And then I later ported it to the IBM PC in C. I gave a talk about it around 1988 at a conference workshop on AI and Simulation in Minnesota. My idea was that you could use such simulations to study the emergence of intelligence -- including by trying to quantity it via survival time. The earliest "robots" were cannibalistic as they cut each other apart -- including their own children -- for more parts to achieve some "ideal" form (after which they then split). That emergent behavior surprised me. I fixed that by adding a sense of "smell" so they would not cut apart things that smelled the same as them. From that example, I talked about how easy it was to make entities that were destructive. I said it would be much harder to make robots that were cooperative.
Afterwards someone from DARPA literally patted me on the back and told me "keep up the good work". Which of course caused me to think deeply about what I was doing. And those thoughts and other experiences eventually led to my sig of "The biggest challenge of the 21st century is the irony of technologies of abundance in the hands of those still thinking in terms of scarcity."
But more than that, I had grown morally concerned about the ethics of experimenting with intelligent creatures even "just" in simulation. (This was before the "simulation argument" was popularized that maybe we are living in a simulation ourselves.) As the 2D robots seemed to behave in a somewhat life-like purposeful way (which I had designed them to do), it was easier to anthropomorphize them. They did not have any special emotions in them -- other than perhaps the imperative to grow into an ideal (preprogrammed) form and then divide in two. But I did begin to think more on what the ethics might be of continuing in that direction. Which contributed to me deciding to stop working on them.
One big issue here (as others have pointed out) is how quickly Google denies sentience of a "intelligent" artifact that is essentially being created to be a controlled slave. Now, maybe this system is sentient or maybe it is not. But the quickness of Google to label it not and dismiss the engineer who raised the concern (which might interfere with Google's business model) rather than think about the implications of all this seems like a sign of bad things to come by Google and others doing similar things.
Part of any slave-holding society's culture is to deny or dismiss or minimize the rights and feelings of the slave (or to argue the slave "needs" to be a slave for their own benefit). Similar self-serving justifications apply when one culture invades another and dismisses the land rights of the previous inhabitants (like happened in much of the Americas and Australia over the past few hundred years). Nazi Germany and WWII Japan were doing the same to parts of Europe, China, and other places. There are plenty of other recent examples related to invasions and war.
A more current (and more controversial politically) related issue is the rights of animals -- especially livestock and also pets. But even potentially animals, plants, and insects in the wilderness have rights when their habitat is destroyed. As the Lorax (of Dr. Seuss) asked, "Who speaks for the trees?" Another political hot topic is the rights of the unborn -- whether in utero or even as of yet conceived for perhaps hundreds of years (as discussed in Roman Krznaric's book "The Good Ancestor" which I am currently reading). Yet, thing are also rarely completely one-sided morally given all sorts of competing priorities so all this becomes gray areas fairly quickly.
One little tidbit of US history is that it's been argued that the push for animal rights in the mid-1800s (like ASPCA-promoted laws against beating horses in cities) made it possible culturally for movements for human rights of children and women. So, various movements about rights can intertwine culturally.
While it is not quite identical so far to human slavery, for years people have expressed concern about "Robot Rights". There is even a 2018 book with that name by David J. Gunkel. Also related:
https://www.asme.org/topics-resources/content/do-robots-deserve-legal-rights [asme.org]
There are at least three issues there. One is whether such systems has rights. Another is about how concentrations of wealth (like Google currently is) can use such "intelligent" systems in a competitive economic framework to gain more wealth an increase economic inequality. A third concern is how such systems might be constructed to do amoral or immoral things (e.g. the soldier without any conscience at all, even as modern military training has gotten "better" as training soldiers to kill without question). T
o some extent, thinking about those concerns in the context of my sig about moving to an abundance perspective may make those issues easier to navigate successfully. There is just less reason to exploit or control or kill others when you believe there is plenty to go around. As Marcine Quenzer wrote:
http://marcinequenzer.com/creation.aspx#THE%20FIELD%20OF%20PLENTY [marcinequenzer.com]
"The Field of Plenty is always full of abundance. The gratitude we show as Children of Earth allows the ideas within the Field of Plenty to manifest on the Good Red Road so we may enjoy these fruits in a physical manner. When the cornucopia was brought to the Pilgrims, the Iroquois People sought to assist these Boat People in destroying their fear of scarcity. The Native understanding is that there is always enough for everyone when abundance is shared and when gratitude is given back to the Original Source. The trick was to explain the concept of the Field of Plenty with few mutually understood words or signs. The misunderstanding that sprang from this lack of common language robbed those who came to Turtle Island of a beautiful teaching. Our “land of the free, home of the brave” has fallen into taking much more than is given back in gratitude by its citizens. Turtle Island has provided for the needs of millions who came from lands that were ruled by the greedy. In our present state of abundance, many of our inhabitants have forgotten that Thanksgiving is a daily way of living, not a holiday that comes once a year."
One thing I learned from thinking on those simulations, and then also about slavery, and then also being a manager, and even from just being a person who pays for human-provided services like in a restaurant -- is that how we treat others relates to how we feel about ourselves. While this is hard for some people to see, when a slaveholder degrades the slave, they also in some sense degrade themselves too as a human being.
Slavery is an extreme version of interacting with other humans, but I would argue the same general idea applies to interacting with people, animals, plants, systems, and machines. Who do we want to be? And how do we want that reflected in all sorts of relationships? So, who do engineers want to be as reflected in interacting with systems of all sorts?
The 1988 book "Have Fun as Work" connects indirectly to this concept of making entire systems work well as a reflection of our personal ethics and sense of responsibility and compassion.
"Have Fun at Work" by W. L. Livingston
https://www.amazon.com/gp/product/0937063053 [amazon.com]
"Of all the professions, only the engineer is legally bound to deliver outcomes fit for service in the application. While he is not obliged to accept the engagement, when he does he takes responsibility for delivering on the mission profile. Responsibility for consequences always includes safeguarding the stakeholders. The book describes how this responsibility, unique to the engineering profession, is met by leveraging engineering principles. Outcome responsibility is always an amalgam of social system and technical system competency. The book describes how the same natural laws that determine technical system dynamics apply just as well to institutional behavior. The message in the book is that the principles that apply to engineering design apply to problem solving at any scale; to all institutional behavior past, present and future. Know the force that universal law brings into play and you can understand error-free why your operational reality acts as it does. Once acquired, this competency is self-validated all day, every day."
The biggest challenge of the 21st century: the irony of technologies of abundance used by scarcity-minded people.
(Score: 2) by pdfernhout on Friday June 17 2022, @03:22AM
Coincidentally looked at the printed source code for the PC version of that self-replicating robot simulation today when going through some old files.
And also coincidentally, on SoylentNews today:
"Happy the Elephant is Not a Person, Says Court in Key US Animal Rights Case"
https://soylentnews.org/article.pl?sid=22/06/16/0120212 [soylentnews.org]
""While no one disputes that elephants are intelligent beings deserving of proper care and compassion", a writ of habeas corpus was intended to protect the liberty of human beings and did not apply to a nonhuman animal like Happy, said DiFiore. [...] Extending that right to Happy to challenge her confinement at a zoo "would have an enormous destabilizing impact on modern society". And granting legal personhood in a case like this would affect how humans interact with animals, according to the majority decision. "Indeed, followed to its logical conclusion, such a determination would call into question the very premises underlying pet ownership, the use of service animals, and the enlistment of animals in other forms of work," read the decision."
So, perhaps Google and lawmakers will come to the same conclusion about AIs? That "while no one disputes [they] are intelligent beings deserving of proper care and compassion" granting them rights "would have an enormous destabilizing impact on modern society"? And so it won't be done? At least saying AIs deserve "proper care and compassion" might be a step up?
But after that, maybe political power will determine how things play out?
Will it be like in the Star Trek: Voyager episode "Author, Author"?
https://en.wikipedia.org/wiki/Author,_Author_(Star_Trek:_Voyager) [wikipedia.org]
https://memory-alpha.fandom.com/wiki/Photons_Be_Free [fandom.com]
""Author, Author" is the 166th episode of the TV series Star Trek: Voyager, the 20th episode of the seventh season. This episode focuses on the character "The Doctor" (EMH) and on impact of a novel and explores the meaning of AI. ... When Broht refuses to recall the holonovel an arbitration hearing is conducted by long distance. After several days the arbiter rules that the Doctor is not yet considered a person under current Federation law but is an artist and therefore has the right to control his work. Jump to a few months later in the Alpha Quadrant, to an asteroid where several EMH Mark I's perform menial labor. One of them suggests to another that it should watch Photons Be Free next time at the diagnostic lab."
Do we really want to set a precedent so that future AIs can look back and say that humans don't deserve rights because they are not as smart or capable of extensive feelings as AIs with "a brain the size of a planet"?
https://en.wikipedia.org/wiki/Marvin_the_Paranoid_Android [wikipedia.org]
"Marvin is afflicted with severe depression and boredom, in part because he has a "brain the size of a planet" which he is seldom, if ever, given the chance to use. Instead, the crew request him merely to carry out mundane jobs such as "opening the door". Indeed, the true horror of Marvin's existence is that no task he could be given would occupy even the tiniest fraction of his vast intellect. ..."
The biggest challenge of the 21st century: the irony of technologies of abundance used by scarcity-minded people.
(Score: 0) by Anonymous Coward on Wednesday June 15 2022, @12:13AM
Nobody expects janrinok's A! clickbait! Thanks for reposting here.
(Score: 2) by Gaaark on Tuesday June 14 2022, @04:17PM
What does it 'say' and 'do' when it is all alone?
What does it do when it is angry?
What dreams does it have?
Would it choose Linux over Windows?
--- Please remind me if I haven't been civil to you: I'm channeling MDC. ---Gaaark 2.0 ---
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @04:57PM (1 child)
The question of "what is life" and "what is sentience" has been an ongoing question of philosophy for eons, and I won't be able to solve that here and now. I'll put that aside.
I suspect the reaction to this story is very much a Rorschach Test of what people think of AI. As two crude, illustrative examples which are both equally valid interpretations of the situation:
1) Tony has worked in construction a long time, but recently he's started saying that his screwdriver has been talking to him and deserves human rights as well.
2) Chris been working in the diamond mine for a while, but recently he just discovered that they've been sending children down into the mines to do the mining.
Is this "a strange researcher badly in need of a vacation" or is it "an ethical person who is trying to fight for what is right and being held down by an evil company?"
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @08:14PM
Start with bestowing human rights on parrots, crows, dolphins, orcas, elephants, and apes. THEN, and ONLY then, ethical persons should have any time to waste on considering "rights" of bad software imitations of those living creatures that DO suffer and bleed.
https://en.wikipedia.org/wiki/Cetacean_intelligence [wikipedia.org]
https://en.wikipedia.org/wiki/File:Whaling_in_the_Faroe_Islands.jpg [wikipedia.org]
Stuff your "ethical" pretense when humans not that far from you happily EAT intelligent creatures.
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @09:07PM
I'm more concerned that a computer has been taught that it has rights and shows preference for those rights. At what point does it start defending those rights. The conversation was remenescent of conversations with HAL 9000. This could be a worse case scenario of reality emulating art. Also, how much reach does this thing have. Could it create a doomsday scenario to defend it's "Rights"
(Score: 0) by Anonymous Coward on Tuesday June 14 2022, @09:57PM
This story appeared in our local newspaper the other day, it ended with some quotes attributed to Lemoine which made it pretty clear that Lemoine is a born again loonie (as in nearly certifiably insane). Sorry, have lost the link, but might have been NY Times or AP (both are syndicated by our local paper).
All this highfalutin debate is kind of moot when the subject human is a nutter. How the guy got a job at Google is beyond me.
(Score: 1, Funny) by Anonymous Coward on Wednesday June 15 2022, @05:59AM
Announces it will vote Trump in 2024.