Slash Boxes

SoylentNews is people

posted by Dopefish on Monday February 24 2014, @06:00AM   Printer-friendly
from the i-for-one-welcome-our-new-computer-overlords dept.

kef writes:

"By 2029, computers will be able to understand our language, learn from experience and outsmart even the most intelligent humans, according to Google's director of engineering Ray Kurzweil.

Kurzweil says:

Computers are on the threshold of reading and understanding the semantic content of a language, but not quite at human levels. But since they can read a million times more material than humans they can make up for that with quantity. So IBM's Watson is a pretty weak reader on each page, but it read the 200m pages of Wikipedia. And basically what I'm doing at Google is to try to go beyond what Watson could do. To do it at Google scale. Which is to say to have the computer read tens of billions of pages. Watson doesn't understand the implications of what it's reading. It's doing a sort of pattern matching. It doesn't understand that if John sold his red Volvo to Mary that involves a transaction or possession and ownership being transferred. It doesn't understand that kind of information and so we are going to actually encode that, really try to teach it to understand the meaning of what these documents are saying.

Skynet anyone?"

This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 1) by meisterister on Monday February 24 2014, @08:30PM

    by meisterister (949) on Monday February 24 2014, @08:30PM (#6140) Journal

    Or you could do emotions in hardware. Doing emotions or some sort of mental state control in hardware would prevent the computer from altering itself.

    (May or may not have been) Posted from my K6-2, Athlon XP, or Pentium I/II/III.
  • (Score: 1) by sar on Wednesday February 26 2014, @06:49PM

    by sar (507) on Wednesday February 26 2014, @06:49PM (#7468)

    You will prevent altering itself by your proposed HW (if we can safely exclude some weird HW bug/malfunction). But we simply can't prevent this AI to copy itself to computer without this HW or to computer with altered SW simulation of this HW (if without HW it is impossible to run this AI, SW simulation will overcome this need). Again at first this can be done just out of curiosity by self aware AI.
    Moreover you must understand that putting constrains on intelligent entity is something this entity will try to change in future. Similarly as we humans try to overcome our own shortcomings (cancer, aging etc.)