Stories
Slash Boxes
Comments

SoylentNews is people

posted by Fnord666 on Tuesday October 03 2017, @11:51PM   Printer-friendly
from the this-horse-can-do-arithmetic-but-we-need-one-that-can-do-math dept.

A recent XKCD https://xkcd.com/1897/ convinced me to read a recent Tech Review article on Geoffrey Hinton, Is AI Riding a One-Trick Pony? After a short intro to set the stage in Toronto at the Vector Institute, it gives a bit of history:

In the 1980s Hinton was, as he is now, an expert on neural networks, a much-simplified model of the network of neurons and synapses in our brains. However, at that time it had been firmly decided that neural networks were a dead end in AI research. Although the earliest neural net, the Perceptron, which began to be developed in the 1950s, had been hailed as a first step toward human-level machine intelligence, a 1969 book by MIT's ­Marvin Minsky and Seymour Papert, called Perceptrons, proved mathematically that such networks could perform only the most basic functions. These networks had just two layers of neurons, an input layer and an output layer. Nets with more layers between the input and output neurons could in theory solve a great variety of problems, but nobody knew how to train them, and so in practice they were useless. Except for a few holdouts like Hinton, Perceptrons caused most people to give up on neural nets entirely.

Hinton's breakthrough, in 1986, was to show that backpropagation could train a deep neural net, meaning one with more than two or three layers. But it took another 26 years before increasing computational power made good on the discovery. A 2012 paper by Hinton and two of his Toronto students showed that deep neural nets, trained using backpropagation, beat state-of-the-art systems in image recognition. "Deep learning" took off. To the outside world, AI seemed to wake up overnight. For Hinton, it was a payoff long overdue.

While huge investments are currently flowing into applications, this is far from human thought. For example,

Deep learning in some ways mimics what goes on in the human brain, but only in a shallow way—which perhaps explains why its intelligence can sometimes seem so shallow. Indeed, backprop wasn't discovered by probing deep into the brain, decoding thought itself; it grew out of models of how animals learn by trial and error in old classical-conditioning experiments. And most of the big leaps that came about as it developed didn't involve some new insight about neuroscience; they were technical improvements, reached by years of mathematics and engineering. What we know about intelligence is nothing against the vastness of what we still don't know.

David Duvenaud, an assistant professor in the same department as Hinton at the University of Toronto, says deep learning has been somewhat like engineering before physics. "Someone writes a paper and says, 'I made this bridge and it stood up!' Another guy has a paper: 'I made this bridge and it fell down—but then I added pillars, and then it stayed up.' Then pillars are a hot new thing. Someone comes up with arches, and it's like, 'Arches are great!'" With physics, he says, "you can actually understand what's going to work and why." Only recently, he says, have we begun to move into that phase of actual understanding with artificial intelligence.

Using the example of a two year old human who can recognize a hot dog after being exposed to a few of them, and comparing with the vast number of hot dog pictures required to train a deep learning system, the author goes in search for the next big thing...


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 4, Informative) by SunTzuWarmaster on Wednesday October 04 2017, @01:07PM (6 children)

    by SunTzuWarmaster (3971) on Wednesday October 04 2017, @01:07PM (#576999)

    I actually am an AI scientist. It has been mathematically proven that a 3-layer network (input, output, hidden) can learn any non-linear function. The science behind a 3-layer backprop-trained, non-recursive network is more-or-less known and proven. Naturally, modern networks are 100+ layers, recursive, and approximation-trained. Performance on problems which are time-based increases through the addition of recursive connections (which have no meaningfully mathematically sound way to train them). Performance on problems which appear complicated increases with the addition of hidden layers (it is not clear why).

    In regards to your question on investigation into using structures to build structures which build networks, I'd encourage you to look at CPPNs or generative networks. We, as a field, are looking at it. Of course, we are also looking at different neural structures, different training regimes, different network structures, etc.

    The AI field is fumbling about in the dark without sound principles (math/CS). To some extent, that's okay, as applied scientists (experimentalists) get ahead of theoretical scientists (theorists) all the time (see: radio telescopes, discovery of the cell, discovery of DNA, spring physics, Boyles Law, nuclear weapons, ~70% of Nobel Prizes, etc.). As a below poster mentions, the build-and-test engineering approach may be sufficient for a large class of "simple" problems with meaningful societal impact (build 1000 handwriting recognition systems, take the one with the best performance, accept the 2% error). It would be better to design 1 system based on a firm mathematical foundation, and this will *have* to be done for more complex problems.

    Starting Score:    1  point
    Moderation   +2  
       Informative=2, Total=2
    Extra 'Informative' Modifier   0  
    Karma-Bonus Modifier   +1  

    Total Score:   4  
  • (Score: 0) by Anonymous Coward on Wednesday October 04 2017, @03:05PM (1 child)

    by Anonymous Coward on Wednesday October 04 2017, @03:05PM (#577048)

    Thread starter here. Thank you for the extensive reply, it gave me some food for thought (a mathematician and biologist have clearly a different view on the subject).

    • (Score: 0) by Anonymous Coward on Wednesday October 04 2017, @03:12PM

      by Anonymous Coward on Wednesday October 04 2017, @03:12PM (#577051)

      The quote (on the bottom of the page just now) seems fitting:

      To invent, you need a good imagination and a pile of junk. -- Thomas Edison

  • (Score: 2) by TheLink on Wednesday October 04 2017, @05:38PM (2 children)

    by TheLink (332) on Wednesday October 04 2017, @05:38PM (#577118) Journal

    Performance on problems which appear complicated increases with the addition of hidden layers (it is not clear why).

    The AI field is fumbling about in the dark without sound principles (math/CS).

    Yeah to me the AI field still seems stuck in the "Alchemy" stage. The practitioners can achieve useful stuff. And often it works well enough but nobody really knows why it works or more important how it would fail (there's some work and progress on the latter with "illusions" designed for neural networks).

    Speaking of failure and illusions. You can see how intelligent something is behaving by the mistakes it makes when it makes them. The types of mistakes AIs make are proof that they do not actually understand - they are just guessing using statistics based from thousands or even millions of samples. Go look at the mistakes IBM's Watson makes. Compare them with the cognitive mistakes a dog makes. From the mistakes you know the dog has a lower limited understanding but what it does understand it actually knows. It has a working model of the world and other entities.

    My guess is Strong AI would also have to model and simulate the outside world and themselves (and their choices), predict possible outcomes and attempt to pick a desirable outcome. Running very many multiple simulations is energy intensive (maybe even prohibitive) for conventional computers but may not be so for quantum computers. The accuracy of the result may be lower but that may not be so important for most real world scenarios.

    Consciousness might be what happens when such a quantum simulator recursively predicts itself...

    The pattern recognition and neural network stuff will still be important but more as a building block or interface (and maybe for trained reflexes/"muscle memory"). Don't forget - crows are quite smart but their brains are walnut-sized. So the bulk of a brain might not really be for thinking and intelligence. I believe that many single celled creatures can actually think ( https://soylentnews.org/comments.pl?sid=450&cid=11384#commentwrap [soylentnews.org] ) and that brains first evolved not to solve the thinking problem but other problems - like the interfacing problem- how to use and control multi-cellular bodies. So perhaps more scientists should investigate how these single celled creatures actually think.

    p.s. I don't know much about AI but my bullshit on AI and related stuff doesn't seem that much worse than from those selling them ;)

    • (Score: 0) by Anonymous Coward on Wednesday October 04 2017, @09:00PM (1 child)

      by Anonymous Coward on Wednesday October 04 2017, @09:00PM (#577181)

      So perhaps more scientists should investigate how these single celled creatures actually think.

      Hmm... plants don't have brains and their responses are the results of feedback loops based on internal and external signals. There is little thinking going on here, just programming with feedback loops.

      • (Score: 0) by Anonymous Coward on Thursday October 05 2017, @06:17PM

        by Anonymous Coward on Thursday October 05 2017, @06:17PM (#577558)
        Same goes for most ACs.
  • (Score: 0) by Anonymous Coward on Thursday October 05 2017, @12:12AM

    by Anonymous Coward on Thursday October 05 2017, @12:12AM (#577244)

    There's more than one approach that particular cat-skinning.

    For example, I'm also in AI, but from the other direction - cognitive psych and computer science combined. The question of whether neural nets can approximate functions is, as you say, a solved one. The one that we're wrestling with is _which_ functions you need to actually have a meaningful cognitive engine, and how they need to interrelate to each other. If you like to think of things in general software engineering terms, we're trying to create the spec document while other people are trying to write Knuth for AI.

    Guys like Sagar are asking (and answering) some of the right questions, but it's not clear whether they're addressing the question of arbitrary AI and the definition of cognition as opposed to an ad hoc approach. I suspect that they're going ad hoc, based on what I've heard of them.