Stories
Slash Boxes
Comments

SoylentNews is people

SoylentNews is powered by your submissions, so send in your scoop. Only 11 submissions in the queue.
posted by Fnord666 on Sunday July 07 2019, @05:10AM   Printer-friendly
from the ¯\_(ツ)_/¯ dept.

Submitted via IRC for Bytram

That this AI can simulate universes in 30ms is not the scary part. It's that its creators don't know why it works so well

Neural networks can build 3D simulations of the universe in milliseconds, compared to days or weeks when using traditional supercomputing methods, according to new research.

To study how stuff interacts in space, scientists typically build computational models to simulate the cosmos. One simulation approach – known as N-body simulation – can be used to recreate phenomena ranging from smaller events, such as the collapse of molecular clouds into stars, to a giant system, such as the whole universe, obviously to varying levels of accuracy and resolution.

The individual interactions between each of the millions or billions of particles or entities in these models have to be repeatedly calculated to track their motion over time. This requires heavy amounts of compute power, and it takes several days or weeks for a supercomputer to return the results.

For impatient boffins, there's now some good news. A group of physicists, led by eggheads at the Center for Computational Astrophysics at the Flatiron Institute in New York, USA, decided to see if neural networks could speed things up a bit.

[...] The accuracy of the neural network is judged by how similar its outputs were to the ones created by two more traditional N-body simulation systems, FastPM and 2LPT, when all three are given the same inputs. When D3M was tasked with producing 1,000 simulations from 1,000 sets of input data, it had a relative error of 2.8 per cent compared to FastPM, and a 9.3 per cent compared to 2LPT for the same inputs. That's not too bad, considering it takes the model just 30 milliseconds to crank out a simulation. Not only does that save time, but it's also cheaper too since less compute power is needed.

To their surprise, the researchers also noticed that D3M seemed to be able to produce simulations of the universe from conditions that weren't specifically included in the training data. During inference tests, the team tweaked input variables such as the amount of dark matter in the virtual universes, and the model still managed to spit out accurate simulations despite not being specifically trained for these changes.

"It's like teaching image recognition software with lots of pictures of cats and dogs, but then it's able to recognize elephants," said Shirley Ho, first author of the paper and a group leader at the Flatiron Institute. "Nobody knows how it does this, and it's a great mystery to be solved.

"We can be an interesting playground for a machine learner to use to see why this model extrapolates so well, why it extrapolates to elephants instead of just recognizing cats and dogs. It's a two-way street between science and deep learning."

The source code for the neural networks can be found here. ®


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 4, Insightful) by Rosco P. Coltrane on Sunday July 07 2019, @06:00AM (1 child)

    by Rosco P. Coltrane (4757) on Sunday July 07 2019, @06:00AM (#864038)

    That's the thing with AI: you program the neural net, let it learn and do its thing, then you don't know what it's doing exactly to produce results - only that the results are an approximation of the exact outcome you want, and that you need to keep training the AI until they're deemed "good enough".

    Kind of like a wetware brain really...

    Starting Score:    1  point
    Moderation   +2  
       Insightful=2, Total=2
    Extra 'Insightful' Modifier   0  
    Karma-Bonus Modifier   +1  

    Total Score:   4  
  • (Score: 2) by legont on Monday July 08 2019, @02:29AM

    by legont (4179) on Monday July 08 2019, @02:29AM (#864294)

    Assuming the neural net AI has access to the same level of hardware as the "researcher", the way it works can't be known in principle. It takes an order of magnitude or two more computing power to analyze neural net than to run it. Yes, similar to wetware.

    --
    "Wealth is the relentless enemy of understanding" - John Kenneth Galbraith.