Stories
Slash Boxes
Comments

SoylentNews is people

SoylentNews is powered by your submissions, so send in your scoop. Only 9 submissions in the queue.

Submission Preview

Link to Story

Mathematicians Propose New Way of Using Neural Networks to Work With Noisy, High-dimensional Data

Accepted submission by upstart at 2020-02-21 17:50:57
News

████ # This file was generated bot-o-matically! Edit at your own risk. ████

Mathematicians propose new way of using neural networks to work with noisy, high-dimensional data [phys.org]:

February 20, 2020

Mathematicians propose new way of using neural networks to work with noisy, high-dimensional data

Mathematicians from RUDN University and the Free University of Berlin have proposed a new approach to studying the probability distributions of observed data using artificial neural networks. The new approach works better with so-called outliers, i.e., input data objects that deviate significantly from the overall sample. The article was published in the journal Artificial Intelligence.

The restoration of the probability distribution of observed data by artificial neural networks is the most important part of machine learning. The probability distribution not only allows us to predict the behaviour of the system under study, but also to quantify the uncertainty with which forecasts are made. The main difficulty is that, as a rule, only the data are observed, but their exact probability distributions are not available. To solve this problem, Bayesian and other similar approximate methods are used. But their use increases the complexity of a neural network and therefore makes its training more complicated.

RUDN University and the Free University of Berlin mathematicians used deterministic weights in neural networks, which would help overcome the limitations of Bayesian methods. They developed a formula that allows one to correctly estimate the variance of the distribution of observed data. The proposed model was tested on different data: synthetic and real; on data containing outliers and on data from which the outliers were removed. The new method allows restoration of probability distributions with accuracy previously unachievable.

The mathematicians of RUDN University and the Free University of Berlin used deterministic weights for neural networks and used the networks outputs to encode the distribution of latent variables for the desired marginal distribution. An analysis of the training dynamics of such networks allowed them to obtain a formula that correctly estimates the variance of observed data, despite the presence of outliers in the data. The proposed model was tested on different data: synthetic and real. The new method allows restoring probability distributions with higher accuracy compared with other modern methods. Accuracy was assessed using the AUC method (area under the curve is the area under the graph that allows making assessment of the mean square error of the predictions depending on the sample size estimated by the network as "reliable"; the higher the AUC score, the better the predictions).

Explore further

Neuroscience opens the black box of artificial intelligence [techxplore.com]More information: Pavel Gurevich et al. Gradient conjugate priors and multi-layer neural networks, Artificial Intelligence (2019). DOI: 10.1016/j.artint.2019.103184 [doi.org] Provided by RUDN University [phys.org]Citation: Mathematicians propose new way of using neural networks to work with noisy, high-dimensional data (2020, February 20) retrieved 21 February 2020 from https://phys.org/news/2020-02-mathematicians-neural-networks-noisy-high-dimensional.html [phys.org] This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only. 48 shares [javascript]

Feedback to editors


Original Submission