A New Decision Making Method for Selection of Optimal Data Using the Von Neumann-Morgenstern Theorem
Metadata
Show full item recordAuthor
García Cabello, JuliaEditorial
Journal Informatica (INST MATHEMATICS & INFORMATICS)
Date
2023-12Referencia bibliográfica
GarcÍa Cabello, J. (2023). A New Decision Making Method for Selection of Optimal Data Using the Von Neumann-Morgenstern Theorem. Informatica, 1-24.
Abstract
The quality of the input data is amongst the decisive factors affecting the speed and effectiveness of recurrent neural network (RNN) learning. We present here a novel methodology to
select optimal training data (those with the highest learning capacity) by approaching the problem
from a decision making point of view. The key idea, which underpins the design of the mathematical
structure that supports the selection, is to define first a binary relation that gives preference to inputs
with higher estimator abilities. The Von Newman Morgenstern theorem (VNM), a cornerstone of
decision theory, is then applied to determine the level of efficiency of the training dataset based on
the probability of success derived from a purpose-designed framework based on Markov networks.
To the best of the author’s knowledge, this is the first time that this result has been applied to data
selection tasks. Hence, it is shown that Markov Networks, mainly known as generative models, can
successfully participate in discriminative tasks when used in conjunction with the VNM theorem.
The simplicity of our design allows the selection to be carried out alongside the training. Hence,
since learning progresses with only the optimal inputs, the data noise gradually disappears: the result
is an improvement in the performance while minimising the likelihood of overfitting.