A repeated imitation model with dependence between stages: decision strategies and rewards
Metadata
Show full item recordEditorial
Sciendo
Materia
Adversarial decision making Imitation Strategies State dependence Reward Inteligencia artificial Artificial intelligence
Date
2015-09-30Referencia bibliográfica
Villacorta,P. & Pelta,D.(2015).A repeated imitation model with dependence between stages: Decision strategies and rewards. International Journal of Applied Mathematics and Computer Science,25(3) 617-630. [https://doi.org/10.1515/amcs-2015-0045]
Sponsorship
Ministry of Science and Innovation, Spain (MICINN) Spanish Government TIN2011-27696-C02-01; Andalusian government P11-TIC-8001; CEI-Biotic Granada GENIL-PYR-2014-9; European Commission; German Research Foundation (DFG)Abstract
Adversarial decision making is aimed at determining strategies to anticipate the behavior of an opponent trying to learn from
our actions. One defense is to make decisions intended to confuse the opponent, although our rewards can be diminished.
This idea has already been captured in an adversarial model introduced in a previous work, in which two agents separately
issue responses to an unknown sequence of external inputs. Each agent’s reward depends on the current input and the
responses of both agents. In this contribution, (a) we extend the original model by establishing stochastic dependence
between an agent’s responses and the next input of the sequence, and (b) we study the design of time varying decision
strategies for the extended model. The strategies obtained are compared against static strategies from theoretical and
empirical points of view. The results show that time varying strategies outperform static ones.