811
Views
83
Downloads
3
Crossref
N/A
WoS
2
Scopus
N/A
CSCD
Agents are always in an interactive environment. With time, the intelligence of agents will be affected by the interactive environment. Agents need to coordinate the interaction with different environmental factors to achieve the optimal intelligence state. We consider an agent’s interaction with the environment as an action-reward process. An agent balances the reward it receives by acting with various environmental factors. This paper refers to the concept of interaction between an agent and the environment in reinforcement learning and calculates the optimal mode of interaction between an agent and the environment. It aims to help agents maintain the best intelligence state as far as possible. For specific interaction scenarios, this paper takes food collocation as an example, the evolution process between an agent and the environment is constructed, and the advantages and disadvantages of the evolutionary environment are reflected by the evolution status of the agent. Our practical case study using dietary combinations demonstrates the feasibility of this interactive balance.
Agents are always in an interactive environment. With time, the intelligence of agents will be affected by the interactive environment. Agents need to coordinate the interaction with different environmental factors to achieve the optimal intelligence state. We consider an agent’s interaction with the environment as an action-reward process. An agent balances the reward it receives by acting with various environmental factors. This paper refers to the concept of interaction between an agent and the environment in reinforcement learning and calculates the optimal mode of interaction between an agent and the environment. It aims to help agents maintain the best intelligence state as far as possible. For specific interaction scenarios, this paper takes food collocation as an example, the evolution process between an agent and the environment is constructed, and the advantages and disadvantages of the evolutionary environment are reflected by the evolution status of the agent. Our practical case study using dietary combinations demonstrates the feasibility of this interactive balance.
D. L. Dowe and J. Hernández-Orallo, IQ tests are not for machines, yet, Intelligence, vol. 40, no. 2, pp. 77–81, 2012.
A. M. Turing, Computing machinery and intelligence, Mind, vol. 59, no. 236, pp. 433–460, 1950.
J. Hernández-Orallo and D. L. Dowe, Measuring universal intelligence: Towards an anytime intelligence test, Artif. Intell., vol. 174, no. 18, pp. 1508–1539, 2010.
W. Ji, J. Liu, Z. W. Pan, J. C. Xu, B. Liang, and Y. Q. Chen, Quality-time-complexity universal intelligence measurement, Int. J. Crowd Sci., vol. 2, no. 1, pp. 18–26, 2018.
S. S. Xiao, C. P. Wei, and M. Dong, Crowd intelligence: Analyzing online product reviews for preference measurement, Inf. Manage., vol. 53, no. 2, pp. 169–182, 2016.
A. W. Woolley, C. F. Chabris, A. Pentland, N. Hashmi, and T. W. Malone, Evidence for a collective intelligence factor in the performance of human groups, Science, vol. 330, no. 6004, pp. 686–688, 2010.
J. Duncan, R. J. Seitz, J. Kolodny, D. Bor, H. Herzog, A. Ahmed, F. N. Newell, and H. Emslie, A neural basis for general intelligence, Science, vol. 289, no. 5478, pp. 457–460, 2000.
V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski, et al., Human-level control through deep reinforcement learning, Nature, vol. 518, no. 7540, pp. 529–533, 2015.
This work was supported by the National Natural Science Foundation of China (No. 62072440) and the Beijing Natural Science Foundation (No. 4202072).
The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).