• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Li, Min (Li, Min.) [1] | Zhu, William (Zhu, William.) [2] | Wang, Shiping (Wang, Shiping.) [3]

Indexed by:

EI

Abstract:

Most reinforcement learning methods describe sequential decision-making as a Markov decision process where the effect of action is only decided by the current state. But this is reasonable only if the state is correctly defined and the state information is sufficiently observed. Thus the learning efficiency of reinforcement learning methods based on Markov decision process is limited when the state information is insufficient. Partially observable Markov decision process and history-based decision process are respectively proposed to describe sequential decision-making with insufficient state information. However, these two processes are easy to ignore the important information from the current observed state. Therefore, the learning efficiency of reinforcement learning methods based on these two processes is also limited when the state information is insufficient. In this paper, we propose a multi-view reinforcement learning method to solve this problem. The motivation is that the interaction information between the agent and its environment should be considered from the views of history, present, and future to overcome the insufficiency of state information. Based on these views, we construct a multi-view decision process to describe sequential decision-making with insufficient state information. A multi-view reinforcement learning method is proposed by combining the multi-view decision process and the actor-critic framework. In the proposed method, multi-view clustering is performed to ensure that each type of sample can be sufficiently exploited. Experiments illustrate that the proposed method is more effective than the compared state-of-the-arts. The source code can be downloaded from https://github.com/jamieliuestc/MVRL. © The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2023.

Keyword:

Behavioral research Decision making Efficiency Learning algorithms Learning systems Markov processes Reinforcement learning

Community:

  • [ 1 ] [Li, Min]Institute of Fundamental and Frontier Sciences, University of Electronic Science and Technology of China, Chengdu, China
  • [ 2 ] [Zhu, William]Institute of Fundamental and Frontier Sciences, University of Electronic Science and Technology of China, Chengdu, China
  • [ 3 ] [Wang, Shiping]College of Computer and Data Science, Fuzhou University, Fuzhou, China

Reprint 's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

International Journal of Machine Learning and Cybernetics

ISSN: 1868-8071

Year: 2024

Issue: 4

Volume: 15

Page: 1533-1552

3 . 1 0 0

JCR@2023

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 0

Affiliated Colleges:

Online/Total:112/10147862
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1