Journal of Shanghai University >
Feature selection based on reinforcement learning and its application in material informatics
Received date: 2020-03-20
Online published: 2022-05-27
Owing the rapid development of big data, artificial intelligence, and high-performance computing, the research and development of data-driven materials has intensified. During data mining and the machine learning of material data, the feature set must be preprocessed by reducing redundant and irrelevant features, which can not only avoid model overfitting, but also improve the model interpretability. Herein, a feature selection method based on reinforcement learning, known as FSRL, is proposed. By abstracting the encapsulated feature selection method into the interaction between the machine learning model and environment, the corresponding features are selected based on the maximum reward and then incorporated to the feature subset. In addition, we propose a feature construction method based on symbolic transformation to generate new high-order features to improve the prediction accuracy of the model. Subsequently, we apply the abovementioned method to the classification task of amorphous alloy materials and the regression task of aluminum matrix composite materials. Experiments show that our proposed method not only successfully achieve feature transformation in the FSRL, but also afford a 2.8% prediction improvement in the classification task and a 22.9% prediction improvement in the regression task respectively.
ZHANG Peng, ZHANG Rui . Feature selection based on reinforcement learning and its application in material informatics[J]. Journal of Shanghai University, 2022 , 28(3) : 463 -475 . DOI: 10.12066/j.issn.1007-2861.2375
| [1] | 周腾, 宋真, Sundmacher K. 大数据为材料研究创造新机遇--材料设计的机器学习方法与应用综述[J]. Engineering, 2019, 5(6): 1017-1026. |
| [2] | Rajan K. Materials informatics: the materials "gene" and big data[J]. Annual Review of Materials Research, 2015, 45: 153-169. |
| [3] | Virshup A M, Contreras-García J, Wipf P, et al. Stochastic voyages into uncharted chemical space produce a representative library of all possible drug-like compounds[J]. Journal of the American Chemical Society, 2013, 135(19): 7296-7303. |
| [4] | Karak S K, Chatterjee S, Bandopadhyay S. Mathematical modelling of the physical and mechanical properties of nano-Y$_2$O$_3$ dispersed ferritic alloys using evolutionary algorithm-based neural network[J]. Powder Technology, 2015, 274: 217-226. |
| [5] | 计智伟, 胡珉, 尹建新. 特征选择算法综述[J]. 电子设计工程, 2011, 19(9): 46-51. |
| [6] | Huang L L, Tang J, Sun D D, et al. Feature selection algorithm based on multi-label ReliefF[J]. Journal of Computer Applications, 2012, 32(10): 2888-2890. |
| [7] | 吴迪, 郭嗣琮. 改进的 Fisher Score 特征选择方法及其应用[J]. 辽宁工程技术大学学报(自然科学版), 2019, 38(5): 472-479. |
| [8] | 崔鸿雁, 徐帅, 张利锋, 等. 机器学习中的特征选择方法研究及展望[J]. 北京邮电大学学报, 2018, 41(1): 1-12. |
| [9] | Guyon I, Weston J, Barnhill S, et al. Gene selection for cancer classification using support vector machines[J]. Machine Learning, 2002, 46: 389-422. |
| [10] | 姚登举, 杨静, 詹晓娟. 基于随机森林的特征选择算法[J]. 吉林大学学报(工学版), 2014, 44(1): 137-141. |
| [11] | 姚旭, 王晓丹, 张玉玺, 等. 特征选择方法综述[J]. 控制与决策, 2012, 27(2): 161-166. |
| [12] | 毛勇, 周晓波, 夏铮, 等. 特征选择算法研究综述[J]. 模式识别与人工智能, 2007, 20(2): 211-218. |
| [13] | Bellman R, Kalaba R. Dynamic programming and statistical communication theory[J]. Proceedings of the National Academy of Science of the United States of America, 1957, 43(8): 749-751. |
| [14] | Andreae J H. STELLA: a scheme for a learning machine[J]. IFAC Proceedings Volumes, 1963, 1(2): 497-502. |
| [15] | Sutton R S. Reinforcement learning: past, present and future[C]// Asia-Pacific Conference on Simulated Evolution and Learning. 1999: 195-197. |
| [16] | Watkins C. Learning from delayed rewards[D]. Cambridge: University of Cambridge, 1989. |
| [17] | Mnih V, Kavukcuoglu K, Silver D, et al. Playing Atari with deep reinforcement learning[C]// NIPS. 2013. |
| [18] | Silver D, Lever G, Heess N, et al. Deterministic policy gradient algorithms[C]// International Conference on Machine Learning. 2014: 387-395. |
| [19] | Van Hasselt H, Guez A, Silver D. Deep reinforcement learning with double $q$- learning[C]// Proceedings of the AAAI Conference on Artificial Intelligence. 2016. |
| [20] | 唐振韬, 邵坤, 赵冬斌, 等. 深度强化学习进展: 从 AlphaGo 到 AlphaGo Zero[J]. 控制理论与应用, 2017, 34(12): 1529-1546. |
| [21] | Watkins C, Dayan P. Technical note: $Q$-learning[J]. Machine Learning, 1992, 8(3/4): 279-292. |
| [22] | 高阳, 陈世福, 陆鑫. 强化学习研究综述[J]. 自动化学报, 2004(1): 86-100. |
| [23] | 刘路放, 冯博琴. 符号回归的枚举原型算法及其匹配算法研究[J]. 西安交通大学学报, 2000, 34(3): 1-4. |
| [24] | 谢锦涛, 何丰, 王如岑. Double DQN 在坦克对战游戏中的应用[J]. 福建电脑, 2020, 36(5): 120-121. |
| [25] | 胡壮麒, 张海峰. 块状非晶合金及其复合材料研究进展[J]. 金属学报, 2010, 46(11): 1391-1421. |
| [26] | 欧阳求保, 张国定, 张荻. 非连续增强铝基复合材料的研究与应用进展[J]. 中国材料进展, 2010, 29(4): 36-40. |
/
| 〈 |
|
〉 |