SCIENTIA SINICA Informationis, Volume 50 , Issue 4 : 540-550(2020) https://doi.org/10.1360/N112019-00048

Intelligent technologies of human-computer gaming

More info
  • ReceivedFeb 27, 2019
  • AcceptedJun 6, 2019
  • PublishedApr 10, 2020


At the frontier of artificial intelligence research, human-computer gaming (HCG) technology has become a research hotspot. It provides an effective experimental environment and approach to exploring the intrinsic growth mechanism and verifying key technologies of machine intelligence. Considering the mounting challenges in intelligent decision-making posed by the complex, high-dynamic, and inconclusive environment coupled with strong confrontation, this paper analyzes the research status and dissects the key elements and intrinsic gaming mechanisms of HCG. This work also proposes a game learning-based theoretical research framework for HCG. Based on these analyses, we discussed HCG's key models: gaming representation and modeling, situation assessment and reasoning, strategy generation, and optimization, as well as action coordination and control. The proposed research framework has laid a foundation for the modeling, computing, and interpreting solutions of complex cognition and decision problems. Finally, this paper summarizes the current application status and looks to future directions of development.


[1] Tan T N. Artificial intelligence - building an intelligent future with AI technologies. Beijing: China Science and Technology Press, 2019. Google Scholar

[2] Li D, Dong Y. Deep Learning: Methods and Applications. Boston: Now Publishers Inc., 2014. Google Scholar

[3] Huang K Q, Ren W Q, Tan T N. A review on image object classification and detection. Chin J Comput, 2014, 37: 1--18[黄凯奇, 任伟强, 谭铁牛. 图像物体分类与检测算法综述 计算机学报, 2014, 37: 1--18]. Google Scholar

[4] Li D, Yang L. Deep Learning in Natural Language Processing. Switzerland: Springer Press, 2018. Google Scholar

[5] Silver D, Huang A, Maddison C J. Mastering the game of Go with deep neural networks and tree search. Nature, 2016, 529: 484-489 CrossRef PubMed ADS Google Scholar

[6] Brown N, Sandholm T. Superhuman AI for heads-up no-limit poker: Libratus beats top professionals.. Science, 2018, 359: 418-424 CrossRef PubMed Google Scholar

[7] Turing A M. Computing machinery and intelligence. In: Parsing the Turing Test. Dordrecht: Springer. 2009. Google Scholar

[8] Zhang W Y. Gaming and Scocity. Beijing: Peking University Press, 2013. Google Scholar

[9] Hinton G E, Salakhutdinov R R. Reducing the Dimensionality of Data with Neural Networks. Science, 2006, 313: 504-507 CrossRef PubMed ADS Google Scholar

[10] Huang K Q, Chen X T, Kang Y F, et al. Intelligent visual surveillance: a review. Chin J Comput, 2015, 38: 1093--1118 [黄凯奇, 陈晓棠, 康运锋, 等. 智能视频监控技术综述. 计算机学报, 2015, 38: 1093--1118]. Google Scholar

[11] Silver D, Schrittwieser J, Simonyan K. Mastering the game of Go without human knowledge. Nature, 2017, 550: 354-359 CrossRef PubMed ADS Google Scholar

[12] Moravcík M, Schmid M, Burch N. DeepStack: Expert-level artificial intelligence in heads-up no-limit poker. Science, 2017, 356: 508-513 CrossRef PubMed ADS arXiv Google Scholar

[13] Vinyals O, Ewalds T, Bartunov S, et al. StarCraft II: A New Challenge for Reinforcement Learning. 2017,. arXiv Google Scholar

[14] Tian Y, Gong Q, Shang W, et al. ELF: An Extensive, Lightweight and Flexible Research Platform for Real-time Strategy Games. 2017,. arXiv Google Scholar

[15] Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets. In: Proceedings of Neural Information Processing Systems, 2014. 2672--2680. Google Scholar

[16] Brown N, Sandholm T, and Amos B. Depth-Limited Solving for Imperfect-Information Games. In: Proceedings of Neural Information Processing Systems, 2018. Google Scholar

[17] Zeiler M D, Fergus R. Visualizing and understanding convolutional networks. In: Proceedings of European Conference on Computer Vision, 2014. Google Scholar

[18] Zhou P, Feng J. Empirical risk landscape analysis for understanding deep neural networks. In: Proceedings of Internaltion Conference on Learning Represetnations, 2018. Google Scholar

[19] Simon H, 著. 杨砾, 译. 管理行为-管理组织决策过程的研究. 北京: 北京经济学院出版社, 1988. Google Scholar

[20] Zeng P, Wu L D, Wei Y M. Analysis, description and design for a tactical plan recognition model. Comput Digi Eng, 2006, 34: 1--4. Google Scholar

[21] Shi C, Li Y T, Zhang J W. A survey of heterogeneous information network analysis. IEEE Trans Knowl Data Eng, 2017, 29: 17-37 CrossRef Google Scholar

[22] Zhu F, Hu X F, Wu L, et al. From situation cognition stepped into situation intelligent cognition. J Sys Simul, 2018, 30: 761--771. Google Scholar

[23] Wang D X, Cui P, Zhu W W. Structural Deep Network Embedding. In: Proceedings of the 22nd ACM SIGKDDInternational Conference on Knowledge Discovery and Data Mining. New York: ACM, 2016. 1225. Google Scholar

[24] Martinage R. 郑伟, 译. 迈向新抵消战略: 利用美国的长期优势恢复美国全球力量投送能力. 江阴: 知远战略与防务研究所, 2014. Google Scholar

[25] Wang Q, Mao Z D, Wang B. Knowledge Graph Embedding: A Survey of Approaches and Applications. IEEE Trans Knowl Data Eng, 2017, 29: 2724-2743 CrossRef Google Scholar

[26] Racaniere S, Weber T, Reichert D P, et al. Imagination-augmented agents for deep reinforcement learning. In: Proceedings of Neural Information Processing Systems, 2017. Google Scholar

[27] Huang W Z, Zhang J G, Huang K Q. Bootstrap estimated uncertainty of the environment model for model-based reinforcement learning. In: Proceedings of the 33rd AAAI Conference on Artificial Intelligence (AAAI-19), 2019. Google Scholar

[28] Hu X F, He X Y, Tao J Y. AlphaGo's breakthrough and challenges of wargaming. Sci Technol Rev 2017, 35: 49--60. Google Scholar

[29] Hu Y, Li J T, Li X, et al. Knowledge-guided agent-tactic-aware Learning for StarCraft Micromanagement. In: Proceedings of Internaltion Joint Conference on Artificial Intelligence, 2018. Google Scholar

[30] Lowe R, Wu Y, Tamar A, et al. Multi-agent actor-critic for mixed cooperative-competitive environments. In: Proceedings of Neural Information Processing Systems, 2017. Google Scholar

[31] Shen S Q, Yin Q Y, Zhang J G, et al. Application and research of assistant command and control decision making method based on deep situation verification. China Conf Command Control, 2018. Google Scholar

[32] Yin Q Y, Zhan D, Zhang J G, et al. Combat Unit Prediction under the Fog of War. China Conf Command Control, 2017. Google Scholar

[33] Han X, Yan H P, Zhang J G, et al. ACM: learning dynamic multi-agent cooperation via attentional communication model. In: Proceedings of Internatial Confernce on Artificial Neural Network, 2018. 219--229. Google Scholar

[34] Wu H K, Zhang J G, Huang K Q. MSC: A dataset for macro-management in StarCraft. 2017,. arXiv Google Scholar

[35] Peng P, Xing J, Cao L, et al. Learning deep decentralized policy network by collective rewards for real-time combat game. In: Proceedings of Internaltion Joint Conference on Artificial Intelligence, 2019. Google Scholar

[36] Churchill D, Buro M. Build order optimization in StarCraft. In: Proceedings of AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, 2011. Google Scholar

[37] Garcia-Sanchez P, Tonday A, Mora A M, et al. Towards automatic starcraft strategy generation using genetic programming. In: Proceedings of IEEE Internaltion Conference on Computer Games, 2015. Google Scholar

[38] Busoniu L, Babuska R, De Schutter B. A Comprehensive Survey of Multiagent Reinforcement Learning. IEEE Trans Syst Man Cybern C, 2008, 38: 156-172 CrossRef Google Scholar

[39] Peng P, Yuan Q, Wen Y, et al. Multiagent Bidirectionally-Coordinated Nets for Learning to Play StarCraft Combat Games. 2017,. arXiv Google Scholar

[40] Sukhbaatar S, Szlam A, Fergus R. Learning multiagent communication with backpropagation. In: Proceedings of Advances in Neural Information Processing Systems, 2016. 2244--2252. Google Scholar

[41] Zhang K Q, Yang Z R, Liu H, et al. Fully decentralized multi-agent reinforcement learning with networked agents. In: Proceedings of International Conference on Machine Learning, 2018. Google Scholar

[42] Li Y X. Deep reinforcement learning: an overview. 2017,. arXiv Google Scholar

[43] Jaderberg M, Czarnecki W M, Dunning I. Human-level performance in 3D multiplayer games with population-based reinforcement learning. Science, 2019, 364: 859-865 CrossRef PubMed ADS arXiv Google Scholar

[44] Qian R Y. Command Decisions and Strategies. Beijing: Kunlun Press, 1999. Google Scholar

Copyright 2020  CHINA SCIENCE PUBLISHING & MEDIA LTD.  中国科技出版传媒股份有限公司  版权所有

京ICP备14028887号-23       京公网安备11010102003388号