Welcome to Journal of Beijing Institute of Technology
Volume 27Issue 4
.
Turn off MathJax
Article Contents
Wei Gao, Dunbo Cai. Using Vector Representation of Propositions and Actions for STRIPS Action Model Learning[J]. JOURNAL OF BEIJING INSTITUTE OF TECHNOLOGY, 2018, 27(4): 485-492. doi: 10.15918/j.jbit1004-0579.18072
Citation: Wei Gao, Dunbo Cai. Using Vector Representation of Propositions and Actions for STRIPS Action Model Learning[J].JOURNAL OF BEIJING INSTITUTE OF TECHNOLOGY, 2018, 27(4): 485-492.doi:10.15918/j.jbit1004-0579.18072

Using Vector Representation of Propositions and Actions for STRIPS Action Model Learning

doi:10.15918/j.jbit1004-0579.18072
  • Received Date:2018-03-20
  • Action model learning has become a hot topic in knowledge engineering for automated planning. A key problem for learning action models is to analyze state changes before and after action executions from observed "plan traces". To support such an analysis, a new approach is proposed to partition propositions of plan traces into states. First, vector representations of propositions and actions are obtained by training a neural network called Skip-Gram borrowed from the area of natural language processing (NLP). Then, a type of semantic distance among propositions and actions is defined based on their similarity measures in the vector space. Finally, k-means and k-nearest neighbor (kNN) algorithms are exploited to map propositions to states. This approach is called state partition by word vector (SPWV), which is implemented on top of a recent action model learning framework by Rao et al. Experimental results on the benchmark domains show that SPWV leads to a lower error rate of the learnt action model, compared to the probability based approach for state partition that was developed by Rao et al.
  • loading
  • [1]
    Fikes R E, Nilsson N J. STRIPS:A new approach to the application of theorem proving to problem solving[J]. Artificial Intelligence, 1971, 2(3-4):189-208.
    [2]
    Fox M, Long D. PDDL2.1:An extension to PDDL for expressing temporal planning domains[J]. Journal of Artificial Intelligence Research (JAIR), 2003, 20:61-124.
    [3]
    Wang X. Learning by observation and practice:An incremental approach for planning operator acquisition[C]//Proceedings of the 12th International Conference on Machine Learning (ICML'95), 1995:549-557.
    [4]
    Blythe J, Kim J, Ramachandran S, et al. An integrated environment for knowledge acquisition[C]//Proceedings of the Sixth International Conference on Intelligent User Interfaces(IUI 2001),2001:13-20.
    [5]
    Yang Q, Wu K, Jiang Y. Learning action models from plan examples with incomplete knowledge[C]//Proceedings of the Fifteenth International Conference on Automated Planning and Scheduling(ICAPS 2005), 2005:241-250.
    [6]
    Yang Q, Wu K, Jiang Y. Learning action models from plan examples using weighted MAX-SAT[J]. Artificial Intelligence,2007,171(2-3):107-143.
    [7]
    Zhuo H, Yang Q, Hu D H, et al. Learning complex action models with quantifiers and logical implications[J]. Artificial Intelligence,2010,174(18):1540-1569.
    [8]
    Amir E, Chang A. Learning partially observable deterministic action models[J]. Journal of Artificial Intelligence Research, 2008, 33:349-402.
    [9]
    Rao D N, Jiang Z H, Jiang Y F, et al. Learning partially observable non-deterministic action models[J]. Journal of Software,2014,25(1):51-63. (in Chinese)
    [10]
    Gregory P, Lindsay A. Domain model acquisition in domains with action costs[C]//Proceedings of Twenty-Sixth International Conference on Automated Planning and Scheduling (ICAPS 2016), 2016:149-157.
    [11]
    Lanchas J, Jiménez S, Fernández F, et al. Learning action durations from executions[C]//Working Notes of ICAPS'07 Workshop on AI Planning and Learning, Rhode Island, Providence, USA, 2007.
    [12]
    Mikolov T, Chen K, Corrado G, et al. Efficient estimation of word representations in vector space[C]//Proc of ICLR, 2013.
    [13]
    Lample G, Ballesteros M, Subramanian S, et al. Neural architectures for named entity recognition[C]//Proc of NAACL-HLT, 2016:260-270.
    [14]
    Caliskan A, Bryson J J, Narayanan A. Semantics derived automatically from language corpora contain human-like biases[J]. Science, 2017, 356(6334):183-186.
    [15]
    Niu J, Chen H, Zhao Q, et al. Multi-document abstractive summarization using chunk-graph and recurrent neural network[C]//Proc of 2017 IEEE International Conference on Communications (ICC), IEEE, 2017:1-6.
    [16]
    Luo S, Chen Q, Guo J, et al. Microblog summarization via enriching contextual features based on sentence-level semantic analysis[J]. Journal of Beijing Institute of Technology,2017,26(4):505-516.
    [17]
    Michalski R S, Carbonell J G, Mitchell T M. Machine learning:an artificial intelligence approach[M]. New York:Springer Science & Business Media, 2013.
    [18]
    McDermott D. PDDL, the planning domain definition language[R]. New Haven, USA:Yale Center for Computational Vision and Control, 1998.
    [19]
    Le Q V, Mikolov T. Distributed representations of sentences and documents[C]//Proc of 31th International Conference on Machine Learning (ICML 2014), 2014:1188-1196.
    [20]
    Wang Fang, Lu Heng, Li Yunpeng, et al. Combined transmission interference spectrum of no core fiber and BP neural network for concentration sensing research[J]. Journal of Beijing Institute of Technology, 2017,26(2):267-275.
  • 加载中

Catalog

    通讯作者:陈斌, bchen63@163.com
    • 1.

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Article Metrics

    Article views (424) PDF downloads(266) Cited by()
    Proportional views
    Related

    /

      Return
      Return
        Baidu
        map