Page 244 - 《软件学报》2025年第12期
P. 244
郑修林 等: 知识图谱补全技术及应用 5625
of the Association for Computational Linguistics and the 7th Int’l Joint Conf. on Natural Language Processing. Beijing: Association for
Computational Linguistics, 2015. 687–696. [doi: 10.3115/v1/P15-1067]
[83] Yoon HG, Song HJ, Park SB, Park SY. A translation-based knowledge graph embedding preserving logical property of relations. In:
Proc. of the 2016 Conf. of the North American Chapter of the Association for Computational Linguistics: Human Language
Technologies. San Diego: Association for Computational Linguistics, 2016. 907–916. [doi: 10.18653/v1/n16-1105]
[84] Do K, Tran T, Venkatesh S. Knowledge graph embedding with multiple relation projections. In: Proc. of the 24th Int’l Conf. on Pattern
Recognition. Beijing: IEEE, 2018. 332–337.
[85] Fan M, Zhou Q, Chang E, Zheng TF. Transition-based knowledge graph embedding with relational mapping properties. In: Proc. of the
28th Pacific Asia Conf. on Language, Information and Computing. Phuket: Department of Linguistics, Chulalongkorn University, 2014.
328–337.
[86] Xie QZ, Ma XZ, Dai ZH, Hovy E. An interpretable knowledge transfer model for knowledge base completion. In: Proc. of the 55th
Annual Meeting of the Association for Computational Linguistics. Vancouver: Association for Computational Linguistics, 2017.
950–962. [doi: 10.18653/v1/P17-1088]
[87] Nguyen DQ, Sirts K, Qu LZ, Johnson M. Neighborhood mixture model for knowledge base completion. In: Proc. of the 20th SIGNLL
Conf. on Computational Natural Language Learning. Berlin: Association for Computational Linguistics, 2016. 40–50. [doi: 10.18653/v1/
K16-1005]
[88] Wang PF, Han JL, Li CL, Pan R. Logic attention based neighborhood aggregation for inductive knowledge graph embedding. In: Proc.
of the 33rd AAAI Conf. on Artificial Intelligence. Honolulu: AAAI, 2019. 7152–7159. [doi: 10.1609/aaai.v33i01.33017152]
[89] Li WD, Zhang XY, Wang YQ, Yan ZH, Peng R. Graph2Seq: Fusion embedding learning for knowledge graph completion. IEEE
Access, 2019, 7: 157960–157971. [doi: 10.1109/ACCESS.2019.2950230]
[90] Zhang Z, Zhuang FZ, Zhu HS, Shi ZP, Xiong H, He Q. Relational graph neural network with hierarchical attention for knowledge graph
completion. In: Proc. of the 34th AAAI Conf. on Artificial Intelligence. New York: AAAI, 2020. 9612–9619. [doi: 10.1609/aaai.v34i05.
6508]
[91] Oh B, Seo S, Lee KH. Knowledge graph completion by context-aware convolutional learning with multi-hop neighborhoods. In: Proc.
of the 27th ACM Int’l Conf. on Information and Knowledge Management. Torino: ACM, 2018. 257–266. [doi: 10.1145/3269206.
3271769]
[92] Ferré S. Link prediction in knowledge graphs with concepts of nearest neighbours. In: Proc. of the 16th Int’l Conf. on the Semantic
Web. Portorož: Springer, 2019. 84–100. [doi: 10.1007/978-3-030-21348-0_6]
[93] Borrego A, Ayala D, Hernández I, Rivero CR, Ruiz D. CAFE: Knowledge graph completion using neighborhood-aware features.
Engineering Applications of Artificial Intelligence, 2021, 103: 104302. [doi: 10.1016/j.engappai.2021.104302]
[94] Yin WP, Yaghoobzadeh Y, Schütze H. Recurrent one-hop predictions for reasoning over knowledge graphs. In: Proc. of the 27th Int’l
Conf. on Computational Linguistics. Santa Fe: Association for Computational Linguistics, 2018. 2369–2378.
[95] Lao N, Mitchell T, Cohen WW. Random walk inference and learning in a large scale knowledge base. In: Proc. of the 2011 Conf. on
Empirical Methods in Natural Language Processing. Edinburgh: Association for Computational Linguistics, 2011. 529–539.
[96] Gardner M, Talukdar P, Krishnamurthy J, Mitchell T. Incorporating vector space similarity in random walk inference over knowledge
bases. In: Proc. of the 2014 Conf. on Empirical Methods in Natural Language Processing (EMNLP). Doha: Association for
Computational Linguistics, 2014. 397–406. [doi: 10.3115/v1/D14-1044]
[97] Neelakantan A, Roth B, McCallum A. Compositional vector space models for knowledge base completion. In: Proc. of the 53rd Annual
Meeting of the Association for Computational Linguistics and the 7th Int’l Joint Conf. on Natural Language Processing. Beijing:
Association for Computational Linguistics, 2015. 156–166. [doi: 10.3115/v1/P15-1016]
[98] Lin YK, Liu ZY, Luan HB, Sun MS, Rao SW, Liu S. Modeling relation paths for representation learning of knowledge bases. In: Proc.
of the 2015 Conf. on Empirical Methods in Natural Language Processing (EMNLP). Lisbon: Association for Computational Linguistics,
2015. 705–714.
[99] Garcia-Durán A, Bordes A, Usunier N, Grandvalet Y. Combining two and three-way embeddings models for link prediction in
knowledge bases. arXiv:1506.00999, 2015.
[100] Xiong SW, Huang WT, Duan PF. Knowledge graph embedding via relation paths and dynamic mapping matrix. In: Proc. of the 2018 Int’l
Conf. on Conceptual Modeling. Xi’an: Springer, 2018. 106–118. [doi: 10.1007/978-3-030-01391-2_18]
[101] Kuželka O, Davis J. Markov logic networks for knowledge base completion: A theoretical analysis under the MCAR assumption. In:
Proc. of the 35th Uncertainty in Artificial Intelligence Conf. Tel Aviv: AUAI Press, 2019. 1138–1148.
[102] Yang F, Yang ZL, Cohen WW. Differentiable learning of logical rules for knowledge base reasoning. In: Proc. of the 31st Int’l Conf. on

