TY - GEN
T1 - Large margin nearest neighbor embedding for knowledge representation
AU - Fan, Miao
AU - Zhou, Qiang
AU - Zheng, Thomas Fang
AU - Grishman, Ralph
N1 - Funding Information:
This work is supported by China Scholarship Council, National Program on Key Basic Research Project (973 Program) under Grant 2013CB329304, National Science Foundation of China (NSFC) under Grant No.61433018 and No.61373075, Proteus Project of NYU. The first author conducted this research while he was a joint-supervision Ph.D. student of Tsinghua University and New York University.
Publisher Copyright:
© 2015 IEEE.
PY - 2016/2/2
Y1 - 2016/2/2
N2 - Traditional way of storing facts in triplets (headentity, relation, tail entity), abbreviated as (h, r, t), allows the knowledge to be intuitively displayed and easily acquired by human beings, but hardly computed or even reasoned about by AI machines. Inspired by the success in applying Distributed Representations to AI-related fields, recent studies expect to represent each entity and relation with a unique lowdimensional embedding, which is different from the symbolic and atomic framework of displaying knowledge in triplets. In this way, the knowledge computing and reasoning can be essentially facilitated by means of a simple vector calculation, i.e. h + r ≈ t. We thus contribute an effective model to learn better embeddings satisfying the formula by pulling the positive tail entities t+ together and close to h + r (Nearest Neighbor), and simultaneously pushing the negatives t- away from the positives t+ via keeping a Large Margin. We also design a corresponding learning algorithm to efficiently find the optimal solution based on Stochastic Gradient Descent in iterative fashion. Quantitative experiments illustrate that our approach can achieve the state-of-the-art performance, compared with several recent methods on some benchmark datasets for two classical applications, i.e. Link prediction and Triplet classification. Moreover, we analyze the parameter complexities among all the evaluated models, and analytical results indicate that our model needs fewer computational resources while outperforming the other methods.
AB - Traditional way of storing facts in triplets (headentity, relation, tail entity), abbreviated as (h, r, t), allows the knowledge to be intuitively displayed and easily acquired by human beings, but hardly computed or even reasoned about by AI machines. Inspired by the success in applying Distributed Representations to AI-related fields, recent studies expect to represent each entity and relation with a unique lowdimensional embedding, which is different from the symbolic and atomic framework of displaying knowledge in triplets. In this way, the knowledge computing and reasoning can be essentially facilitated by means of a simple vector calculation, i.e. h + r ≈ t. We thus contribute an effective model to learn better embeddings satisfying the formula by pulling the positive tail entities t+ together and close to h + r (Nearest Neighbor), and simultaneously pushing the negatives t- away from the positives t+ via keeping a Large Margin. We also design a corresponding learning algorithm to efficiently find the optimal solution based on Stochastic Gradient Descent in iterative fashion. Quantitative experiments illustrate that our approach can achieve the state-of-the-art performance, compared with several recent methods on some benchmark datasets for two classical applications, i.e. Link prediction and Triplet classification. Moreover, we analyze the parameter complexities among all the evaluated models, and analytical results indicate that our model needs fewer computational resources while outperforming the other methods.
UR - http://www.scopus.com/inward/record.url?scp=85028348338&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85028348338&partnerID=8YFLogxK
U2 - 10.1109/WI-IAT.2015.125
DO - 10.1109/WI-IAT.2015.125
M3 - Conference contribution
AN - SCOPUS:85028348338
T3 - Proceedings - 2015 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology, WI-IAT 2015
SP - 53
EP - 59
BT - Proceedings - 2015 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology, WI-IAT 2015
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2015 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology Workshops, WI-IAT Workshops 2015
Y2 - 6 December 2015 through 9 December 2015
ER -