Adapting Neural Link Predictors for Data-Efficient Complex Query Answering

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Standard

Adapting Neural Link Predictors for Data-Efficient Complex Query Answering. / Arakelyan, Erik; Minervini, Pasquale ; Daza, Daniel ; Cochez, Michael ; Augenstein, Isabelle.

37th Conference on Neural Information Processing Systems (NeurIPS 2023).. OpenReview.net, 2024.

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Harvard

Arakelyan, E, Minervini, P, Daza, D, Cochez, M & Augenstein, I 2024, Adapting Neural Link Predictors for Data-Efficient Complex Query Answering. in 37th Conference on Neural Information Processing Systems (NeurIPS 2023).. OpenReview.net, 37th Conference on Neural Information Processing Systems - NeurIPS 2023, New Orleans., United States, 10/12/2023. <https://openreview.net/forum?id=1G7CBp8o7L>

APA

Arakelyan, E., Minervini, P., Daza, D., Cochez, M., & Augenstein, I. (2024). Adapting Neural Link Predictors for Data-Efficient Complex Query Answering. In 37th Conference on Neural Information Processing Systems (NeurIPS 2023). OpenReview.net. https://openreview.net/forum?id=1G7CBp8o7L

Vancouver

Arakelyan E, Minervini P, Daza D, Cochez M, Augenstein I. Adapting Neural Link Predictors for Data-Efficient Complex Query Answering. In 37th Conference on Neural Information Processing Systems (NeurIPS 2023).. OpenReview.net. 2024

Author

Arakelyan, Erik ; Minervini, Pasquale ; Daza, Daniel ; Cochez, Michael ; Augenstein, Isabelle. / Adapting Neural Link Predictors for Data-Efficient Complex Query Answering. 37th Conference on Neural Information Processing Systems (NeurIPS 2023).. OpenReview.net, 2024.

Bibtex

@inproceedings{b2232ea9f11140e88d47af3a41336d26,
title = "Adapting Neural Link Predictors for Data-Efficient Complex Query Answering",
abstract = "Answering complex queries on incomplete knowledge graphs is a challenging task where a model needs to answer complex logical queries in the presence of missing knowledge. Prior work in the literature has proposed to address this problem by designing architectures trained end-to-end for the complex query answering task with a reasoning process that is hard to interpret while requiring data and resource-intensive training. Other lines of research have proposed re-using simple neural link predictors to answer complex queries, reducing the amount of training data by orders of magnitude while providing interpretable answers. The neural link predictor used in such approaches is not explicitly optimised for the complex query answering task, implying that its scores are not calibrated to interact together. We propose to address these problems via CQD, a parameter-efficient score \emph{adaptation} model optimised to re-calibrate neural link prediction scores for the complex query answering task. While the neural link predictor is frozen, the adaptation component -- which only increases the number of model parameters by -- is trained on the downstream complex query answering task. Furthermore, the calibration component enables us to support reasoning over queries that include atomic negations, which was previously impossible with link predictors. In our experiments, CQD produces significantly more accurate results than current state-of-the-art methods, improving from to Mean Reciprocal Rank values averaged across all datasets and query types while using of the available training query types. We further show that CQD is data-efficient, achieving competitive results with only of the complex training queries and robust in out-of-domain evaluations. Source code and datasets are available at https://github.com/EdinburghNLP/adaptive-cqd.",
author = "Erik Arakelyan and Pasquale Minervini and Daniel Daza and Michael Cochez and Isabelle Augenstein",
year = "2024",
language = "English",
booktitle = "37th Conference on Neural Information Processing Systems (NeurIPS 2023).",
publisher = "OpenReview.net",
note = "37th Conference on Neural Information Processing Systems - NeurIPS 2023 ; Conference date: 10-12-2023 Through 16-12-2023",

}

RIS

TY - GEN

T1 - Adapting Neural Link Predictors for Data-Efficient Complex Query Answering

AU - Arakelyan, Erik

AU - Minervini, Pasquale

AU - Daza, Daniel

AU - Cochez, Michael

AU - Augenstein, Isabelle

PY - 2024

Y1 - 2024

N2 - Answering complex queries on incomplete knowledge graphs is a challenging task where a model needs to answer complex logical queries in the presence of missing knowledge. Prior work in the literature has proposed to address this problem by designing architectures trained end-to-end for the complex query answering task with a reasoning process that is hard to interpret while requiring data and resource-intensive training. Other lines of research have proposed re-using simple neural link predictors to answer complex queries, reducing the amount of training data by orders of magnitude while providing interpretable answers. The neural link predictor used in such approaches is not explicitly optimised for the complex query answering task, implying that its scores are not calibrated to interact together. We propose to address these problems via CQD, a parameter-efficient score \emph{adaptation} model optimised to re-calibrate neural link prediction scores for the complex query answering task. While the neural link predictor is frozen, the adaptation component -- which only increases the number of model parameters by -- is trained on the downstream complex query answering task. Furthermore, the calibration component enables us to support reasoning over queries that include atomic negations, which was previously impossible with link predictors. In our experiments, CQD produces significantly more accurate results than current state-of-the-art methods, improving from to Mean Reciprocal Rank values averaged across all datasets and query types while using of the available training query types. We further show that CQD is data-efficient, achieving competitive results with only of the complex training queries and robust in out-of-domain evaluations. Source code and datasets are available at https://github.com/EdinburghNLP/adaptive-cqd.

AB - Answering complex queries on incomplete knowledge graphs is a challenging task where a model needs to answer complex logical queries in the presence of missing knowledge. Prior work in the literature has proposed to address this problem by designing architectures trained end-to-end for the complex query answering task with a reasoning process that is hard to interpret while requiring data and resource-intensive training. Other lines of research have proposed re-using simple neural link predictors to answer complex queries, reducing the amount of training data by orders of magnitude while providing interpretable answers. The neural link predictor used in such approaches is not explicitly optimised for the complex query answering task, implying that its scores are not calibrated to interact together. We propose to address these problems via CQD, a parameter-efficient score \emph{adaptation} model optimised to re-calibrate neural link prediction scores for the complex query answering task. While the neural link predictor is frozen, the adaptation component -- which only increases the number of model parameters by -- is trained on the downstream complex query answering task. Furthermore, the calibration component enables us to support reasoning over queries that include atomic negations, which was previously impossible with link predictors. In our experiments, CQD produces significantly more accurate results than current state-of-the-art methods, improving from to Mean Reciprocal Rank values averaged across all datasets and query types while using of the available training query types. We further show that CQD is data-efficient, achieving competitive results with only of the complex training queries and robust in out-of-domain evaluations. Source code and datasets are available at https://github.com/EdinburghNLP/adaptive-cqd.

M3 - Article in proceedings

BT - 37th Conference on Neural Information Processing Systems (NeurIPS 2023).

PB - OpenReview.net

T2 - 37th Conference on Neural Information Processing Systems - NeurIPS 2023

Y2 - 10 December 2023 through 16 December 2023

ER -

ID: 381236392