References(49)
[1]
H. Wang, C. Focke, R. Sylvester, N. Mishra, and W. Wang, Fine-tune Bert for DocRED with two-step process, arXiv preprint arXiv: 1909.11898, 2019.
[2]
H. Tang, Y. Cao, Z. Zhang, J. Cao, F. Fang, S. Wang, and P. Yin, HIN: Hierarchical inference network for document-level relation extraction, in Proc. of the Advances in Knowledge Discovery and Data Mining: 24th Pacific-Asia Conf., Singapore, 2020, pp. 197–209.
[3]
X. Han and L. Wang, A novel document-level relation extraction method based on BERT and entity information, IEEE Access, vol. 8, pp. 96912–96919, 2020.
[4]
M. Eberts and A. Ulges, An end-to-end model for entity-level relation extraction using multi-instance learning, in Proc. 16th Conf. European Chapter of the Association for Computational Linguistics: Main Volume, Virtual event, 2021, pp. 3650–3660.
[5]
H. Wang, K. Qin, G. Lu, J. Yin, R. Y. Zakari, and J. W. Owusu, Document-level relation extraction using evidence reasoning on RST-GRAPH, Knowl.-Based Syst., vol. 228, p. 107274, 2021.
[6]
C. Yuan, H. Huang, C. Feng, G. Shi, and X. Wei, Document-level relation extraction with Entity-Selection Attention, Inform. Sci., vol. 568, pp. 163–174, 2021.
[7]
R. Li, J. Zhong, Z. Xue, Q. Dai, and X. Li, Heterogenous affinity graph inference network for document-level relation extraction, Knowl.-Based Syst., vol. 250, p. 109146, 2022.
[8]
R. C. Bunescu and R. J. Mooney, A shortest path dependency kernel for relation extraction, in Proc. Human Language Technology Conf. and Conf. Empirical Methods in Natural Language Processing, Vancouver, Canada, 2005, pp. 724–731.
[9]
Y. Xu, L. Mou, G. Li, Y. Chen, H. Peng, and Z. Jin, Classifying relations via long short term memory networks along shortest dependency paths, in Proc. 2015 Conf. Empirical Methods in Natural Language Processing, Lisbon, Portugal, 2015, pp. 1785–1794.
[10]
P. Gupta, S. Rajaram, H. Schütze, and T. Runkler, Neural relation extraction within and across sentence boundaries, in Proc. AAAI Conf. Artif. Intell., vol. 33, no. 1, pp. 6513–6520, 2019.
[11]
Y. Yao, D. Ye, P. Li, X. Han, Y. Lin, Z. Liu, Z. Liu, L. Huang, J. Zhou, and M. Sun, DocRED: A large-scale document-level relation extraction dataset, in Proc. 57th Annu. Meeting of the Association for Computational Linguistics, Florence, Italy, 2019, pp. 764–777.
[12]
W. C. Mann and S. A. Thompson, Rhetorical structure theory: Toward a functional theory of text organization, Text-Interdisciplinary Journal for the Study of Discourse, vol. 8, no. 3, pp. 243–281, 1988.
[13]
Z. Guo, Y. Zhang, Z. Teng, and W. Lu, Densely connected graph convolutional networks for graph-to-sequence learning, Transactions of the Association for Computational Linguistics, vol. 7, pp. 297–312, 2019.
[14]
K. Xu, Y. Feng, S. Huang, and D. Zhao, Semantic relation classification via convolutional neural networks with simple negative sampling, in Proc. 2015 Conf. Empirical Methods in Natural Language Processing, Lisbon, Portugal, 2015, pp. 536–540.
[15]
Y. Xu, R. Jia, L. Mou, G. Li, Y. Chen, Y. Lu, and Z. Jin, Improved relation classification by deep recurrent neural networks with data augmentation, in Proc. COLING 2016, the 26th Int. Conf. Computational Linguistics: Technical Papers, Osaka, Japan, 2016, pp. 1461–1470.
[16]
R. Cai, X. Zhang, and H. Wang, Bidirectional recurrent convolutional neural network for relation classification, in Proc. 54thAnnu. Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Berlin, Germany, 2016, pp. 756–765.
[17]
C. Zhang, C. Cui, S. Gao, X. Nie, W. Xu, L. Yang, X. Xi, and Y. Yin, Multi-gram CNN-based self-attention model for relation classification, IEEE Access, vol. 7, pp. 5343–5357, 2019.
[18]
W. Zhou, K. Huang, T. Ma, and J. Huang, Document-level relation extraction with adaptive thresholding and localized context pooling, Proc. AAAI Conf. Artif. Intell., vol. 35, no. 16, pp. 14612–14620, 2021.
[19]
F. Xue, A. Sun, H. Zhang, J. Ni, and E. S. Chng, An embarrassingly simple model for dialogue relation extraction, in Proc. of ICASSP 2022–2022 IEEE Int. Conf. Acoustics, Speech and Signal Processing (ICASSP), Singapore, 2022, pp. 6707–6711.
[20]
G. Nan, Z. Guo, I. Sekulic, and W. Lu, Reasoning with latent structure refinement for document-level relation extraction, in Proc. 58th Annu. Meeting of the Association for Computational Linguistics, Virtual event, 2020, pp. 1546–1557.
[21]
S. Zeng, R. Xu, B. Chang, and L. Li, Double graph based reasoning for document-level relation extraction, in Proc. 2020 Conf. Empirical Methods in Natural Language Processing (EMNLP), Virtual event, 2020, pp. 1630–1640.
[22]
B. Li, W. Ye, Z. Sheng, R. Xie, X. Xi, and S. Zhang, Graph enhanced dual attention network for document-level relation extraction, in Proc. 28th Int. Conf. Computational Linguistics, Barcelona, Spain, 2020, pp. 1551–1560.
[23]
F. Christopoulou, M. Miwa, and S. Ananiadou, Connecting the dots: Document-level neural relation extraction with edge-oriented graphs, in Proc. 2019 Conf. Empirical Methods in Natural Language Processing and the 9th International Joint Conf. Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 2019, pp. 4925–4936.
[24]
F. Xue, A. Sun, H. Zhang, and E. S. Chng, GDPNet: Refining latent multi-view graph for relation extraction, Proc. AAAI Conf. Artif. Intell., vol. 35, no. 16, pp. 14194–14202, 2021.
[25]
R. C. Bunescu and R. J. Mooney, Subsequence kernels for relation extraction, in Proc. 18th Int. Conf. Neural Information Processing Systems, Vancouver, British Columbia, Canada, 2005, pp. 171–178.
[26]
F. Ren, D. Zhou, Z. Liu, Y. Li, R. Zhao, Y. Liu, and X. Liang, Neural relation classification with text descriptions, in Proc. 27th Int. Conf. Computational Linguistics, Santa Fe, NM, USA, 2018, pp. 1167–1177.
[27]
H. Wang, K. Qin, G. Lu, G. Luo, and G. Liu, Direction-sensitive relation extraction using Bi-SDP attention model, Knowl.-Based Syst., vol. 198, p. 105928, 2020.
[28]
X. Wang, H. Wang, C. Li, T. Huang, and J. Kurths, Improved consensus conditions for multi-agent systems with uncertain topology: The generalized transition rates case, IEEE Trans. Netw. Sci. Eng., vol. 7, no. 3, pp. 1158–1169, 2020.
[29]
C. Quirk and H. Poon, Distant supervision for relation extraction beyond the sentence boundary, in Proc. 15th Conf. European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), Valencia, Spain, 2017, pp. 1171–1182.
[30]
N. Peng, H. Poon, C. Quirk, K. Toutanova, and W. T. Yih, Cross-sentence n-ary relation extraction with graph LSTMs, Transactions of the Association for Computational Linguistics, vol. 5, pp. 101–115, 2017.
[31]
T. Hirao, Y. Yoshida, M. Nishino, N. Yasuda, and M. Nagata, Single-document summarization as a tree knapsack problem, in Proc. 2013 Conf. Empirical Methods in Natural Language Processing, Seattle, WA, USA, 2013, pp. 1515–1520.
[32]
P. Jansen, M. Surdeanu, and P. Clark, Discourse complements lexical semantics for non-factoid answer reranking, in Proc. 52nd Annu. Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Baltimore, MD, 2014, pp. 977–986.
[33]
Y. Yoshida, J. Suzuki, T. Hirao, and M. Nagata, Dependency-based discourse parser for single-document summarization, in Proc. 2014 Conf. Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 2014, pp. 1834–1839.
[34]
Z. Liu and N. Chen, Exploiting discourse-level segmentation for extractive summarization, in Proc. 2nd Workshop on New Frontiers in Summarization, Hong Kong, China, 2019, pp. 116–121.
[35]
X. Tan, L. Zhang, F. Kong, and G. Zhou, Towards discourse-aware document-level neural machine translation, in Proc. 31st Int. Joint Conf. Artificial Intelligence, Vienna, Austria, 2022, pp. 4383–4389.
[36]
W. Zhou and M. Chen, An improved baseline for sentence-level relation extraction, arXiv preprint arXiv: 2102.01373, 2022.
[37]
R. Jia, C. Wong, and H. Poon, Document-level n-ary relation extraction with multiscale representation learning, in Proc. 2019 Conf. North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, 2019, pp. 3693–3704.
[38]
Y. Ji and J. Eisenstein, Representation learning for text-level discourse parsing, in Proc. 52nd Annu. Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Baltimore, MD, USA, 2014, pp. 13–24.
[39]
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, Attention is all you need, in Proc. 31st Int. Conf. Neural Information Processing Systems, Long Beach, CA, USA, 2017, pp. 6000–6010.
[40]
N. Zhang, X. Chen, X. Xie, S. Deng, C. Tan, M. Chen, F. Huang, L. Si, and H. Chen, Document-level relation extraction as semantic segmentation, in Proc. 30th Int. Joint Conf. Artificial Intelligence, Montreal, Canada, 2021, pp. 3999–4006.
[41]
J. Li, Y. Sun, R. J. Johnson, D. Sciaky, C. H. Wei, R. Leaman, A. P. Davis, C. J. Mattingly, T. C. Wiegers, and Z. Lu, Biocreative V CDR task corpus: A resource for chemical disease relation extraction, Database, vol. 2016, p. baw068, 2016.
[42]
Y. Wu, R. Luo, H. C. M. Leung, H. F. Ting, and T. W. Lam, RENET: A deep learning approach for extracting gene-disease associations from literature, in Proc. 23rd Annu. Int. Conf. Research in Computational Molecular Biology, Washington, DC, USA, 2019, pp. 272–284.
[43]
Y. Gu, R. Tinn, H. Cheng, M. Lucas, N. Usuyama, X. Liu, T. Naumann, J. Gao, and H. Poon, Domain-specific language model pretraining for biomedical natural language processing, ACM Trans. Comput. Healthcare, vol. 3, no. 1, p. 2, 2022.
[44]
Z. Guo, Y. Zhang, and W. Lu, Attention guided graph convolutional networks for relation extraction, in Proc. 57th Annu. Meeting of the Association for Computational Linguistics, Florence, Italy, 2019, pp. 241–251.
[45]
D. Ye, Y. Lin, J. Du, Z. Liu, P. Li, M. Sun, and Z. Liu, Coreferential reasoning learning for language representation, in Proc. 2020 Conf. Empirical Methods in Natural Language Processing (EMNLP), Virtual event, 2020, pp. 7170–7186.
[46]
B. Xu, Q. Wang, Y. Lyu, Y. Zhu, and Z. Mao, Entity structure within and throughout: Modeling mention dependencies for document-level relation extraction, Proc. AAAI Conf. Artif. Intell., vol. 35, no. 16, pp. 14149–14157, 2021.
[47]
P. Verga, E. Strubell, and A. McCallum, Simultaneously self-attending to all mentions for full-abstract biological relation extraction, in Proc. 2018 Conf. North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), New Orleans, LA, 2018, pp. 872–884.
[48]
I. Beltagy, K. Lo, and A. Cohan, SciBERT: A pretrained language model for scientific text, in Proc. 2019 Conf. Empirical Methods in Natural Language Processing and the 9th Int. Joint Conf. Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 2019, pp. 3615–3620.
[49]
D. Wang, W. Hu, E. Cao, and W. Sun, Global-to-local neural networks for document-level relation extraction, in Proc. 2020 Conf. Empirical Methods in Natural Language Processing (EMNLP), Virtual event, 2020, pp. 3711–3721.