[1] RAJPURKAR P, JIA R, LIANG P. Know what you don’t know: unanswerable questions for SQuAD[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2018: 784-789.
[2] ZELLERS R, BISK Y, SCHWARTZ R, et al. SWAG: a large-scale adversarial dataset for grounded commonsense inference[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018: 93-104.
[3] GREFF K, SRIVASTAVA R K, KOUTNíK J, et al. LSTM: a search space odyssey[J]. IEEE Transactions on Neural Networks & Learning Systems, 2016, 28(10): 2222-2232.
[4] VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017: 6000-6010.
[5] DEVLIN J, CHANG M W, LEE K, et al. BERT: pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 2018: 4171-4186.
[6] MIKOLOV T, CORRADO G, KAI C, et al. Efficient estimation of word representations in vector space[C]//International Conference on Learning Representations, 2013.
[7] PENNINGTON J, SOCHER R, MANNING C D. Glove: global vectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2014: 1532-1543.
[8] TALMOR A, HERZIG J, LOURIE N, et al. CommonsenseQA: a question answering challenge targeting commonsense knowledge[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 2019: 4149-4158.
[9] SPEER R, CHIN J, HAVASI C. ConceptNet 5.5: an open multilingual graph of general knowledge[C]//Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 2017: 4444-4451.
[10] XU Y, ZHU C, WANG S, et al. Human parity on commonsenseQA: augmenting self-attention with external attention[C]//Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence Main Track, 2021: 2762-2768.
[11] CLARK K, LUONG M T, LE Q V, et al. ELECTRA: pre-training text encoders as discriminators rather than generators[C]//2020 International Conference on Learning Representations, 2020.
[12] LIN B Y, CHEN X, CHEN J, et al. KagNet: knowledge-aware graph networks for commonsense reasoning[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019: 2829-2839.
[13] YASUNAGA M, REN H, BOSSELUT A, et al. QA-GNN: reasoning with language models and knowledge graphs for question answering[C]//Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021: 535-546.
[14] CUI W, XIAO Y, WANG H, et al. KBQA: learning question answering over QA corpora and knowledge bases[J]. Proceedings of the VLDB Endowment, 2017, 10(5): 565-576.
[15] WU Y, LIU X, FENG Y, et al. Relation-aware entity alignment for heterogeneous knowledge graphs[C]//Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, 2019: 5278-5284.
[16] YASUNAGA M, BOSSELUT A, REN H, et al. Deep bidirectional language?knowledge graph pretraining[C]//36th Conference on Neural Information Processing Systems (NeurIPS 2022), 2022.
[17] LV S, GUO D, XU J, et al. Graph-based reasoning over heterogeneous external knowledge for commonsense question answering[C]//Proceedings of the AAAI Conference on Artificial Intelligence, 2020: 8449-8456.
[18] RAJANI N F, MCCANN B, XIONG C, et al. Explain Yourself! Leveraging language models for commonsense reasoning[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 4932-4942.
[19] ZELIKMAN E, WU Y, MU J, et al. Star: bootstrapping reasoning with reasoning[C]//Advances in Neural Information Processing Systems, 2022: 15476-15488.
[20] WU S, IRSOY O, LU S, et al. Bloomberggpt: a large language model for finance[J]. arXiv:2303.17564, 2023.
[21] LAN Z, CHEN M, GOODMAN S, et al. ALBERT: a lite bert for self-supervised learning of language representations[C]//2020 International Conference on Learning Representations, 2019.
[22] ZESCH T, MüLLER C, GUREVYCH I. Extracting lexical semantic knowledge from wikipedia and wiktionary[C]//Proceedings of the International Conference on Language Resources and Evaluation (LREC 2008), Marrakech, Morocco, 2008: 1646-1652.
[23] SU J, CAO J, LIU W, et al. Whitening sentence representations for better semantics and faster retrieval[J]. arXiv:2103.15316, 2021. |