456
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Exploring graph representation strategies for text classification

ORCID Icon, ORCID Icon & ORCID Icon
Article: 2289832 | Received 16 Apr 2023, Accepted 27 Nov 2023, Published online: 21 Dec 2023

References

  • Adel'son-Velskii, G. M., & Landis, E. M. (1962). An algorithm for the organization of information. Doklady Akademii Nauk, 146, 263–266.
  • Ayed, R., Labidi, M., & Maraoui, M. (2017). Arabic text classification: New study. 2017 International Conference on Engineering & mis (Icemis), 1–7.
  • Bayer, R. (1972). Symmetric binary b-trees: Data structure and maintenance algorithms. Acta Informatica, 1(4), 290–306. https://doi.org/10.1007/BF00289509
  • Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent dirichlet allocation. Journal of Machine Learning Research, 3(Jan), 993–1022.
  • Broder, A. Z., Glassman, S. C., Manasse, M. S., & Zweig, G. (1997). Syntactic clustering of the web. Computer Networks and ISDN Systems, 29(8-13), 1157–1166. https://doi.org/10.1016/S0169-7552(97)00031-7
  • Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., & Agarwal, S. (2020). Language models are few-shot learners. In H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, & H. Lin (Eds.), Advances in neural information processing systems (Vol. 33, pp. 1877–1901). Curran Associates, Inc. https://proceedings.neurips.cc/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf
  • Bugueño, M., & de Melo, G. (2023). Connecting the dots: What graph-based text representations work best for text classification using graph neural networks?
  • Cañete, J., Chaperon, G., Fuentes, R., Ho, J.-H., Kang, H., & Pérez, J. (2020). Spanish pre-trained bert model and evaluation data. In Pml4dc at iclr 2020.
  • Cao, W., Yan, Z., He, Z., & He, Z. (2020). A comprehensive survey on geometric deep learning. IEEE Access, 8, 35929–35949. https://doi.org/10.1109/ACCESS.2020.2975067
  • Chan, B., Schweter, S., & Möller, T. (2020, December). German’s next language model. In Proceedings of the 28th international conference on computational linguistics (pp. 6788–6796). International Committee on Computational Linguistics. https://aclanthology.org/2020.coling-main.598
  • Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019, June). BERT: Pre-training of deep bidi- rectional transformers for language understanding. In Proceedings of the 2019 conference of the north American chapter of the association for computational linguistics: Human language technolo- gies, volume 1 (long and short papers) (pp. 4171–4186). Association for Computational Linguistics. https://www.aclweb.org/anthology/N19-1423
  • Ding, K., Wang, J., Li, J., Li, D., & Liu, H. (2020, November). Be more with less: Hypergraph attention networks for inductive text classification. In In Proceedings of the 2020 conference on empirical methods in natural language processing (emnlp) (pp. 4927–4936). Association for Computational Linguistics. https://aclanthology.org/2020.emnlp-main.399
  • Ehrenfried, H. V., Todt, E., (2022). Should i buy or should i pass: E-commerce datasets in Portuguese. In V. Pinheiro et al. (Ed.), Computational processing of the Portuguese language (pp. 420–425). Springer International Publishing.
  • Gori, M., Monfardini, G., & Scarselli, F. (2005). A new model for learning in graph domains. In Proceedings. 2005 Ieee International Joint Conference on Neural Networks, 2005. (Vol. 2, p. 729-734 vol. 2). https://doi.org/10.1109/IJCNN.2005.1555942.
  • Gupta, P., Pagliardini, M., & Jaggi, M. (2019). Better word embeddings by disentangling contextual n-gram information. In NAACL-HLT (1) (pp. 933–939). Association for Computational Linguistics.
  • Hakimi, S. L. (1962). On realizability of a set of integers as degrees of the vertices of a linear graph. i. Journal of the Society for Industrial and Applied Mathematics, 10(3), 496–506. Retrieved 2022-11-22, from http://www.jstor.org/stable/2098746
  • Haonan, L., Huang, S. H., Ye, T., & Xiuyan, G. (2019). Graph star net for generalized multi-task learning.
  • Harris, Z. S. (1954). Distributional structure. WORD, 10(2-3), 146–162. https://doi.org/10.1080/00437956.1954.11659520
  • Havel, V. (1955). Poznámka o existenci konecných grafu. Casopis pro peˇstování matematiky, 080(4), 477-480. http://eudml.org/doc/19050
  • Honnibal, M., Montani, I., Van Landeghem, S., & Boyd, A. (2020). spaCy: Industrial-strength Natural Language Processing in Python. Zenodo.
  • Hosmer, D. W., Lemeshow, S., & Sturdivant, R. X. (2013). Assessing the fit of the model. In Applied logistic regression (pp. 153–225). John Wiley & Sons, Ltd. https://onlinelibrary.wiley.com/doi/abs/10.10029781118548387.ch5
  • Huang, L., Ma, D., Li, S., Zhang, X., & Wang, H. (2019, November). Text level graph neural network for text classification. In Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (emnlp-ijcnlp) (pp. 3444–3450). Hong Kong, People’s Republic of China: Association for Computational Linguistics. https://aclanthology.org/D19-1345
  • Huang, Y.-H., Chen, Y.-H., & Chen, Y.-S. (2022, October). ConTextING: Granting document-wise contextual embeddings to graph neural networks for inductive text classification. In Proceedings of the 29th International Conference on Computational Linguistics (pp. 1163–1168). Gyeongju, Republic of Korea: International Committee on Computational Linguistics. https://aclanthology.org/2022.coling-1.100
  • Joachims, T. (1998). Text categorization with support vector machines: Learning with many relevant features. In C. Nédellec, & C. Rouveirol (Eds.), Machine learning: Ecml-98 (pp. 137–142). Springer Berlin Heidelberg.
  • Joshi, M., Chen, D., Liu, Y., Weld, D. S., Zettlemoyer, L., & Levy, O. (2020). Spanbert: Improving pre-training by representing and predicting spans. Transactions of the Association for Computational Linguistics, 8, 64–77. https://doi.org/10.1162/tacl_a_00300
  • Joulin, A., Grave, E., Bojanowski, P., & Mikolov, T. (2016). Bag of tricks for efficient text classification. arXiv preprint arXiv:1607.01759.
  • Khodak, M., Saunshi, N., Liang, Y., Ma, T., Stewart, B., & Arora, S. (2018, July). A la carte embedding: Cheap but effective induction of semantic feature vectors. In Proceedings of the 56th annual meeting of the association for computational linguistics (volume 1: Long papers) (pp. 12–22). Association for Computational Linguistics. https://aclanthology.org/P18-1002
  • Kipf, T. N., & Welling, M. (2017). Semi-supervised classification with graph convolutional networks. International Conference on Learning Representations (iclr).
  • Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., & Soricut, R. (2019). Albert: A lite bert for self-supervised learning of language representations. arXiv. https://arxiv.org/abs/1909.11942
  • Lin, Y., Meng, Y., Sun, X., Han, Q., Kuang, K., Li, J., & Wu, F. (2021). BertGCN: Transductive Text Classification by Combining GCN and BERT. arXiv preprint arXiv:2105.05727. http://arxiv.org/abs/2105.05727
  • Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., & Stoyanov, V. (2019). Roberta: A robustly optimized bert pretraining approach. arXiv. https://arxiv.org/abs/1907.11692
  • Lu, Z., Du, P., & Nie, J.-Y. (2020). Vgcn-bert: augmenting bert with graph embedding for text classification. In Advances in information retrieval: 42nd European conference on ir research, ecir 2020, lisbon, portugal, april 14–17, 2020, proceedings, part i 42 (pp. 369–382).
  • Malliaros, F. D., & Skianis, K. (2015, aug). Graph-based term weighting for text categorization. In Proceedings of the 2015 Ieee/acm International Conference on Advances in Social Networks Analysis and Mining, Asonam 2015 (pp. 1473–1479). Association for Computing Machinery, Inc.
  • Mei, X., Cai, X., Yang, L., & Wang, N. (2021). Graph transformer networks based text representation. Neurocomputing, 463, 91–100. https://www.sciencedirect.com/science/article/pii/S0925231221012169
  • Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.
  • Moschitti, A., & Basili, R. (2004). Complex linguistic features for text classification: A comprehensive study. In S. McDonald, & J. Tait (Eds.), Advances in information retrieval (pp. 181–196). Springer Berlin Heidelberg.
  • Nguyen, D. Q., Nguyen, T. D., & Phung, D. (2022). Universal graph transformer self-attention networks. In Companion Proceedings of the web Conference 2022 (www ‘22 companion).
  • Pang, B., & Lee, L. (2005). Seeing stars: Exploiting class relationships for sentiment categorization with respect to rating scales. In Proceedings of the 43rd annual meeting on association for computational linguistics (pp. 115–124). Association for Computational Linguistics. https://doi.org/10.3115/1219840.1219855.
  • Pennington, J., Socher, R., & Manning, C. D. (2014). Glove: Global vectors for word representation. In emnlp (Vol. 14, pp. 1532–1543). https://nlp.stanford.edu/pubs/glove.pdf
  • Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., & Zettlemoyer, L. (2018, June). Deep contextualized word representations. In Proceedings of the 2018 conference of the north American chapter of the association for computational linguistics: Human language technologies, volume 1 (long papers) (pp. 2227–2237). New Orleans, Louisiana: Association for Computational Linguistics. https://aclanthology.org/N18-1202
  • Risch, J., Stoll, A., Wilms, L., & Wiegand, M. (2021, September). Overview of the GermEval 2021 shared task on the identification of toxic, engaging, and fact-claiming comments. In Proceedings of the germeval 2021 shared task on the identification of toxic, engaging, and fact-claiming comments (pp. 1–12). Association for Computational Linguistics. https://aclanthology.org/2021.germeval-1.1
  • Rousseau, F., & Vazirgiannis, M. (2013). Graph-of-word and tw-idf: New approach to ad hoc ir. In Proceedings of the 22nd acm international conference on information; knowledge management (pp. 59–68). Association for Computing Machinery. https://doi.org/10.1145/2505515.2505671.
  • Salton, G. (1989). Automatic text processing: The transformation, analysis, and retrieval of information by computer. Addison-Wesley Longman Publishing Co., Inc.
  • Shukri, M. (2019). How the embedding layers in bert were implemented. https://medium.com/@_init_/why-bert-has-3-embedding-layers-and-their-implementation-details-9c261108e28a (Accessed in July 16th, 2021).
  • Souza, F., Nogueira, R., & Lotufo, R. (2020). Bertimbau: Pretrained bert models for Brazilian Portuguese. In R. Cerri, & R. C. Prati (Eds.), Intelligent systems (pp. 403–417). Springer International Publishing.
  • Sun, Y., Wang, S., Li, Y., Feng, S., Chen, X., Zhang, H., Tian, X., Zhu, D., Tian, H., & Wu, H. (2019). ERNIE: enhanced representation through knowledge integration. CoRR, abs/1904.09223. http://arxiv.org/abs/1904.09223
  • Tesnière, L. (1959). Eléments de syntaxe structurale. Klincksieck.
  • Wang, B., Sun, Y., Chu, Y., Min, C., Yang, Z., & Lin, H. (2023, May 29). Local discriminative graph convolutional networks for text classification. Multimedia Systems, https://doi.org/10.1007/s00530-023-01112-y
  • Wang, Y., Feng, L., Liu, A., Wang, W., & Hou, Y. (2023a). Dual bigru-cnn-based sentiment classification method combining global and local attention. The Journal of Supercomputing, https://doi.org/10.1007/s11227-023-05558-9
  • Wang, Y., Wang, C., Zhan, J., Ma, W., & Jiang, Y. (2023b). Text fcg: Fusing contextual information via graph learning for text classification. Expert Systems with Applications, 219, 119658. https://www.sciencedirect.com/science/article/pii/S0957417423001598
  • Wiegand, M., Siegel, M., & Ruppenhofer, J. (2018). Overview of the germeval 2018 shared task on the identification of offensive language. Verlag der Österreichischen Akademie der Wissenschaften.
  • Windley, P. F. (1960). Trees, forests and rearranging. The Computer Journal, 3(2), 84–88. https://doi.org/10.1093/comjnl/3.2.84
  • Wu, F., Souza, A., Zhang, T., Fifty, C., Yu, T., & Weinberger, K. (2019). Simplifying graph convolutional networks. In International Conference on Machine Learning, 6861–6871.
  • Xu, L., Xie, H., Li, Z., Wang, F. L., Wang, W., & Li, Q. (2023). Contrastive learning models for sentence representations. ACM Transactions on Intelligent Systems and Technology, 14(4), https://doi.org/10.1145/3593590
  • Yang, Y., & Cui, X. (2021). Bert-enhanced text graph neural network for classification. Entropy, 23(11), https://www.mdpi.com/1099-4300/23/11/1536
  • Yao, L., Mao, C., & Luo, Y. (2019). Graph convolutional networks for text classification. In Proceedings of the aaai conference on artificial intelligence (Vol. 33, pp. 7370–7377). AAAI Press.
  • Zhang, Y., Yu, X., Cui, Z., Wu, S., Wen, Z., & Wang, L. (2020, July). Every document owns its structure: Inductive text classification via graph neural networks. In Proceedings of the 58th annual meeting of the association for computational linguistics (pp. 334–339). Association for Computational Linguistics. https://aclanthology.org/2020.acl-main.31
  • Zhu, H., & Koniusz, P. (2021). Simple spectral graph convolution. International Conference on Learning representations.