147
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Event Assignment Based on KBQA for Government Service Hotlines

, &
Article: 2348162 | Received 26 Apr 2023, Accepted 18 Apr 2024, Published online: 16 May 2024

ABSTRACT

Government service hotlines are closely related to people’s lives, and the hotline events assignment task is very important in China, which affects people’s satisfaction with the Chinese government. To address the challenge of improving hotline event assignment accuracy, which is currently hindered by the lack of prior knowledge in the traditional direct departmental assignment approach, we introduce a novel hotline event assignment model based on Knowledge Base Question Answering (KBQA) to leverage prior knowledge and enhance assignment performance. The model extracts event key information by using event extraction module, and the key information of event extracte by this module is used to construct historical events knowledge graph. Then we employ subgraph retrieval module and text retrieval module to obtain relevant prior knowledge from the knowledge graph and “sanding” text respectively. The integrated prior knowledge is then used to predict department scores, guiding the selection of the optimal assignment department through ranking. The experiment results show that the proposed method outperforms existing baseline methods. Meanwhile, the Knowledge Graph (KG) edge elimination experiment also shows that the proposed model is better than other baseline models in terms of incomplete KG.

Introduction

Government service hotline is directly oriented to businesses and the public, and serves as an important channel to report problems, make suggestions, facilitate government decision-making and help solve public service issues. It has played a positive role in addressing pressing concerns for enterprises and people in recent years (Meng, Huang, and Zhang Citation2021). In the past, citizens in China had to dial 12,369 when encountering noise pollution, 12,315 when encountering consumer rights protection, and 12,319 when encountering occupational behaviors. In the past, when reporting problems to relevant departments, the general public need to make different calls by category. In order to simplify the government service hotline system to improve service efficiency and quality, Chinese government has canceled or merged several existing service hotlines at the end of 2021. Nowadays, the public only need to dial only one phone number to perform operations such as consultation, complaints, reflections, and suggestions. This phone number is called “12345,” which is a 24-hour government service hotline. The hotline is an important window for the municipal government in China to connect with the people, and aims to promote the openness of government affairs, to improve administrative efficiency (Ma, Li, and Shen Citation2020).

Chinese government service hotline adopts the manual assignment mode. When receiving a call from a citizen, a hotline operator need to keep textual record of event information contained in a hotline such as event time, event address, event content, and event classification, and then assign each event to a corresponding government department (i.e., determine a department to deal with the event) immediately. Because of the shortage of the service seats compared with the developed countries, the quality of Chinese government service hotline assignment still needs to be improved. For example, at present there is only about 50 operators who work for the Wuhu (a city in China) government hotline day and night, and they deal with 600,000 hotline events every month averagely. After merging several existing service hotlines, it puts forward higher requirements for a hotline operator. The accuracy of manual event assignment depends not only on his exact judgment on the classification of citizens’ demands but also on an in-depth understanding of the functions of various government departments. The textual contents recorded by a hotline operator for different hotline events vary in length; most of them have problems such as unclear description and incomplete elements. Thus, each hotline operator requires strong business skills and personal comprehensive quality, needing not only proficiency in communication and summarization skills, but also familiarity with the division of responsibilities of government departments. Unfortunately, due to the imperfect mechanism of operator training, it is difficult for a hotline operator to select the correct responsible department from dozens of departments to complete event assignment on the basis of fully understanding of citizens’ demands. The second hotline event assignment will occur when it is assigned to a wrong department. It will cause disputes between departments as well as reduce public satisfaction with the government. With more and more calls accepted by Chinese government hotline, it is significant to develop an automatic event assignment approach that can accurately determine the corresponding responsible department with the purpose of helping the hotline operators to improve service quality.

With respect to most government hotline event texts, the sentences in an event text are often described around the same event subject. However, the text often contains multiple types of key elements, and such elements are often discrete and distributed in different locations of the description text. In order to better integrate these types of key information, language models such as BERT-like model are used to discover important semantic information in the event text. Directly encoding the event text often results in the model not being able to capture these key information effectively. Even if the key information in the event text can be captured, using a classification-based model to directly establish associations between events and departments may still fail to perform well, because the semantic information contained in the department name itself is relatively limited, and information about the responsibilities of some departments is not available from their names. Thus, the model cannot effectively find the potential association. Further, after considering the above factors together to build an event assignment model, there is a possibility that by focusing only on the current event, more effective potential information cannot be obtained from the associated events, thus affecting the accuracy of the model.

As a new form of knowledge storage and representation, knowledge graph (KG) has been effectively applied in many actual industrial usage scenarios, and is one of the key technologies for the next wave of technology development (Wang et al. Citation2017). Thus we introduce knowledge graph as a way of encoding and storing historical prior knowledge, and propose a KBQA-based event assignment model for Chinese government hotline (Zhou et al. Citation2017). As a structure that can effectively store and represent external knowledge, knowledge graph can provide more valuable decision support for constructing relationships between events and departments. Automatic and accurate assignment of the events can not only reduce operating costs of the government hotlines, but also optimize the process and resource allocation of government hotlines. This will allow the government hotlines to be connected faster, assigned more accurately, and handled more effectively, and enhance public recognition of the government. Our model achieved optimal results on two hotline assignment datasets, and on the Dataset1 Hits@1 index achieves 88.66%, F1 index achieves 91.32%, on the Dataset2 Hits@1 index achieves 85.52%, and the F1 index achieves 88.77%. Meanwhile, knowledge graphs’ incomplete experimental results show that our model also has a high effect on incomplete knowledge graphs. The main contributions of this paper include the following:

  1. Constructing an event extraction module to extract key information about hotline events and designing a graph construction scheme to construct an event knowledge graph based on historical event texts.

  2. Performing subgraph retrieval based on three types of information obtained from event extraction module and designing a subgraph fusion scheme to fuse the three types of subgraph retrieval results to obtain the final retrieved subgraph.

  3. Fusing the graph-based prior knowledge extracted by fused-GCN and BERT with the “sanding” prior knowledge and using the KBQA-based approach to predict the probability distribution of the departments, and finally ranking the distribution to output the optimal assignment department.

Related Work

Government Hotline

In order to realize automatic hotline event classification, methods based on machine learning are generally adopted (Chen and Guestrin Citation2016). Firstly, the features are extracted from the text description of a hotline event, then multiple features are connected to form a high-dimensional feature vector, and finally the classifier is used to complete the classification. These methods require a lot of feature engineering. Selection and analysis of these features are complex which need to spend a lot of energy to conceive. In order to solve the problem that single neural network (e.g., CNN, RNN and LSTM) is difficult to discover the tiny difference among Chinese hotline events, Chen proposed event classification method fusing RoBERTa and feature extraction (Chen et al. Citation2022). The method obtains context-aware semantic feature vectors from textual descriptions of a work-order by RoBERTa and uses a hybrid neural network combined with CNN and BiGRU to explore the local and global features of the work-order semantic encodings. The fused features are sent into the classifier to finish work-order classification. Due to the complexity of events, it is difficult to accurately assign events by simply using the text corpus of 12,345 government hotline. In follow-up researches, the public data of government departments such as department description or “sanding” (containing the description of department responsibilities, organization settings and staffing) is also often fed into the task as additional knowledge to strengthen the correlation between events and government and improve the assignment accuracy of hotline events. Liu proposed a multi-label classification framework based on GCN, BERT, and memory network for Chinese government hotline (Liu Citation2022). The framework constructs the event graph with the abstract meaning representation and extracts the event topic information vector with GCN, and uses the label count-based multi-label selection to sort the event label candidate set and guide to output the optimal multi-label set of the event contained in Chinese government hotline. However, it cannot perform event assignment. She proposed a joint learning method for event text classification and event assignment for Chinese government hotline (She, Chen, and Chen Citation2022). The method uses GCN and BERT to process the event text of the Chinese government hotline, and the result vector is fused to achieve the event classification task. At the same time, the attention mechanism is used to process the GCN vector and BERT vector, and complete the event allocation based on the assignment record in the historical event database. However, the paper only used keywords in sentences as entities when performing graph construction, without considering the significance of the keywords for event assignment, as well as not defining relationships based on assignment, and only used whether different entities appeared in the same sentence as associated edges. Based on the event extraction method (Chen Citation2022), Chen proposed an automatic event assignment method for Chinese government hotline (Chen et al. Citation2022). The method uses the CRF network to solve the problem of sequence annotation for event texts and design an external knowledge embedding layer which integrates “rights and responsibilities lists” with the historical information of the event to assist assignment. However, the method cannot use historical information to enhance the accuracy of event assignment.

Event Extraction

Event extraction is an important part of many downstream applications such as information retrieval, Question-Answer (QA) systems, and decision support, so this field has been a hot spot in natural language processing research. Traditional event extraction methods are mostly based on manual experience, but when dealing with unstructured text, it is difficult for such methods to mine the deep semantic features in it. Notably, deep learning models such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have shown considerable advantages in capturing complex event patterns in unstructured text data (Lyu et al. Citation2021). Chen et al. introduced a word representation model DMCNN in (Chen et al. Citation2015). This method captures meaningful semantic rules for words and uses a CNN-based framework to capture semantic features at the sentence level. Similarly, JRNN (Nguyen, Cho, and Grishman Citation2016) is a joint-paradigm based bidirectional recurrent neural network for event extraction, which consists of encoding phase and prediction phase using RNN to summarize contextual information. However, while such methods can capture local semantic information of text to some extent, they are still inadequate in global semantic capture. With the proposal of the attention mechanism, the number of models that incorporate the attention mechanism has gradually increased, such as joint multiple event extraction (JMEE) (Liu, Luo, and Huang Citation2018) which uses attention-based graph convolutional networks for joint modeling to extract multiple event trigger words and thesis elements, and optimizes the biased loss function when jointly extracting the event trigger words and thesis elements in order to solve the dataset imbalance problem. For complex situations in real-world scenarios, research has been conducted to enhance the corresponding extraction effect using various techniques, such as using grammatical relations between words to improve the efficiency of event extraction (Cao et al. Citation2022; Li et al. Citation2022), or optimizing the effect of event extraction on a less-sample dataset based on syntactic structural information (Fei et al. Citation2022, 15460–75). Pre-trained language models such as BERT with other BERT-based derivatives have become effective tools for natural language understanding tasks including event extraction (Fincke et al. Citation2022; Liu et al. Citation2022). These models utilize large amounts of unlabeled textual data to learn rich contextual representations and can be fine-tuned for specific downstream tasks. Yang et al (Citation2019) proposed a pre-trained language model-based event extractor (PLMEE) framework, which automatically generates labeled data through “editing prototypes” and filters development samples according to quality ranking. Fei et al. proposed a method combining large-scale biomedical knowledge graph to enrich the contextual language model and improve the effect of biomedical information extraction (Fei et al. Citation2022, bbaa110). The pre-trained language model effectively improves the accuracy of the extracted semantic information, but in the real-world scenario of hotline event assignment, it often faces comprehension bias due to the lack of external knowledge.

KBQA

Question Answering (QA) is an important research area in artificial intelligence and natural language processing. Its goal is to enable computers to understand the questions asked by users and give accurate answers. According to the different data sources, technical methods, and application scenarios it uses, it can be categorized into different ways, such as textual QA systems (QA) that use natural language processing techniques to retrieve answers from textual data or to generate answers based on textual data, and visual QA systems(VQA) that use computer vision techniques to process image and video content and combine them with techniques of natural language processing, information retrieval (Li et al. Citation2023), etc., to textually generate answers to questions (Cao and Li Citation2023a, Citation2023b). These question answering systems usually generate answers based on input text, images, and video content, whereas knowledge-based QA (KBQA) incorporates prior knowledge from a knowledge base to generate more accurate answers.

KBQA aims to answer a question over a knowledge base (KB) that is a structured database containing a collection of facts (alias triples) in the form (subject, relation, object). According to different execution paradigms, KBQA can be divided into two categories: the Semantic parsing-based (SP-based) method and Information retrieval-based (IR-based) method (Lan et al. Citation2021). An essential step is getting a question representation as accurate as possible in KBQA. To represent long-range dependencies between the answer and the topic entity in question, Luo et al. (Citation2018) extracted the dependency path between them. By encoding the directional dependency path, they concatenated both syntactic and local semantic features to form a global question representation. In order to alleviate the inaccurate syntactic parsing of complex questions, Sun et al. (Citation2020) leveraged the skeleton-based parsing to obtain the trunk of a complex question, which is a simple question with several branches (i.e., head word of original text spans) to be expanded. Under such a skeleton structure, it is more accessible to obtain accurate parsing results only by further analyzing simple sentences. It is also crucial to analyze the syntax of queries and ensure that the generated logic forms can meet the complex syntax of questions. But this method has a greater likelihood of poorer accuracy for QA in complex language scenarios. Maheshwari et al. (Citation2019) proposed a novel ranking model which exploits the structure of query graphs and uses attention weights to explicitly compare predicates with natural language questions. Specifically, they proposed a fine-grained slot matching mechanism to conduct hop wise semantic matching between the question and each predicate in the core reasoning chain. Instead of capturing semantic correlations between a question and a simple relation chain, Zhu, Cheng, and Su (Citation2020) focused on structure properties of query and conducted KBQA with query-question matching. They employed a structure-aware encoder to model entity or relation context in a query, promoting the matching between queries and questions. Instead of enumerating logic forms with a single pass, researchers try to propose methods to generate the complex queries with multiple steps. However, this method has very high requirements for retrieval accuracy and efficiency. Wei et al. (Citation2018) and Nikita, Xinyi, and Jagadish (Citation2019) proposed to first decompose a complex question into multiple simple questions, where each simple question was parsed into a simple logic form. The final answers are obtained with either the conjunction or composition of the partial logic forms. This decompose-execute-join strategy can effectively narrow down the search space. Unlike decomposing a complex question to sub-questions, a number of studies adopted the expand-and-rank strategies to reduce the search space by expanding the logic forms in an iterative manner. Yunshi and Jing (Citation2020) defined three expansion actions for each iteration, which are extending, connecting, and aggregating to correspond to multi-hop reasoning, constrained relations, and numerical operations, respectively. To solve the problem of missing graph structure in the IR-based method, Haitian et al. (Citation2018) proposed to complement the graph with extra question-relevant sentences as nodes and reason on the augmented heterogeneous graph. According to the entities mentioned in sentences, they linked them to corresponding entities on the graph and viewed them as nodes. Most of the current KBQA is offline training and online deployment. However, due to the unbalanced data distribution and insufficient amount during the training process, it is impossible to learn the unsolvable problems that appear online in the actual situation. Two research directions will be how to solve the imbalance of data and the processing of unseen data online.

Methodology

We proposed a KBQA-based method to obtain the corresponding department of the event for event assignment task, which can effectively use the event extraction results and the knowledge graph built based on historical events, and use the QA method to achieve event assignment. The event assignment model in this paper mainly consists of three parts: event association information retrieval module, event association feature extraction module and event allocation department prediction, and the specific model structure is shown in . For a given input event text, we first used event extraction algorithm to obtain event trigger words, event types, and argument roles, then carried out subgraph retrieval to get subgraph, and used Fusion-GCN model to obtain subgraph representation features. Secondly, we use GloVe-based text retrieval method to obtain the most relevant “sanding” text, and use BERT to obtain the “sanding” representation vector. Then, we use another BERT to extract the event text features and generate the fused event vector set by combining the set of departmental entities in the subgraph. Finally, the fused event vector set elements vector, graph representation features vector, and the “sanding” representation vector are fused to predict the department scores, and the optimal assignment department is output according to the score results.

Figure 1. Model structure.

Figure 1. Model structure.

Event Extraction

The input of the event extraction module for historical events is a complete event description text, and the output of the module is the corresponding event trigger words, event type, and event argument roles. To make full use of the contextual relationships in the event description text and the key information contained therein, this paper uses a joint-based model for event extraction. Considering that argument roles are more dependent on the key information in the event description text (Ding et al. Citation2019), the event extraction module has argument roles extraction as an end task. The input of this task depends on the extraction results of event trigger words and event types, and the three tasks are jointly optimized as a whole module. The specific structure is shown in . First, the event description text is sent into the trigger extraction layer to generate a list of trigger words for the event, and then the event trigger words embedding vectors are input into the event classification layer to output the corresponding event type prediction results, and finally the event classification result encoding vector is input into the argument roles extraction layer together with the trigger words embedding vectors and words embedding vectors to generate the corresponding event argument roles prediction results.

Figure 2. Event extraction module structure.

Figure 2. Event extraction module structure.

Considering the fact that Chinese has more semantic levels and contains more complex information, we firstly perform word segmentation on the event text before processing. Specifically, the Chinese word segmentation tool is used to divide the input text into word sequenceX={x1,x2,,xn} (where n is the length of the word sequence obtained by word segmentation tool). The content of the description of government hotline events often involves government entities, which are often composed of multiple juxtaposed phrases, and the noun as a determiner modifies the noun, which makes the length of the entity larger, and the difficulty of entity boundary identification is relatively high, e.g., “芜湖市人民政府国有资产监督管理委员会” (SASAC of Wuhu Municipal People’s Government). On the other hand, the vocabulary that constitutes the governmental entities has obvious official text characteristics; the vocabulary that cannot constitute the governmental entities also has significant textual characteristics. In order to make full use of the domain characteristics of the constituent words of governmental entities and improve the recognition accuracy of governmental entities, we use “standard phrase dictionaries” and “nonstandard phrase dictionaries” to adjust the segmentation results. The “standard phrase dictionary” consists of high-frequency words that are involved in the formation of governmental entities, while the “nonstandard phrase dictionary” consists of high-frequency words that are difficult to form governmental entities. When performing word segmentation, the results are compared with the positive and negative lexicon, and sequences that satisfy the positive words are merged, while sequences that satisfy the negative words are split. Then the word2vec model is used to encoding word sequence to obtain the initialized semantic encoding vectorInit={init1,init2,,initn}. Meanwhile, for each word, the corresponding position vector posiand segment vector segmentiare constructed, both of which are one-dimensional vectors with the length of the word sequence, where the position vector posi is set to 1 at the corresponding position of the word and 0 at the remaining positions, while the segment vector is used as the text vector, with 1 in the position corresponding to the sentence in which the word is to be encoded and 0 in the rest. After that, the position vector and the segment vector are input into the pre-trained language model together with the initialization embedding vector to obtain the corresponding semantic embedding vector:

(1) ei=LM(initi,posi,segmenti)(1)

The semantic encoding vectors constitute the set E={e1,e2,,en}. To obtain the trigger words in the text, the trigger word extraction task is converted into a binary classification problem of whether the words in the word sequence are trigger words or not. A classification model is constructed and the word embedding vectors E in are input to the classification successively to obtain the trigger-or-not probability distribution:

(2) Pitri=softmax(Wtriei+btri)(2)

where Wtri,btri are the learnable parameters. After the classification probability distribution of the sequence is obtained, the word sequence corresponding to top-n in the probability distribution is selected as the final trigger word prediction result. For the trigger extraction layer, cross-entropy loss function is used to optimize this layer:

(3) Ltri=[yitrilogyˆitri+(1yitri)log(1yˆitri)](3)

where yitri denotes the actual trigger classification result and yˆitri denotes the predicted trigger classification result.

After completing the trigger extraction, the sequence of trigger words T={t1,t2,,tk} corresponding to the event text is generated. Then, the semantic vector corresponding to the trigger words in the sequence is fused, and then input into the multi-classification network to obtain the corresponding probability distribution of event types Pitype, and the highest probability result is selected as the prediction result of the event types.

(4) Pitype=softmax(Wtypee+btype)(4)
(5) e=max(et1,et2,,etk)(5)

where Wtype,btype is the learnable parameter, etk indicates semantic embedding vector and max denotes the maximum pooling function, taking the maximum value in each dimension respectively. The event classification layer is optimized using multi-class cross-entropy function:

(6) Ltype=1Jic=1LCyiclog(pic)(6)

where LC is the number of event types, yic is symbolic function that takes 1 if the true event type of the sample event isc and 0 otherwise. pic denotes the prediction probability of sample eventias typec.

The argument roles extractor layer aims to extract arguments and their roles given the trigger words. Considering the high dependence of arguments on the key information carried by the trigger words and the fact that most of the arguments are noun phrases, this paper use a multi-classification to predict the argument roles corresponding to the word sequence of event description text. Meanwhile, since the event set contains many different event types, the classification network contains the roles corresponding to multiple event types in order to make full use of the event type information. After the trigger extraction layer, the key information position vector postri is constructed based on the result of the trigger words sequence, and the key information position vector is a one-dimensional vector, which is marked as 1 at the word position corresponding to the trigger words and 0 at the rest positions to inform the encoding model where the key information is located. Then, the words embedding vectors E are input to the pre-trained language model together with the position vectors pos and key information position vectors postri to obtain embedding vectors e i.

(7) e i=LM(initi,posi,postri)(7)

Then the embedding vectors are fed into the argument roles classifier together with the event type embedding vector qtype to generate the probability distribution of argument roles:

(8) Pirolej=softmax(Wrolee ′′qtype+brole)(8)

where Wrole,brole are learnable parameters, Event type embedding vector qtype is a one-dimensional vector constructed by the generated results of the event type classification layer, and its length is the number of event types and 1 is set on the position corresponding to the predicted event type result, and 0 is set on the rest positions. After the probability distribution is obtained, the top-2 probability is selected for comparison. If the predicted probabilities of the first two argument roles are similar, then the two prediction results will be output at the same time; otherwise, only the prediction result with the highest probability will be output. The element role extraction module is optimized by multi-classification cross entropy function:

(9) Lrole=1Nic=1Myiclog(pic)(9)

where M is the number of argument roles and yic is a symbolic function that takes 1 if the true roles of sample i is c and 0 otherwise. pic denotes the probability of predicting the outcome argument rolec for samplei.

For the overall event extraction module, the joint function of three loss functions is used as the loss function of event extraction module:

(10) Loss=Ltri+Ltype+Lrole(10)

Knowledge Graph Construction

When constructing a knowledge graph, in order to ensure accurate and efficient extraction of key information from complex event descriptions and to gradually improve the graph structure, the model in this paper adopts a bottom-up construction method. This method starts from specific events, gradually abstracts and generalizes, and finally forms a comprehensive and fine knowledge system. Through the event extraction module, three key information including trigger words, event types, and argument roles is extracted from the event description text, and in the knowledge graph construction module, we construct the corresponding event knowledge graph in a bottom-up manner. The bottom-up knowledge graph construction method is an iterative update process, and each round of update includes two parts: information extraction and knowledge fusion. The specific process for each step in the iteration is described next.

Information Extraction

As the data source for the construction of event knowledge graph, the process of information extraction in this paper includes two parts: entity construction and relationship extraction.

Specifically, the entity construction part mainly consists of the results of the event extraction module, but in addition to the event trigger words, event types, and argument roles obtained from the event extraction module, this paper also uses the event ID information and assignment department information of each historical event as the other two source of entity information needed to construct the knowledge graph.

In the relationship extraction part, the entity information obtained from the entity extraction module is associated by constructing the graph relational model to generate the basic triples that build the event knowledge graph. Relational schema construction establishes association schemas among entities to standardize the final output of fact representation patterns from information extraction, while the database generated by schema relationship construction can ensure less redundancy of relationships. The event knowledge graph in this paper includes five types of entities, and according to the cognitive framework of the basic elements structure of events based on prior knowledge, the corresponding knowledge model framework of these five types of entities is constructed, and the relationship schema diagram is shown in .

Figure 3. The relationship schema diagram.

Figure 3. The relationship schema diagram.

This knowledge mapping relationship pattern graph contains 5 types of entities and 4 types of relations. Considering that there is no significant ”entity-attribute” attribution relationship among the elements extracted in the event extraction module, multiple entity nodes in each type of entity are directly associated with events. Due to the diversity of event types, there are many ways to describe logical relationship between multiple argument roles or multiple trigger words and the same event, and in order to unify the descriptions, a unified relationship description is used between multiple argument roles or multiple trigger words and the same event. Take the event ”(3471, ‘上周日,我在某车站预约的车牌号ABCDE的出租车司机直接拒载 Last Sunday, the cab driver with the license plate number ABCDE I booked at a station refused to take me directly),’ ‘公安局 (Public Security Bureau)’” as an example, according to this event text, the examples of relational triples formed are shown in .

Table 1. Examples of relational triples.

According to the results shown in , after completing the above pattern construction, the extraction results around the same event will form n + 1 triples around the event entity, where n indicates the number of key elements obtained by the event extraction module.

Knowledge Fusion

After the information extraction completes the extraction and collation, the corresponding entities, relations, and corresponding triples are output, but these triples are independent of each other and do not form an association relationship, which still cannot form a hierarchical and logical knowledge graph. Meanwhile, considering that the extracted entity and relationship information may have redundant and wrong information, the knowledge fusion is used to integrate the entity, relationship, and triad information. The knowledge fusion mainly includes two parts: entity link and knowledge merge. The main processing flow of the knowledge fusion is described as shown in Algorithm 1.

Entity link mainly corrects and aligns the redundant referents or ambiguous referents existing between entities, so that a part of the constructed triples forms a valid association relationship. Entity link is mainly containing two key elements: entity disambiguation, which eliminates homonymous ambiguity, and co-reference disambiguation, which eliminates different referents to the same object. The entity disambiguation part for any two entities composed of the same name entity pairs in the entity collection obtained from information extraction, the two entities, and their surrounding words are associated, and then input into the pre-trained language model to obtain the corresponding embedding vector, after which the semantic similarity between the two entities is calculated, merging the entities whose similarity is greater than a certain threshold and distinguishing the entities with the same name that have different references (Mueller and Thyagarajan Citation2016):

(11) G(ei,ej)=ei,ejeiejd(11)
(12) Sim(Ei,Ej)=G(Ei,Ej)G(Ei,Ei)G(Ej,Ej)(12)

where Ei denotes the embedding vector of entity i and d denotes the length of the entity embedding vector.

Since the co-reference disambiguation occurs mainly in the argument roles, thus this paper only use co-reference disambiguation on the entities set of the event argument roles. Considering that the co-reference disambiguation between argument roles is closely related to the context of event description text, after obtaining the set of argument role entities, the semantic embedding vector of the argument role and its sentence are fused to form the fusion embedding vector of argument role. After that, the fusion embedding vectors are clustered, and the fusion vector similarity is calculated in the set of argument roles embedding vectors aggregated in the same cluster, and the pairs of arguments with similarity greater than a certain threshold are merged to form a unified entity referent.

(13) Srolei=μerolei+(1μ)esent(13)
(14) Rel(rolem,rolen)=Sim(Srolem,Srolen)(14)

where Srolei denotes the fusion embedding of the argument role i, erolei,esent are the embedding vectors of the argument role and its sentence, and μ is the learnable weight.

After the completion of entity disambiguation and co-reference disambiguation, the entity information of triples has been integrated and aligned, and then n subgraphs {g1,g2,,gn} are constructed from these triples generated from each event. Then, the duplicate entities are fused and the connections between the subgraphs are established to form the complete event knowledge graph.

KBQA-Based Event Assignment

Event Information Retrieval

We use two parallel pipelines to obtain event-related features: one is subgraph retrieval based on the knowledge graph, i.e., retrieval on the knowledge graph, which returns the set of entities; the other is text retrieval based on the “sanding” responsibilities, i.e., text retrieval on the “sanding” text corpus, which returns the document set. The retrieved entities and documents are then fed into the prediction model as fusion information of events.

Subgraph Retrieval

In order to retrieve entities related to input events from the knowledge graph, we first perform event extraction on the input event text and obtained three types of entities, namely event trigger word, event type, and argument role, as seed entities, denoted as set {Se}. Then, the Personalized PageRank (PPR) algorithm (Taher Citation2002) is applied to the elements in the set {Se} to identify entities that may be related to the event. For the entity identified by the PPR algorithm, the inner product method is adopted to calculate the score of the entity word vector and event sentence vector, and the score is taken as the edge weight of the entity and the seed entity.

Through this method, we can first obtain the subgraphs of three different types of entities associated with trigger words, event types, and argument roles, and merge the same entities in the subgraphs of the same entities (trigger words, event types and event roles) to form three different type subgraphs. Therefore, we also adopt the same method to merge the same entities in the subgraph of different types to form the final subgraph.

Specifically, we first merge the same entities in n subgraphs of n seed entities of the same type to form three subgraphs (G1, G2 and G3) of the same type. Then the same entities in three subgraphs of the same type are merged to get the final subgraph G. For these merged entities, we trace back to the root (i.e., the seed entity) in each subgraph and trace forward to the leaves. We then retain only the entities and relations on all trace path to form the subgraph. As shown in , given the input event text “车牌号ABCDE的出租车司机直接拒载 (the cab driver with the license plate number ABCDE refused to take a passenger directly),” the event extraction algorithm is used to obtain the seed entities “拒载,” “出租车拒载” and “车牌号 ABCDE.” And then we use the seed entities to retrieve three types of subgraphs {G1, G2, G3}, and merge the same entities (the entities in red boxes in ) to obtain the final subgraph G.

Figure 4. Subgraph retrieval procedure.

Figure 4. Subgraph retrieval procedure.

The subgraphs G1, G2, and G3, derived from the three entity types related to events (event trigger words, event types, and argument roles), along with the final subgraph G obtained through subgraph merging, are input into the Subgraph Reader module to extract subgraph features.

Document Retrieval

We used 355 “sanding” text data as the document retrieval corpus, and carried out sentence-level document retrieval, i.e., we retrieved each sentence in the “sanding” retrieval corpus. The execution steps of our “sanding” document retrieval are as follows:

Firstly, we vectorize all sentences in the “sanding” retrieval corpus using GloVe word table to obtain a vector representation set of all sentences, denoted as {Visandin},i=1,2,.

Secondly, we vectorize the input event text with the same GloVe word table to obtain the event text representation vector Ve, and calculate the inner product of Ve between the elements in {Visandin}, then sort all the sentences according to the inner product result to get top-5 associated sentences.

Finally, trace back to the source according to the top-5 associated sentences, the scores of the “sanding” text corresponding to the top-5 sentences are calculated, and the “sanding” text with the highest score is selected as the text retrieval result.

The “sanding” original text calculated by the aforementioned Text Retrieval algorithm is used as one of the inputs for the Text Reader module to extract “sanding” features.

Event Feature Extraction

Subgraph Reader

For the constructed subgraphs G1, G2, G3, and G, we use Fusion-GCN (Kipf and Welling Citation2019) for subgraph feature extraction, which uses multi-layer residual fusion for single-graph feature extraction on the basis of GCN, and uses attention mechanism for fusion of multi-graph features to accumulate knowledge of multiple subgraphs.

To extract the graph representation features of each key vertex (trigger word, event type, argument role, and department), we pair G1, G2, G3, and G respectively to form pairs, namely (G1, G), (G2, G), and (G3, G). For the key input vertex V, we define graph convolution operation in two vertex domains to aggregate the local and global features of this vertex. This graph convolution operation for feature extraction is calculated as follows:

(15) H(l)=ReLU(A˜(H(l1)+H(l2))W(l)+b(l)),l>1ReLU(A˜H(l1)W(l)+b(l)),l=1(15)

where A˜=D12A D12, A =A+I, A is the adjacency matrix of the input graph, I is the identity matrix, Dii=jA ij, W(l) and b(l) are learnable parameters, H(0)=V. To further preserve the graph feature information, we fuse the graph convolution output using a multi-layer aggregation, which is calculated as shown below:

(16) H(out)=λ1max(H(1),H(2),,H(L))+λ2H(L)(16)

whereλ1+λ2=1, L is the number of graph convolution layers, in this paper we set L = 3.

According to the above formula, the vertex representation vector set of each subgraph can be calculated, and the representation vector of three different subgraphs can be obtained by weighted fusion of its seed entity vectors. Different types of subgraph representation vectors and final subgraph representation vector constitute graph vector pair (HGi(out),HG(out)), where HGi(out) represents the local feature representation vector and HG(out) represents the global feature representation vector. Then two vectors are fused using the attention mechanism to get the fusion representation vector VGi:

(17) VGi=softmax(W1HG(out)(W1HGi(out))Td)W2HGi(out)(17)

where W1, W2 are learnable parameters, d is the number of attention heads.

According to the above formula, we calculated the representation vectors of three types: trigger, event type, and argument role respectively, and then concatenate them to obtain the final graph representation vector VG=concat(VG1;VG2;VG3). And all the department vertex representation vector of the final subgraph, denoted as {Vid},i=1,2,3,.

Text Reader

The text reader consists of two parts: the problem text representation (i.e., the vectorized representation of the input event text) and the “sanding” text representation.

For the problem text representation, we fuse the input event text with the elements in the department vertex representation vector set {Vid} as the problem representation vector. The specific way is as follows, A BERT is used to extract feature from input event text and get event representation vector V0q:

(18) V0q=f(q)(18)

where q is the input event text and f is the BERT pre-trained model. The vector V0q is fused with each element in the department vertex representation vector set {Vid} to obtain the fused problem vector set {Viq},i=1,2,3,.

(19) Viq=γiV0q+φitanh(Wiq[V0q,Vid,V0qVid])(19)
(20) γi=sigmoid(W˜iq[V0q,Vid,V0qVid])(20)

where Wiq and W˜iq are learnable parameters, and γi+φi=1.

For the retrieved “sanding” text, we use another BERT model to extract features directly to obtain the “sanding” feature representation vector Vsanding,

(21) Vsanding=fˆ(Tsanding)(21)

where, fˆ is the BERT pre-trained model.

Event Assignment Department Prediction

After we get the event knowledge graph feature, the “sanding” feature and the input event feature, the graph representation vector VG is concatenated with the “sanding” representation vector Vsanding, and then the probability prediction of the department entity is made by matching the problem vector Viq and the concatenate vector:

(22) pi=softmax((Viq)TWi[Vg;Vsanding]+bi)(22)

where, Wi and bi are learnable parameters. In this paper, we use the binary cross-entropy as the training loss of the model and use the final output probability as the score of the department.

All elements in the vector set {Viq} are predicted to obtain the score set {Si},i=1,2,3, of the departments. Finally, set {Si} is sorted and the optimal department is selected as the event assignment department.

Experiment

Experiment Settings

Dataset

Table 2. Dataset description.

We construct two experimental datasets based on the real government hotline event assignment result. After counting on the full event dataset, more than 98% of the data are less than 300 words and 0.32% of the data are less than 10 words, and most of this data are wrong calls without proper semantic information, so in the pre-processing stage of the data, we removed the data with more than 500 words and less than 10 words. Firstly, the data was preprocessed to delete the events of more than 500 words and less than 10 words. After processing, the Dataset1 contains 27,691 events of 30 municipal departments, and the Dataset2 contains 35,642 events of 36 municipal departments. Secondly, based on the real event assignment results of the government hotline 27,691 of Dataset1 and 35,642 of Dataset2 positive sample event data corresponding to “event-department” were constructed respectively. Finally, 27691 positive samples (matched) were randomly selected from 29 unmatched departments to construct 55,382 negative samples (unmatched) to generate Dataset1. 35642 positive samples (matched) were randomly selected from 35 unmatched departments to construct 71,284 negative samples (unmatched) to generate Dataset2. The description of the data set is shown in .

Hyperparameter

In the Event Extraction section, BERT-wwm-ext is used as pre-trained language model in the Event Extraction module as the base encoder for the Trigger Word Classifier and the Argumentative Role Classifier to encode the input hotline event text. Meanwhile, Adam optimizer with learning rate of 10e-5 is used as the model optimizer, the training epoch is set to 30 and the batch size is set to 32. In the KBQA-based Event Assignment section, for the GloVe model used in the Event Information Retrieval module, the historical case data is used for training to obtain the vector dimension of 128. We also use BERT-wwm-ext as pre-trained language model to process the text retrieval results and the input hotline event text. In Fusion-GCN, GCN uses a 3-layer graph convolution operation to obtain the corresponding graph feature vectors. We use Adam optimizer to train the model with 20 epochs, batch size 16 and learning rate 10–5.

Evaluation Metrics

We evaluate the performance of the subgraph retriever by the coverage rate of the departments corresponding to the events. We calculate the proportion of the subgraph retrieval results containing the real processing departments of the events, which is calculated as follows:

(23) coveragerate=NsggtNsg(23)

where Nsggt represents the number of subgraphs containing the actual assignment department of the event and Nsg represents the total number of retrieval subgraphs. Since the final assignment department is generated from the subgraph retrieval results, this metric reflects the upper bound of the model performance.

For the overall model’s assignment performance, we use Hits@1 to evaluate whether the top-1 predicted answer is correct. Meanwhile, we further evaluated the model effect by calculating the Precision, Recall, and F1 values of the model.

Baselines

We compare the model with three types of baseline models on the constructed dataset:

  1. Text classification based model, i.e., using text classification models to directly process event text for event assignment, including: HAN (Yang et al. Citation2016), XLNet (Yang et al. Citation2019).

  2. Answer selection based model, i.e., using event text as the question and “sanding” text as the answer to perform answer selection for event assignment, including: BERT-BiGRU-based (Nikita, Xinyi, and Jagadish Citation2019), EEuEK (Chen Citation2022), BGMA (Devlin et al. Citation2019), DeBERTa-SSP (Liello et al. Citation2022).

  3. KBQA based model, i.e., building knowledge graphs, using event text as questions, and using KBQA method for event assignment, including: Pullnet (Haitian, Tania, and William Citation2019), EmbedKGQA (Apoorv, Aditay, and Partha Citation2020), RnG-KBQA (Ye et al. Citation2022), UniK-QA (Oguz Citation2022), SKP (Dong et al. Citation2023).

Comparisons of Baselines

In order to verify the effectiveness of the model in this paper, we compared the model with the baseline models on the constructed dataset, and the experimental results are shown in , where “no-sanding” denotes the model with text retrieval and “sanding” text embedding removed based on our model, and only the knowledge graph features are used for event assignment.

Table 3. Comparisons of different baselines result.

As can be seen from , our model achieves better experimental results in two datasets. The text classification based model is significantly less effective than the answer selection based model and KBQA-based model for event assignment, which is due to the fact that the text classification model is the worst because it only considers the relevant information of the event itself and cannot obtain more additional information. However, there is also a gap between the two text classification models, which is due to the gap in the effectiveness of the models for event text feature extraction, with the HAN model using BiGRU combined with the attention mechanism to extract text features, while XLNet is a language model trained on a large-scale corpus, thus having a better feature extraction capability and thus achieving a better result than HAN.

The effect of the answer selection based model is significantly lower than our model, because the answer selection based model only takes into account the text features of the “sanding” closely related to the department, and does not take into account the a priori knowledge of historical events, which illustrates the effectiveness of the knowledge graph based on historical events. Meanwhile, the BERT-BiGRU-based model, EEuEK and BGMA also have a gap, which is also due to the gap between the text extraction effect of different text feature extractors. Compared with BERT-BiGRU-based model, which uses BERT model as the feature extraction model of event text, EEuEK uses RoBERTa model, which is more effective, and thus achieves better results. BGMA simultaneously uses BERT and GCN as feature extraction models, thus achieving a better model effect than EEuEK. DeBERTa SSP uses the better pre trained model DeBERTa as the feature extraction model, while adding the Spans in Same Paragraph (SSP) strategy, thus achieving good model performance.

Compared with our model, the effects of Pullnet, EmbedKGQA, RnG-KBQA, UniK-QA and SKP are also significantly lower than it, because these models only consider the a priori knowledge of the historical event knowledge graph, but do not use the text features of the “sanding” closely related to the department. From the perspective of coverage rate, the subgraph retrieval effect of the baseline model is inferior to our model, so the model in this paper has achieved the optimal effect. From the results of “No-sanding” model, the effect of this model is only slightly lower than that of RnG-KBQA model, which can reach the same level as those most advanced KBQA models, which also fully illustrates the performance of the model in this paper. It is worth noting that the experimental results show that the effect of KBQA-based models is slightly higher than that of answer selection based models, which indicates that the a priori knowledge of the historical event is better than the “sanding” text features.

According to the comparison results between “No-sanding” model and our model, the effect of other evaluation indicators except coverage rate is lower than that of the model in this paper, which indicates that the retrieval embedding based on “sanding” text can effectively improve the assignment performance of the model. Since the subgraph retrieval process of the two models is the same, the coverage rate is the same. Since the “sanding” text is closely related to the department, the embedded “sanding” text can add the feature related to the department, so the retrieval and embedding based on the “sanding” text can effectively improve the performance of the model.

The horizontal comparison of the results of the models on the two data sets shows that with the increase of the number of assignment departments, the effects of all models decline to a certain extent. However, the effect of our model is still optimal, and the decrease of the model effect is also the smallest, which shows that our model certain advantages in dealing with the hotline assignment problem, and the stability and generalization performance of the model are relatively high.

Based on the above experimental results, our model achieves optimal results in both longitudinal baseline model comparisons and cross-sectional comparisons on different datasets. This is due to the fact that our model uses the knowledge graph based on historical events and also incorporates the retrieval embedding based on “sanding” text, and when using the knowledge graph, the subgraph retrieval is utilized to narrow down the candidate set of answers in order to improve the accuracy of the model’s prediction, and when using “sanding” text, the text retrieval algorithm is designed to obtain the most relevant text, thus further improving the model’s prediction accuracy. When using the “sanding” text, the text retrieval algorithm is designed to obtain the most relevant text to further improve the model performance.

Incomplete KG Analysis

The model in this paper needs to retrieve the path between the head entity and the answer entity in the knowledge graph to answer the question, and the answer is restricted to retrieve the entities in the subgraph. Therefore, in order to verify the effectiveness of our model in the case of incomplete KG, we perform random elimination of KG edges, and tested the assignment performance of our model under the condition of retaining 30%, 50%, and 80% edges of the knowledge graph.

show the main experimental results with different settings of incomplete KG, comparing Pullnet, EmbedKGQA, RnG-KBQA, UniK-QA, SKP and “No-sanding” models, we can see that the Hit@1 performance of RnG-KBQA model outperforms the other models regardless of the KG completeness setting when only KG information is used. However, our model outperforms the RnG-KBQA model in the case of incomplete KB due to the use of KG and the “sanding” text information, which indicates that the “sanding” text features closely related to the department can effectively improve the model performance.

Table 4. KG incomplete experimental results of Dataset1.

Table 5. KG incomplete experimental results of Dataset2.

As shown in , from the perspective of coverage rate, the subgraph retrieval effect of our model is optimal for all kinds of incomplete KGs, so the optimal model effect is achieved. In addition, according to the results of the cross-sectional comparison between our model and other KBQA models on different settings of incomplete KGs, it can be seen that the indicators of the model are decreasing as the number of edge removal increases, which shows that the KG can effectively improve the model effect, and the more complete the KG model is, the better the effect is.

Figure 5. KG incomplete experimental comparison of Dataset1.

Figure 5. KG incomplete experimental comparison of Dataset1.

Figure 6. KG incomplete experimental comparison of Dataset2.

Figure 6. KG incomplete experimental comparison of Dataset2.

As shown in , the performance of our model can be improved by using the “sanding” text features under different settings of incomplete KGs, and the effect on the model performance is increasing with the increase of KG completeness (except for 30% KB, which may be due to the effect of excessive edge connection removal). Based on general common sense, as the KG completeness increases, the impact of other types of features on the performance of the model becomes smaller and smaller, while the performance of our model is significantly enhanced by using “sanding” text with the condition of KG completeness increase. This may be due to the fact that the KG built in this paper is currently incomplete enough. However, from the experimental results, regardless of the KG completeness setting, the effect of our model is optimal, which fully illustrates the effectiveness of the model. At the same time, it can be proved that the performance of KBQA is effectively improved by using the “sanding” text method in the case of incomplete knowledge base.

Inference Time Analysis

The assignment of hotline events in the proposed model consists of two sequential processes: the construction process of a knowledge graph based on event extraction and the “event-department” matching process based on subgraph retrieval and text retrieval. The experimental results are presented in .

In the experimental environment, the average time taken by the knowledge graph construction process, based on event extraction, is approximately 3.8e-03 seconds, while the average time taken by the “event-department” matching process, based on subgraph retrieval and text retrieval, is about 6.0e-03 seconds. In the same experimental environment, the overall time for SKP to complete the inference and assignment task is 30.7e-03 seconds, while the time required by our model is 16.2e-03 seconds. It can be observed that the inference speed of the model in this paper is in milliseconds, In the case of similar performance, the proposed model can also have better efficiency than the compared baseline model on meeting the increasing demand for 12,345 hotline events assignment at a rapid pace.

Table 6. Comparison of time consumption of models.

Conclusions

Due to the problem that the existing government hotline event assignment methods cannot effectively utilize a priori knowledge and only assign events based on the textual content of events, resulting in low assignment accuracy, we proposed a government hotline event assignment model based on KBQA method to effectively utilize the a priori knowledge of historical events and the a priori knowledge of “sanding” to improve the performance of event assignment. Firstly, we use the event extraction model to obtain the key information of events and construct the knowledge graph based on historical events. Secondly, using the event trigger, event type, and argument role obtained from the event extraction to retrieve the subgraph, and the fusion-GCN model is used to obtain the subgraph representation vectors. Then, text retrieval method is used to get the most relevant “sanding” text, and a BERT is used to obtain the “sanding” representation vector and another BERT is used to obtain the event text feature representation vector set. Finally, the event text feature representation vector set, the graph representation feature vector and the “sanding” representation vector are fused to predict the department score, and the optimal assignment department is output according to the score ranking results. The experimental results also prove the effectiveness of our model, and the KG incompleteness analysis also shows that the KG completeness of this paper needs to be further improved.

However, it should be noted that our model also has some limitations; for example, the amount of information contained in the constructed graph may have a large impact on the prediction results due to the relatively heavy reliance of our model on the completeness and accuracy of the constructed knowledge graph. In the future, we will consider completing missing event-related entities in the knowledge graph to further improve the KG completeness and thus improve the assignment performance of the model. Meanwhile, we can design more efficient graph algorithms and model structures to further improve the model’s inference speed. Currently, our model is mainly applied to the Chinese government hotline field; for broader application scenarios, we will conduct further research on English-related data in the future.

Acknowledgements

We would like to thank the anonymous reviewers for their valuable comments and suggestions that helped improve the quality of this manuscript.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement

The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.

Additional information

Funding

The work was supported by the Key Technologies Research and Development Program of Anhui Province [202104a05020071].

References

  • Apoorv, S., T. Aditay, and T. Partha. 2020. Improving multi-hop question answering over knowledge graphs using knowledge base embeddings. In Proceedings of the 58th annual meeting of the association for computational linguistics(ACL), 4498–30, Seattle, Washington, USA.
  • Cao, R., and Z. Li. 2023a. Balancing and contrasting biased samples for debiased visual question answering. In 2023 IEEE international conference on data mining (ICDM), 962–67. Shanghai, China.
  • Cao, R., and Z. Li. 2023b. Overcoming language priors for visual question answering via loss rebalancing label and global context. In Proceedings of the Thirty-ninth conference on uncertainty in artificial intelligence (UAI), 249–59. Pittsburgh, Pennsylvania, USA.
  • Cao, H., J. Li, F. Su, F. Li, H. Fei, S. Wu, and D. Ji. 2022. OneEE: A one-stage framework for fast overlapping and nested event extraction. In Proceedings of the 29th International conference on computational linguistics (COLING), 1953–64. Gyeongju, Republic of Korea.
  • Chen, G. 2022. Government hotline work-order classification fusing RoBERTa and feature extraction. Journal of Computer Modernization 32 (6):21–27.
  • Chen, C., and C. Guestrin. 2016. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd international conference on knowledge discovery and data mining (SIGKDD), 785–94. San Francisco, California, USA.
  • Chen, G., X. R. She, J. P. Chen, J. Chen, and J. Qin. 2022. Automatic work-order assignment method for Chinese government hotline. Engineering Reports 5 (3):1–16. doi:10.1002/eng2.12580.
  • Chen, Y., L. Xu, K. Liu, D. Zeng, and J. Zhao. 2015. Event extraction via Dynamic Multi-Pooling Convolutional Neural Networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (ACL-IJNLP), 167–76. Beijing, China.
  • Devlin, J., MW. Chang, K. Lee, and K. Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 4171–86. Minneapolis, USA.
  • Ding, N., Z. Li, Z. Liu, H. Zheng, and Z. Lin. 2019. Event detection with trigger-aware lattice neural network. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 347–56. Hong Kong, China.
  • Dong, G., R. Li, S. Wang, Y. Zhang, Y. Xian, and W. Xu. 2023. Bridging the KB-Text gap: Leveraging structured knowledge-aware pre-training for KBQA. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM), 3854–59. Birmingham, UK.
  • Fei, H., Y. Ren, Y. Zhang, D. Ji, and X. Liang. 2022. Enriching contextualized language model from knowledge graph for biomedical information extraction. Briefings in Bioinformatics 22 (3):bbaa110. doi:10.1093/bib/bbaa110.
  • Fei, H., S. Wu, J. Li, B. Li, F. Li, L. Qin, and T.-S. Chua. 2022. LasUIE: Unifying information extraction with latent adaptive structure-aware generative language Model. Advances in Neural Information Processing Systems 35:15460–75.
  • Fincke, S., S. Agarwal, S. Miller, and E. Boschee. 2022. Language Model priming for cross-lingual event extraction. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 10627–35. Washington D.C. USA.
  • Haitian, S., D. Bhuwan, Z. Manzil, M. Kathryn, S. Ruslan, and C. William. 2018. Open domain question answering using early fusion of knowledge bases and text. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), 4231–42. Brussels, Belgium.
  • Haitian, S., B. W. Tania, and W. C. William. 2019. Pullnet: Open domain question answering with iterative retrieval on knowledge bases and text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2380–90. Hong Kong, China.
  • Kipf, T. N., and M. Welling. 2019. Semi-supervised classification with graph convolutional networks. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) ,11305–12. Long Beach, California, USA.
  • Lan, Y. S., G. He, J. H. Jiang, J. Jiang, W. X. Zhao, and J. R. Wen. 2021. A survey on complex knowledge base question answering: Methods, challenges and solutions. In Proceedings of the 30th International Joint Conference on Artificial Intelligence (IJCAI). Chengdu, China.
  • Liello, L. D., S. Garg, L. Soldaini, and A. Moschitti. 2022. Pre-training transformer models with sentence-level objectives for answer sentence selection. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP), 11806–16. ABU Dhabi, United Arab Emirates
  • Li, J., H. Fei, J. Liu, S. Wu, M. Zhang, C. Teng, and F. Li. 2022. Unified named entity recognition as word-word relation classification. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 10965–73. Vancouver, Canada
  • Li, B., and Z. Li. 2023. Large-scale cross-modal hashing with unified learning and multi-object regional correlation reasoning. Neural Networks: The Official Journal of the International Neural Network Society 171:276–92. doi:10.1016/j.neunet.2023.12.018.
  • Liu, B. 2022. GCN-BERT and memory network based multi-label classification for event text of the Chinese government hotline. IEEE Access 10:109267–76. doi:10.1109/ACCESS.2022.3213978.
  • Liu, X., H. Huang, G. Shi, and B. Wang. 2022. Dynamic prefix-tuning for generative template-based event extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL), 1:5216–28. Dublin, Ireland.
  • Liu, X., Z. Luo, and H. Huang. 2018. Jointly multiple events extraction via attention-based graph information aggregation. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), 1247–56. Brussels, Belgium.
  • Luo, K. Q., F. L. Lin, X. S. Luo, and K. Zhu. 2018. Knowledge base question answering via encoding of complex query graphs. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2185–94. Brussels, Belgium.
  • Lyu, Q., H. Zhang, E. Sulem, and D. Roth. 2021. Zero-shot event extraction via transfer learning: Challenges and insights. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL-IJCNLP), 322–32. Bangkok, Thailand.
  • Maheshwari, G., P. Trivedi, D. Lukovnikov, N. Chakraborty, A. Fischer, and J. Lehmann. 2019. Learning to rank query graphs for complex question answering over knowledge graphs. In The Semantic Web–ISWC 2019: 18th International Semantic Web Conference, 18:487–504. Auckland, New Zealand.
  • Ma, X. L., Y. C. Li, and B. Shen. 2020. Operation strategy analysis of Guangzhou 12345 government service hotline in COVID-19 prevention and control period. Guangdong Communication Technology 40 (9):2–4.
  • Meng, T. G., Z. B. Huang, and X. J. Zhang. 2021. Social governance innovation driven by Government Hot-Line in megacity-A case study on Beijing’s reform of responding to whistles. Journal of Public Management 18 (2):1–12.
  • Mueller, J., and A. Thyagarajan. 2016. Siamese recurrent architectures for learning sentence similarity. In Proceedings of the AAAI conference on artificial intelligence (AAAI), Phoenix, Arizona, USA.
  • Nguyen, T. H., K. Cho, and R. Grishman. 2016. Joint event extraction via recurrent neural networks. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), 300–09. San Diego, California, USA.
  • Nikita, B., Z. Xinyi, and H. V. Jagadish. 2019. Learning to answer complex questions over knowledge bases with query composition. In Proceedings of the 28th ACM international conference on information and knowledge management (CIKM), 739–48. Beijing, China.
  • Oguz, B. 2022. UniK-QA: Unified representations of structured and unstructured knowledge for open-domain question answering. In Findings of the Association for Computational Linguistics (NAACL), 1535–46. Seattle, Washington, USA.
  • She, X. R., J. P. Chen, and G. Chen. 2022. Joint learning with BERT-GCN and multi-attention for event text classification and event assignment. Institute of Electrical and Electronics Engineers Access 10:27031–40. doi:10.1109/ACCESS.2022.3156918.
  • Sun, Y. W., L. L. Zhang, G. Cheng, and Y. Z. Qu. 2020. Sparqa: Skeleton-based semantic parsing for complex questions over knowledge bases. In Proceedings of the AAAI conference on artificial intelligence (AAAI), 34:8952–59. Manhattan, New York, USA.
  • Taher, H. H. 2002. Topic-sensitive pagerank. In Proceedings of the 11th international conference on World Wide Web (WWW), 517–26. Honolulu, Hawaii, USA.
  • Wang, Q., Z. D. Mao, B. Wang, and L. Guo. 2017. Knowledge graph embedding: A survey of approaches and applications. IEEE Transactions on Knowledge and Data Engineering 29 (12):2724–43. doi:10.1109/TKDE.2017.2754499.
  • Wei, G. Z., X. Y. Jeffrey, Z. Lei, and C. Hong. 2018. Question answering over knowledge graphs: Question understanding via template decomposition. Findings on Stakeholder Engagement 11 (11):1373–86. Rio de Janeiro, Brazil. doi:10.14778/3236187.3236192.
  • Yang, Z., Z. Dai, Y. Yang, J. Carbonell, RR. Salakhutdinov, and QV. Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. In Proceedings of the 31th Conference on Advances in Neural Information Processing Systems (NeurlIPS), 5754–64. Vancouver, Canada.
  • Yang, S., D. Feng, L. Qiao, Z. Kan, and D. Li. 2019. Exploring pre-trained language models for event extraction and generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), 5284–94. Florence, Italy.
  • Yang, Z., D. Yang, C. Dyer, X. He, A. Smola, and E. Hovy. 2016. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference on North American Chapter of the Association for Computational Linguistics (NAACL), 1480–89. San Diego, California, USA.
  • Ye, X., S. Yavuz, K. Hashimoto, Y. Zhou, and C. Xiong. 2022. RnG-KBQA: Generation augmented iterative ranking for knowledge base question answering. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, 6032–43. Dublin, Ireland.
  • Yunshi, L., and J. Jing. 2020. Query graph generation for answering multi-hop complex questions from knowledge bases. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), 969–74. Seattle, Washington, USA.
  • Zhou, B., C. Sun, L. Lin, and B. Liu. 2017. InsunKBQA: A question-answering system over knowledge base. Intelligent Computer and Applications 7 (5):150–54.
  • Zhu, S. G., X. Cheng, and S. Su. 2020. Knowledge-based question answering by tree-to-sequence learning. Neurocomputing 372:64–72. doi:10.1016/j.neucom.2019.09.003.