6,426
Views
14
CrossRef citations to date
0
Altmetric
New Topics/Others

Relation extraction with weakly supervised learning based on process-structure-property-performance reciprocity

, & ORCID Icon
Pages 649-659 | Received 14 Mar 2018, Accepted 12 Jul 2018, Published online: 19 Sep 2018

ABSTRACT

In this study, we develop a computer-aided material design system to represent and extract knowledge related to material design from natural language texts. A machine learning model is trained on a text corpus weakly labeled by minimal annotated relationship data (~100 labeled relationships) to extract knowledge from scientific articles. The knowledge is represented by relationships between scientific concepts, such as {annealing, grain size, strength}. The extracted relationships are represented as a knowledge graph formatted according to design charts, inspired by the process-structure-property-performance (PSPP) reciprocity. The design chart provides an intuitive effect of processes on properties and prospective processes to achieve the certain desired properties. Our system semantically searches the scientific literature and provides knowledge in the form of a design chart, and we hope it contributes more efficient developments of new materials.

Graphical Abstract

This article is part of the following collections:
Materials Informatics

1. Introduction

Machine learning and data science for knowledge extraction are studied in a wide variety of field. Knowledge extraction is to find desired knowledge from text. For example, relationships among scientific knowledge are extracted from scientific literature in ScienceIE [Citation1], and a knowledge base is extracted from Web text in TAC.Footnote 1 The impressive performance of machine learning appears promising for knowledge extraction in material design as well.

Material design is a process of developing new materials with specific properties. In most practical cases, the desired process cannot be envisioned, and instead it is constructed by trial and error. In this approach, a trial is a time-consuming experiment. Minimizing the number of trials is critical for an efficient development. On the contrary, such efficient development is challenging because 1) The relation between a process and a property is unclear and indirect; and 2) The search space (the set of possible processes) is too large to look up.

In practice, researchers find these processes relying on their end-to-end knowledge including effects of processes to the properties. Such knowledge is technical and might not be well formalized, so they spend long time to obtain such knowledge. We believe it is beneficial to provide the end-to-end knowledge for accelerating material developments.

1.1. Knowledge representation

The processing-structure-property-performance (PSPP) reciprocity [Citation2] explains effect of processes on properties in three stages. The first stage is ‘process’ that can be controlled to develop a new material. The second stage is ‘structure’ of the material that the processes build. The third stage is ‘property’ that the structure gives. The properties in the third stage give the total performance of the new material.

The PSPP design chart [Citation3] represents end-to-end knowledge in the form of relationships among factors, as shown in [Citation4]. A factor represented by a node is an important phenomenon or concept for material design, such as annealing, grain size, and strength. A factor is classified into one of the three stages, ‘process’, ‘structure’, and ‘property’, where the factor performs. For example, annealing is an important concept in ‘process’. Following the PSPP reciprocity, a factor is influenced by connected factors to its left, and influences factors connected to its right (i.e. a process builds structures, and a structure influences properties of the material). These relations are represented by their connectivities. The chart intuitively represents end-to-end knowledge in the form of relationships between processes and properties mediated by structures [Citation5].

1.2. Data limitation

Despite major developments in machine learning, the technology suffers from limited data availability for model training in practical cases. For example, AtomWork [Citation6], one of the largest databases available for material design, contains records of more than 55,000 properties of materials. However, the knowledge in the database unlikely covers all knowledge needed for material design. For example, AtomWorks covers crystal structures and related properties such as lattice constant and space group but microstructures of materials are unlisted. Such limited data likely lead to over-fitting during training, and thus to poor performance.

In this study, we aim to overcome the problem of limited data availability using Natural Language Processing (NLP), an application of machine learning technologies to natural language resources, such as scientific articles and Web texts. Natural language resources are widely available and are machine readable, for a variety of fields including material design. For example, Elsevier’s APIFootnote 2 provides access to over 250,000 fully digitized scientific articles. More importantly, natural language is the most popular means of representing knowledge, and our desired knowledge is thus likely to be present.

1.3. Weakly supervised learning

We leverage weakly supervised learning to identify relationships in the PSPP chart. Weakly supervised learning is used to train a model with a minimal number of annotations for relation identification [Citation7]. In a typical supervised setting, the training data for relation identification is a sentence labeled with entities in the sentence and the relations among them. However, labeling sentences is expensive because an annotator must read a sentence, and label the entities and relations described therein. This renders each label clean and strong. On the contrary, in a weakly supervised setting, a knowledge base produces a pair of entities and their relations and all sentences containing these entities are weakly labeled with the given relation. For instance, for the given entities, Elevation Partners and Roger McNamee, and their relation founded_by, both of the following sentences are labeled with founded_by,

Elevation Partners, the $1.9 billion private equity group that was founded by Roger McNamee;

Roger McNamee, a managing director at Elevation Partners,…;

where the first sentence describes the relation founded_by but the second sentence does not. Weak labels do not require any annotations but the knowledge base, however; they are noisy, and a model needs to overcome the noisy labels.

In recent years, convolutional neural network (CNN) models have surpassed feature-based models [Citation7Citation11]. CNNs are a class of neural networks with convoluted neural units. Residual learning is used to help the deep CNN network [Citation12]. Zeng et al. [Citation13] split a sentence into three parts, and then applied max pooling to each part of the sentence over a CNN layer.

Sentence-level attention is introduced for selecting a key sentence. In this approach, a network takes a set of sentences for a relation between two entities. Each sentence contains both entities. An attention mechanism over a CNN allows the network to automatically select a key sentence, which is likely describing the desired relation. It seems helpful to overcome the problem of noisy labels [Citation14Citation16].

1.4. Our contribution

We develop a computer-aided material design (CAMaD) system with the aim of generating a PSPP design chart for desired properties from text. The PSPP chart suggests prospective processes to achieve given arbitrarily desired properties. The system is based on a machine learning model for relation extraction from text. The model is efficiently trained with weakly supervised learning, which minimizes the annotation cost of the training data. We believe the PSPP chart helps more efficient material developments by suggesting a prospective process.

Our contribution in this study is twofold. First, we proposed a novel knowledge graph based on PSPP charts and developed a system to build the knowledge graph from text using NLP technologies. Second, we experimentally verified that such technical knowledge can be extracted from text using machine learning models. Our target knowledge is relations in PSPP design charts. These relations appear rather technical and significantly different from typical relations in NLP such as ‘has_a’ and ‘is_a’. Extraction of these relations from text only is difficult and might need other knowledge resource such as equations and properties of materials. We, however, experimentally verified that a state-of-the-art machine learning model can extract these relations from text.

In the following sections, we formalize our task with three subtasks in Section 2 and describe our pipeline system for each subtask in Section 3. We are especially interested in the second subtask; we evaluated our system for the subtask in Section 4, and present the results in Section 5. We also briefly describe the end-to-end implementation in Section 6, and future works in Section 7.

2. Our approach and task definition

Our approach is knowledge graph population and graph search. The knowledge graph represents knowledge concerning material design. It consists of factors and their relations following PSPP reciprocity, and a PSPP chart is considered as a part of the graph with factors related to desired properties. Thus in this approach, we first extract the structure of the graph from text (knowledge graph population) and then find a PSPP chart from the graph for a desired material (graph search).

The PSPP knowledge graph consists of nodes and edges between them. Each node represents a factor, an important concept in material design. A factor is classified into one of ‘process’, ‘structure’, and ‘property’. Each edge is a relation between factors represented by the nodes. The nodes are connected by the edge if and only if the corresponding factors are related in the PSPP chart. Therefore, there is no edge between processing factors and structural factors.

A PSPP chart is considered part of the PSPP knowledge graph for a developing material. For instance, considering a typical material development scenario, where a new material is desired with specific properties, the PSPP chart is composed of factors related to the desired properties. The PSPP chart is part of the PSPP knowledge graph around the nodes of the desired property factors.

Following this approach, the task is decomposed into three subtasks: factor collection, relation identification, and branching.

The first subtask is to collect factors for nodes in the PSPP knowledge graph, and these factors are classified into each PSPP class. A factor is an important scientific concept for material design and is classified into ‘process’, ‘structure’, or ‘property’ following PSPP reciprocity. For example, process factors include ‘tempering’ and ‘hot working’; structural factors include ‘grain refining’ and ‘austenite dispersion’; and property factors include ‘strength’ and ‘cost’. Each factor is represented by a node in the PSPP knowledge graph, i.e. a node represents a processing, structural, or property factor.

The second subtask is relation identification, where relations among nodes in the PSPP knowledge graph are identified by reading text. In this subtask, for two given nodes and sentences mentioning the factor represented by these nodes, their relation is identified. Following PSPP charts, the relation is labeled in binary manner, i.e. positive or negative. A positive relation between factors A and B indicates that ‘factor A affects factor B’, and a negative relation between factors A and C indicates that ‘factor A occurs independently of factor C’. In a chart, two nodes are connected if their factors of the nodes have a positive relation, and are otherwise unconnected. Thus, denoting a pair of factors by (f1,f2), the desired relation is

(1) h(f1,f2)=pos/neg.(1)

The third subtask is to obtain a PSPP chart by branching the PSPP knowledge graph. We assume a scenario where a scientist is developing a new material with certain desired properties and looking for factors related to the properties in a PSPP chart. In this scenario, the PSPP chart is part of the PSPP knowledge graph, with certain factors around the desired properties. Thus, the subtask is to find part of the PSPP knowledge graph given a set of properties.

3. System description

Our system is a pipeline system consisting of three components for each subtask. The first component collects factors from a keyword list (Section 3.1). In the second component, a relation between two nodes is identified by reading sentences containing the factors represented by the nodes in the CNN model (Section 3.2). In the third component, part of the PSPP knowledge graph is extracted for the given desired properties by a simplified maximum flow algorithm (Section 3.3).

3.1. Factor collection

First, factors are collected from the keyword list of the journal Scripta Materialia.Footnote 3 Such keywords help identify the topics of each article. The keyword list is divided into five sections: 1) Synthesis and Processing; 2) Characterization; 3) Material Type; 4) Properties and Phenomena; and 5) Theory, Computer Simulations, and Modeling. In this approach, keywords in synthesis and processing are collected as processing factors, those in material type are collected as structural factors, and keywords in properties and phenomena as property factors.

Second, structural factors are collected from text using linguistic rules. From a material science standpoint, the number of structural factors is significantly greater than those of processing and property factors, and the keyword list is not long enough to cover structural factors. Candidate phrases, noun phrases consisting of multiple NNs (singular nouns, or mass nouns), are collected from a corpus described in Section 4.2 using Stanford CoreNLP [Citation17]. Each candidate phrase is classified into structural factors if it does not contain any words in the keyword list. The phrase containing a keyword is classified as the class of the keyword. For instance, lists two sentences with noun phrases. Here ‘phrase_transition’ is classified as a structural factor, but ‘hardness_distribution’ is classified as a property factor, as ‘hardness’ is in the keyword list.

Figure 1. The process-structure-property-performance reciprocity.

Figure 1. The process-structure-property-performance reciprocity.

Figure 2. Sentences containing noun phrases.

Figure 2. Sentences containing noun phrases.

All keywords and the n most frequent candidate phrases are collected, and each word/phrase is assigned a node in the PSPP knowledge graph. The total number of factors was 500, 500, and 1000 for process, property, and structural factors, respectively. lists samples of the n most frequent phrases.

Table 1. Samples of factors obtained by the linguistic rules.

3.2. Relation identification

In this section, we describe our CNN model for identifying the relation of a factor pair by weakly supervised learning. We also describe the linguistic resource for the factor pair where the model was trained.

The linguistic resource of a factor pair is a set of sentences mentioning both factors. A mention is a part of a sentence referring to a factor. A factor is mapped to the mention in the sentence by max-span string matching, i.e. a factor is mapped to the mention if the mention is the factor name, and no other mention overlaps the given mention. For instance,

• Within each phase, the properties are …

• When a substance undergoes a phase transition …

The phase in the first sentence is mapped to a factor, ‘phase’, but phrase transition is mapped to ‘phase_transition’ instead of ‘phase’ in the second sentence. A sentence in the linguistic resource of a factor pair is a sentence mentioning both factors.

The CNN model proposed by Huang et al. [Citation12] is a state-of-the-art deep neural network model for weakly supervised relation extraction. For each sentence in a linguistic resource, the network takes word embeddings and the relative position embeddings toward factors in the sentence. Convolutional units with a deep residual learning framework then embed the sentence into a vector representation. The sigmoid layer over the vector representation produces the probability distribution of the binary relation. shows the overall structure of the network.

The input to the model is a sentence and the output is a relation rpos,neg. Let (f1,f2) be factors of the relation and sSf1,f2 be the sentence, i.e. the linguistic resource, Sf1,f2 is a set of sentences mentioning the factors. Each sentence s is padded to a fixed length L.

A token embedding is a vector representation of a token in a sentence, denoting the sentence s=t0,...,ti,..., where ti is the ith token. The token representation is xi, which is a concatenation of a word embedding and two position embeddings. A word embedding W() is a vector representation of the word of the token whereas a position embedding P() gives a vector representation of the relative position of each factor:

(2) xi=[W(ti);P1(k1i);P2(k2i)](2)

where W()Rdw, P()Rdp and k1,k2 are the position indices of each factor. Note that any relative distance greater than Dmax is treated as Dmax.

A convolution layer takes embeddings around position i, and maps them into ciRdc:

(3) ci=g(wxi:i+h+b)(3)

where xi:i+h=[xi;xi+1;...;xi+h1], wRdc×h(dw+2dp) and bRdc is a bias. g is an element-wise non-linear function, ReLU.

Following the first convolutional layer, the other layers are stacked with residual learning connections that directly transmit a signal from a lower to a higher layer while skipping the middle layers. Thus, the kth residual CNN block consists of two CNN layers, with one taking signals from the two lower layers:

(4) cˆik=g(wˆk(c˜i:i+hk1+c˜i:i+hk2)+bˆk)(4)
(5) c˜ik=g(w˜kcˆi:i+hk+b˜k)(5)

where c˜0=c. The first CNN layer cˆik takes a signal from the immediately lower layer c˜i:i+hk1 and another signal from the lower block c˜i:i+hk2.

Max pooling is performed over the output of the last CNN units, c˜KRLh+1×dc.

(6) z=maxpoolic˜iK(6)

Then, two fully connected layers and a sigmoid function yield the probability distribution of the desired relation given in the sentence P(r=pos/neg|s):

(7) z1=g(wg1z+bg1)(7)
(8) z2=g(wg2z1+bg2)(8)
(9) P(r=pos|s)=σ(vrz2)(9)

where wgRdc×dc and bgRdc.

The desired probability P(r=pos| f1,f2) is the maximum of the probabilities over sentences. This is

(10) P(r=pos| f1,f2)=maxsSf1,f2P(r=pos|s)(10)

On the contrary, the model is trained on a weakly supervised approach, where the objective function is maximized for each sentence.

(11) maxΦ(f1,f2,r)DtrainsSf1,f2logP(r|s)(11)

where Dtrain is the training data, a set of tuples of factors f1,f2 and relation rpos,neg. The parameters Φ=W,P1,P2,w,b

3.3. Branching

The PSPP knowledge graph is branched for the given desired properties without losing related factors. We consider the branching of a max-flow problem, where the flow occurs from the given property factors to the processing factors. The inlets are all nodes of the processing factors and the outlets are those of the given properties. The capacity of each edge is the score of the relation, i.e. P(r=pos|f1,f2). We maximize the amount of flow with a limited number of nodes in the graph.

We compute the capacity of a node in the graph, which is the amount of flow that it can accept. Recalling that nodes of structural factors are connected to property and processing factors, and no processing factor and property factor are connected, all flow pass through the nodes of the structural factors. The capacity of the node of a structural factor fstr is

(12) Cfstr=minfPRCP(r=pos| f,fstr),fPRPP(r=pos| fstr,f)(12)

where PRC represents processing factors and PRP’ is the desired property factors. Similarly, the capacity of a node of a processing factor fprc is

(13) Cfprc=fSTRP(r=pos| fprc,f)(13)

where STR’ represents structural factors that are not branched.

The desired PSPP chart is composed of n processing factors, m structural factors, and the desired property factors, where n and m are the given hyper-parameters. The nodes of the processing/structural factors are the n and m most capable nodes. For efficiency, the nodes are greedily searched such that optimality is not guaranteed. The PSPP chart shows the processing/structural factors related to the desired properties.

4. Experiment for relation identification

Relation identification is a challenging subtask in this research. The system performance on the subtask was evaluated in a weakly supervised relation extraction setting. In this evaluation, the relation was identified as positive/negative for each factor pair.

The training data consisted of relationship data and a corpus. The relationship data consisted of a pair of factors and its relation label. The corpus, scientific literature, was a set of sentences describing the factors. A model was trained on part of the relationship data and the corpus and was tested on held-out data and the corpus by predicting relationships in the held-out data.

Our model was trained using stochastic gradient descent and dropout. Dropout randomly drops some signals in the network that are thought to help the generalization capabilities of the network. We employed an Adam optimizer with a learning rate of 0.00005 and randomly dropped signals from max pooling during training with a probability of 20%. The word embeddings were initialized with GloVe vectors [Citation18]. Other hyper-parameters are listed in .

Table 2. Hyper-parameters of the CNN model.

4.1. Relationship data

The relationship data consisted of tuples of two factors and their binary relation (pos/neg). From four design charts [Citation5], 104 factor pairs were collected as shown in and .

Table 3. Factors in the relationship data.

Table 4. Relations in the relationship data.

For evaluation, our model was trained on part of the relationship data and tested on the held-out data. The training data consisted of relationships from three arbitrary charts out of the four, and the test data consisted of relationships in the fourth chart. Thus, four pairs of training and test data were prepared for the evaluation. For accurate evaluation, the likelihoods of relationships in the test data were computed by a model trained using the corresponding training data. Precision and recall curves were then computed for the overall relationship data to obtain a smooth curve.

4.2. Corpus

Our corpus consisted of publicly accessible scientific articles on ScienceDirect.Footnote 4 ScienceDirect is an Elsevier platform providing access to articles in journals in a variety of fields, such as social sciences and engineering. Approximately 3400 articles were collected using the keyword search on ScienceDirect. The keywords were ‘material’ and ‘microstructure’, i.e. each article was related to both ‘material’ and ‘microstructure’.

The CNN models were trained on a pair of factors and sentences. As described in Section 3.2, each sentence mentions both factors. For the relationship data, about 5000 sentences were founded in the corpus in total, roughly 50 sentences for each pair of factors on average.

4.3. Baseline models

A baseline model is a text-classification-based binary classifier where for each factor pair, each classifier takes a set of sentences mentioning the factors and classifies the text into a positive or negative relation. The problem setting and the set of sentences were exactly the same as the one in the CNN model.

Logistic regression and SVM with bag-of-words features were employed for the binary classifier. These are standard machine learning binary classifiers. Bag-of-words is a feature that indicates whether a word is in a set of sentences. The feature is represented by a sparse binary vector, where an element is one if the corresponding word is in the sentences and zero otherwise.

Stop words removal and n-gram features are explored in and ; however, the effect was limited. Note that the radial basis function (RBF) kernel was used in all SVM models.

Figure 3. Structure of the CNN model. The convolutional layers embed a sentence, and the max pooling and two fully connected layers give a binary probability distribution with a sigmoid function.

Figure 3. Structure of the CNN model. The convolutional layers embed a sentence, and the max pooling and two fully connected layers give a binary probability distribution with a sigmoid function.

Figure 4. Precision-recall curve of the logistic regression model. The features are ‘bag of words’, ‘bag of words + stop word removal’ and ‘bag of unigram + bigram + trigram’.

Figure 4. Precision-recall curve of the logistic regression model. The features are ‘bag of words’, ‘bag of words + stop word removal’ and ‘bag of unigram + bigram + trigram’.

4.4. Evaluation metric

The evaluation metrics were precision and recall, which are the standard metrics for information extraction tasks. Precision is the ratio of correctly predicted positive factor pairs to all predicted positive factor pairs and gives the accuracy of the prediction. Recall is the ratio of correct predictions to all positive factor pairs in the test data and gives the coverage of the prediction. A positive factor pair is a pair whose relation is positive. We obtain high precision and low recall if a system returns only a small number of high confidence predictions, and low precision and high recall if a system returns many low confidence predictions. Typically, these are balanced by a hyper-parameter (confidence) of system prediction. Thus, the trajectory of precision and recall pairs is computed with various values of the hyper-parameter and is called as a precision-recall curve.

In this evaluation, the hyper-parameter was an integer t, the number of positive factor pairs in the prediction. For a given t and a set of factor pairs in the test relationship data, the system predicts a binary relation, pos/neg, for each pair. It predicts the t most likely positive pairs, and the other pairs are predicted as negative.

The factor pairs in the test relationship data were scored by a machine learning model trained on the corresponding training relationship data, where the score was P(r=pos|f1,f2). A test data corresponded with a training data, unaware of the relationships in the test data (Section 4.1). A model was trained on the corresponding training data and scored a pair in the test data to avoid letting the model know the true relationships during training.

Then, a precision and recall pair for a given hyper-parameter t was computed as follows:

(14) Precisiont=RtRtestt(14)
(15) Recallt=RtRtestRtest(15)

where Rtest is the set of factor pairs with positive relations in all test relationship data, and Rt represents the t most likely positive factor pairs. The likelihood was a score given by the model.

5. Results of relation identification

Precision-recall curves for the baseline models are shown in and . These figures show various feature representation schemes, such as stop words and n-grams (Section 4.3) on the logistic and SVM models. The logistic model performed well on low recall space, i.e. most confidently predicted positive factor pairs were actually positively related. On the contrary, the performance of the SVM model was poorer in the space but better overall than the logistic model. In both models, the effects of the feature representation schemes were limited.

Figure 5. Precision-recall curve of the SVM model. The features are ‘bag of words’, ‘bag of words + stop word removal’ and ‘bag of unigram + bigram + trigram’.

Figure 5. Precision-recall curve of the SVM model. The features are ‘bag of words’, ‘bag of words + stop word removal’ and ‘bag of unigram + bigram + trigram’.

The performance of the CNN model is shown in . The precision was one when the recall was about 0.4, i.e. roughly speaking half the positive factor pairs were perfectly identified. The performance was superior to that of the baseline models.

Figure 6. Precision-recall curve over the relationship data of the CNN model.

Figure 6. Precision-recall curve over the relationship data of the CNN model.

shows some representative sentences scored by the CNN model. A representative sentence is the highest scored sentence in a sentence set Sf1,f2 for each factor pair, i.e. a representative sentence is s=argmaxsSf1,f2P(r=pos|s). The score is a likelihood where a positive relation is described in the sentence. The sentence with the highest score in each sentence set most likely represents the positive relation of each factor pair, according to the CNN model.

Table 5. Sample representative sentences scored by the CNN model. Label P indicates that the factors are positively related in the test relationship data and label N indicates a negative relation. Factors in each sentence are underlined. The score is the vrz2 of each sentence. See Appendix for the source articles.

Appendix. Source articles.

Representative highly scoring sentences seem to describe the desired relations (sentences 4 and 8) and, interestingly, relations described in the equation were also discovered by the model (sentences 2, 3, and 6). This implies that some important relations tend to be described in an equation. This result also indicates that the relations in which we are interested are significantly different from typical relations in other NLP tasks like ‘has_a’, ‘is_a’.

6. End-to-end system

An end-to-end demo system was developed to test our CAMaD system on Apache TomcatFootnote 5 as . The demo system worked in a typical scenario, where a scientist was looking for factors related to certain desired properties. The demo system provides a PSPP design chart for the desired properties that the scientist provided.

Figure 7. The end-to-end demo system. a) Desired properties and a base material were selected. b) A sample of the generated PSPP design chart. The desired properties were toughness and creep strength, and ‘steel’ was selected as base material. c) The representative sentence describing the relation between toughness and carbon content.

Figure 7. The end-to-end demo system. a) Desired properties and a base material were selected. b) A sample of the generated PSPP design chart. The desired properties were toughness and creep strength, and ‘steel’ was selected as base material. c) The representative sentence describing the relation between toughness and carbon content.

The system input consisted of the desired properties along with a base material. The desired properties were selected from a list of properties collected as in Section 3.1. The base material was the target material, such as aluminum or titanium. It was important to obtain the desired knowledge. For example, the relationship between strength and matrix in titanium alloys might have been different from this relationship in aluminum alloys.

Thus, relationships were extracted from the scientific literature describing the base material. A set of studies was specified by the base material to find the desired knowledge. As in Section 4.2, the literature was collected by keyword search in ScienceDirect. All relations among the factors collected in Section 3.1 were scored by the CNN model as in Section 3.2 and some relationships were branched as described in Section 3.3.

The system output was a PSPP design chart suggesting the required structures and processes. The chart formed by three columns – process, structure, and property – suggested relations from the processes to the desired properties. Moreover, for each relation, a representative sentence for each factor pair was provided to justify the relation and aid the researcher’s understanding.

7. Conclusions and future work

In this study, we developed and tested a CAMaD system, a progressive knowledge extraction and representation system intended to support material design, by representing knowledge as relationships. Knowledge was represented as relationships in PSPP design charts. We leveraged weakly supervised learning for relation extraction. The end-to-end system proved our concept, and its relation identification performance was superior to that of other baseline models.

Further evaluation is the major feature of our work. In spite of the impressive results of relation extraction and brightness of the end-to-end system, a natural evaluation metric for the end-to-end task remains unclear. Additionally, system performances on other subtasks, factor collection and branching, are not evaluated in this study. Evaluations of these subtasks are not trivial. In factor collection, there are infinite number of factors in material design, and it is difficult to see the coverage of factors a system collected. In branching, there is no natural metric to compare or evaluate PSPP design charts. Thus, end-to-end evaluations are even more challenging.

We consider factor collection and mapping as the bottleneck of our system. The factor collection described in Section 3.1 and these factors were mapped into sentences that refer each factor in Section 3.2. Unlike in previous works [Citation7,Citation11], the factors were not named entities. Any noun phrase can be a factor, and factors were not predefined. At present, our system recognizes factors in a sentence using string matching. The obtained factors appear to be noisy and are not correctly categorized in some cases.

A natural extension of this task is multi-labeling. The relation we use at present is binary (pos/neg), and only identifies whether two factors are related. This label might be too abstract. In multi-labeling, a relation between factors is described with a label such as produce, depress, and independent. We believe multi-labeling renders the task more natural and informative.

Acknowledgments

We are grateful to Dr. Yutaka Sasaki and Dr. Makoto Miwa from Toyota Technological Institute for insightful comments.

Disclosure statement

No potential conflict of interest was reported by the authors. The underlying research materials for this article can be accessed at https://bitbucket.org/0024takeshi/pspp_relation.

Notes

References

  • Augenstein I , Das M , Riedel S , et al. SemEval 2017 Task 10: ScienceIE -Extracting Keyphrases and Relations from Scientific Publications. In: Proceedings of the 11th International Workshop on Semantic Evaluations (SemEval-2017); Vancouver (Canada); 2017. p. 546–555.
  • Cohen M. Unknowables in the essence of materials science and engineering. Mater Sci Eng. 1976;25:3–4.
  • Olson GB . Genomic materials design: the ferrous frontier. Acta Mater. 2013;61(3):771–781.
  • Olson GB . Designing a new material world. Science. 2000;288(5468):993–998.
  • Xiong W , Olson GB . Cybermaterials: Materials by design and accelerated insertion of materials. npj Computational Materials. 2016;2:15009 .
  • Xu Y , Yamazaki M , Villars P . Inorganic materials database for exploring the nature of material. Jpn J Appl Phys. 2011;50(11S):11RH02.
  • Mintz M , Bills S , Snow R , et al. Distant supervision for relation extraction without labeled data. In: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP; Suntec (Singapore); 2009. p. 1003–1011.
  • Hoffmann R , Zhang C , Ling X , et al. Knowledge-based weak supervision for information extraction of overlapping relations. In: Proceedings of 49th Annual Meeting of the Association for Computational Linguistics; Portland, OR; 2011. p. 541–550.
  • Surdeanu M , Tibshirani J , Nallapati R , et al. Multi-instance multi-label learning for relation extraction. In: Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning; Jeju Island (Korea); 2011. p. 455–465.
  • Fan M , Zhao D , Zhou Q , et al. Distant supervision for relation extraction with matrix completion. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics; Baltimore, MD; 2014. p. 839–849.
  • Riedel S , Yao L , McCallum A . Modeling relations and their mentions without labeled text. In: Proceedings of the 2010 European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases; Catalonia (Spain); 2010. p. 148–163.
  • Huang YY , Wang WY. Deep residual learning for weakly-supervised relation extraction. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing; Copenhagen (Denmark); 2017. p. 1803–1807.
  • Zeng D , Liu K , Chen Y , et al. Distant supervision for relation extraction via piecewise convolutional neural networks. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing; Lisbon (Portugal); 2015. p. 1753–1762.
  • Lin Y , Shen S , Liu Z , et al. Neural relation extraction with selective attention over instances. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics; Berlin (Germany); 2016. p. 2124–2133.
  • Ji G , Liu K , He S , et al. Distant supervision for relation extraction with sentence-level attention and entity descriptions. In: Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence; San Francisco (CA); 2017.
  • Liu T , Wang K , Chang B , et al. A soft-label method for noise-tolerant distantly supervised relation extraction. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing; Copenhagen (Denmark); 2017. p. 1791–1796.
  • Manning CD, Surdeanu M , Bauer J , et al. The Stanford CoreNLP Natural Language Processing Toolkit. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations; Baltimore (MD); 2014. p. 55–60.
  • Pennington J , Socher R , Manning CD. GloVe: Global Vectors for Word Representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing; Doha (Qatar); 2014. p. 1532–1543.