[go: up one dir, main page]

Next Article in Journal
Unraveling Hidden Major Factors by Breaking Heterogeneity into Homogeneous Parts within Many-System Problems
Previous Article in Journal
Rotating Minimal Thermodynamic Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Word-Granular Adversarial Attacks Framework for Causal Event Extraction

1
Colledge of Computer Science and Technology, Jilin University, Changchun 130015, China
2
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130015, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(2), 169; https://doi.org/10.3390/e24020169
Submission received: 26 November 2021 / Revised: 12 January 2022 / Accepted: 22 January 2022 / Published: 24 January 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
As a data augmentation method, masking word is commonly used in many natural language processing tasks. However, most mask methods are based on rules and are not related to downstream tasks. In this paper, we propose a novel masking word generator, named Actor-Critic Mask Model (ACMM), which can adaptively adjust the mask strategy according to the performance of downstream tasks. In order to demonstrate the effectiveness of the method, we conducted experiments on two causal event extraction datasets. Experiment results show that, compared with various rule-based masking methods, the masked sentences generated by our proposed method can significantly enhance the generalization of the model and improve the model performance.

1. Introduction

Constructing a causal knowledge graph has essential impact on natural language reasoning tasks. In addition, how to extract causal event pairs from texts is a fundamental problem. In the early studies, causal event extraction was formulated as a pipeline process [1]. It identifies all potential event pairs and performs causal relation classification for each of them. However, this extraction method tends to suffer from the error propagation problem, which means the error from the extracted event pair will be passed to the classification process. One way to mitigate this effect is a sort of sequence labeling task [2]. The task performs a multi-class identification on each word in sentence. Through such a sequence labeling task, the problem in the pipeline method can be solved. In general, this joint sequence labeling method focuses on how to obtain word semantic representations to achieve better results.
With the emergence of large scale pre-trained language models [3,4,5], multiple downstream tasks have achieved SOTA results. In our experiments, the representation of pre-trained language models is also suitable for causal event extraction task. On the other hand, the current causal event extraction datasets are relatively small. A language model, such as BERT, has very large amount of parameters. It is unsuitable to add more components to the model, which is very prone to overfitting problems. With this in mind, data augmentation on a small corpus is an effective method.
For causality extraction, the target is to extract event pairs with causal relations. Some sentences contain explicit causal connectives, such as “cause” and “due to”. These connectives clearly show the causal relationship between events. Most of the causal event pairs are on the left and right sides of the connective, as shown in the Figure 1. Furthermore, there are some sentences that contain causal events with no obvious connectives. To identify implicit causal relation, we need to pay more attention to the semantic connection between causal event pairs rather than connectives. How to identify event pairs in implicit causal relationships is a difficult problem. In this paper, our model uses adaptive masking of causal connectives to make the model more focused on the relationship between events. This method allows the model to extract events that have causal relationships without excessive causal conjunctions.
On the contrary, the original method of masking word is a random process. The performance depends on multiple random selections, which is unstable. In this paper, we propose a mask word generation model based on reinforcement learning. It generates a task-adaptive mask word distribution through confrontation with the BERT model, which can mask words in a targeted manner to capture the semantic relation of causal events better.
Specifically, we adopt the A2C framework to obtain a mask word distribution generator. It consists of a BERT-based encoder, a relational classifier and a mask position generator. Experiments show that, for the two datasets, the performance of using a large pre-training model has surpassed the SOTA model of the sequence labeling task. However, a model with adaptive mask generator can achieve a better performance. At the same time, we also compared with the rule-based method, and it also shows the superiority of our model.
The main contributions of our work are as follows:
  • We propose an adaptive mask model that utilizing reinforcement learning to obtain a word mask distribution. In addition, through adversarial training methods, the generalization performance of large models on small datasets can be improved. Our method proposes a new perspective to combine reinforcement learning and deep learning to solve problems.
  • We calibrate the causal dataset extracted from a benchmark dataset and obtain a more accurate one with causal event pair annotation.
  • We conduct extensive experiments on the causal event extraction datasets. Experiments show that the model with ACMM is better than its basic version.

2. Related Works

2.1. Causality Extraction

As an essential task, the causality extraction task has applications in many upstream tasks, such as event prediction [6], scenario generation [7], question and answering [8], and so forth. In some real-world tasks, such as medicine [9], there are also applications of causality extraction. At a higher level, the task of causal extraction and causal reasoning is inextricably linked. In some viewpoints, artificial intelligence will be fundamentally restricted if there is no causal reasoning ability [10].
At present, the research on causality extraction based on deep learning is mainly divided into three categories. The first is to use the method of pattern matching, which uses semantic features and vocabulary symbol features to extract causality [11]. This method, that completely relies on pattern matching, has poor cross-domain adaptability and requires expert knowledge. The second method is based on a combination of pattern matching and machine learning [8,12]. This method generally uses a pipeline method. First, the pattern matching method is used to extract the potential causal entity pairs, and then the machine learning method is used to judge the candidate causal pairs. However, this method also requires domain knowledge and has the problem of error transmission, i.e., errors in the first stage will affect the effect of the second stage. The third method is based on deep learning. Because it does not require domain knowledge and can effectively capture ambiguous causality, it has become the most commonly used method in recent years. Some commonly used models have applications in causality extraction, such as CNN [13], LSTM [14], etc. At the same time, there are some studies on the introduction of external knowledge for causal extraction [15].

2.2. Pre-Trained Language Model

Using a representation language model pre-trained from large-scale unlabeled text is a universal and effective method in most natural language understanding tasks [16]. Most of these language models use self-supervised training methods. With the success of MLM pre-training [17], a large number of MLM-based pre-training methods and models have been proposed [4,18,19,20]. However, these tasks are limited to applying the mask strategy to the pre-training stage. There is no attempt to learn a mask strategy for downstream tasks and data, which is the problem we want to solve in this work.

2.3. Reinforcement Learning-Based NLP

Deep reinforcement learning is another system different from neural networks. It currently performs well on many decision-based problems. For example, in games or other fields, it has many applications [21,22,23]. At the same time, we believe that the application of reinforcement learning to NLP still has great potential. In some NLP tasks involving serialization decision-making, reinforcement learning has many applications, such as multi-document information extraction [24], reference resolution [25], text classification [26], text summarization [27], etc. In most NLP tasks, reinforcement learning is used as an auxiliary task. For example, reinforcement learning is used as a model for deleting irrelevant words and judging the structure of the current phrase in text classification tasks [26]. In addition, in text summarization, the reinforcement learning model distinguish important paragraphs from a large amount of text to filter irrelevant information. Then, the deep learning model uses the condensed text for text summarization tasks. RL method is also used in large-scale pre-training language models. For instance, a masking word strategy is learned through reinforcement learning to enhance the convergence rate [28] or task adaptation capabilities of the pre-training model [29]. The difference between this article and the above work is that we focus on the fine-tuning stage of the task and design a masking word strategy against the model to improve the generalization of the model.

3. Materials and Methods

3.1. Datasets

We evaluated our framework on the public datasets SemEval-2010 Task 8 [30] and Causal TB (Causal TimeBank) [31] (only the entity annotations for causality are extracted). We only kept the causal annotations for the Causal TB dataset and transformed them into a sequence annotation dataset. For the SemEval dataset, some parts of the causal event pair are not accurately labeled. Therefore, in order to identify the causal relationship, we performed the following processing on the dataset.
  • Each word pair with a causal relationship is labeled as one of the “B-Cause”, “I-Cause”, “B-Effect”, “I-Effect”; other relationships and irrelevant words are labeled as “O”.
  • The SemEval dataset only contains a pair of causal relationship entity tags. Yet, there is a variety of one-to-many, many-to-many causal relationships (Examples are shown in Figure 2). We have labeled these data.
The details of the two datasets are shown in the Table 1. After reprocessing, the causal event pair of the SemEval dataset is expanded from 904 to 1053. Due to the lack of datasets for causal event relations, our data volume is relatively small, which can illustrate the improved generalization performance of our adversarial model.

3.2. Methods

Our framework is mainly divided into two parts: language model warm-up; and mask model and language model confrontation training. In the initial stage, we use conventional methods to warm-up the language model, which means we fine-tune the model on the train dataset. After obtaining a better language model, we regard it as an environment model to warm-up the mask model. Finally, after obtaining a stable mask model, we use the adversarial training method to train two models simultaneously. The main structure of the frame is shown in Figure 3.

3.2.1. Encoder Model Warm-Up

We employ pre-trained BERT as the primary language model and use the word-piece sequence after word segmentation as input to obtain a high-dimensional representation of the sentence. In subsequent training step, it is used as the original input into each model. Before introducing the model, let us briefly clarify the mathematical notations we used. We use X = ( x 1 , , x n ) to represent an input sample, where x i represents the index of each word. E represents the hidden state generated by the encoder model, where E = ( e 1 , , e n ) . θ e n c o d e r represents the parameters of the context encoder model used (BERT), θ c l s represents the parameters of the classifier model, and θ a 2 c represents the parameters that mask word generator model used. The masked position generated by the masked word generator is replaced with [ M A S K ] .
First, we use the encoder model to obtain the hidden representation of the sentence and then obtain the classification of each word through the classifier. Then, we calculate the loss through the cross-entropy loss function (1).
L c l s ( θ e n c o d e r , θ c l s | X ) = 1 N i = 1 N c = 1 M z j c log ( p j l a b e l = c ) ,
z j c = I ( y j c = y j l a b e l ) .
N is the number of words. M is the number of tags, which is 5 in this article. z j c is an indicator function. z j c = 1 means the predicted label is equal to the actual label. In the warm-up stage, we use the BERT model and classifier to fine-tune the training set. In the experiment, we use 10 epochs for warm-up, which is a hyperparameter.

3.2.2. Environment Model

For most reinforcement learning algorithms, we need to obtain the current state score and the next state. Both of them are taken from the environment. In order to make it suitable for the training of the mask model, we adopt the following method:
We utilize the high-dimensional vector obtained by the BERT encoder model to represent the state. For ease of use, we have performed dimension reduction operations on it.
S = E n c o d e r X ; θ e n c o d e r W env + b env .
We use X mask to represent the sequence masked by the mask model, and the next state is represented as:
S n e x t = E n c o d e r X mask ; θ e n c o d e r W env + b env .
For each state S, we regarded the loss produced by classifier as the score.
s c o r e = L c l s ( θ e n c o d e r , θ c l s | X ) .

3.2.3. Mask Model Overview

For the mask model, we use the advantage-actor-critic [32] framework. Since this task cannot be modeled as a split-screen Markov process, we transform this task into a TD(0) method. Specifically, we use an actor model to generate the action distribution p a of current state. In order to obtain the more extensive exploration space of reinforcement learning, we sampled p a repeatedly and obtained multiple next states through different actions s t a t e i = s t a t e 1 , , s t a t e n . Through the environment model, we obtain the scores for these states, and then the reward of each action is calculated as the formula:
r i = R e L U ( s c o r e s t a t e i / s c o r e s t a t e 1 ) .
As the environment model is trained, its score will become smaller. To avoid the impact of the decreased score value, we use the form of percentage to represent the improvement of the score, which indicates how much the score has improved compared to the previous state. At the same time, we limit the number of covered words, which is a hyperparameter. It is obvious that the more words are covered, the vaster the losses are generated. All in all, Reward indicates that the model hopes to mask the word that has the most significant impact on predicting. Finally, we will obtain a mask distribution p a for the task. Subsequently, we will introduce the details of the actor model and the critic model, respectively.

3.2.4. Critic Model

In the A2C framework, the critic model is generally used to calculate the state value. First, in order to make effective use of BERT’s powerful language representation capabilities, the critic model and the actor model share the dynamic word vector E generated by the BERT encoder, yet it does not perform backpropagation. In the low layer, the critic model uses a simple linear layer to estimate the current state’s value. Then, we employ the Mean Square Error function to calculate the difference between the estimated value and the reward.
V π s   = P θ c r i t i c | s t a t e ,
L c r i t i c = M S E R , V π s , a .

3.2.5. Actor Model

We use the advantage function to update the policy gradient of the actor model. Specifically, we use A to represent the advantage function, which able to reduce the gradient for stable training. The calculation of the advantage function is as follows:
A π s t , a t   = r s t , a t + V π s t + 1 V π s t .
Therefore, using the policy gradient method [33], our gradient to be updated is
J θ a c t o r   = θ a c t o r log π θ a c t o r a t , s t A π s t , a t .
Update the strategy parameters θ a c t o r :
θ a c t o r = θ a c t o r + α J θ a c t o r .
The A2C model is not a method of calculating value functions, and its on-policy feature makes it impossible to use Experience Replay operations. For the purpose of making the model converge faster, we use a function similar to Experience Replay, which is different from the function in DQN. We store each ( s t a t e , a c t i o n , r e w a r d , n e x t _ s t a t e ) quadruple in the reply buffer and, when the reply buffer is full, randomly select a batch of data for training. We do this because we need the predicted probability and its calculation graph corresponding to each action during training. It is unrealistic to save. However, our purpose is only to obtain a distribution that covers each word. To this end, we make use of each state to re-pass the model to obtain a calculation graph of the current strategy and then use the previously explored action to update the parameters. In this way, we can meet its on-policy property and explore a larger space for action.

3.3. Adversarial Training

In pace with training of environment, the scores corresponding to the same state will also change. Thence, after the hot start, we need to train the reinforcement learning part and the deep learning part at the same time. We adopt a kind of adversarial training method, and the details as shown in Algorithm 1.
Algorithm 1 Adversarial training.
Require: environment model E N V , mask model A 2 C , dataset D, Sampling threshold ϵ
1:
Fixed environment model parameters θ e n c o d e r , θ c l s ;
2:
while done do
3:
   for  d a t a in D do
4:
     Train the mask model and update the parameters θ a 2 c ;
5:
   end for
6:
   fixed mask model parameters θ a 2 c
7:
   for  d a t a in D do
8:
     Sampling random number r 0 , 1 ;
9:
     if  r > ϵ  then
10:
        Get new data n d a t a from A 2 C model;
11:
        Train the E N V model and update the parameters θ e n c o d e r , θ c l s with n d a t a ;
12:
     else
13:
        Train the E N V model and update the parameters θ e n c o d e r , θ c l s with d a t a ;
14:
     end if
15:
   end for
16:
end while

4. Results and Discussion

4.1. Experimental Setting

4.1.1. Evaluation method

In this part, we will introduce our evaluation indicators: P (Precision), R (Recall), and F1 (F1-score). Their calculation formula is as follows:
P = T P T P + F P , R = T P T P + F N , F 1 = 2 × P × R P + R .
Among them, TP and FP indicate how many labels predicted to be Positive are correct and incorrect, respectively. FN means the number of labels predicted to be negative is actually wrong. In short, P represents how many of the predicted true labels are truly true labels, and R represents how many truly true labels are predicted.
For this sequence labeling task, the number of “O” labels is far greater than the number of causal labels. Therefore, we first calculate the F1 value of each label, and then average it as the final indicator. The method is called Macro-F1 and can reduce the impact of too many “O” labels.

4.1.2. Implementation Details

We adopt mini-bach mechanism to train our model. As seen in the two pictures above, in Figure 4, we have conducted many experiments by setting the batch size to [ 4 , 8 , 16 , 32 , 64 ] , and, finally, we have chosen a better batch size of 8. As for the learning rate, we found that it is a better choice to choose different learning rates for different parameters through experiments. We set the learning rate of the θ e n c o d e r parameters to 1 × 10 5 , and the learning rate of the parameters of θ a 2 c to 1 × 10 4 . As for the learning rates of the models BiLSTM [34] and TARGER [35], we refer to the corresponding values in the paper, which are 0.1 and 0.01, respectively.
We first use 10 epochs hot start to fine-tune the encoder model. As shown in the last two pictures of Figure 4, the model to be stabilize at about 10–15 epochs. In order to choose a more stable language model, we think the model in 10 epochs is a better choice. We also use an early stopping mechanism, and the training is terminated when the performance on validation set has not improved in the last 20 epochs. We use the Adam optimizer for training, and the parameters selected the recommended options in the paper [36].

4.2. Baselines

To illustrate the effect of our method in the causal event extraction task, we choose both BiLSTM-based and BERT-based methods that achieve state-of-the-arts results as baselines in our experiments.
BiLSTM-CRF [34] is a classic sequence labeling method, including a BiLSTM encoder and a CRF classifier.
TARGER [35] is a sequence labeling model that uses CRF classifier and CNN-BiLSTM encoder to learn Char-level features and Word level features.
BERT [17] is a large pre-trained language model composed of multiple transformer blocks, performing well in multiple downstream tasks.
BERT-CRF and BERT-BiLSTM-CRF are extensions of the BERT model, using pre-trained word embedding generated by BERT as input.
BERT*-ACMM is our proposed model, among which ACMM (Actor-Critic Mask Model) is a model for confrontation with the BERT model and extensions of BERT model. All the parameters of the above BERT model use the huggingface’s bert-base-uncased version (https://huggingface.co/models, accessed on 13 March 2020).

4.3. Experimental Results

In Table 2, we report the performance of the baseline and our model on the task of causality extraction. The first part of the table is the traditional model that does not use the BERT pre-trained model, the second part is the model that uses BERT as encoder, and the third part is the second part of the model with our adversarial model components.
In the first part of the Table 2, we chose traditional neural network models, such as CNN and RNN. The experimental results show that on this small dataset, the generalization performance of the model is not good. As a recently popular pre-training model in NLP, we use BERT and its variants as a representative of this type of model. The experimental results show that this model is better than the traditional network structure. There are 12 points and 3 points of F1-value improvement on SemEval and Causal TB, respectively. This is due to its large amount of language pre-training knowledge and complex structure. However, this improvement is not endless. On the smaller dataset Causal TB, we can see that, due to the increase in model complexity, its generalization is weakened. From BERT-CRF to BERT-BiLSTM-CRF, the F1-value of the model is reduced by 2.52.
Since our model only interacts with the language model at word granularity, it does not increase its model complexity. However, after adding our adversarial component to each model, its performance has improved. For example, the BERT model has increased the F1 value by 0.57 and 3.63, respectively. Similarly, the boost values of the BERT-BiLSTM-CRF model on the two datasets are 0.51 and 1.98. This is very exciting because our adversarial component works better on small datasets (Causal TB). These experiments show that, when increasing model complexity is no longer effective, we provide an alternative method to continue to improve model performance.

4.4. Analysis of Different Mask Strategy

In this section, we compared the masking strategy we generated with various rule-based masking strategies. The following is a detailed introduction of each word masking method:
No Mask uses the basic language model (BERT) and linear classifier for fine-tuning without using any mask strategy.
Whole Word takes the entire word as a unit and randomly mask a specified number of words [37].
Word Piece is used during BERT pre-training. For the word sequence that has been segmented, randomly mask the specified proportion of tokens.
Span randomly selects multiple consecutive tokens for masking for each sentence [19].
As shown in Table 3, we used three random word masking strategies to compare with our strategy. It can be seen that, although this data enhancement method has obvious effects on the Causal TB dataset, it is not effective for the SemEval dataset. It is nearly 3 points in Causal TB, and not improved in SemEval. This result shows that, in large datasets, this data augmentation method is not particularly effective. The reason is that, as many recent studies have shown, this simple noise is limited to the improvement of the model [38]. However, our reinforcement learning strategy is different from it. We use the propagation of loss to generate difficult samples to obtain better results. As you can see, our strategy has achieved the best results compared to other strategies. Even in the SemEval dataset, the best recall appears in the Whole Word strategy, but its precision is significantly reduced. This causes its F1 value to be inferior to our strategy, the same as Causal TB.
At the same time, another point that shows our model is superior to rule-based strategies is its stability. To demonstrate this, we performed 5 repeated experiments for each model using random training seeds. As shown in Figure 5, we use box plots to show the mean and variance of various strategies. It can be seen that the variance of the model without any masking strategy is the smallest. When the random strategy is adopted, the variance of the model becomes larger. This is because, in addition to the randomness of the model parameters, the randomness of the masking words is added. This result makes the model more unstable, and the result of each experiment depends on the random result. However, as can be seen from the figure, although our model adds randomness, it is more stable. This is because our strategy is based on the probability distribution p a output by the ACMM model, which has a higher probability of deleting important words. This kind of method can not only generate difficult samples but also ensure stability.

4.5. Analysis of Masked Sentence

In Figure 6, we have selected three sentences to show the effect of the mask model. It can be seen that, whether it is explicit causality or implicit causality, the causal entity and the words in the causal entity always have a higher probability of being masked (For the stability of the model, we adopted some heuristics, that is, the causal entity is not masked.). The possible reason is that obliterating causal entities is the most likely to reduce the accuracy of model, and the words between entities also have a high probability of indicating whether the entities have causal relationships.
In explicit causality, the model can accurately distinguish causal connectives whether in long sentences or short sentences. In implicit causality, although the word between the entity does not clearly indicate causal relation (e.g., “after”), it is more likely to cause prediction errors in the model than other words in the sentence. From the perspective of information theory, our model is more willing to delete words that contain more information for predicting causality. For example, our concealment probability for meaningless conjunctions (such as “a”, “the”, and “of”) is much lower than words with actual meaning (such as “city”, “pentagon”, “yards”, etc.).

5. Conclusions

We propose a novel framework that can automatically generate adversarial attacks on language models at word granularity. The adversarial attacks we generate for causality extraction tasks can make the model achieve better generalization performance on small datasets. To this end, we use the actor-critic model to generate adversarial examples in a targeted manner. We performed empirical studies on multiple models and different rule-based mask strategies, which show the effectiveness of our optimal mask strategy.

Author Contributions

Conceptualization, W.Z. and X.Z.; methodology, Y.Z. (Yu Zhao) and S.L.; software, Y.Z. (Yu Zhao); validation, X.Y. and Y.Z. (Yijia Zhang); formal analysis, Y.Z. (Yijia Zhang); investigation, X.Y.; resources, W.Z.; data curation, X.Z.; writing—original draft preparation, Y.Z. (Yu Zhao); writing—review and editing, W.Z. and S.L.; visualization, X.Y.; supervision, W.Z.; project administration, X.Z.; funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is sponsored by the National Natural Science Foundation of China (61976103, 61872161), the Scientific and Technological Development Program of Jilin Province (20190302029GX, 20180101330JC, 20180101328JC). All opinions, findings, conclusions, and recommendations in this paper are those of the authors and do not necessarily reflect the views of the funding agencies.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/zy234/ACMM/tree/main/data, accessed on 15 June 2021.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACMMActor-Critic Mask Model
SOTAstate of the art
A2CAdvantage Actor-Critic
MLMMask Language Model
NLPNatural Language Processing
DQNDeep Q-Learning
Causal TBCausal TimeBank
TDTemporal-Difference

References

  1. Dmitry, Z.; Chinatsu, A.; Anthony, R. Kernel methods for relation extraction. J. Mach. Learn. Res. 2003, 3, 1083–1106. [Google Scholar]
  2. Suncong, Z.; Feng, W.; Hongyun, B.; Yuexing, H.; Peng, Z.; Bo, X. Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; Volume 1: Long Papers. [Google Scholar]
  3. Clark, K.; Luong, M.-T.; Le Quoc, V.; Manning Christopher, D. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  4. Liu, Y.; Myle, O.; Naman, G.; Jingfei, D.; Mandar, J.; Danqi, C.; Omer, L.; Mike, L.; Luke, Z.; Veselin, S. Roberta: A robustly optimized bert pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  5. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL-HLT, Minneapolis, MN, USA, 6–7 June 2019. [Google Scholar]
  6. Kira, R.; Sagie, D.; Shaul, M. Learning causality for news events prediction. In Proceedings of the 21st International Conference on World Wide Web, New York, NY, USA, 16–20 April 2012. [Google Scholar]
  7. Chikara, H.; Kentaro, T.; Julien, K.; Motoki, S.; Istv’an, V.; Jong-Hoon, O.; Yutaka, K. Toward future scenario generation: Extracting event causality exploiting semantic relation, context, and association features. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 22–27 June 2014; Volume 1. [Google Scholar]
  8. Roxana, G. Automatic detection of causal relations for question answering. In Proceedings of the ACL 2003 Workshop on Multilingual Summarization and Question Answering, Sapporo, Japan, 7–12 July 2003. [Google Scholar]
  9. Lee, D.; Shin, H. Disease causality extraction based on lexical semantics and document-clause frequency from biomedical literature. BMC Med. Inform. Decis. Mak. 2017, 17, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Judea, P.; Dana, M. The Book of Why: The New Science of Cause and Effect; Basic Books: New York, NY, USA, 2018. [Google Scholar]
  11. Khoo, C.S.G.; Kornfilt, J.; Oddy, R.N.; Myaeng, S.H. Automatic extraction of cause-effect information from newspaper text without knowledge-based inferencing. Lit. Linguist. Comput. 1998, 13, 177–186. [Google Scholar] [CrossRef]
  12. Gordon, A.S.; Bejan, C.A.; Sagae, K. Commonsense causal reasoning using millions of personal stories. In Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 7–11 August 2011. [Google Scholar]
  13. Kruengkrai, C.; Torisawa, K.; Hashimoto, C.; Kloetzer, J.; Oh, J.; Tanaka, M. Improving event causality recognition with multiple background knowledge sources using multi-column convolutional neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  14. Dasgupta, T.; Saha, R.; Dey, L.; Naskar, A. Automatic extraction of causal relations from text using linguistically informed deep neural networks. In Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue, Melbourne, Australia, 12–14 July 2018; pp. 306–316. [Google Scholar]
  15. Li, P.; Mao, K. Knowledge-oriented convolutional neural network for causal relation extraction from natural language texts. Expert Syst. Appl. 2019, 115, 512–523. [Google Scholar] [CrossRef]
  16. Schomacker, T. Tropmann-Frick M. Language Representation Models: An Overview. Entropy 2021, 23, 1422. [Google Scholar] [CrossRef] [PubMed]
  17. Lan, Z.; Chen, M.; Sebastian, G.; Kevin, G.; Piyush, S.; Radu, S. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  18. Li, D.; Nan, Y.; Wenhui, W.; Furu, W.; Xiaodong, L.; Yu, W.; Jianfeng, G.; Ming, Z.; Hsiao-Wuen, H. Unified language model pre-training for natural language understanding and generation. arXiv 2019, arXiv:1905.03197. [Google Scholar]
  19. Mandar, J.; Chen, D.; Liu, Y.; Weld, D.S.; Zettlemoyer, L.; Levy, O. Spanbert: Improving pre-training by representing and predicting spans. Trans. Assoc. Comput. Linguist. 2020, 8, 64–77. [Google Scholar]
  20. Wei, W.; Bin, B.; Ming, Y.; Chen, W.; Jiangnan, X.; Zuyi, B.; Liwei, P.; Luo, S. StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  21. Chen, S.; Jiang, C.; Li, J.; Xiang, J.; Xiao, W. Improved Deep Q-Network for User-Side Battery Energy Storage Charging and Discharging Strategy in Industrial Parks. Entropy 2021, 23, 1311. [Google Scholar] [CrossRef] [PubMed]
  22. Xia, K.; Feng, J.; Yan, C.; Duan, C.B. BeiDou Short-Message Satellite Resource Allocation Algorithm Based on Deep Reinforcement Learning. Entropy 2021, 23, 932. [Google Scholar] [CrossRef] [PubMed]
  23. Wan, K.; Wu, D.; Zhai, Y.; Li, B.; Gao, X.; Hu, Z. An Improved Approach towards Multi-Agent Pursuit–Evasion Game Decision-Making Using Deep Reinforcement Learning. Entropy 2021, 23, 1433. [Google Scholar] [CrossRef] [PubMed]
  24. Karthik, N.; Adam, Y.; Regina, B. Improving information extraction by acquiring external evidence with reinforcement learning. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP, Austin, TX, USA, 1–5 November 2016. [Google Scholar]
  25. Hongliang, F.; Xu, L.; Dingcheng, L.; Ping, L. End-to-end deep reinforcement learning based coreference resolution. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar]
  26. Zhang, T.; Huang, M.; Zhao, L. Learning Structured Representation for Text Classification via Reinforcement Learning. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  27. Romain, P.; Caiming, X.; Richard, S. A Deep Reinforced Model for Abstractive Summarization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  28. Chen, L.; Zhang, T.; He, D.; Ke, G.; Wang, L.; Liu, T.-Y. Variance-reduced language pretraining via a mask proposal network. arXiv 2020, arXiv:2008.05333. [Google Scholar]
  29. Minki, K.; Moonsu, H.; Ju, H.S. Neural Mask Generator: Learning to Generate Adaptive Word Maskings for Language Model Adaptation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020. [Google Scholar]
  30. Hendrickx, I.; Kim, S.N.; Kozareva, Z.; Nakov, P.; Séaghdha, D.Ó.; Padó, S.; Pennacchiotti, M.; Romano, L.; Szpakowicz, S. SemEval-2010 task 8: Multi-way classification of semantic relations between pairs of nominals. In Proceedings of the 5th International Workshop on Semantic Evaluation, Los Angeles, CA, USA, 15–16 July 2010. [Google Scholar]
  31. Paramita, M.; Rachele, S.; Sara, T.; Manuela, S. Annotating causality in the tempeval-3 corpus. In Proceedings of the EACL 2014 Workshop on Computational Approaches to Causality in Language (CAtoCL), Gothenburg, Sweden, 26–27 April 2014. [Google Scholar]
  32. Volodymyr, M.; Puigdomenech, B.A.; Mehdi, M.; Alex, G.; Timothy, L.; Tim, H.; David, S.; Koray, K. Asynchronous methods for deep reinforcement learning. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016. [Google Scholar]
  33. Sutton, R.S.; McAllester, D.; Singh, S.; Mansour, Y. Policy gradient methods for reinforcement learning with function approximation. NIPs 1999, 99, 1057–1063. [Google Scholar]
  34. Huang, Z.; Xu, W.; Yu, K. Bidirectional LSTM-CRF models for sequence tagging. arXiv 2015, arXiv:1508.01991. [Google Scholar]
  35. Artem, C.; Oleksiy, O.; Philipp, H.; Alexander, B.; Matthias, H.; Chris, B.; Alexander, P. Targer: Neural argument mining at your fingertips. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Florence, Italy, 28–31 July 2019. [Google Scholar]
  36. Ilya, L.; Frank, H. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  37. Cui, Y.; Che, W.; Liu, T.; Qin, B.; Yang, Z. Pre-training with whole word masking for chinese bert. arXiv 2019, arXiv:1906.08101. [Google Scholar] [CrossRef]
  38. Li, B.; Hou, Y.; Che, W. Data Augmentation Approaches in Natural Language Processing: A Survey. arXiv 2021, arXiv:2110.01852. [Google Scholar]
Figure 1. Examples of sentences with pairs of causal events.
Figure 1. Examples of sentences with pairs of causal events.
Entropy 24 00169 g001
Figure 2. Examples of reprocessing data. The red and yellow words in the table are the newly added reason and result labels, respectively.
Figure 2. Examples of reprocessing data. The red and yellow words in the table are the newly added reason and result labels, respectively.
Entropy 24 00169 g002
Figure 3. An overview of the proposed ACMM framework.
Figure 3. An overview of the proposed ACMM framework.
Entropy 24 00169 g003
Figure 4. Experimental diagram of hyperparameter selection. The two pictures above show the results of models under different batch sizes. The remaining two pictures show the loss curve of the model.
Figure 4. Experimental diagram of hyperparameter selection. The two pictures above show the results of models under different batch sizes. The remaining two pictures show the loss curve of the model.
Entropy 24 00169 g004
Figure 5. Box plot of F1 for 5 different strategies. The upper and lower sides represent the maximum and minimum values, and the middle line represents the mean values.
Figure 5. Box plot of F1 for 5 different strategies. The upper and lower sides represent the maximum and minimum values, and the middle line represents the mean values.
Entropy 24 00169 g005
Figure 6. Examples of three masked sentences. Among them, “cas” represents the cause entity, and “eff” represents the result entity. There are two sentences with explicit causality, and one sentence with implicit causality.
Figure 6. Examples of three masked sentences. Among them, “cas” represents the cause entity, and “eff” represents the result entity. There are two sentences with explicit causality, and one sentence with implicit causality.
Entropy 24 00169 g006
Table 1. Experiment dataset statistics. Cause-effect means that sentence contains causal-effect pairs.
Table 1. Experiment dataset statistics. Cause-effect means that sentence contains causal-effect pairs.
SemEvalCausal TB
TrainTestTrainTest
Cause-effect904421220181
Other6574277520278
All74783196422103
Table 2. Results on two causal extraction datasets.
Table 2. Results on two causal extraction datasets.
ModelSemEvalCausal TB
PRF1PRF1
BiLSTM-CRF65.9045.3753.7232.0731.4131.64
TARGER61.3870.2065.4940.6949.5544.69
BERT77.6177.8577.7144.0153.8547.60
BERT-CRF78.6377.9178.2655.2446.8050.66
BERT-BiLSTM-CRF80.4077.2078.7454.6043.6048.14
BERT-ACMM79.1977.4178.2848.3754.4951.23
BERT-CRF-ACMM80.1979.2379.7154.3347.1850.50
BERT-BiLSTM-CRF-ACMM81.5577.0879.2555.3645.7850.12
Table 3. Comparison of our method with rule-based strategies.
Table 3. Comparison of our method with rule-based strategies.
MaskSemEvalCausal TB
FunctionPRF1PRF1
No Mask77.6177.8577.7144.0153.8547.60
Whole Word76.1979.9277.9948.3853.8550.94
Word Piece76.5478.6777.5852.3149.3650.78
Span77.0476.1376.5843.1353.8547.90
RL Mask79.1977.4178.2848.3754.4951.23
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, Y.; Zuo, W.; Liang, S.; Yuan, X.; Zhang, Y.; Zuo, X. A Word-Granular Adversarial Attacks Framework for Causal Event Extraction. Entropy 2022, 24, 169. https://doi.org/10.3390/e24020169

AMA Style

Zhao Y, Zuo W, Liang S, Yuan X, Zhang Y, Zuo X. A Word-Granular Adversarial Attacks Framework for Causal Event Extraction. Entropy. 2022; 24(2):169. https://doi.org/10.3390/e24020169

Chicago/Turabian Style

Zhao, Yu, Wanli Zuo, Shining Liang, Xiaosong Yuan, Yijia Zhang, and Xianglin Zuo. 2022. "A Word-Granular Adversarial Attacks Framework for Causal Event Extraction" Entropy 24, no. 2: 169. https://doi.org/10.3390/e24020169

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop