This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Towards Interpretable Reasoning over Paragraph Effects in Situation

Mucheng Ren1∗†  Xiubo Geng2†,  Tao Qin3,  Heyan Huang1    Daxin Jiang2‡
1School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China
2STCA NLP Group, Microsoft
3Microsoft Research Asia
{renm,hhy63}@bit.edu.cn
{xigeng,taoqin,djiang}@microsoft.com
Abstract

We focus on the task of reasoning over paragraph effects in situation, which requires a model to understand the cause and effect described in a background paragraph, and apply the knowledge to a novel situation. Existing works ignore the complicated reasoning process and solve it with a one-step “black box” model. Inspired by human cognitive processes, in this paper we propose a sequential approach for this task which explicitly models each step of the reasoning process with neural network modules. In particular, five reasoning modules are designed and learned in an end-to-end manner, which leads to a more interpretable model. Experimental results on the ROPES dataset demonstrate the effectiveness and explainability of our proposed approach.

11footnotetext: Work done during internship at STCA NLP Group, Microsoft.22footnotetext: Equal Contribution33footnotetext: Corresponding author

1 Introduction

As a long-standing fundamental task of natural language processing, machine reading comprehension (MRC) has attracted remarkable attention recently and different MRC datasets have been studied (Rajpurkar et al., 2018; Dua et al., 2019b; Choi et al., 2018; Yang et al., 2018), among which reasoning over paragraph effects in situation (ROPES for short) is a very challenging scenario that needs to understand knowledge from a background paragraph and apply it to answer questions in a novel situation. Table 1 shows an example of the ROPES dataset (Lin et al., 2019), where the background passage states that developmental difficulties could usually be treated by using iodized salt, the situation passage describes two villages using different salt, and questions about which village having more/less people experiencing developmental difficulties need to be answered.

Background
Before iodized salt was developed, some people experienced a number of developmental difficulties, including problems with thyroid gland function and mental retardation. In the 1920s, we learned that these conditions could usually be treated easily with the addition of iodide anion to the diet. One easy way to increase iodide intake was to add the anion to table salt.
Situation
People from two villages ate lots of salt. People from Salt village used regular salt, while people from Sand village people used iodized salt in their diets, after talking to specialists.
Q&A
Q: Which village had more people experience developmental difficulties? A: Salt
Q: Which village had less people experience developmental difficulties? A: Sand
Table 1: An example from the ROPES dataset. Effect property tokens are highlighted in blue, cause property tokens in orange, and world tokens in green.

Almost all existing works Lin et al. (2019); Khashabi et al. (2020); Dua et al. (2019a); Gardner et al. (2020) for this task adopt a standard MRC approach based on deep learning in one step: the question and a pseudo passage constructed by concatenating the background and situation are fed into a large pre-trained model (e.g. RoBERTa large), and the answer is predicted directly by the model. However, the ROPES task is more complicated than traditional MRC since it requires a model to not only understand the causes and effects described in a background paragraph, but also apply the knowledge to a novel situation. Ignoring the understanding and reasoning process hinders such models from achieving their best performance. Consequently, the best F1 (61.6%) achieved so far is far below human performance (89.0%). More importantly, such a one-step approach makes the reasoning process unexplainable, which is of great importance for complicated reasoning tasks.

We observe that human solve this kind of complicated reasoning tasks in a sequential manner with multiple steps Evans (1984); Sloman (1996); Sheorey and Mokhtari (2001); Mokhtari and Reichard (2002); Mokhtari and Sheorey (2002). As shown in Table 1, the background paragraph usually states the relationship between a cause property and an effect property, the situation describes multiple worlds each of which is associated with a specific value in terms of the cause property. Human usually does reasoning in a multi-step process: (1) identifying mentioned worlds, (2) identifying the cause and effect property, (3) understanding the relationship between the cause and effect property, (4) comparing identified worlds in terms of the cause property, and (5) reasoning about the comparison of mentioned worlds in terms of the effect property based on (3) and (4).

Inspired by human cognitive processes, in this paper, we propose a sequential approach that leverages neural network modules to implement each step of the above process111The code is publicly available at https://github.com/Borororo/interpretable_ropes.. Specifically, we define

  • a World Detection module to identify potential worlds,

  • an Effect and Cause Detection module to identify effect and cause property,

  • a Relation Classification module to understand the relationship between effect and cause,

  • a Comparison module to compare identified worlds in terms of the cause property, and

  • a Reasoning module to infer comparison of mentioned worlds in terms of the effect property.

These modules are trained in an end-to-end manner, and auxiliary loss over intermediate latent decisions further boosts the model accuracy.

Explicitly modeling the sequential reasoning process has two advantages. First, it achieves better performance since the complicated reasoning process is decomposed into more manageable sub-tasks and each module only needs to focus on a simple sub-task. Second, intermediate outputs provide a better understanding of the reasoning process, making the learnt model more explainable.

Experimental results on the ROPES dataset demonstrate the effectiveness and explainability of our proposed approach. It surpasses the state-of-the-art model by a large margin (6% absolute difference) in the five-fold cross-validation setting. Furthermore, analyses on intermediate outputs show that each module in our learnt model performs well on its corresponding sub-task and well explains the reasoning process.

2 Related Work

Neural network modules have been studied by several works. Andreas et al. (2016) propose neural module networks with a semantic parser on visual question answering. Jiang and Bansal (2019) apply a self-assembling modular network with only three modules: Find, Relocate and Compare to Hotpot QA (Yang et al., 2018).  Gupta et al. (2019) extend the neural module networks to answer compositional questions against a paragraphs of text as context, and perform symbolic reasoning on the self-pruned subset of DROPS (Dua et al., 2019b). Compared with them, we focus on a more challenging MRC task: reasoning over paragraph effects in situation, which has been rarely investigated and needs more complex reasoning. So far as we know, the only two works (i.e.  (Lin et al., 2019) and (Khashabi et al., 2020)) on this topic uses a one-step “black box” model. Such an approach performs well on some questions at the expense of limited intepretability. Our work solves this task in a logical manner and exposes intermediate reasoning steps which improves performance and interpretability concurrently.

Refer to caption
Figure 1: The left part is the architecture of our model. The middle part is the interpretable reasoning component in our model. The right part is the summary for inputs and outputs flowing between each module. The encoded contextual representations, 𝑯𝒒,𝑯𝒔,𝑯𝒃\bm{H^{q},H^{s},H^{b}}, serve as global variables for the interpretable reasoning component.

3 Methodology

As shown in Figure 1, our approach consists of three components which are contextual encoding, interpretable reasoning, and answer prediction.

3.1 Contextual Encoding

We use RoBERTa (Devlin et al., 2019; Liu et al., 2019) to encode background, situation and question together and generate contextualized embeddings. Specifically, given a background passage B={bi}i=1mB=\left\{b_{i}\right\}_{i=1}^{m}, a situation passage S={sj}j=1nS=\left\{s_{j}\right\}_{j=1}^{n} and a question Q={qk}k=1lQ=\left\{q_{k}\right\}_{k=1}^{l}, we concatenate them with special tokens as sq1,,ql/s/ss1,,sn;b1,,bm/s\left\langle s\right\rangle q_{1},\dots,q_{l}\left\langle/s\right\rangle\left\langle/s\right\rangle s_{1},\dots,s_{n};b_{1},\dots,b_{m}\left\langle/s\right\rangle, which is then fed into a series of successive transformer blocks contained in RoBERTa,

𝑯𝒒,𝑯𝒔,𝑯𝒃=Transformers(Q,S,B),\bm{H^{q},H^{s},H^{b}}=\operatorname*{Transformers}(Q,S,B), (1)

where 𝑯𝒃m×d\bm{H^{b}}\in\mathbb{R}^{m\times d}, 𝑯𝒔n×d\bm{H^{s}}\in\mathbb{R}^{n\times d}, and 𝑯𝒒l×d\bm{H^{q}}\in\mathbb{R}^{l\times d} are contextual embeddings for the background, situation, and question, respectively, dd is the dimension for hidden states.

3.2 Interpretable Reasoning

World Detection

The module aims to identify concerned worlds from situation according to a question. Take Table 1 as an example, the question cares about two worlds, Sand Village and Salt Village. To achieve that, we apply a multilayer perceptron (MLP) over the situation representations 𝑯𝒔\bm{H^{s}} and normalize the projected logits (using a softmax function) to get attention over all situation tokens for each world,

𝒑𝒘𝟏𝒔\displaystyle\bm{p_{w_{1}}^{s}} =softmax(MLP(𝑯𝒔;θw1))n,\displaystyle=\operatorname*{softmax}(\operatorname*{MLP}(\bm{H^{s}};\theta_{w_{1}}))\in\mathbb{R}^{n}, (2)
𝒑𝒘𝟐𝒔\displaystyle\bm{p_{w_{2}}^{s}} =softmax(MLP(𝑯𝒔;θw2))n,\displaystyle=\operatorname*{softmax}(\operatorname*{MLP}(\bm{H^{s}};\theta_{w_{2}}))\in\mathbb{R}^{n}, (3)

where 𝒑𝒘𝟏𝒔\bm{p_{w_{1}}^{\bm{s}}} and 𝒑𝒘𝟐𝒔\bm{p_{w_{2}}^{s}} are the attention vectors over situation for the first and second world, θ\theta’s are learnable parameters of MLP. Note that since most examples in the ROPES dataset are related to two concerned worlds, we identify two worlds in our model. However, we can handle multiple worlds by simply extending the module with more MLPs.

Effect and Cause Detection

This module aims to identify effect and cause properties described in the background. To achieve that, another MLP is used to identify the effect property,

𝒑𝒆𝒃=softmax(MLP(𝑯𝒃;θe))m.\displaystyle\bm{p_{e}^{b}}=\operatorname*{softmax}(\operatorname*{MLP}(\bm{H^{b}};\theta_{e}))\in\mathbb{R}^{m}. (4)

Here 𝒑𝒆𝒃\bm{p_{e}^{b}} is the attention vector over background tokens in terms of the effect property, which attends more to tokens of effect property. Take Table 1 as an example, 𝒑𝒆𝒃\bm{p_{e}^{b}} is the attention over background tokens, whose value is much larger for developmental difficulties than other tokens.

Next, we apply a relocate operation which re-attends to the background based on the situation and is used to find the cause property in the background (e.g., shifting the attention from developmental difficulties to iodized salt in Table 1). This is achieved with the help of a situation-aware background-to-background attention matrix 𝑹m×m\bm{R}\in\mathbb{R}^{m\times m},

𝑹ij=𝒘𝒓𝒆𝒍𝒐T[(𝒔+𝑯𝒊𝒃);𝑯𝒋𝒃;(𝒔+𝑯𝒊𝒃)𝑯𝒋𝒃],\displaystyle\bm{R}_{ij}={\bm{w_{relo}}}^{T}\left[(\bm{s}+\bm{H_{i}^{b}});\bm{H_{j}^{b}};(\bm{s}+\bm{H_{i}^{b}})\odot\bm{H_{j}^{b}}\right], (5)
𝒔=1nin𝑯𝒊𝒔d,\displaystyle\bm{s}=\frac{1}{n}\sum_{i}^{n}\bm{H_{i}^{s}}\in\mathbb{R}^{d}, (6)

where [;] denotes the concatenation operation and \odot is Hadamard product. 𝒘𝒓𝒆𝒍𝒐3d\bm{w_{relo}}\in\mathbb{R}^{3d} is a learnable parameter vector, 𝒔\bm{s} can be viewed as an embedding of the whole situation. Then each row of 𝑹\bm{R} is normalized using the softmax operation. Finally we get the attention vector over background tokens in terms of the cause property 𝒑𝒄𝒃\bm{p_{c}^{b}},

𝒑𝒄𝒃=𝑹T𝒑𝒆𝒃m.\displaystyle\bm{p_{c}^{b}}=\bm{R}^{T}\bm{p_{e}^{b}}\in\mathbb{R}^{m}. (7)

Here 𝒑𝒄𝒃\bm{p_{c}^{b}} should attend more to the tokens of effect property. For example, iodized salt will get larger attention value than other tokens in the background.

Relation Classification

This module aims to predict the qualitative relation between effect and cause property. Take Table 1 as an example, the cause property iodized salt and the effect property developmental difficulty is negatively correlated. To achieve that, we first derive and concatenate representations of cause and effect property by averaging background representation 𝑯𝒃\bm{H^{b}} weighted by according attention vector, 𝒑𝒄𝒃\bm{p_{c}^{b}} and 𝒑𝒆𝒃\bm{p_{e}^{b}}. Next, we adopt another MLP stacked with softmax to get corresponding probabilities,

𝒑𝒓𝒆𝒍=softmax(MLP((𝑯𝒃T𝒑𝒆𝒃;𝑯𝒃T𝒑𝒄𝒃);θrel)),\displaystyle\bm{p_{rel}}=\operatorname*{softmax}(\operatorname*{MLP}((\bm{{H^{b}}}^{T}\bm{p_{e}^{b}};\bm{{H^{b}}}^{T}\bm{p_{c}^{b}});\theta_{rel})), (8)

where 𝒑𝒓𝒆𝒍=[prel\bm{p_{rel}}=[p_{rel-}, prel+]p_{rel+}] denotes probability of negative and positive relation, θrel\theta_{rel} is a learnable parameter in the MLP. In the example shown in Table 1, prelp_{rel-} is supposed to be larger than prel+p_{rel+}.

Comparison

This module aims to compare the worlds in terms of the cause property. For example, world 1 (salt village) is more relevant to iodized salt than world 2 (sand village) in Table 1 since people in salt village use iodized salt while people in sand village use regular salt.

This is achieved by three steps. First, we derive the attention of cause property over situation 𝒑𝒄𝒔\bm{p_{c}^{s}} from 𝒑𝒄𝒃\bm{p_{c}^{b}} with a similarity matrix 𝑴n×m\bm{M}\in\mathbb{R}^{n\times m} between situation and background,

𝑴ij=𝑯𝒊𝒔𝑾sb𝑯𝒋𝒃T,\displaystyle\bm{M}_{ij}={\bm{H_{i}^{s}}}\bm{W}_{sb}{\bm{{H_{j}^{b}}}^{T}}, (9)
𝒑𝒄𝒔=𝑴𝒑𝒄𝒃n,\displaystyle\bm{p_{c}^{s}}=\bm{M}\bm{p_{c}^{b}}\in\mathbb{R}^{n}, (10)

where 𝑾sbd×d\bm{W}_{sb}\in\mathbb{R}^{d\times d} are learnable parameters.

Second, we use 𝒑𝒘𝒔\bm{p_{w}^{s}} to mask out irrelevant cause property for each world. This part ensures the alignment between each world and its cause property, which is critical when one situation contains multiple worlds.

𝒑𝒄𝒘𝟏𝒔=softmax(𝒑𝒘𝟏𝒔𝒑𝒄𝒔),\displaystyle\bm{p_{c_{w_{1}}}^{s}}=\operatorname*{softmax}(\bm{p_{w_{1}}^{s}}\odot\bm{p_{c}^{s}}), (11)
𝒑𝒄𝒘𝟐𝒔=softmax(𝒑𝒘𝟐𝒔𝒑𝒄𝒔).\displaystyle\bm{p_{c_{w_{2}}}^{s}}=\operatorname*{softmax}(\bm{p_{w_{2}}^{s}}\odot\bm{p_{c}^{s}}). (12)

Third, each world’s cause property is evaluated by a bilinear function in terms of its relevance to the cause property in background, which is further normalized into a probability with softmax,

logitw1=(𝑯𝒃T𝒑𝒄𝒃)T𝑾com(𝑯𝒔T𝒑𝒄𝒘𝟏𝒔),\displaystyle logit_{w_{1}}=(\bm{{H^{b}}}^{T}\bm{p_{c}^{b}})^{T}\bm{W}_{com}(\bm{{H^{s}}}^{T}\bm{p_{c_{w_{1}}}^{s}}), (13)
logitw2=(𝑯𝒃T𝒑𝒄𝒃)T𝑾com(𝑯𝒔T𝒑𝒄𝒘𝟐𝒔),\displaystyle logit_{w_{2}}=(\bm{{H^{b}}}^{T}\bm{p_{c}^{b}})^{T}\bm{W}_{com}(\bm{{H^{s}}}^{T}\bm{p_{c_{w_{2}}}^{s}}), (14)
𝒑𝒄𝒐𝒎𝒘=softmax(logitw1,logitw2),\displaystyle\bm{p_{com_{w}}}=\operatorname*{softmax}(logit_{w_{1}},logit_{w_{2}}), (15)

where 𝑾𝒄𝒐𝒎d×d\bm{W_{com}}\in\mathbb{R}^{d\times d} is a learnable matrix, 𝑯𝒃T𝒑𝒄𝒃\bm{{H^{b}}}^{T}\bm{p_{c}^{b}} represents expected embedding of cause property in background, 𝑯𝒔T𝒑𝒄𝒘𝒊𝒔\bm{{H^{s}}}^{T}\bm{p_{c_{w_{i}}}^{s}} represents expected embedding of cause property for world ii, pcomwip_{com_{w_{i}}} denotes the probability that world ii is relevant to cause property.

Reasoning

Given the relationship between effect property and cause property, prel+p_{rel+} and prelp_{rel-}, and the comparison between worlds in terms of cause property pcomwip_{com_{w_{i}}}, this module infers comparison between identified worlds in terms of the effect property. Take Table 1 as an example, given the negative relationship between developmental difficulties and iodized salt, and salt village uses more iodized salt than sand village, we infer that people in sand village are more likely to have developmental difficulties.

To this end, we have

pew1=pcomw1×prel++pcomw2×prel,\displaystyle p_{e_{w_{1}}}=p_{com_{w_{1}}}\times p_{rel+}+p_{com_{w_{2}}}\times p_{rel-}, (16)
pew2=pcomw1×prel+pcomw2×prel+,\displaystyle p_{e_{w_{2}}}=p_{com_{w_{1}}}\times p_{rel-}+p_{com_{w_{2}}}\times p_{rel+}, (17)

where pewip_{e_{w_{i}}} is the probability that world ii is more relevant to effect property.

3.3 Answer Prediction

Given intermediate outputs from the interpretable reasoning component, this module predicts the final answer for a question. Specifically, we first convert these intermediate outputs into text spans or 0/1 class as follow.

  • We take two steps to convert an attention vector output by World Detection or Effect and Cause Detection into a text span. First, the token with the highest probability is selected. Then it is expanded with left and right neighbors which are continuous spans and the probability of each token is larger than threshold tt. In our experiment we set t=1lt=\frac{1}{l}, where ll is the length of the paragraph.

  • For Comparison, Relation Classification, and Reasoning, we select the class with the highest probability.

Then we synthesize a sentence s^\hat{s} in the format of [World 1] has [larger/smaller] [Effect Property] than [World 2], where choosing “larger” or “smaller” depends on results from the Reasoning module. Take Table 1 as an example, the synthetic sentence is Salt village has larger developmental difficulties than Sand village. Such synthetic text explicitly expresses comparison between the identified worlds in terms of the effect property. Finally, we concatenate it with the situation ss and question qq as sq;s/s/ss^/s\left\langle s\right\rangle q;s\left\langle/s\right\rangle\left\langle/s\right\rangle\hat{s}\left\langle/s\right\rangle, and feed them into RoBERTa which directly predicts the starting and end position of the final answer.

3.4 Model Training

Two models (i.e. interpretable reasoning model; and answer prediction model) are learned in our approach.

Interpretable Reasoning

The final loss function for interpretable reasoning is defined as

lintp=𝒙𝑿α𝒙𝒙~Tlog(𝒙).\displaystyle l_{intp}=-\sum_{\bm{x}\in\bm{X}}\alpha_{\bm{x}}\bm{\tilde{x}}^{T}log(\bm{x}). (18)

Here 𝑿={𝒑𝒘𝟏𝒔n,𝒑𝒘𝟐𝒔n,𝒑𝒆𝒃m,𝒑𝒄𝒃m,𝒑𝒄𝒘𝟏𝒔n,𝒑𝒄𝒘𝟐𝒔n,𝒑𝒓𝒆𝒍2,𝒑𝒄𝒐𝒎𝒘2,𝒑𝒆𝒘2}\bm{X}=\{\bm{p_{w_{1}}^{s}}\in\mathbb{R}^{n},\bm{p_{w_{2}}^{s}}\in\mathbb{R}^{n},\bm{p_{e}^{b}}\in\mathbb{R}^{m},\bm{p_{c}^{b}}\in\mathbb{R}^{m},\bm{p_{c_{w_{1}}}^{s}}\in\mathbb{R}^{n},\bm{p_{c_{w_{2}}}^{s}}\in\mathbb{R}^{n},\bm{p_{rel}}\in\mathbb{R}^{2},\bm{p_{com_{w}}}\in\mathbb{R}^{2},\bm{p_{e_{w}}}\in\mathbb{R}^{2}\} are predictions of different modules, 𝒙~T{0,1}n\bm{\tilde{x}}^{T}\in\left\{0,1\right\}^{n} or 𝒙~T{0,1}m\bm{\tilde{x}}^{T}\in\left\{0,1\right\}^{m} or 𝒙~T{0,1}2\bm{\tilde{x}}^{T}\in\left\{0,1\right\}^{2} are corresponding gold labels, and αx\alpha_{x} is the weight for module xx.

Answer Prediction

The training objective of the answer prediction model is defined as

lans=(𝒔~log(𝒔)+𝒆~log(𝒆)),\displaystyle l_{ans}=-(\bm{\tilde{s}}\log(\bm{s})+\bm{\tilde{e}}\log(\bm{e})), (19)

where 𝒔,𝒆m+n+k\bm{s},\bm{e}\in\mathbb{R}^{m+n+k} are predicted probabilities of the starting and end position, kk is the length of the synthetic sentence s^\hat{s}, and 𝒔~,𝒆~{0,1}m+n+k\bm{\tilde{s}},\bm{\tilde{e}}\in\left\{0,1\right\}^{m+n+k} are corresponding gold labels.

4 Experimental Setup

4.1 Dataset

We evaluate our proposed approach on the ROPES Lin et al. (2019) dataset222https://leaderboard.allenai.org/ropes/submissions/get-started. So far as we know, it is the only dataset that requires reasoning over paragraph effects in situation. Given a background paragraph that contains knowledge about relations of causes and effects and a novel situation, questions about applying the knowledge to the novel situation need to be answered. Table 1 shows an example and Table 2 presents the statistics. To be noticed, different from other extractive MRC datasets, train/dev/test set in ROPES is split based on annotators instead of context (Geva et al., 2019; Lin et al., 2019). This might pose a large data bias in each set. For example, as can be seen in Table 2, dev and test sets have similar numbers of questions, while the vocabulary size of background and situation in test set is 2×2\times and 2.7×2.7\times larger than that in dev set. The same thing happens on the size of question vocabulary, which indicates the existence of the distribution gap between train/dev and test sets and it might lead to underestimate/overestimate the performance of a model.

Statistics Train Dev Test
background vocabulary size 8,616 2,008 3,988
situation vocabulary size 6,949 1,077 2,736
question vocabulary size 1,457 1,411 1,885
avg. background length 121.6 90.7 123.1
avg. situation length 49.1 63.4 55.6
avg. question length 10.9 12.4 10.6
No. of questions 10,924 1,688 1,710
No. of annotators 7 2 2
Table 2: ROPES statistics
Cross Validation

Because of the limited size of the official dev set and potential data bias between train/dev and test, we conduct 5-fold cross-validation to verify the effectiveness of the proposed approach. K-fold cross-validation assesses the predictive performance of the models and judges how they perform outside the sample to a new data set. Therefore it can assure unbiased results, avoid over-fitting, and testify the generalization capability of a model (Burman, 1989; Browne, 2000; Raschka, 2018). Specifically, we first exclude the labeled 1074 questions from the training data, and then split the remaining training plus dev data into five folds based on background, which ensures each subset has independent vocabulary space and they do not look through each other. For each split, we directly apply modules trained from auxiliary supervision data, and use the training data to train a model for answer prediction. Averaged results on the 5-fold cross-validation setting are reported.

Auxiliary Supervision

We randomly sampled 10% (1,074 questions) of the original training data and labeled them for training the proposed modules. For each example, we label two concerned worlds, effect property in background and situation, values of cause property for two worlds in situation, comparison between two worlds in terms of cause property and effect property, and relationship between cause property and effect property. More detailed guidelines and labeled examples are in Appendix B. Note the neural network modules are trained only on the labeled 1,074 questions.

4.2 Implementation Details

Our model is evaluated based on the pretrained language model RoBERTa large in Pytorch version333https://github.com/huggingface/transformers. We train the five modules on one P100 16GB GPU and use four GPUs for predicting final answer. We tune the parameter αx\alpha_{x}’s according to the averaged performance of all modules, and set it to be 0.05 for span-based loss, 0.2 for the Comparison and Relation prediction, and 0.3 for the Reasoning prediction. Evaluation metrics are EM and F1 which are same as the ones used in SQuAD444https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py. The detailed hyperparameters are described in the Appendix A.

4.3 Baseline

We re-implemented the best model (RoBERTa) in the leaderboard555https://leaderboard.allenai.org/ropes/submissions/public and achieved similar performance (Our implemented baseline achieves EM 69.0 / F1 71.1 on dev and EM 55.2 / F1 61.0 on test while the official one achieves EM 59.7 / F1 70.2 on dev and EM 55.4 / F1 61.1 on test). The basic idea is to leverage the RoBERTa large model (Liu et al., 2019) to encode the concatenation of question, background and situation, and then apply a linear layer to predict the starting and end position of an answer directly.

5 Experimental Results

5.1 Question Answering Performance

Table 3 shows question answering performance of different models, where our approach outperforms the RoBERTa large model by 8.4% and 6.4% in terms of EM and F1 scores respectively. These results show that compared to one-step “black box” model, our interpretable approach which mimics the human reasoning process has a better capability of conducting such complex reasoning.

Furthermore, we also list the performance of our approach and the baseline model when using only randomly sampled 10% of training data in Table 3. That is, both the neural network modules and answer prediction model in our approach are trained with only 1074 questions. As seen in the table, our model learned from 10% of training examples achieves competitive performance to the baseline model learned from full data (71.8% v.s. 71.1% in terms of F1 score). In contrast, the performance of the baseline model drops dramatically by 32%. This indicates that traditional black-box approach requires much more training data while our approach has better generalization ability and can learn the reasoning capability with much fewer examples.

We also implement a rule-based answer prediction approach (detailed descriptions in the Appendix E), which are generated based on the same 10% of training examples as in interpretable reasoning components. As shown in Table 3 the rule-based approach performs worse than the RoBERTa model, indicating better generalization ability of pre-trained models.

Model Dev Test
EM F1 EM F1
RoBERTaLarge{}_{Large}^{*} 61.4 68.4 64.0 71.1
Ours 73.0 78.1 72.4 77.5
RoBERTaLarge{}_{Large}^{*} (10%10\% data) - - 43.1 53.9
Ours (10%10\% data) - - 60.9 71.8
Ours (rule-based) - - 54.3 65.5
Table 3: Performance of different models on the ROPES dataset under cross-validation setting.

5.2 Case Study of Interpretability

The most remarkable difference between our model and the one-step “black box” model is that our model outputs multiple intermediate predictions which well explains the reasoning process. Note all modules in our model output probabilities or attention on input text, which are further fed into downstream ones for end-to-end learning. In order to explicitly visualize the output of each module, we take a similar approach to §3.3 to convert these probabilities into a text span or a 0/1 classification.

We demonstrate the reasoning process of our model with a running example shown in Table 4. Please see more examples in Appendix D. Here the background states the relationship between CPU load and data volume, i.e. CPU load goes up when processing larger volume of data. The situation describes that Tory stored different sizes of data at a different time. For example, he stored 301 Gigabytes at 8 AM and went to sleep at 1 PM. Finally, the question asks to compare CPU loads between 8 AM and 1 PM.

Background
Storing large volumes of data - When storing XML to either file or database, the volume of data a system produces can often exceed reasonable limits, with a number of detriments: the access times go up as more data is read, CPU load goes up as XML data takes more power to process, and storage costs go up….
Situation
Tory had a busy day storing XML. At 7 AM, he stored 201 Gigabytes to the database. At 8 AM, he stored 301 Gigabytes to the database. At 9 AM,… At 10 AM, … At 11 AM, … At 12 PM, … At 1 PM, he went to sleep to finish storing XML later on that day.
Q&A
Q: What time did CPU load go up: 8 AM or 1 PM? A: 8 AM
Predictions
Worlds: [8 AM,1 PM] Effect: CPU load goes up CauseB\text{Cause}^{B}: Storing large volumes of data CauseWorld1S\text{Cause}_{\text{World1}}^{S}: 301 Gigabytes CauseWorld2S\text{Cause}_{\text{World2}}^{S}: sleep Cause Cmp: World 1 Relation: positively related Effect Cmp: World 1 Final Answer: 8 AM
Table 4: A running example with visualized intermediate outputs of our approach.

As shown in Table 4, our model outputs several intermediate results. First, it identifies two concerned worlds, 8 AM and 1 PM from the situation. Then it predicts the effect property, CPU load goes up, given which the cause property in the background (i.e. storing large volumes of data) and according values for the two worlds (i.e. 301 Gigabytes and sleep) are predicted. Next, it compares the two worlds in terms of cause property and predicts that world 1 is larger than world 2. Also it predicts that the cause property and effect property is positively related, i.e. the relation is classified as 1. Finally, it reasons that world 1 takes higher CPU loads than world 2. This example demonstrates that our approach not only predicts the final answer for the question, but also provides detailed explanations for the reasoning process.

F1 Fuzzy F1 Accuracy
World 1 83.5 86.8 Comparison 83.8%
World 2 84.4 86.1 Relation 84.5%
Effect 67.8 83.6 Reasoning 74.0%
CauseB\text{Cause}^{B} 57.6 70.1
CauseWorld1S\text{Cause}_{\text{World1}}^{S} 69.4 81.3
CauseWorld2S\text{Cause}_{\text{World2}}^{S} 58.2 71.9
Table 5: Performance of Each Module

5.3 Neural Network Module Performance

Taking the same approach as in §5.2, we convert the output of each module into a text span or a predicted class. We manually sampled another 5% of the training data, labeled them with outputs for each module, and evaluate the visualized results of all modules. Table 5 summarizes the performance for each module, where the predicted text span is measured by F1 score and classification prediction is measured by accuracy.

World Detection

This module implements a similar capability as traditional extractive MRC, since both require to detect concerned text spans from a passage according to a question. Consequently, it achieves similar performance to top models of the popular SQuAD dataset666https://rajpurkar.github.io/SQuAD-explorer/, where our World Detection module reaches about 83% F1 score and single RoBERTa large model on SQuAD gets about 89%. The gap might come from different modeling styles. Our model predicts the probability of each token being concerned, while SQuAD models directly predict the starting and end position of an answer, which performs better on boundary detection.

Effect and Cause Detection

Compared with World Detection, the F1 score for this module decreases but actually still acceptable (F1=67.6 for EffectB\text{Effect}^{B}, 57.6 for CauseB\text{Cause}^{B}). The most possible reason is that, effects and causes are usually longer than world names. For example, the average length of world name is 1.2, while those of effect and cause are 2.7 and 2.2 respectively. Longer text span increases the difficulty of prediction.

For above two span-related modules, we argue that since our model leverages the attention score in a soft way, it is less sensitive to accuracy of boundary changes. Therefore, we added another fuzzy F1 score for them. The fuzzy F1 of each question is set to 1 as long as its original F1 is larger than 0. As shown in Table 5, the fuzzy F1 scores of these two modules increase to 70%\sim86%, indicating good reasoning capability of them.

Comparison, Relation Classification

These two modules essentially requires the capability of classification. The high accuracy (83.8% and 84.5%) indicates that our modeling approach can effectively leverages the prediction of upstream modules and does a good job on them.

Reasoning

Given the high accuracy of the Comparison and Relation Classification modules, the Reasoning model achieves 74% of accuracy, which provides high-quality input for final answer prediction.

5.4 Error Analysis

We randomly sampled 200 wrongly predicted questions to do error analysis and find that they fall into two major types, which are described below (Please see complete description of questions, backgrounds and situations in the Appendix C).

5.4.1 Type One Error

Type one errors are caused by wrong model predictions, most of which occur in below three modules.

Wrong Predicted Worlds

Such errors are mainly caused by length imbalance between question and situation. Since situation is usually much longer than question, the World Detection module might make the same predictions for different questions of same situation.

Wrong Predicted Cause Property in Situation

Such errors are mainly caused by imbalanced descriptions for different worlds, where a situation describes details for one world but mentions another world with very few words. In such cases, the Comparison module might assign the same cause property for different worlds in situation.

Wrong Predicted Comparison Results

Such errors often occur when two worlds are described with similar words, e.g. “high” v.s. “higher”, or “smoking” vs. “not smoking”, in which case the Comparison module might be confused by similar expressions of two worlds and fail to compare them.

5.4.2 Type Two Error

Type two errors occur when the proposed framework is not suitable to solve the questions. Here we list some example cases.

Missing Knowledge

Background paragraph does not provide sufficient knowledge for reasoning. For example, a background paragraph only describes information about fish while the questions asks fertilization take place inside or outside of mother’s body for a mammal creature.

Implicit Worlds

Concerned worlds in a question are not explicitly described in the situation. For example, a situation paragraph says that Mattew does intensive worksouts while the question asks that his strengths will increase or decrease when he stops working out. In such a case, the world that Mattew stop working out is not explicitly described in the situation.

Additional Math Computation

Answering such questions requires additional math computation. For example, a background states the speed of sound waves in air/water/iron and the question asks how much faster a channel (with water) would be than another channel (with air). Answering such questions requires additional math computation (i.e. subtraction, addition etc.)

6 Conclusion and Future Work

In this paper, we aim to answer ROPES questions in an interpretable way by leveraging five neural network modules. These modules are trained in an end-to-end manner and each module provides transparent intermediate outputs. Experimental results demonstrate the effectiveness of each module, and analysis on intermediate outputs presents good interpretability for the inference process in contrasted with “black box” models. Moreover, we find that with explicitly designed compositional modeling of inference process, our approach with a few training examples achieves similar accuracy to strong baselines with full-size training data which indicates a better generalization capability. Meanwhile, extending these models to a larger scope of question types or more complex scenarios is still a challenge, and we will further investigate the trade-off between explainability and scalability.

Acknowledgement

We acknowledge this work is supported by National Natural Science Foundation of China (No.61751201) and National Key R&D Plan (No.2016QY03D0602), We would also like to thank the anonymous reviewers for their insightful suggestions.

References

  • Andreas et al. (2016) Jacob Andreas, Marcus Rohrbach, Trevor Darrell, and Dan Klein. 2016. Neural module networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 39–48.
  • Browne (2000) Michael W Browne. 2000. Cross-validation methods. Journal of mathematical psychology, 44(1):108–132.
  • Burman (1989) Prabir Burman. 1989. A comparative study of ordinary cross-validation, v-fold cross-validation and the repeated learning-testing methods. Biometrika, 76(3):503–514.
  • Choi et al. (2018) Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, and Luke Zettlemoyer. 2018. QuAC: Question answering in context. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2174–2184, Brussels, Belgium. Association for Computational Linguistics.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Dua et al. (2019a) Dheeru Dua, Ananth Gottumukkala, Alon Talmor, Sameer Singh, and Matt Gardner. 2019a. Orb: An open reading benchmark for comprehensive evaluation of machine reading comprehension. arXiv preprint arXiv:1912.12598.
  • Dua et al. (2019b) Dheeru Dua, Yizhong Wang, Pradeep Dasigi, Gabriel Stanovsky, Sameer Singh, and Matt Gardner. 2019b. DROP: A reading comprehension benchmark requiring discrete reasoning over paragraphs. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 2368–2378, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Evans (1984) Jonathan St BT Evans. 1984. Heuristic and analytic processes in reasoning. British Journal of Psychology, 75(4):451–468.
  • Gardner et al. (2020) Matt Gardner, Yoav Artzi, Victoria Basmova, Jonathan Berant, Ben Bogin, Sihao Chen, Pradeep Dasigi, Dheeru Dua, Yanai Elazar, Ananth Gottumukkala, et al. 2020. Evaluating nlp models via contrast sets. arXiv preprint arXiv:2004.02709.
  • Geva et al. (2019) Mor Geva, Yoav Goldberg, and Jonathan Berant. 2019. Are we modeling the task or the annotator? an investigation of annotator bias in natural language understanding datasets. arXiv preprint arXiv:1908.07898.
  • Gupta et al. (2019) Nitish Gupta, Kevin Lin, Dan Roth, Sameer Singh, and Matt Gardner. 2019. Neural module networks for reasoning over text. arXiv preprint arXiv:1912.04971.
  • Jiang and Bansal (2019) Yichen Jiang and Mohit Bansal. 2019. Self-assembling modular networks for interpretable multi-hop reasoning. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4474–4484, Hong Kong, China. Association for Computational Linguistics.
  • Khashabi et al. (2020) Daniel Khashabi, Tushar Khot, Ashish Sabharwal, Oyvind Tafjord, Peter Clark, and Hannaneh Hajishirzi. 2020. Unifiedqa: Crossing format boundaries with a single qa system. arXiv preprint arXiv:2005.00700.
  • Lin et al. (2019) Kevin Lin, Oyvind Tafjord, Peter Clark, and Matt Gardner. 2019. Reasoning over paragraph effects in situations. In Proceedings of the 2nd Workshop on Machine Reading for Question Answering, pages 58–62, Hong Kong, China. Association for Computational Linguistics.
  • Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
  • Mokhtari and Reichard (2002) Kouider Mokhtari and Carla A Reichard. 2002. Assessing students’ metacognitive awareness of reading strategies. Journal of educational psychology, 94(2):249.
  • Mokhtari and Sheorey (2002) Kouider Mokhtari and Ravi Sheorey. 2002. Measuring esl students’ awareness of reading strategies. Journal of developmental education, 25(3):2–11.
  • Rajpurkar et al. (2018) Pranav Rajpurkar, Robin Jia, and Percy Liang. 2018. Know what you don’t know: Unanswerable questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 784–789, Melbourne, Australia. Association for Computational Linguistics.
  • Raschka (2018) Sebastian Raschka. 2018. Model evaluation, model selection, and algorithm selection in machine learning. arXiv preprint arXiv:1811.12808.
  • Sheorey and Mokhtari (2001) Ravi Sheorey and Kouider Mokhtari. 2001. Differences in the metacognitive awareness of reading strategies among native and non-native readers. System, 29(4):431–449.
  • Sloman (1996) Steven A Sloman. 1996. The empirical case for two systems of reasoning. Psychological bulletin, 119(1):3.
  • Yang et al. (2018) Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William Cohen, Ruslan Salakhutdinov, and Christopher D. Manning. 2018. HotpotQA: A dataset for diverse, explainable multi-hop question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2369–2380, Brussels, Belgium. Association for Computational Linguistics.

Appendices

A Parameters List

Interpretable Reasoning Search Space(Bounds) Best Assignment
No. of GPU(P100) 1 1
Average runtime (mins) 45 45
No. of params.(include LM) 361661706 361661706
No. of Layers in MLP 3 3
No. of Search trials 38 38
learning rate optimizer Adam Adam
Max Seq. Length choice[384,512] 512
Doc stride choice[64,128] 64
Learning Rate uniform-float[5e-6,3e-5] 2e-5
Batch Size per GPU choice[1,2,4] 2
Gradient Accumulation Step choice[1,2] 1
No. of Epoch uniform-integer[1,5] 4
Fixed Length for Q,S,B uniform-integer[20,30],[150,250],[350,450] 30,200,400
Table 6: Detailed parameters used in Interpretable Reasoning, we provide search bounds for each hyperparameter and list out the hyperparameters combination for out best model. Other unmentioned parameters keep same as the one used in BERT.
Answer Prediction Search Space(Bounds) Best Assignment
No. of GPU(P100) 4 4
Average runtime (mins) 60 60
No. of params.(include LM) 355361794 355361794
No. of Search trials 64 64
learning rate optimizer Adam Adam
Max Seq. Length choice[384,512] 384
Doc stride choice[64,128] 128
Learning Rate uniform-float[5e-6,3e-5] 1.5e-5
Batch Size per GPU uniform-integer[4,8] 4
Answer Length Limit uniform-integer[5,30] 9
Gradient Accumulation Step choice[1,2] 1
No. of Epoch uniform-integer[1,5] 4
Table 7: Detailed parameters used in Answer Prediction, we provide search bounds for each hyperparameter and list out the hyperparameters combination for out best model and baseline model. Other unmentioned parameters keep same as the one used in BERT.

B Auxiliary Supervision Instruction

Background
As a cell grows, its volume increases more quickly than its surface area. If a cell was to get very large, the small surface area would not allow enough nutrients to enter the cell quickly enough for the cell’s needs…Such cell types are found lining your small intestine, where they absorb nutrients from your food through protrusions called microvilli .
Situation
There are two cells inside a Petri dish in a laboratory, cell X and cell Z. These cells are from the same organism, but are not the same age. Cell X was created two weeks ago, and cell Z was created one month ago. Therefore, cell Z has had two extra weeks of growth compared to cell X.
Q&A
Q: Which cell has a larger volume?
A: cell Z
Labels
World1: Cell X [142,148] World2: cell Z [180,186] Effect: volume [21,27] CauseB\text{Cause}^{B}: cell grows [5,15] CauseWorld1S\text{Cause}_{\text{World1}}^{S}: two weeks ago [161,174] CauseWorld2S\text{Cause}_{\text{World2}}^{S}: one month ago [199,212] Cause Cmp: cell Z 1 Relation: positively related 1 Effect Cmp: World 2 1
Table 8: An example with auxiliary supervision labels.

Table 8 shows one labelled example, and the process of adding auxiliary supervision label contain the following steps:

  1. 1.

    Annotate the samples manually: For the selected examples, we find the spans for the worlds, cause property and effect property in the background, cause property for the Worlds in the situation and decide the results for comparison, relation and reasoning modules.

  2. 2.

    Generate machine-readable labels automatically: Then we use scripts to automatically transform the annotations to the machine-readable form, i.e. we record the start and end character index for all spans and keep the results for comparison, relation and reasoning modules as binary form.

C Error Cases for Modules

Table 9 lists out several type 1 error cases mentioned in the Error Analysis part, while Table 10 lists out type 2 error cases which beyonds the scope of our model.

D More Examples

We present more examples that correctly answered by our model in Table 11.

E Heuristic Rules for Answer Prediction

We also conduct a rule-based approached to predict the final answer which contains the following steps:

  1. 1.

    Categorize questions based on the type of answer: By looking at the labeled train dataset, we can summarize that the types of answer can be divided into two types:1.World Type, answer is one of the compared worlds; 2. Comparative Word Type, like ”more” or ”less”.

  2. 2.

    For World Type, we filter out such type of questions by searching question keywords, for example, questions started with { What, Which, Who, Where, When} usually have world type of answers. Then we determine the results based on the prediction obtained in Reasoning Module.

  3. 3.

    For Comparative Word Type, we further filter out this type of questions from the rest questions by defining a list of comparative word pair like {’more’:’less’,’higher’:’lower’… }. Then we identify the primary world that being compared in the question and associate it with our identified worlds from Group Detection module, then determine the comparative word for the primary compared world by using the results from Reasoning module.

  4. 4.

    For the remaining questions, we simply return the world with higher effect property probability from Reasoning module as the final answer.

Examples Prediction
ID:1867731649 & 350571026
Background: Fish mortality is a parameter used in fisheries population dynamics to account for the loss of fish in a fish stock through death. The mortality can be divided into two types:Natural mortality: the removal of fish from the stock due to causes not associated with fishing. Such causes can include disease, competition, cannibalism, old age, predation, pollution or any other natural factor that causes the death of fish. In fisheries models natural mortality is denoted by (M).[1]Fishing mortality: the removal of fish from the stock due to fishing activities using any fishing gear It is denoted by (F) in fisheries models.
Situation: Tony is about to go on two fishing trips in the up coming week. On Friday, he is going to Bear Lake, which is located near a factory that has been known to dump waste into the lake. On Saturday, he is going to Fox Lake, which is in a secluded valley.
Q&A:Which day will Tony visit a lake that more likely has more fish? Saturday
Q&A:Which lake probably has more fish in it? Fox Lake
Wrong Predicted Worlds
Worlds: [Bear Lake, Fox Lake] Effect: fishing CauseB\text{Cause}^{B}: Natural mortality CauseWorld1S\text{Cause}_{\text{World1}}^{S}: dump waste CauseWorld2S\text{Cause}_{\text{World2}}^{S}: secluded valley Cause Cmp: World 1 Relation: negatively related Effect Cmp: World 2 Final Answer: Fox Lake
ID:4215374242
Background: Making these healthy lifestyle choices can also help prevent some types of cancer. In addition, you can lower the risk of cancer by avoiding carcinogens , which are substances that cause cancer. For example, you can reduce your risk of lung cancer by not smoking. You can reduce your risk of skin cancer by using sunscreen. How to choose a sunscreen that offers the most protection is explained below ( Figure below ). Some people think that tanning beds are a safe way to get a tan. This is a myth. Tanning beds expose the skin to UV radiation. Any exposure to UV radiation increases the risk of skin cancer. It doesn’t matter whether the radiation comes from tanning lamps or the sun.
Situation: Steve and Bill are really good friends with each other. The other day they were talking about some habits they have. Steve likes to work out and stay in shape, eats healthy and does not smoke. Bill said he wants to be more like Steve. Bill currently smokes, and loves to go out tanning in the sun without out sunscreen.
Q&A:Who has a more likely chance to get lung cancer in the future? Bill
Wrong Predicted Comparison Results
Worlds: [Steve, Bill] Effect: risk of lung cancer CauseB\text{Cause}^{B}: not smoking. CauseWorld1S\text{Cause}_{\text{World1}}^{S}: does not smoke CauseWorld2S\text{Cause}_{\text{World2}}^{S}: smokes Cause Cmp: World 2 Relation: negatively related Effect Cmp: World 1 Final Answer: Steve
ID: 4035582237
Background: Sometimes muscles and tendons get injured when a person starts doing an activity before they have warmed up properly. A warm up is a slow increase in the intensity of a physical activity that prepares muscles for an activity. Warming up increases the blood flow to the muscles and increases the heart rate. Warmed-up muscles and tendons are less likely to get injured. For example, before running or playing soccer, a person might jog slowly to warm muscles and increase their heart rate. Even elite athletes need to warm up ( Figure below ).
Situation: Greg and Carl and about to do a marathon. Greg sees Carl doing some warm ups and laughs to himself and thinks it is silly. They both want to get a good time, and are both avid runners.
Q&A:Who is more likely to get an injury during the race? Greg
Wrong Predicted Cause Property in Situation
Worlds: [Greg,Carl] Effect: get injured. CauseB\text{Cause}^{B}: Warmed-up CauseWorld1S\text{Cause}_{\text{World1}}^{S}: warm ups CauseWorld2S\text{Cause}_{\text{World2}}^{S}: warm ups Cause Cmp: World 1 Relation: negatively related Effect Cmp: World 2 Final Answer: Carl
Table 9: Type 1 error cases made by our model
Examples Prediction
ID:710693196
Background: Fish reproduce sexually. They lay eggs that can be fertilized either inside or outside of the body. In most fish, the eggs develop outside of the mother’s body. In the majority of these species, fertilization also takes place outside the mother’s body. The male and female fish release their gametes into the surrounding water, where fertilization occurs. Female fish release very high numbers of eggs to increase the chances of fertilization.
Situation: All marine creatures are not fish. There are some mammals, for example, whales, also live in the water. Rob wants to know more about differences between fish and other non fish creatures in the water. He divided them into two groups, group A and group B. Group A consists of fish, and group B consists of non fish creatures in the water. He started to see the differences between these two groups
Q&A:In group B, would fertilization most likely take place inside or outside of mother’s body? inside
Missing Knowledge
Worlds: [Group A, group B] Effect: fish CauseB\text{Cause}^{B}: fertilization CauseWorld1S\text{Cause}_{\text{World1}}^{S}: fish CauseWorld2S\text{Cause}_{\text{World2}}^{S}: non fish creatures Cause Cmp: World 1 Relation: positively related Effect Cmp: World 1 Final Answer: Group A
ID:3339143431
Background: In exercises such as weight lifting, skeletal muscle contracts against a resisting force (see Figure below ). Using skeletal muscle in this way increases its size and strength. In exercises such as running, the cardiac muscle contracts faster and the heart pumps more blood. Using cardiac muscle in this way increases its strength and efficiency. Continued exercise is necessary to maintain bigger, stronger muscles. If you don’t use a muscle, it will get smaller and weaker–so use it or lose it.
Situation: A study was done in the town of Greenwich comparing muscle strength to the amount a person exercises. Mathew goes to the gym 5 times a week and does very intensive workouts. Damen on the other hand does not go to the gym at all and lives a mostly sedentary lifestyle.
Q&A:Given Mathew suffers an injury while working out and cannot go to the gym for 3 months, will Mathews strength increase or decrease? decrease
Implicit Worlds
Worlds: [Mathew, Damen] Effect: strength and efficiency CauseB\text{Cause}^{B}: exercises such as running CauseWorld1S\text{Cause}_{\text{World1}}^{S}: goes to the gym CauseWorld2S\text{Cause}_{\text{World2}}^{S}: does not go to the gym Cause Cmp: World 1 Relation: positively related Effect Cmp: World 1 Final Answer: Mathew
ID: 2918297602
Background: In common everyday speech, speed of sound refers to the speed of sound waves in air. However, the speed of sound varies from substance to substance: sound travels most slowly in gases; it travels faster in liquids; and faster still in solids. For example, (as noted above), sound travels at 343 m/s in air; it travels at 1,480 m/s in water (4.3 times as fast as in air); and at 5,120 m/s in iron (about 15 times as fast as in air). In an exceptionally stiff material such as diamond, sound travels at 12,000 metres per second (27,000 mph);[1] (about 35 times as fast as in air) which is around the maximum speed that sound will travel under normal conditions.
Situation: John and Keith are neighbors. They have been pondering about how to communicate with each other in a doomsday scenario when all the electronic devices would be useless. They connected their houses with three ducts. One of the ducts is filled with air; they called it channel A. Another duct is filled with water; they called it channel B. And the last duct is filled with iron; they called it channel C. They can now transmit sound with these channels of communication; in case, disaster strikes
Q&A:How much faster would be channel B than channel A in m/s, 1130 m/s or 1137 m/s? 1137 m/s
Additional Math Computation
Worlds: [channel A,channel B] Effect: 5,120 m/s CauseB\text{Cause}^{B}: water CauseWorld1S\text{Cause}_{\text{World1}}^{S}: air CauseWorld2S\text{Cause}_{\text{World2}}^{S}: water Cause Cmp: World 1 Relation: positively related Effect Cmp: World 1 Final Answer: channel A
Table 10: Type 2 error cases could not be solved by our model
Examples Prediction
ID:1629973236
Background:One result of air pollution is acid rain. Acid rain is precipitation with a low (acidic) pH. This rain can be very destructive to wildlife. When acid rain falls in forests, freshwater habitats, or soils, it can kill insects and aquatic life. It causes this damage because of its very low pH. Sulfur oxides and nitrogen oxides in the air both cause acid rain to form ( Figure below ). Sulfur oxides are chemicals that are released from coal-fired power plants. Nitrogen oxides are released from motor vehicle exhaust.
Situation: Bill is planning on moving soon. He wants to move to a city that has fresher air and more wildlife to see. His two options that he must choose from are St. Louis and Seattle. Recently,Seattle has installed a new wind farm, and zero emission solar farm to generate power, while St. Louis recently installed a coal fired power plant. Both cities have similar commercial and industrial sectors, and only differ in how they generate power.
Q&A: Which city will more likely have more vibrant wildlife? Seattle
Module Output:
Worlds: [Seattle,St. Louis] Effect: wildlife CauseB\text{Cause}^{B}: acid rain. CauseWorld1S\text{Cause}_{\text{World1}}^{S}: wind farm, CauseWorld2S\text{Cause}_{\text{World2}}^{S}: coal fired power plant. Cause Cmp: World 2 Relation: negatively related Effect Cmp: World 1 Final Answer: Seattle
ID:802123740
Background:One result of air pollution is acid rain. Acid rain is precipitation with a low (acidic) pH. This rain can be very destructive to wildlife. When acid rain falls in forests, freshwater habitats, or soils, it can kill insects and aquatic life. It causes this damage because of its very low pH. Sulfur oxides and nitrogen oxides in the air both cause acid rain to form ( Figure below ). Sulfur oxides are chemicals that are released from coal-fired power plants. Nitrogen oxides are released from motor vehicle exhaust.
Situation: Bill is planning on moving soon. He wants to move to a city that has fresher air and more wildlife to see. His two options that he must choose from are St. Louis and Seattle. Recently, Seattle has installed a new wind farm, and zero emission solar farm to generate power, while St. Louis recently installed a coal fired power plant. Both cities have similar commercial and industrial sectors, and only differ in how they generate power.
Q&A: Will Seattle have more or less sulfur oxides in the air than St. Louis? less
Module Output:
Worlds: [Seattle,St. Louis] Effect: Sulfur oxides CauseB\text{Cause}^{B}: coal-fired power plants. CauseWorld1S\text{Cause}_{\text{World1}}^{S}: wind farm, CauseWorld2S\text{Cause}_{\text{World2}}^{S}: coal fired power plant. Cause Cmp: World 2 Relation: positively related Effect Cmp: World 2 Final Answer: St. Louis
ID:2133492859
Background:Turner et al (2006) derived crash prediction models for this report’s predecessor and found a pronounced ’2018safety in numbers’ effect in the models. Using the crash prediction model for mid-block locations, generic motorist and cyclist volumes can be used to demonstrate the impacts on the expected crash rate of varying motor vehicle and cycle volumes. As shown in figure 2.20, an increase in the proportion of cyclists to the overall traffic volume causes an increase in expected crashes at mid-block locations, but the crash rate increases at a decreasing rate. That is to say, the crash rate per cyclist goes down as the cycle volume increases.
Situation: There were a lot of motorcycles on Interstate 17 last week. On Monday, there were 1355 motorcyclists. On Tuesday, there were 2355 motorcyclists. On Wednesday, there were 3351 motorcyclists. On Thursday, there were 4351 motorcyclists. On Friday, there were 5351 motorcyclists. On Saturday, there were 6351 motorcyclists. On Sunday, there were 7351 motorcyclists.
Q&A: What day had a lower crash rate per cyclist: Thursday or Sunday? Sunday
Module Output:
Worlds: [Thursday,Sunday] Effect: crash rate per cyclist CauseB\text{Cause}^{B}: the cycle volume CauseWorld1S\text{Cause}_{\text{World1}}^{S}: 4351 motorcyclists CauseWorld2S\text{Cause}_{\text{World2}}^{S}: 7351 motorcyclists Cause Cmp: World 2 Relation: negatively related Effect Cmp: World 1 Final Answer: Thursday
ID:3099625752
Background:The example of someone having a positive experience with a drug is easy to see how drug dependence and the law of effect works. The tolerance for a drug goes up as one continues to use it after having a positive experience with a certain amount the first time. It will take more and more to get that same feeling. This is when the controlled substance in an experiment would have to be modified and the experiment would really begin. The law of work for psychologist B. F. Skinner almost half a century later on the principles of operant conditioning, ä learning process by which the effect, or consequence, of a response influences the future rate of production of that response.
Situation: The Speed Squad met to discuss their experiences. They all said they always had a great experience using speed and used the same amount each time. They told how many times they used speed. Todd used it 36 times, Jesse used it 40 times, Craig used it 44 times, Alan used it 56 times, Shawn used it 69 times, Clarence used it 78 times, and Sean used it 86 times.
Q&A: Who has a higher tolerance for speed: Alan or Clarence? Clarence
Module Output:
Worlds: [Alan,Clarence] Effect: tolerance CauseB\text{Cause}^{B}: positive experience with CauseWorld1S\text{Cause}_{\text{World1}}^{S}: used it 56 times CauseWorld2S\text{Cause}_{\text{World2}}^{S}: used it 78 times Cause Cmp: World 2 Relation: positively related Effect Cmp: World 2 Final Answer: Clarence
Table 11: Examples correctly answered by our model in an intepretable manner.