Skip to main content
  • Research article
  • Open access
  • Published:

Evaluating the use of different positional strategies for sentence selection in biomedical literature summarization

Abstract

Background

The position of a sentence in a document has been traditionally considered an indicator of the relevance of the sentence, and therefore it is frequently used by automatic summarization systems as an attribute for sentence selection. Sentences close to the beginning of the document are supposed to deal with the main topic and thus are selected for the summary. This criterion has shown to be very effective when summarizing some types of documents, such as news items. However, this property is not likely to be found in other types of documents, such as scientific articles, where other positional criteria may be preferred. The purpose of the present work is to study the utility of different positional strategies for biomedical literature summarization.

Results

We have evaluated three different positional strategies: (1) awarding the sentences at the beginning of the document, (2) preferring those at the beginning and end of the document, and (3) weighting the sentences according to the section in which they appear. To this end, we have implemented two summarizers, one based on semantic graphs and the other based on concept frequencies, and evaluated the summaries they produce when combined with each of the positional strategies above using ROUGE metrics. Our results indicate that it is possible to improve the quality of the summaries by weighting the sentences according to the section in which they appear (≈17% improvement in ROUGE-2 for the graph-based summarizer and ≈20% for the frequency-based summarizer), and that the sections containing the more salient information are the Methods and Material and the Discussion and Results ones.

Conclusions

It has been found that the use of traditional positional criteria that award sentences at the beginning and/or the end of the document are not helpful when summarizing scientific literature. In contrast, a more appropriate strategy is that which weights sentences according to the section in which they appear.

Introduction and Motivation

The amount of biomedical literature being published is growing rapidly in recent years, making it difficult for researchers to find the information they need. In this context, text automatic techniques may help alleviate the information overload problem. First, automatic summaries may be useful in anticipating the contents of the original documents, so that users may decide which of the documents to read further. As stated in [1], even in the presence of the author’s abstract, there are two main reasons for wanting to generate text summaries from a full-text: (1) the abstract, which is usually limited to around 200 words, may be missing relevant content, and (2) there is not a single ideal summary, but rather, the ideal summary depends on the user’s information needs. Moreover, automatic summaries have been shown to improve indexing and categorization of biomedical literature when used as substitutes for the articles’ abstracts [2, 3], since they help to filter non relevant and noisy information.

Text summarization refers to the process of generating a brief summary of one or several documents [4]. Summaries may be extractive or abstractive. Extractive summaries are created by identifying salient textual units (i.e., sentences or paragraphs) in the sources, while abstractive summaries are built by paraphrasing the information in the original documents. In other words, while extractive summarization is mainly concerned with what the summary content should be, abstractive summarization puts the emphasis on the form [5]. Although human summaries are typically abstracts, most existing systems produce extracts largely because extractive summarization has been demonstrated to report better results than abstractive summarization [6]. This is due to the difficulties that the abstraction process entails, which usually involves the identification of the most prevalent concepts in the source, the appropriate semantic representation of them, a minimum level of inference and the rewriting of the summary through Natural Language Generation techniques.

Extractive methods typically construct summaries based on a superficial analysis of the text. The most popular approaches include statistical techniques and graph-based methods (see [4, 7] for a more detailed study of summarization techniques). Graph-based methods represent the text as a graph, where the nodes correspond to word, sentences or even concepts, and the edges represent various types of syntactic and semantic relations among them. Different clustering methods are then applied to identify salient nodes within the graph and to extract the sentences for the summary.

Statistical approaches are based on simple heuristics to rank the sentences for the summary, such as the position of the sentences in the document [8-12], the frequency of their terms [9, 10, 13-17], the presence of certain cue words [14, 18, 19], and the word overlap between the sentences and the document title and headings [11, 14]. Despite their simplicity, these features are commonly used in the most recent works on extractive summarization, usually in combination with other more complex approaches, such as graph-based or template-based [20, 21].

Focusing on the position of sentences in a document, this has been traditionally considered an important factor in finding the sentences that are most related to the central topic of the document, and used in different NLP tasks [8, 22, 23]. Baxendale [8] examined 200 paragraphs to find that in 85% of the paragraphs the topic sentence came as the first one and in 7% of the time it was the last sentence. Other works argue that the sentences close to the beginning and/or the end of the document are supposed to deal with the main theme of the document, and so more weight is assigned to them [11, 12]. This criterion has showed to be very effective when summarizing some types of documents, such as news items, where the information is placed following the inverted pyramid structure (i.e., the most important information is placed first but, as the article continues, the less important details are presented) [24]. However, as stated in [5, 22], even though texts generally follow a predictable discourse structure, and the sentences of greater topic centrality tend to occur in certain specifiable locations, this structure significantly varies over domains and the importance of the sentence position must be evaluated ad hoc for each domain and type of document.

Regarding summarization of biomedical literature, our previous work [20] showed that using a positional function that attaches greater relevance to sentences close to the beginning and end of the document together with a semantic graph-based summarization approach decreases performance compared with not using any positional information. This result was not surprising because scientific papers are not (a priori) expected to present the core information at the beginning and end of the document. In contrast, the first sentences in scientific papers usually introduce the motivation of the study, whereas the last sentences provide conclusions and future work. The most important information is expected to be found in the middle sentences, as part of the method, results, and discussion sections. Therefore, it seems that a more appropriate positional criterion would be one that gives priority to sentences belonging to such central sections. This intuition, however, needs to be empirically evaluated.

Following this idea, in the present work we study if the use of different positional criteria may be of help when summarizing scientific biomedical articles. In particular, three strategies are examined: (1) awarding the sentences at the beginning of the document, (2) preferring those at the beginning and end of the document, or (3) weighting the sentences according to the section (or section group) in which they appear. To this end, we have implemented two different summarizers, one based on semantic graphs and the other based on concept frequencies, and evaluated the summaries they produce when combined with each of the positional strategies above. Our results show that it is possible to improve the quality of the summaries that are generated by weighting the sentences according to the section in which they appear and giving priority to sentences from the Method and Material and Results and Discussion sections of the article. We believe these results to be of great interest since they may guide NLP tasks involving extraction of salient information in biomedical literature.

The paper is structured as follows. We first present some related work in biomedical summarization. We next describe several positional strategies for sentence selection, along with the two summarizers. We evaluate the summaries generated by both summarizers using the different positional strategies and present the evaluation results. These results are then discussed. We finally draw the main conclusions of the study and outline future work.

Background

Even though the first works in automatic text summarization date from the middle of the last century [13], research in biomedical summarization has started only recently. Biomedical summarization works typically adapt existing methods from domain-independent summarization to deal with the highly specialized biomedical terminology. To this end, they make use of external knowledge sources to represent the texts as sets of domain concepts and relations. This produces a richer representation than the one provided by traditional term-based models and results in better quality summaries.

A pioneer work in biomedical summarization is found in [25]. They propose the use of semantic predications provided by SemRep [26] and information from the Unified Medical Language System (UMLS)Ⓡ[27] to extract biomedical entities and relations, and generate semantic-level abstracts, which are presented in graphical format. Ling et al. [28] focus on a narrower domain, genomic, and present a gene summary system that ranks sentences according to three features: the relevance of six gene aspects, such as the DNA sequence, the relevance of the documents where the sentences are taken from, and the position of the sentences in the document. Reeve et al. [1] use the frequency of the UMLS Metathesaurus concepts found in the text and adapt the lexical chaining approach [29] to deal with concepts instead of terms. Their system is used to produce single-document extracts of biomedical articles.

More sophisticated is the work of Yoo et al. [30] for multi-document summarization. They represent a corpus of documents as a graph, where the nodes are the MeSHⓇs[31] descriptors found in the corpus and the edges represent hypernymy and co-occurrence relations between them. They cluster the MeSH concepts in the corpus to identify sets of documents dealing with the same topic and then generate a summary from each document cluster. BioSquash [32] is a question-oriented extractive system for biomedical multi-document summarization. It constructs a semantic graph that contains concepts of three types: ontological concepts (general ones from WordNet [33] and specific ones from the UMLS), named entities and noun phrases.

More recent is the work of Shang et al. [34], where the aim is to combine information retrieval techniques with information extraction methods to generate text summaries of sets of documents describing a certain topic. To do this, they use SemRep to extract relations among UMLS Metathesaurus concepts and a relation-level retrieval method to select the relations more relevant to a given query concept. Finally, they extract the most relevant sentences for each topic based on the previous ranking of relations and the location of the sentences in different sections of the document. However, no details are given about how the location scores are calculated.

Methods

In this section, we first present the different positional strategies for sentence selections. We next describe the two different summarizers (one based on semantic graphs and the other based on concept frequencies) that have been developed to test such positional strategies.

Positional strategies for sentence selection

In order to test our hypothesis that the position of a sentence in the different sections of the document is an indication of the importance of the sentence for inclusion in a summary, and that traditional positional strategies are not appropriate for summarizing biomedical literature, we have defined the following positional criteria:

  • Distance to the beginning ( Begin-Pos ): Sentences close to the beginning of the document are supposed to the deal with the central topic of the document, and so more weight is assigned to them. Thus, a score B e g i n _P o s(S j )∈(0,1] is calculated as the reciprocal to the position of the sentence in the document, as shown in equation 1, where m j represents the position of the sentence, S j , within the document. In this way, for instance, for the first sentence in the document Begin_Pos( S 1 )= 1 1 =1, for the second sentence Begin_Pos( S 2 )= 1 2 =0.5, and so on.

    Begin_Pos( S j )= 1 m j
    (1)
  • Distance to the beginning and end ( Begin-End-Pos): Sentences close to the beginning and the end of the document are considered highly relevant. Therefore, a score B e g i n _E n d _P o s(S j )∈(0,1] is calculated for each sentence as shown in equation 2, where M represents the number of sentences in the document and m j represents the position of the sentence, S j , within the document. This equation returns the maximum of the reciprocal to the position of the sentence from the beginning and the end of the document. In this way, for instance, if M=10, then for the first sentence in the document Begin_End_Pos( S 1 )=max{ 1 1 , 1 10 − 1 + 1 }=1, for the second sentence Begin_End_Pos( S 2 )=max{ 1 2 , 1 10 − 2 + 1 }= 1 2 =0.5, and for the last sentence Begin_End_Pos( S 10 )=max{ 1 10 , 1 10 − 10 + 1 }=1.

    Begin_End_Pos( S j )=max 1 m j , 1 M − m j + 1
    (2)
  • Section in the document ( Section-Pos ): We consider the following section classes or clusters: (1) Introduction, (2) Background, (3) Materials and Methods, (4) Results and Discussion, and (5) Conclusions and Future Work. Following the methodology used in [2], we have first analyzed the typical structure of biomedical articles and manually grouped the section headers in these five section groups based on the similarity of their contents. Lexical variants of the same section header are included in the same class. For example, Method and Methods are clustered in the Materials and Methods class. Similarly, sections that differ in their title but refer to the same content are included in the same class (e.g., Experimental Procedures is also included in the class Material and Methods). We calculate a score S e c t i o n _P o s(S j )∈(0,1] according to the equation 3, where the functions I n t r o(S j ), B a c k G(S j ), M&M(S j ), R&D(S j ), and C&F(S j ) are equal to 1 if the sentence S j belongs to each of the five section groups, respectively, and 0 otherwise. The values of γ, δ, θ, σ, and Π vary between 0 and 1, and need to be empirically determined.

    Section_ Pos ( S j ) = γ × Intro ( S j ) + δ × BackG ( S j ) + θ × M & M ( S j ) + σ × R & D ( S j ) + Π × C & F ( S j )
    (3)

Graph-based summarizer

We use the graph-based summarization method presented in [20]. This method is based on the representation of the document as a conceptual graph, using the UMLS [27] as the knowledge source, and the use of a degree-based clustering algorithm for detecting salient concepts within the graph. The original summarizer has been modified to incorporate more advanced positional strategies in the sentence selection step. The system architecture is illustrated in Figure 1.

Figure 1
figure 1

Architecture of the graph-based summarization system. The figure illustrates the different steps in the summarization process: (i) concept identification, (ii) document representation, (iii) topic recognition and (iv) sentence selection.

The method consists of the 4 main steps, which are briefly explained below (see [20] for a detailed explanation):

  • The first step, concept identification, is to map the document to concepts from the UMLS Metathesaurus and semantic types from the UMLS Semantic Network. We run the MetaMap [35] program over the body section of the document to obtain the Metathesaurus concepts that are found within the text. MetaMap is invoked using the word sense disambiguation option (-y flag). This flag implements the Journal Descriptor Indexing (JDI) methodology described in [36]. UMLS concepts belonging to very general semantic types are discarded since they have been found to be excessively broad and do not contribute to summarization. These types are Quantitative concept, Qualitative concept, Temporal concept, Functional concept, Idea or concept, Intellectual product, Mental process, Spatial concept and Language.

  • The second step, document representation, is to construct a graph-based representation of the document. To do this, we first extend the UMLS concepts with their complete hierarchy of hypernyms (is_a relations) and merge the hierarchies of all the concepts in the same sentence to construct a sentence graph. The upper levels of these hierarchies are removed, since they represent concepts with excessively broad meanings. Next, all the sentence graphs are merged into a single document graph. This graph is extended with two further types of relations: relations between concepts in the UMLS Metathesaurus and relations between semantic types in the UMLS Semantic Network (see [37] for a description of the different relationships in the UMLS). Finally, each edge is assigned a weight in [0, 1] as shown in equation 4. The weight of an edge e representing an is_a relation between two vertices, v i and v j (where v i is a parent of v j ), is calculated as the ratio of the depth of v i to the depth of v j from the root of their hierarchy. The weight of an edge representing any other relation (i.e. associated with and other related) between pairs of leaf vertices is always 1.

    weight(e, v i , v j )=β
    (4)
    where β = depth ( v i ) depth ( v j ) if e represents an is _ a relation β = 1 otherwise
  • To illustrate this process, Figure 2 shows the document graph for the following text from [38]:

Interactions among LRF-1, JunB, c-Jun, and c-Fos define a regulatory program in the G1 phase of liver regeneration. In regenerating liver, a physiologically normal model of cell growth, LRF-1, JunB, c-Jun, and c-Fos among Jun/Fos/LRF-1 family members are induced posthepatectomy. In liver cells, high levels of c-Fos/c-Jun, c-Fos/JunB, LRF-1/c-Jun, and LRF-1/JunB complexes are present for several hours after the G0/G1 transition, and the relative level of LRF-1/JunB complexes increases during G1. We provide evidence for dramatic differences in promoter-specific activation by LRF-1- and c-Fos-containing complexes. LRF-1 in combination with either Jun protein strongly activates a cyclic AMP response element-containing promoter which c-Fos/Jun does not activate.

  • The third step, topic recognition, consists of clustering the UMLS concepts in the document graph using a degree-based clustering method similar to that used by [30]. The aim is to construct sets of concepts strongly related in meaning, based on the assumption that each of these clusters represents a different topic in the document. We first compute the salience of each vertex in the graph, as the sum of the weights of the edges that are linked to it. Next, the nodes are ranked according to its salience. The n vertices with a highest salience are labeled as hub vertices. The clustering algorithm then groups the hub vertices into hub vertex sets (HVS). These can be interpreted as sets of concepts strongly connected and will represent the centroids of the final clusters. The remaining vertices (i.e. those not included in the HVS) are iteratively assigned to the cluster to which they are more connected. The output of this step is, therefore, a number of clusters of UMLS concepts, each cluster represented by the set of most highly connected concepts within it (the so-called HVS). In this way, for instance, the top five salient concepts in the document graph represented in Figure 2 are: cells, LRF, cJUN, c-fos, and growth.

  • The last step, sentence selection, consists of computing the similarity between each sentence graph and each cluster, and selecting the sentences for the summary based on these similarities. To compute sentence-to-cluster similarity, we use a non-democratic vote mechanism so that each vertex of a sentence assigns a vote to a cluster if the vertex belongs to its HVS, half a vote if the vertex belongs to it but not to its HVS, and no votes otherwise. The similarity between the sentence graph and the cluster is computed as the sum of the votes assigned by all the vertices in the sentence graph to the cluster. A single score for each sentence is calculated, as the sum of its similarity to each cluster adjusted to the cluster’s size (equation 5).

    Sem_Sim( S j )= ∑ C i similarity ( C i , S j ) | C i |
    (5)

Finally, this semantic similarity is normalized in the [0,1] interval and combined with each of the positional criteria explained in the previous section using a linear function (see equation 6). The N sentences with higher score are then selected for the summary.

Score( S j )=α×Sem_Sim( S j )+β×Position( S j );
(6)

α and β can be assigned different weights between 0 and 1. Their optimal values must be determined empirically.

Figure 2
figure 2

Example document graph. Dashed black lines represent hypernymy relations; red lines represent Metathesaurus relations; and blue lines represent Semantic Network relations.

Concept frequency-based summarizer

The second summarizer is based on the frequency of UMLS concepts in the document. It consists of 4 steps:

  • The first step, concept identification, is to map the document to concepts from the UMLS Metathesaurus and semantic types from the UMLS Semantic Network using MetaMap, as explained for the graph-based summarizer.

  • Concept frequency representation: Following Luhn’s theory, we assume that the more times a word (or concept) appears in a document, the more relevant become the sentences that contain this word. In this way, if {C1,C2,...,C n } is the set of n Metathesaurus concepts that appear in the document d, and f i (d) is the number of times that C i appears in it, then the document may be represented by the vector D={f1(d),f2(d),...,f n (d)}. Similarly, we build the vectors representing each sentence in the document, S j , and compute a concept frequency score, C F(S j ), as the sum of the frequency of all the concepts in the sentence multiplied by the frequency of those concepts in the document. This C F(S j ) score is normalized in the [0,1] interval.

  • In this way, the text modeled as a conceptual graph in Figure 2 is represented by the following document and sentence vectors:

    D = { LRF = 7 , c − Jun = 3 , c − fos = 3 , Liver = 3 , Cell = 2 , Promoter = 2 , Complexes = 2 , Program = 1 , Regeneration = 1 , growth = 1 , LRF − 1 = 1 , Transition = 1 , C − Complex = 1 , Combination = 1 , JUNProtein = 1 , Protein = 1 , Element = 1 } S 1 = { LRF = 1 , c − Jun = 1 , c − fos = 1 , Program = 1 , Liver = 1 , Regeneration = 1 } S 2 = { LRF = 2 , c − Jun = 1 , c − fos = 1 , Liver = 1 , Cell = 1 , growth = 1 } S 3 = { LRF = 2 , c − Jun = 1 , c − fos = 1 , Liver = 1 , Cell = 1 , LRF − 1 = 1 , Transition = 1 , Complexes = 2 } S 4 = { LRF = 1 , Promoter = 1 , C − Complex = 1 } S 5 = { LRF = 1 , Promoter = 1 , Combination = 1 , JUNProtein = 1 , Protein = 1 , Element = 1 }
  • The concept frequency scores of the sentences are, respectively, C F(S1)=18, C F(S2)=26, C F(S3)=31, C F(S4)=10 and C F(S5)=13.

  • Sentence position: Different position scores are calculated for each sentence in the document, according to the positional strategies described previously in the article.

  • The last step, sentence selection, consists of extracting the most important sentences for the summary up to the desired length. Having computed the different weights for each sentence, the final score for a sentence (Score(S j )) is calculated according to the equation 7. Finally, the N sentences with higher score are extracted for the summary.

    Score( S j )=α×CF( S j )+β×Position( S j )
    (7)
  • α and β can be assigned different weights between 0 and 1. Their optimal values must be determined empirically.

Evaluation methods

The most common approach to evaluating automatically generated summaries of a document (also known as peers) is to compare them against manually-created summaries (called reference or model summaries) and measure the similarity between their content. The more content that is shared between the peer and reference summaries, the better the peer summary is assumed to be. To the authors’ knowledge, no corpus of model summaries exists for biomedical articles. For this reason, in this work we use a collection of 100 biomedical scientific articles randomly selected from the PMC Open Access Subset [39]. When collecting the articles, we made sure that they present, at least, the five main following sections: Introduction, Background, Methods, Results and Discussion, and Conclusions and Future Work. The abstracts for the articles are used as model summaries, since they condensate the most relevant content in the articles and have been written manually.

The ROUGE metrics [40] are used to quantify the content similarity between the automatic summaries and the reference ones. ROUGE is a commonly used evaluation method for summarization which uses the proportion of n-grams between a peer and one or more reference summaries to compute a value within [0,1]. Higher values of ROUGE are preferred, since they indicate a greater content overlap between the peer and the model. The following ROUGE metrics are used: ROUGE-2 and ROUGE-SU4. ROUGE-2 counts the number of bigrams that are shared by the peer and reference summaries and computes a recall-related measure. Similarly, ROUGE-SU4 measures the overlap of skip-bigrams (i.e., pairs of words in their sentence order, allowing for arbitrary gaps), using a skip distance of 4. Both ROUGE-2 and ROUGE-SU4 have shown high correlation with the human judges gathered from the Document Understanding Conferences [41]. However, it must be noted that ROUGE metrics present two important limitations: (1) they depend on the length of the peer summaries (i.e., the longer is the peer with respect to the model, the higher are expected to be the ROUGE scores), and (2) since they use lexical matching instead of semantic matching, peer summaries that are worded different but have the same semantic information may be assigned different ROUGE scores. Thus, these metrics should only be used in a comparative fashion on the same dataset and should not be interpreted as absolute measures.

Automatic summaries are generated by selecting sentences until each summary reaches the same number of sentences than its corresponding model summary (i.e., the article’s abstract). We generate summaries using both the graph-based and the frequency-based summarizers and the three positional strategies for sentence selection, and assigning different weights to the different parameters of the summarizers. For these experiments, different combinations of values for α, β, γ, δ, θ, σ, and Π were tested. However, for the sake of brevity, only the combinations that produced the best ROUGE scores are presented. A Wilcoxon Signed Ranks Test with a 95% confidence interval is used to test statistical significance of the results.

Results

We first evaluate the adequacy of the Begin-Pos positional strategy. The results of evaluating the automatic summaries generated when different weights are assigned to such criterion are shown in Table 1. As it may seen from this table, giving greater weight to sentences at the beginning of the document improves the quality of the automatic summaries compared with not using positional information, but only when the weight assigned to the position of the sentences is low (from 0.1 to 0.2). The improvement achieved is higher for the frequency-based summarizer than for the graph-based one, and it is only statistically significant for the first one.

Table 1 ROUGE scores for the summaries generated using the Begin-Pos strategy

We next evaluate the effect of the Begin-End-Pos criterion, which attaches greater weight to sentences close to the beginning and end of the document. It may be seen from Table 2 that this strategy does not benefit the quality of the graph-based summaries, regardless of the weights assigned to the different criteria, but slightly improves the performance of the concept-frequency based summarizer when β is set to 0.1. However, once again, the improvement achieved is not significant.

Table 2 ROUGE scores for the summaries generated using the Begin-End-Pos strategy

We finally examine the summaries generated when different weights are assigned to sentences depending on the section in which they appear (Section-Pos strategy). These results are shown in Table 3. For both summarizers, there exist a combination of weights that produces significantly better summaries compared with the non-positional information summaries (β=0.0). In particular, the best ROUGE scores are reported when the Introduction and the Conclusions and Future Work sections are given a weight of 0.2, no weight is assigned to sentences from the Related Work section, and the sentences from the Methods and Material and the Results and Discussion sections are given a weight of 1.0. This configuration allows for improvements of over 17% in ROUGE-2 for the graph-based summarizer and over 20% for the frequency-based summarizer.

Table 3 ROUGE scores for the summaries generated using the Section-Pos strategy

It is also worth mentioning that the experiments showed that weights for the Background and Conclusion and Future Work sections (i.e., δ and Π values) above 0.1 produce very poor summarization results, and that γ values (i.e., the weight for the Introduction section) upper 0.2 decrease performance as well. In contrast, the best results are reported when the Methods and Material and the Results and Discussion sections are assigned high weights.

Finally, Table 4 compiles the best results for each positional strategy and summarizer. For comparison purposes, this table also shows the ROUGE scores for the summaries generated using LexRank [42]. LexRank is the best-known graph-based method for summarization. It models documents as undirected graphs in which each node corresponds to a sentence, represented by its TF-IDF vector, and the edges are labeled with the cosine similarity between the sentences. It may be seen from Table 4 that the best ROUGE scores are obtained when the graph-based summarizer is combined with information about the position of the sentences in the different document sections. These scores are significantly better than those of the frequency-based summarizer and than those of LexRank.

Table 4 Comparison of summarization approaches

Discussion

The results in Tables 1, 2, and 3 confirm our hypothesis that traditional positional strategies are not appropriate when summarizing biomedical scientific articles, as opposed to summarizing other types of documents, such as news items. Our experiments have shown that awarding sentences close to the end of the document decreases summarization performance comparing with not using any positional information, while awarding sentences close to the beginning of the document only improves the quality of the summaries for the frequency-based summarizer.

In contrast, we have found that it is possible to improve summarization by taking into account the section in the article in which the different sentences appear, and attaching greater relevance to sentences from the appropriate sections. In particular, it has been found that sentences from the Methods and Material and Results and Discussion sections are more relevant for inclusion in the summary, since they are more related to the main topic of the document that sentences in other sections, such as Introduction and Conclusions and Future Work. Sentences in the Related Work section seem to be secondary, and therefore are not usually included in a summary. These results confirm what may be observed in the abstracts of the articles. If we examine such abstracts we realize that the information considered as important by the authors of the articles is mostly related to the method and results being described in the article, while the remaining sections are given less credit.

An interesting finding is that, in general, the frequency-based approach takes more advantage from the information about the position of sentences in the document than the graph-based one. We think this is due to the fact that the use of the frequency of the concepts alone is not enough to capture the salience of the sentences, and so the use of the positional criteria helps to bias the selection of sentences toward the most relevant information.

In contrast, the graph-based method captures better the importance of the different sentences, and thus produces better quality summaries even when no positional information is used. However, it still presents some limitations that will be addressed in future work. The first limitation is to do with the coverage of the UMLS. While general clinical terms are quite well covered, other vocabulary, specially that related to genomic, is not well supported [43]. As a results, automatic summaries of genetic and proteomic articles present low ROUGE values.

The second limitation is to do with the accuracy of the MetaMap mappings and the ambiguity in the UMLS Metathesaurus. Even using the -y disambiguation option, the precision of the JDI algorithm is reported to be around 0.78 when evaluated against a set of 45 ambiguous terms from the NLM-WSD corpus [44]. This precision, however, is expected to be lower for genomic entities, such as protein and gene names, where ambiguity is more frequent.

Third, our summarization and evaluation methods assume that all users have the same information needs, and that these needs are reflected in the authors’ abstracts. However, different users may have different interests. In future work, we plant to extend the summarizer to produce query-based summaries that take into account the readers’ information needs as specified in a user’s query. To this end, the similarity of each sentence in the document to the user’s query may be computed and uses as a feature for sentence selection.

Conclusions

This work explores the utility of the position of the sentences as a feature for automatic summarization of scientific articles. Toward this goal, we have developed two different summarizers, one based on semantic graphs and the other using concept frequencies, which implement three different positional strategies: the first gives more importance to sentences at the beginning of the article, the second prefers sentences both at the beginning and end, and the third weights sentences according to the section in which they appear. The summaries generated are evaluated and compared with non-positional summaries.

Overall, the results suggest that it is possible to improve summarization by taking into account the section in the article in which the different sentences appear, and attaching greater relevance to sentences from the appropriate sections. In contrast, traditional strategies that attach greater weights to sentences at the beginning and end of the document are not suitable when summarizing biomedical scientific articles. We believe that our results are of great interest since they may guide NLP tasks involving extraction of salient information in biomedical literature.

As future work, we plan to investigate the importance of the specific location of sentences within the different sections of the article. In this way, for instance, the last sentences of the Introduction section may be more relevant that the first sentences in the same sections, since they usually anticipate the content of the document.

References

  1. Reeve L, Han H, Brooks A: The use of domain-specific concepts in biomedical text summarization. Inf Process Manage 2007, 43: 1765-1776. 10.1016/j.ipm.2007.01.026

    Article  Google Scholar 

  2. Gay C, Kayaalp M, Aronson A: Semi-automatic indexing of full text biomedical articles. In Proceedings of the AMIA Annual Symposium. Washington, DC; 2005:271-275.

    Google Scholar 

  3. Identification of important text in full text articles using summarization. Tech. rep., National Library of Medicine; 2006.http://ii.nlm.nih.gov/resources/Summarization_and_FullText.pdf []. [Last accessed December, 3rd, 2012]

  4. Mani I: Automatic Summarization. Amsterdam: J. Benjamins Pub. Co.; 2001.

    Book  Google Scholar 

  5. Das D, Martins AF: A survey on automatic text summarization. Literature Survey for the Language and Statistics II course at CMU 2007. [http://www.cs.cmu.edu/~nasmith/LS2/das-martins.07.pdf] []

    Google Scholar 

  6. Mani I, Maybury M: Advances in Automatic Text Summarization. Cambridge: The MIT Press; 1999.

    Google Scholar 

  7. Afantenos S, Karkaletsis V, Stamatopoulos P: Summarization from medical documents: a survey. Artif Intell Med 2005,33(2):157-177. 10.1016/j.artmed.2004.07.017

    Article  PubMed  Google Scholar 

  8. Baxendale P: Machine-made index for technical literature - an experiment. IBM J Res Dev 1958,2(4):354-361.

    Article  Google Scholar 

  9. Kupiec J, Pedersen JO, Chen F: A trainable document summarizer. In Proceedings of the 18th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. Seattle, WA; 1995:68-73.

    Google Scholar 

  10. Teufel S, Moens M: Sentence extraction as a classification task. In Proceedings of the Association for Computational Linguistics, Workshop on Intelligent Scallable Text Summarization. Madrid, Spain; 1997:58-65.

    Google Scholar 

  11. Bawakid A, Oussalah M: A semantic summarization system: University of Birmingham at TAC 2008. In Proceedings of the First Text Analysis Conference. Gaithersburg, MD; 2008.

    Google Scholar 

  12. Bossard A, Genereux M, Poibeau T: Description of the LIPN systems at TAC 2008: summarizing information and opinions. In Proceedings of the 1st Text Analysis Conference. Gaithersburg, MD; 2008.

    Google Scholar 

  13. Luhn HP: The automatic creation of literature abstracts. IBM J Res Dev 1958,2(2):159-165.

    Article  Google Scholar 

  14. Edmundson HP: New methods in automatic extracting. J Assoc Comput Mach 1969,2(16):264-285.

    Article  Google Scholar 

  15. Hovy E, Lin CY: Automated Text Summarization in SUMMARIST. Cambridge: The MIT Press; 1999.

    Google Scholar 

  16. Steinberger J, Poesio M, Kabadjov MA, Jezek K: Two uses of anaphora resolution in summarization. Inf Process Manage 2007,43(6):1663-1180. 10.1016/j.ipm.2007.01.010

    Article  Google Scholar 

  17. Lloret E, Ferrández O, Munoz R, Palomar M: A text summarization approach under the influence of textual entailment. In Proceedings of the 5th International Workshop on Natural Language Processing and Cognitive Science in Conjunction with the 10th International Conference on Enterprise Information Systems. Madeira, Portugal; 2008:22-31.

    Google Scholar 

  18. Rush JE, Zamora A, Salvador R: Automatic abstracting and indexing II. production of abstracts by application of contextual inference and syntactic coherence criteria. J Am Soc Inf Sci 1971,22(4):260-274. 10.1002/asi.4630220405

    Article  CAS  Google Scholar 

  19. Aker A, Gaizauskas R: Generating image descriptions using dependency relational patterns. In Proceedings of the Association of Computational Linguistic. Uppsala, Sweeden; 2010:1250-1258.

    Google Scholar 

  20. Plaza L, Díaz A, Gervás P: A semantic graph-based approach to biomedical summarisation. Artif Intell Med 2011, 53: 1-15. 10.1016/j.artmed.2011.06.005

    Article  PubMed  Google Scholar 

  21. Oakes MP, Paice CD: The automatic generation of templates for automatic abstracting. In BCS-IRSG Annual Colloquium on IR Research. Autrans, France; 1999.

    Google Scholar 

  22. Lin C, Hovy E: Identifying topics by position. In Proceedings of the Fifth Conference on Applied Natural Language Processing. Washington, DC; 1997:283-290.

    Chapter  Google Scholar 

  23. Ko Y, Park J, Seo J: Improving text categorization using the importance of sentences. Inf Process Manage 2004, 40: 65-79. 10.1016/S0306-4573(02)00056-0

    Article  Google Scholar 

  24. Plaza L, Díaz A: Using semantic graphs and word sense disambiguation techniques to improve text summarization. In Proceedings of the XXVII Congreso de la Sociedad Espanola para el Procesamiento del Lenguaje Natural (SEPLN 2011). Huelva, Spain; 2011:97-105.

    Google Scholar 

  25. Fiszman M, Rindflesch TC, Kilicoglu H: Abstraction summarization for managing the biomedical research literature. In Proceedings of the HLT-NAACL Workshop on Computational Lexical Semantics. Boston, MA; 2004:76-83.

    Chapter  Google Scholar 

  26. Rindflesch T, Fiszman M: The interaction of domain knowledge and linguistic structure in natural language processing: interpreting hypernymic propositions in biomedical text. J Biomed Inf 2003, 36: 462-477. 10.1016/j.jbi.2003.11.003

    Article  Google Scholar 

  27. UMLS (Unified medical language System) [http://www.nlm.nih.gov/research/umls/] []

  28. Ling X, Jiang J, He X, Mei Q, Zhai C, Schatz B: Generating gene summaries from biomedical literature: A study of semi-structured summarization. Inf Process Manage 2007,43(6):1777-1791. 10.1016/j.ipm.2007.01.018

    Article  Google Scholar 

  29. Barzilay R, Elhadad M: Using lexical chains for text summarization. In Proceedings of the ACL Workshop on Intelligent Scalable Text Summarization. Madrid, Spain; 1997:10-17.

    Google Scholar 

  30. Yoo I, Hu X, Song IY: A coherent graph-based semantic clustering and summarization approach for biomedical literature and a new summarization evaluation method. BMC Bioinformatics 2007,8(9):S4. 10.1186/1471-2105-8-S9-S4

    Article  PubMed Central  PubMed  Google Scholar 

  31. Medical subject headings [http://www.nlm.nih.gov/mesh/] []

  32. Shi Z, Melli G, Wang Y, Liu Y, Gu B, Kashani MM, Sarkar A, Popowich F: Question answering summarization of multiple biomedical documents. In Proceedings of the Canadian Conference on Artificial Intelligence. Montreal, Canada; 2007:284-295.

    Google Scholar 

  33. Miller GA: WordNet: a lexical database for english. Commun ACM 1995,38(11):39-41. 10.1145/219717.219748

    Article  Google Scholar 

  34. Shang Y, Li Y, Lin H, Yang Z: Enhancing biomedical text summarization using semantic relation extraction. PLoS one 2011,6(8):e23862. 10.1371/journal.pone.0023862

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  35. MetaMap [http://metamap.nlm.nih.gov/] []

  36. Humphrey S, Rogers W, Kilicoglu H, Demner-Fushman D, Rindflesch T: Word sense disambiguation by selecting the best semantic type based on journal descriptor indexing: preliminary experiment. J Am Soc Inf Sci Technol 2006, 57: 96-113. 10.1002/asi.20257

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  37. UMLS Reference Manual [http://www.ncbi.nlm.nih.gov/books/NBK9676/] []

  38. Hsu J, Bravo R, Taub R: Interactions among LRF-1, JunB, c-Jun, and c-Fos define a regulatory program in the G1 phase of liver regeneration. Mol Cell Biol 1992,12(10):4654-4665.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  39. PMC Open Access Subset [http://www.ncbi.nlm.nih.gov/pmc/tools/openftlist/] []

  40. Lin CY: Rouge: A package for automatic evaluation of summaries. In Proceedings of the ACL 2004 Workshop: Text Summarization Branches Out. Barcelona, Spain; 2004:74-81.

    Google Scholar 

  41. Document Understanding Conferences [http://duc.nist.gov/] []

  42. Erkan G, Radev DR: LexRank: graph-based lexical centrality as salience in text summarization. J Artif Intell Res (JAIR) 2004, 22: 457-479.

    Google Scholar 

  43. Bodenreider O, Mitchell J, McCray A: Evaluation of the UMLS as a terminology and knowledge resource for biomedical informatics. In Proceedings of AMIA Symposium. Chicago, IL; 2002:61-65.

    Google Scholar 

  44. Weeber M, Mork J, Aronson A: Developing a test collection for biomedical word sense disambiguation. In Proceedings of the AMIA Symposium. Washington, DC; 2001:746-750.

    Google Scholar 

Download references

Acknowledgements

This research is partially founded by the European Union (FP7-ICT-2011 - Language Technologies - nr 288024 (LiMoSINe)).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Laura Plaza.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

LP designed and developed the graph-based method for automatic summarization and performed the evaluation. JCdA designed and developed the concept frequency-based summarizer and performed the evaluation. LP and JCdA drafted and reviewed the manuscript. Both authors read and approved the final manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Plaza, L., Carrillo-de-Albornoz, J. Evaluating the use of different positional strategies for sentence selection in biomedical literature summarization. BMC Bioinformatics 14, 71 (2013). https://doi.org/10.1186/1471-2105-14-71

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1471-2105-14-71

Keywords