Data Descriptor | Open | Published:

PubMed Phrases, an open set of coherent phrases for searching biomedical literature

Scientific Data volume 5, Article number: 180104 (2018) | Download Citation


In biomedicine, key concepts are often expressed by multiple words (e.g., ‘zinc finger protein’). Previous work has shown treating a sequence of words as a meaningful unit, where applicable, is not only important for human understanding but also beneficial for automatic information seeking. Here we present a collection of PubMed® Phrases that are beneficial for information retrieval and human comprehension. We define these phrases as coherent chunks that are logically connected. To collect the phrase set, we apply the hypergeometric test to detect segments of consecutive terms that are likely to appear together in PubMed. These text segments are then filtered using the BM25 ranking function to ensure that they are beneficial from an information retrieval perspective. Thus, we obtain a set of 705,915 PubMed Phrases. We evaluate the quality of the set by investigating PubMed user click data and manually annotating a sample of 500 randomly selected noun phrases. We also analyze and discuss the usage of these PubMed Phrases in literature search.

Design Type(s)
  • source-based data analysis objective
  • natural language processing objective
Measurement Type(s)
  • Natural Language
Technology Type(s)
  • class discovery data transformation
Factor Type(s)
    Sample Characteristic(s)

      ISA-Tab metadata

      Background & Summary

      Unlike other general domains, the language of biomedicine uses its own terminology to describe scientific discoveries and applications. To understand the semantics of biomedical text, it is important to identify not only the meaning of individual words, but also of multi-word phrases appearing in text1. Finding phrases is a fundamental but often overlooked process. Controlled vocabularies such as dictionaries and ontologies may help, but maintaining those is costly and their coverage is known to be limited. As demonstrated in a study of PubMed search2, meaningful phrases represent a significant fraction of queries in PubMed. This suggests that users, in many cases, have phrase(s) in mind when entering a query, e.g., ‘Central venous pressure’ and ‘familial Mediterranean fever’. While PubMed search interprets these queries as a conjunction of individual terms, an earlier study2 demonstrates that there is a qualitative difference between the results containing all individual terms and those containing the phrase. Therefore, it would be beneficial to interpret such queries as phrases.

      From a corpus linguistic point of view, coherent phrases are similar to collocations. A collocation is an expression consisting of two or more words that correspond to some conventional way of saying things3. This notion emphasizes the necessity of finding multi-word expressions in the biomedical domain. For example, a common colloquial expression such as ‘flu shot’ is rare in PubMed, while ‘influenza vaccine’ is the biomedical term used for the same concept. Our phrases and collocations share some similarities, but the main difference comes from grammatical completeness. Collocations are restricted to noun/adjective phrases or phrasal verbs, whereas we do not limit phrases grammatically, but rather see them as more flexible entities to be used as building blocks to form longer phrases or sentences. Such an interpretation of phrases is better aligned with our goal of using the corpus to analyze queries, as queries may frequently contain incomplete phrases and, in general, are known to differ from traditional forms of written language4. This makes us believe that statistical methods are better suited for our goal.

      While the task of identifying the most useful phrases has been studied extensively1,5,​6,​7,​8,​9, it remains challenging. Several studies have relied on natural language parsers to extract well-formed phrases such as noun, verb and/or prepositional phrases5,7,8. Other studies using statistical methods include noun phrase query segmentation10 and well-formed phrase extraction from MEDLINE®6. Another work developed for multi-word expression part-of-speech (POS) tagging11 utilizes neural networks to identify phrases. However, all the statistical approaches mentioned above are supervised, and thus require training sets to achieve their goal. Furthermore, it is unknown whether the results would improve search performance.

      Our approach is unsupervised, guided first by the hypergeometric test to determine whether a sequence of words is a coherent segment. The strings that pass the hypergeometric test are further treated as queries and evaluated in terms of their effectiveness in document retrieval in a comparative way: by treating them as a phrase versus as individual words. For example, given an extracted phrase ‘lung cancer patients’, we would compare retrieval performance when treating it as ‘lung cancer patients’ or as individual words, ‘lung’, ‘cancer’ and ‘patients’. This step is designed to select those phrases which result in improved information retrieval performance. Here, we use the BM25 ranking function12 for retrieval.

      To compute and compare the retrieval performance in the absence of a manually annotated gold standard, we use a novel pseudo-relevance judgement technique, which is based on the assumption that the documents containing query terms in the titles are more relevant to the query than the documents that do not13. Guided by this evaluation, we collect a set of 705,915 multi-word strings that benefit from being interpreted as phrases rather than individual tokens in terms of retrieval performance. We refer to this set as PubMed Phrases. PubMed Phrases, as an open data collection, represents a rich knowledge base of over seven hundred thousand phrases available to the scientific community. The resource can be readily used for text mining tasks including query/text segmentation, document understanding and beyond. Note that, throughout this paper, the term phrase refers to a coherent chunk of words that are frequently used together.


      In this section, we present our unsupervised framework for computing PubMed Phrases. This method consists of two steps: 1) obtaining candidate phrases from PubMed titles/abstracts and UMLS based on the hypergeometric test and 2) filtering the phrases using the BM25 ranking function. The first step segments text strings from PubMed documents and selects those segments that are likely to be used as a unit. The second step examines the phrases further to validate whether using the candidate as a coherent unit benefits retrieval performance. Figure 1 depicts an overall workflow of our approach.

      Figure 1: Workflow of PubMed Phrase extraction.
      Figure 1

      Identifying candidate phrases

      We start by preprocessing the entire PubMed and compiling a comprehensive list of multi-token text segments from the literature. From titles and abstracts of PubMed documents we collect all multi-word text strings that are bounded by punctuation or stopwords, and appear at least 5 times in PubMed. For this process, hyphens and apostrophes are not used as delimiters. As of late 2017, this resulted in 23.4 million unique text strings. This approach limits our space to phrases without prepositions or other stopwords. It has been shown that the majority of meaningful biomedical phrases that appear in both UMLS® and PubMed do not include stopwords1. However, to avoid missing meaningful phrases with stopwords, we also extracted 5.3 million phrases from UMLS and added them as candidates. This allows to identify some good quality phrases such as ‘activin a’.

      Next, we process the multi-word strings to identify substrings that are observed in the PubMed literature as coherent units. Given two words, word1 and word2, the p-value is computed to determine whether they appear together in PubMed by chance. Let Ns be the number of sentences that contain word1, Nt be the number of sentences that contain word2, Nst be the number of sentences that contain the words word1 and word2, and N be the total number of sentences in PubMed. The random variable Y representing the number of sentences containing word1 and word2 is a hypergeometric random variable with parameters Ns, Nt and N14. The probability distribution of Y is shown as follows: P(Y=y)=(Nty)(NNtNsy)(NNs)

      From Nst we compute the p-value, i.e., the probability of the observed Nst or a higher frequency arising by chance as follows: pvalue=y=Nstmin(Ns, Nt)P(y).

      Since our interest is in computing whether word1 and word2 form a phrase, we place a more stringent requirement on the hypergeometric test. We define N’st as the number of sentences that contain the phrase word1 word2 (a subset of Nst) and replace the value of Nst with N’st when we compute the probability P(y) and the corresponding p-value. If the observed frequency of a phrase is above the expected value for a random incident (we set the p-value threshold to 0.01), we conclude that word1 and word2 form a coherent segment and retain this segment for further processing.

      In this work, we also expand the p-value test by replacing word1 with a text segment, i.e., the hypergeometric distribution is used to test the significance of a text segment and a word appearing together. Once a pair of words has been pooled, the algorithm considers the relationship between the segment word1 word2 and the next token word3, by computing the p-value score between word1 word2 and word3. The process continues until we either reach the end of the multi-word text string (we collect these strings) or encounter a token wordi that does not pass the p-value test when tested against the chunk word1wordi-1 (we discard these strings). Figure 2 shows an example of how text is segmented using the hypergeometric test. In the figure, the statistics for terms ‘early’ and ‘lung’ suggests that they appear together by chance, but ‘lung cancer treatment’ appears as a unit in PubMed and it is statistically significant.

      Figure 2: Example of text segmentation using the hypergeometric test.
      Figure 2

      Since we perform multiple hypergeometric tests for each candidate, we adopt a useful statistical correction for multiple comparisons, the Benjamini-Hochberg correction15, which allows one to achieve the desired confidence level for a phrase when multiple comparisons are performed. Using the Benjamini-Hochberg correction, we eliminated 11 phrases that do not pass the set p-value threshold. As a result, we obtained 8,730,788 candidates from 28.7 million multi-word strings in this first step. We do not apply a multiple testing correction to this set but rather accept a 1% error rate in the 8.7 million candidates.

      Traditional approaches of identifying collocations are mutual information, log likelihood or the chi-square test16,​17,​18. In particular, it has been observed that the hypergeometric test and mutual information provide similar performance19. However, for mutual information, one must set a threshold to determine the significance of co-occurrence of two words. The threshold varies from corpus to corpus, and does not offer an intuitive probabilistic interpretation. For the hypergeometric test one must also set a p-value threshold, but the p-value of 0.01 is a widely accepted standard for significance, while there is no such standard for mutual information. Furthermore, the hypergeometric test returns a value that directly indicates the statistical significance of an observed event.

      Evaluating phrases for their effectiveness in document retrieval

      The resulting candidates obtained from the previous step are statistically significant. However, there is no guarantee that treating these text segments as phrases would improve the performance of document retrieval. To address this issue, we apply a simple but effective procedure that we call the BM25 scoring framework.

      The BM25 scoring framework is designed to detect those text segments that benefit the search when interpreted as phrases. This process is optimized towards retrieval performance and guarantees that each selected phrase is a coherent unit and improves the retrieval performance when treating the phrase as a unit. However, a difficulty with this approach is that the evaluation of retrieval results requires a significant effort in human judgements. To address this problem, we employ an automatic approach to efficiently compare search results without manually creating a gold standard set.

      Our method is based on the following hypothesis:

      As a class, the documents that do not contain all query terms in the title, but do contain all in the abstract (set A) qualitatively differ from the records that do contain all the query terms in the title and all in the abstract (set T). The difference between set A and set T reflects the characteristics of these two sets that make records in T more likely than records in A to be useful to an information seeker.

      There are two reasons why this hypothesis is highly plausible. First, we believe that titles are carefully chosen to be good indicators of the contents of scientific articles. Second, studies analyzing user behavior in response to PubMed retrieval20 have found that an article was more likely to be clicked on if the query terms appeared in the title. If we use a likelihood of being clicked as a pseudo-relevance measure, it justifies our hypothesis that documents with query words in a title are likely to be more relevant than those documents that contain query words only in the abstract. Figure 3 illustrates our hypothesis with two documents, each containing the phrase ‘breast cancer treatments’. One has the phrase in the title (set T) and the other has the same phrase in the abstract (set A). Both are related to ‘breast cancer treatments’ however it is clear the one belonging to set T is more relevant to the query ‘breast cancer treatments’. We further discuss the validity of the hypothesis in Technical Validation.

      Figure 3: Comparison of a document including the phrase, ‘breast cancer treatments’ in the title versus a document including the same phrase in the abstract.
      Figure 3

      For more precise evaluation, synonymous words/phrases should be taken into account to judge the relevance of documents. While this complex approach would work better, it may raise the ambiguity problem among synonymous phrases. Thus, we use a simple approach relying only on exact word/phrase matching and believe that it should provide a reasonable metric for detecting useful phrases.

      Each phrase candidate identified by the hypergeometric test is now used as a query. We first create the appropriate set AT for the phrase. We then re-rank these abstracts based on BM25 scores for two scenarios: 1) The traditional way treating the query as represented by its individual words, which scores the abstracts based on the sum of the weights of individual words and 2) Computing BM25 scores by treating a query (or phrase) as a single multi-word term, not using the individual words in the scoring.

      The intuition behind this comparison can be seen from a simple example. Consider the phrase ‘zinc finger’ and consider the individual words, ‘zinc’ and ‘finger’. If ‘zinc’ and ‘finger’ appear in a sentence or a title, it is very likely they would appear as the phrase ‘zinc finger’. If ‘zinc’ and ‘finger’ occur in an abstract, but not as the phrase ‘zinc finger’, it is highly unlikely the phrase ‘zinc finger’ will appear in the abstract or the title. Thus, that document will have been marked as a negative and also will not be retrieved by the phrase retrieval, but will be retrieved by the usual individual word retrieval. It is in this situation that we can see the phrase retrieval outperform the individual word retrieval. This also illustrates why the phrase retrieval can benefit the user. If the user is interested in ‘zinc finger’, they will not likely be interested in documents that mention ‘zinc’ and ‘finger’ separately.

      In our framework, BM25 ranking of candidate documents is performed using only the abstracts of PubMed documents. Titles are used only to assign positive or negative labels. A document is labeled as positive (set T) if its title contains all query tokens, and negative (set A) otherwise. In the BM25 scoring framework, we measure the retrieval performance using average precision21, i.e., the average of precisions across all ranks containing relevant documents.

      We collected PubMed Phrases following the four criteria:

      1. For every candidate phrase, a set of titles containing all individual words in a phrase and a set of abstracts containing the same words should have at least five documents in common.

      2. The search performance should be higher when a query is treated as a phrase as compared to being treated as individual query words.

      3. The search performance of a phrase in terms of average precision should be higher than the baseline performance, which is the score obtained when a set of documents is randomly ranked.

      4. The average precision using the BM25 individual word retrieval must be higher than 0.01. A lower average precision from the search may mean there is insufficient evidence for judging relevance in the set, thus we discard such phrases.

      Filtering the candidate phrases to satisfy the four criteria above resulted in 705,915 PubMed Phrases. If we limit the lower bound of performance improvement to 10% in criterion 2, we obtain the subset of 568,125 phrases that we will refer to as PubMedsmall.

      Data Records

      The PubMed Phrase set is a collection of 705,915 coherent text segments that improve retrieval performance when interpreted as phrases rather than individual words. An interesting property of the PubMed Phrase set is that it is computed using completely data-driven approaches without considering POS tags. As a result, the set includes more data than the traditionally favored well-formed noun phrases. We examined the composition of the set and found that 84.1% of the phrases are noun phrases. The remaining 15.9% of phrases are frequently used coherent segments that are found to improve the retrieval performance despite not being noun phrases such as ‘high resolution 3d’.

      While the statistical filtering step using the hypergeometric test focuses on the likelihood of words appearing as a unit, the empirical filtering using BM25 analyzes how PubMed authors prefer to refer to a concept. For example, the candidates that pass the hypergeometric test but not BM25 include ‘cavoatrial tumor’ and ‘Kentucky farmers’. While these phrases look reasonable, ‘cavoatrial … tumor thrombus’ and ‘farmersin Kentucky’ are also commonly used to express the same concepts.

      Table 1 shows the evaluation results of PubMed Phrases using the BM25 scoring framework. The performance of the selected PubMed Phrase set achieves significantly higher performance in phrase-based search for both PubMedall and PubMedsmall sets.

      Table 1: Number of phrases and mean average precision performance for individual word-based and phrase-based document retrieval.

      The 705,915 multi-word strings of the PubMed Phrase set can be found in ‘all_dictionary.txt’ [Data Citation 1: Figshare] (Future updates of the PubMed Phrase set may be available at There are four additional datasets released with the list of PubMed Phrases. The first set (‘all_dictionary.pos’) includes POS tags of each phrase. Since POS tags used can differ depending on context, we extracted the two most common POS tags appearing in PubMed abstracts and included them in the POS tag file. The second set (‘’) clusters the PubMed Phrase set by assessing whether a phrase contains other phrases in the set. Each line starts with a phrase and is followed by substrings that also appear in the PubMed Phrase set. For example, the phrase, ‘super heavy oil’, includes two phrases ‘heavy oil’ and ‘super heavy’ from the same set. The third set (‘all_dictionary.pmid’) includes all PMIDs containing a phrase, for each phrase. Finally, the fourth set (‘all_dictionary.sco’) is a list of vertical bar separated lines, where each line contains a phrase in the first column, a set of p-value scores for every subphrase considered in the second column, and two average precision scores in the third column. The first average precision score is based on the BM25 ranking treating each word individually, while the second by treating the phrase as a single unit. For example, given the phrase ‘natural language processing’, the second column contains p-value scores for natural/language and natural language/processing. The third column reports the BM25 scores of search treating the phrase as individual words, ‘natural’, ‘languageandprocessing’ versus as a phrase, ‘natural language processing’. The POS tag, group, PubMed ID, and score files provide a wealth of information that could be useful for analyzing and selecting a subset of PubMed Phrases.

      Technical Validation

      Validity of using PubMed titles for relevance judgment

      Our relevance measure for the phrase filtering step assumes that a document with title containing all query terms is likely to be relevant to the user issuing the query (i.e., phrase). Titles in scientific documents are normally chosen to provide a good summary of the article13. Therefore, it is logical to assume that a document is more likely relevant to a user issuing a query if its title contains all query terms. This is supported by the empirical observation that users prefer to click on documents containing query terms in the title22. We further noticed from the PubMed user logs that, given documents scored by BM25, users are four times more likely to click on a document containing query terms in the title than on a document that does not. This conclusion is further bolstered by several studies documenting that user clicks are strongly correlated with relevance23,​24,​25. Note that we are not making any assertion about the documents that contain query terms in the abstract and not in the title, as a number of them may also be relevant. Moreover, it is less obvious how to judge relevance of a document whose title contains some of the query terms but not all. Therefore, we take a more conservative approach and treat those documents as negative for computing average precision. As a consequence, the average precision we compute is only an approximate lower bound for the true average precision, but when this lower bound is quite high we believe that is evidence that supports our argument.

      BM25 is one of the popular ranking functions and known to be a good performer in document search12. In PubMed, BM25 can be applied to both titles and abstracts for search. However, in our experiments, we apply BM25 to abstracts only, and use titles as the indicator for relevance. Here is the logic of our arguments to use PubMed titles for relevance judgments:

      1. BM25 provides good performance in retrieving relevant documents.

      2. If we assign positive labels to those documents that have all query words in the titles and evaluate BM25 retrieval based on that assignment, we find that BM25 produces average precisions that are significantly higher than a random ranking.

      3. The obvious explanation for #2 is that positive labels are being assigned preferentially to the most relevant documents.

      We believe 1-3 above provide significant support for our assumption that query words in the title are a useful indication of relevance. To experimentally validate item 2 here, we chose 27,870 frequent user queries from PubMed26, and performed BM25 ranking for all abstracts with query words in them. This was then compared with a random order of the same PubMed document set for each query. Using the criterion that documents having the query words in the title are counted as relevant, BM25 achieved 0.3490 average precision, whereas the random document ranking achieved 0.1759 average precision. The difference is substantial and it shows that using titles to create a relevance standard is a practically sound strategy.

      User click-based evaluation of PubMed Phrases

      Although there are various factors to affect user clicks, a consensus of user clicks may provide a clue of how useful PubMed Phrases are. Therefore, we performed a user click-based experiment as follows. We first chose the 100 PubMed Phrases that were also the most frequent queries in 2017. Assuming the document that is clicked is relevant to the query, we examined the top 20 retrieved documents from all PubMed users. 20 is the default number of documents displayed in PubMed on the first page. For each query, we compared click rates between two groups of documents: 1) the documents with query terms appearing as the phrase and 2) the documents with query terms not appearing as the phrase. By click rate of a given set of documents, we mean the fraction of the documents that are clicked. We computed the click rate difference for each query, and averaged these differences over the 100 PubMed Phrases under consideration. The results show that, when a query appears as the phrase in a document, the click rate is 96.5% higher than when it does not appear as the phrase, on average. Four of these 100 queries have lower click rates when the phrase is present, however, these four cases have very few documents that have the query terms but not the phrase, making these cases statistically unreliable. This means using the PubMed Phrases as phrases has the potential to benefit PubMed search.

      Manual evaluation of PubMed Phrases

      To assess whether phrases in the PubMed Phrase set satisfied criteria of well-formedness and completeness, we conducted manual evaluation of a randomly selected subset of phrases. The annotation task was performed in two rounds and conducted by two annotators with backgrounds in biomedical informatics research and experience in annotating biomedical corpora.

      During a preliminary round, we observed that human annotators naturally preferred well-formed noun phrases and it was challenging to come up with clear annotation guidelines for annotating non-noun phrases. For that reason, we decided to sample for manual annotation five hundred noun phrases by considering POS tags and choosing phrases ending in singular noun (NN) or plural noun (NNS) POS tags. In the first round, each annotator was given a phrase and three sample sentences containing the phrase and asked to assign one of the three labels: positive, negative and partial. Figure 4 shows the web interface we used for manual evaluation.

      Figure 4: Manual annotation interface for PubMed Phrases.
      Figure 4

      Each annotator assigned yes (positive), no (negative) or partial by reviewing a given phrase and three PubMed sentences including the phrase.

      The annotators were asked to keep the following questions in mind when assigning a label to a phrase:

      • Does a phrase have a clear meaning on its own?

      • Is a phrase obviously missing terms?

      A phrase was labeled positive if it had a specific meaning and was a complete noun phrase. A phrase was labeled negative if it was judged not to be a phrase but a collection of terms. Finally, a phrase was labeled partial if it had a specific meaning but was missing certain terms.

      After the first round of evaluation, the annotators had agreed upon 476 phrases (95.2% inter-annotator agreement), assigning 463 phrases as positive and 13 phrases as partial. Examples with the partial labels are ‘locus yac’ and ‘mushroom macrolepiota procera’. The ‘locus yac’ was judged partial because in the example sentences, it was always used with ‘beta-globin’, as ‘beta-globin locus yac’. The meaning of ‘mushroom macrolepiota procera’ is quite specific, but usually it is referred to as ‘parasol mushroom’ or ‘macrolepiota procera’.

      During the second round, the annotators were asked to review the phrases for which they had not achieved agreement in the first round. After this stage the annotators had reached an agreement for all 500 phrases, of which 480 phrases were labeled positive, 3 phrases were labeled negative and 17 phrases partial. Four additional partial cases, found during the second round of annotations, are ‘agri sp’, ‘iib myosin heavy chain’, ‘rat skeletal l6’ and ‘vivo footprints’. ‘rat skeletal l6’ is typically used in the form ‘rat skeletal l6’ + myotubes/myoblasts/cells, while other phrases are missing certain words, e.g., ‘sphingomonas agri sp nov’ for ‘agri sp’, ‘type iib myosin heavy chain’ for ‘iib myosin heavy chain’ and ‘in vivo footprints’ for ‘vivo footprints’.

      The three examples annotated as negative were ‘coated multiparticulate’, ‘desert spiny’ and ‘immunoreactive activin’. ‘coated multiparticulate’ is an adjective and usually followed by noun(s) such as ‘systems’ but the POS tagger assigned noun for ‘multiparticulate’. The full noun phrase of ‘desert spiny’ is ‘desert spiny lizards’ in general, but ‘desert spiny’ itself is a vague term. The most common phrase in PubMed that includes ‘immunoreactive activin’ is ‘immunoreactive activin a’, but ‘a’ was treated as a stopword in our process and the same term does not appear in UMLS (Note: ‘activin a’ is in our PubMed Phrase set). Moreover, ‘a’ is not the only noun that can follow the ‘immunoreactive activin’ segment. Although ‘immunoreactive activin’ is labeled as negative, it is still useful as a coherent chunk for improved literature search. The random phrases chosen for annotation and their evaluation results are available in the Supplementary Data.

      Usage Notes

      While our main goal is to obtain a large collection of phrases from PubMed that can be beneficial for literature search, the collection of phrases can be used for other text mining and complex text-based tasks. For instance, document summarization methods27,28 frequently depend on mapping free text to a set of phrases and use these phrases as atomic units to generate summaries. Phrase-based statistical machine translation and paraphrase methods29,30 rely on phrase information and bilingual phrase tables. Document clustering tasks31,​32,​33 have been demonstrated to benefit from using phrase information. Using phrases as atomic search units has also been shown to improve the retrieval performance in a question answering task34.

      Since 2017, the PubMed Phrase set has been used for indexing documents in PubMed’s new relevance search ( In addition, we present two use cases to illustrate the utility of the set.

      Phrase Clouds versus Word Clouds for better knowledge visualization

      Word clouds (or tag clouds) are widely used as a visualization tool that conveniently summarizes textual content. Word clouds are constructed to represent term size proportional to frequency, and they are typically based on single terms. With the availability of PubMed Phrases, an attractive alternative is to construct a word cloud based on phrases. Figures 5a and b compare the word clouds generated by single words versus PubMed Phrases. Both tag clouds are drawn from the same set of PubMed documents that represent ‘deafness28. In Fig. 5a, single words are extracted from the set and weighted by the log of their counts. Figure 5b depicts a word cloud that is based on the phrases available in our PubMed Phrase set. The tag clouds are for the top 50 single words and top 50 phrases. We find that phrase clouds provide a more descriptive and user-friendly representation of the document set.

      Figure 5: Tag clouds for the disease set, ‘deafness’.
      Figure 5

      The clouds are based on top 50 single words (a) and top 50 PubMed Phrases (b) appearing in the set. Each term was weighted by the log of frequency.

      Title generation in clusters

      Clustering is a task of grouping a set of documents in such a way that documents with a similar topic belong to the same group. Clustering is usually an unsupervised process which computes clusters that are not self-descriptive, and one should further sort through the documents to understand the theme of a cluster. Presenting visual cues, such as cluster titles, can significantly improve the user perception of clustering results35,36. One way of describing the subject of a cluster is to select a few keywords or a sentence to represent a cluster37,38. Another way to approach the problem is to apply topic modeling techniques, such as Latent Dirichlet Allocation (LDA)39, which provides a list of topic terms for each topic it finds. Here we show an example of how to select descriptive titles from PubMed Phrases based on topic terms generated by LDA.

      In the experiment, we used a subset of PubMed documents that represent ‘cystic fibrosis’, ‘deafness’, ‘autism’ and ‘hypertrophic cardiomyopathy’ from the OMIM disease set28. We ran LDA on this OMIM set using the gensim toolkit40. The number of topics was set to 10, and default parameters were used without further optimization. Our goal is, for each topic, to select representative phrases from the PubMed Phrase set using top 20 topic terms from LDA. To achieve this, we extracted all noun PubMed Phrases that match with the topic terms. We eliminated phrases that were substrings of longer phrases. Table 2 shows the topic terms and corresponding phrases for certain clusters in the ‘cystic fibrosis’, ‘deafness’, ‘autism’ and ‘hypertrophic cardiomyopathy’ data sets. We find suggested phrases to be reasonable and representative of the given topic single terms. In addition, as supported in the literature36, we find them clearer and easier to comprehend compared to single topic terms. Note that ‘two cases’ in the table is an artifact emerging from ‘two’ and ‘cases’ in the LDA topic terms. This is simply a consequence of the fact that LDA includes all terms in its analysis.

      Table 2: Examples of topic terms and their corresponding noun phrases from PubMed Phrases.

      Additional information

      How to cite this article: Kim, S. et al, PubMed Phrases, an open set of coherent phrases for searching biomedical literature. Sci. Data 5:180104 doi: 10.1038/sdata.2018.104 (2018).

      Publishers note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.


      1. 1.

        , , & Identifying well-formed biomedical phrases in MEDLINE text. Journal of Biomedical Informatics 45, 1035–1041 (2012).

      2. 2.

        , , & How to interpret PubMed queries and why it matters. Journal of the American Society for Information Science 60, 264–274 (2009).

      3. 3.

        & Foundations of statistical natural language processing (MIT Press, 1999).

      4. 4.

        , , & Using search-logs to improve query tagging in Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL 2012). 238-242 (2012).

      5. 5.

        , & NLTK: the Natural Language Toolkit. (2008).

      6. 6.

        & Corpus based statistical screening for phrase identification. Journal of the American Medical Informatics Association 7, 499–511 (2000).

      7. 7.

        & Extracting noun phrases from large-scale texts: a hybrid approach and its automatic evaluation in Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL 1994). 234-241 (1994).

      8. 8.

        , , & Extracting noun phrases for all of MEDLINE in Proceedings of the AMIA Symposium. 671-675 (1999).

      9. 9.

        Phrase detection and the associative memory neural network. Architecture 4, 2599–2603 (2003).

      10. 10.

        & Learning noun phrase query segmentation in Proceedings of the International Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL 2007). 819-826 (2007).

      11. 11.

        & Phrase representations for multiword expressions in Proceedings of the 12th Workshop on Multiword Expressions. 67–71 (2016).

      12. 12.

        & The probabilistic relevance framework: BM25 and beyond. Foundations and Trends in Information Retrieval 3, 333–389 (2009).

      13. 13.

        Relative effectiveness of document titles and abstracts for determining relevance of documents. Science 134, 1004–1006 (1961).

      14. 14.

        Introduction to probability theory and statistical inference. 3rd edn, (John Wiley & Sons, 1982).

      15. 15.

        & Controlling the false discovery rate - a practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B (Methodological) 57, 289–300 (1995).

      16. 16.

        A comparative evaluation of collocation extraction techniques in Proceedings of the International Conference on Language Resources and Evaluation (LREC 2002). 1530–1536 (2002).

      17. 17.

        , , , & TermeX: a tool for collocation extraction in Proceedings of the International Conference on Intelligent Text Processing and Computational Linguistics (CICLing 2009). 149–157 (2009).

      18. 18.

        Normalized (pointwise) mutual information in collocation extraction in Proceedings of the Biennial GSCL Conference 2009. 31-40 (2009).

      19. 19.

        , & Meshable: searching PubMed abstracts by utilizing MeSH and MeSH-derived topical terms. Bioinformatics 32, 3044–3046 (2016).

      20. 20.

        , , & Understanding PubMed user search behavior through log analysis. Database 2009, bap018 (2009).

      21. 21.

        & Modern information retrieval (Addison-Wesley, 1999).

      22. 22.

        & Click-words: learning to predict document keywords from a user perspective. Bioinformatics 26, 2767–2775 (2010).

      23. 23.

        Evaluating retrieval performance using clickthrough data in Proceedings of the SIGIR Workshop on Mathematical/Formal Methods in Information Retrieval. (2002).

      24. 24.

        , , , & Generating labels from clicks in Proceedings of the ACM International Conference on Web Search and Data Mining (WSDM 2009) 172–181 (2009).

      25. 25.

        , , , & Improving quality of training data for learning to rank using click-through data in Proceedings of the ACM International Conference on Web Search and Data Mining (WSDM 2010) 171–180 (2010).

      26. 26.

        , , & Bridging the gap: incorporating a semantic similarity measure for effectively mapping PubMed queries to documents. Journal of Biomedical Informatics 75, 122–127 (2017).

      27. 27.

        , , & Product review summarization by exploiting phrase properties in Proceedings of the 26th International Conference on Computational Linguistics (COLING 2016) 1113–1124 (2016).

      28. 28.

        , & Summarizing topical contents from PubMed documents using a thematic analysis in Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2015) 805–810 (2015).

      29. 29.

        , & Statistical phrase-based translation in Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology (NAACL 2003) 48–54 (2003).

      30. 30.

        & Paraphrasing with bilingual parallel corpora in Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL 2005) 597–604 (2005).

      31. 31.

        , & CorePhrase: keyphrase extraction for document clustering in Proceedings of the International Conference on Machine Learning and Data Mining in Pattern Recognition 265–274 (2005).

      32. 32.

        , & Text clustering based on key phrases in Proceedings of the International Conference on Information Science and Engineering. 986-989 (2009).

      33. 33.

        , , & Retro: concept-based clustering of biomedical topical sets. Bioinformatics 30, 3240–3248 (2014).

      34. 34.

        , & Exact phrases in information retrieval for question answering in Proceedings of the COLING Workshop on Information Retrieval for Question Answering 9–16 (2008).

      35. 35.

        et al. Evaluating visual representations for topic understanding and their effects on manually generated topic labels. Transactions of the Association for Computational Linguistics 5, 1–16 (2017).

      36. 36.

        & Summarizing topics: from word lists to phrases in NIPS Workshop on Modern Machine Learning and Natural Language Processing. (2014).

      37. 37.

        & Automatic keyphrase extraction: a survey of the state of the art in Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL 2014) 1262–1273 (2014).

      38. 38.

        & Recent automatic text summarization techniques: a survey. Artificial Intelligence Review 47, 1–66 (2017).

      39. 39.

        , & Latent Dirichlet Allocation. The Journal of Machine Learning Research 3, 993–1022 (2003).

      40. 40.

        & Software framework for topic modelling with large corpora in Proceedings of the LREC Workshop on New Challenges for NLP Frameworks 46–50 (2010).

      Download references

      Data Citations

      1. 1.

        Kim, S., Yeganova, L., Comeau, D. C., Wilbur, W. J., & Lu, Z Figshare (2018)


      This research was supported by the Intramural Research Program of the NIH, National Library of Medicine.

      Author information


      1. National Center for Biotechnology Information, National Library of Medicine, National Institutes of Health, Bethesda, Maryland 20894, USA

        • Sun Kim
        • , Lana Yeganova
        • , Donald C. Comeau
        • , W. John Wilbur
        •  & Zhiyong Lu


      1. Search for Sun Kim in:

      2. Search for Lana Yeganova in:

      3. Search for Donald C. Comeau in:

      4. Search for W. John Wilbur in:

      5. Search for Zhiyong Lu in:


      S.K. and W.J.W. initially proposed the idea. S.K., L.Y., D.C.C., W.J.W. and Z.L collected and analyzed the data. S.K. and L.Y. wrote the manuscript. W.J.W. and Z.L. supervised and revised the manuscript.

      Competing interests

      The authors declare no competing interests.

      Corresponding author

      Correspondence to Zhiyong Lu.

      Supplementary information

      About this article

      Publication history





      Rights and permissions

      Creative Commons BYOpen Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit The Creative Commons Public Domain Dedication waiver applies to the metadata files made available in this article.