Abstract
Adoption of Electronic Health Record (EHR) systems has led to collection of massive healthcare data, which creates oppor tunities and challenges to study them. Computational phenotyping offers a promising way to convert the sparse and complex data into meaningful concepts that are interpretable to healthcare givers to make use of them. We propose a novel su pervised nonnegative tensor factorization methodology that derives discriminative and distinct phenotypes. We represented cooccurrence of diagnoses and prescriptions in EHRs as a thirdorder tensor, and decomposed it using the CP algorithm. We evaluated discriminative power of our models with an Intensive Care Unit database (MIMICIII) and demonstrated superior performance than stateoftheart ICU mortality calculators (e.g., APACHE II, SAPS II). Example of the resulted phenotypes are sepsis with acute kidney injury, cardiac surgery, anemia, respiratory failure, heart failure, cardiac arrest, metastatic cancer (requiring ICU), endstage dementia (requiring ICU and transitioned to comfortcare), intraabdominal conditions, and alcohol abuse/withdrawal.
Introduction
A phenotype is an outward physical manifestation of a genotype. Investigating the association between phenotypes and genotypes has been a principal genetic research goal^{1}. Electronic health records (EHRs) are increasingly used to identify phenotypes because EHRs encompass several aspects of patient information such as diagnoses, medication, laboratory results, and narrative reports. Given the importance of these efforts, collaborative groups have been created to develop and share phenotypes obtained from EHRs, such as the Electronic Medical Records and Genomics (eMERGE) Network^{2} and the Observational Medical Outcomes Partnership^{3}. Two of the main obstacles to generate phenotypes are the needs for substantial time and domain expert knowledge^{4, 5}. Furthermore, phenotypes created using clinical judgement^{6, 7} or healthcare guidelines^{5, 8} in one institution often cannot be easily ported to the other institutions, reducing generalizability and leading to unstandardized phenotype definitions^{9}.
Consequently, phenotyping based on machine learning has been proposed to facilitate extraction of meaningful phenotypes automatically from EHRs without human supervision through a process called computational phenotyping. The most widely used approach is unsupervised feature extraction that derives meaningful and interpretable characteristics without supervision on data label. Frequent pattern mining defines phenotypes as a pattern that is frequently observed set of ordered items from sequential numerical data such as laboratory^{10, 11}. A natural language processing technique extracts frequent terms from clinical narrative notes and defines phenotypes as a set of relevant and frequent terms^{12,13,14}. These frequent set mining methods are useful but unable to learn underlying latent characteristics. Deep learning methods such as autoencoders or skipgrams represent patient as a vector^{15,16,17}, but it is hard to derive understandable latent concepts due to the nonlinear combinations of multiple layers.
Recently, dimensionality reduction phenotyping methods have been introduced to handle sparse and noisy data from EHRs’ large and heterogeneous features. These methods represent phenotypes as latent medical concepts^{18}. That is, the phenotypes are defined as a probabilistic membership to medical components, and patients also have a probabilistic membership to the phenotypes. For example, Bayesian finite mixture modeling discovers Parkinson’s disease phenotypes as latent subgroups^{19}. Another dimensionality reduction technique, matrix factorization, decomposes timeseries matrix data from EHRs into latent medical concepts^{20,21,22}. Most recently, nonnegative tensor factorization (NTF) is becoming particularly popular due to its ability to capture high dimensional data. It generates latent medical concepts using interaction between components from multiple information source^{23,24,25,26,27}. Ho et al. first introduce NTF for phenotyping^{23, 24}. They define phenotypes as sets of cooccurring diagnoses and prescrptions, and obtain the phenotypes from latent representation of the cooccurrence. They use KullbackLeibler divergence to decompose the observed cooccurrences that follow Poisson distribution based on CP decomposition. Ho et al. also incorporate sparsity constraints by setting thresholds for negligibly small values. Wang et al. enforce orthogonality constraints on NTF to derive less overlapping phenotypes^{25}. Another NTF based on Tucker decomposition discovers (highorder) feature subgroups as decomposing the tensor into a core tensor multiplied by orthogonal factor matrices for each mode. It uses the core tensor to encode interactions among elements in each mode^{26, 28}.
One of important characteristics that phenotypes should have is to be discriminative to a certain clinical outcome of interest such as mortality, readmission, cost, et al. So far, however, there has been little consideration about discriminative phenotypes associated with certain clinical outcomes. The discriminative phenotypes can be beneficial to clinicians because they can directly apply the phenotypes to their daily practice to improve the clinical outcome of interest. For example, clinicians can use our phenotype to evaluate patients’ risk of hospital death like APACHE II or SAPS score does, and improve resource allocation and qualityofcare in ICUs. Membership to the several different phenotypes can provide an insight on the situation of a patient beyond a single score. In addition, another crucial characteristic for phenotypes is to be distinct from each other, because otherwise clinicians cannot interpret and use the phenotypes easily. For example, let us say a patient suffers from hypertension and diabetes. To represent the patient, we can use a mixture of two phenotypes. We prefer Phenotype 1 = {hypertension, ACE inhibitors}, Phenotype 2 = {diabetes, insulin} to Phenotype 1 = {hypertension, ACE inhibitors, insulin}, Phenotype 2 = {diabetes}, because the former is more distinct and meaningful than the latter. Yet another critical concern about phenotypes is the compactness. Generally speaking, compact representation is more preferable than the lengthy one to end users if both have the same discrimination power and distinctness.
This paper proposes a new tensor factorization methodology for generating discriminative and distinct phenotypes. We defined phenotypes as the sets of cooccurring diagnoses and prescriptions. We used a tensor to represent diagnosis and prescription information from EHRs, and decomposed the tensor into latent medical concepts (i.e., phenotypes). To discriminate a highrisk group (high mortality), we incorporated the estimated probability of mortality from logistic regression during the decomposition process. We also found cluster structures of diagnoses and prescriptions using contextual similarity between the components, and absorbed the cluster structure into the tensor decomposition process.
Methods
We first describe a computational phenotyping method that we developed (Fig. 1) and experiment design.
Phenotyping based on tensor factorization
We built a thirdorder tensor \({\mathscr{O}}\) with cooccurrences of patients, diagnoses, and prescriptions from intensive care unit (ICU) EHRs. Detailed tensor construction can be found in Supplementary methods. The cooccurrence is a natural representation of interactions between many diagnoses and prescriptions. We only focused on diagnosis and prescription data as previous phenotyping definition^{29,30,31}, but we can extend the tensor to a high order (>3) to utilize additional data such as laboratory results and procedures. Specifically, we first built a matrix for individual patient to represent association between prescription and diagnosis. For example, let us say patient 1 is diagnosed with acute respiratory failure and hypertension, and is ordered the medicine phenylephrine during his or her admission. Then, each cooccurrence of acute respiratory failure and phenylephrine, and hypertension and phenylephrine is one, respectively (Fig. 2). Again, let us say patient I is diagnosed with Alzheimer’s disease and is ordered medicine morphine sulfate twice. Then, the cooccurrence of Alzheimer’s disease and morphine sulfate is 2. We collected all the matrices from all the patients and built the thirdorder observed tensor \({\mathscr{O}}\). Entries at (i, j, k) of the tensor (i.e., \({\mathscr{O}}\) _{ i,j,k }) is the number of cooccurrence of diagnosis j and prescription k for patient i.
To decompose the tensor, we used CP algorithm^{32, 33}; detailed description of CP can be found in Supplementary methods. Recently, phenotyping based on Tucker model has been proposed^{26, 28}. It provides a more flexible modeling than does CP by allowing subgroups in each mode, but CP has an advantage in that it is computationally cheap and extendable by imposing regularizers. Using CP model, the thirdorder tensor \({\mathscr{O}}\) was decomposed into three factor matrices: A for patient mode, B for diagnosis mode, and C for prescription mode (Fig. 3). A phenotype consisted of diagnoses and prescriptions, and patients were involved in each phenotype. That is, the r th phenotype consisted of J diagnoses and K prescriptions with membership values that describe how much the diagnoses and prescriptions are involved and contribute to the r th phenotype. The membership values were normalized values between 0 and 1, and stored in the normalized vectors \({\overline{{\bf{B}}}}_{:r}\) and \({\overline{{\bf{C}}}}_{:r}\), respectively. Meanwhile, patients were involved in the R phenotypes with membership values that represent how much the patient has the characteristic of the phenotypes. The membership values of patients were also normalized values between 0 and 1, and stored in the normalized vector \({\overline{{\bf{A}}}}_{:r}\). Ability of r th phenotype that can capture and describe the data was stored in \({\lambda }_{r}={{\bf{A}}}_{:r}{}_{F}{{\bf{B}}}_{:r}{}_{F}{{\bf{C}}}_{:r}{}_{F}\), because large values in A _{:r }, B _{:r }, and C _{:r } means that the r th phenotype describes large portion of cooccurrence values in \({\mathscr{O}}\). So, conversely, a phenotype with highly cooccurring diagnosis and prescription may have large λ _{ r }.
For example, ICU survived patients (half of total patients) have Phenotype 1 in Fig. 3, which consists of the first two elements of diagnosis mode and the first one element of prescription mode. The second diagnosis element has higher membership to the Phenotype 1 than the first element does. The patients who died in ICU have Phenotype 2, which consists of the third diagnosis and the second prescription. Similarly, the deceased patients and a few patients who survived have Phenotype R, which consists of the fourth diagnosis and the third prescription. Note that in this example, elements in a phenotype are not overlapped with elements in other phenotypes; thus, we can interpret the phenotype easily. Also, note that phenotypes for the deceased patients and the patients who survived are separated so that we can easily determine which phenotypes are more associated with mortality; consequently, we can further use the phenotypes to evaluate the risk of patients according to the membership to the phenotypes. We introduced two regularizations to make the phenotype discriminative and distinct in the following sections.
Supervised phenotyping for discriminative power
We proposed a supervised approach to encourage the phenotypes separated according to mortality by adding a logistic regression regularization. In the previous section, patients had the membership values to the phenotypes. We used the membership as a feature vector to express patients, and used the feature vector to predict mortality. As a previous work on graphbased phenotyping method^{21}, we added a regularization for supervised term. Let us say y _{ i } is a binary indicator of mortality, i.e., y _{ i } = 1 if i th patient dies during hospital admission and y _{ i } = −1 otherwise. The i th patient in training set L (i ∈ L) was represented as the membership values to the phenotypes, A _{ i:}, which is the i th row vector of A. Given logistic regression parameters θ, a probability of i th patient’s mortality to be y _{ i } is
where δ _{ i } = [A _{ i:}, 1] · θ. We then maximized the logprobability, or minimize the negative logprobability,
Thus, the objective function for updating each row A _{ i:} is
with a weighting constant ω (\(\odot \) refers to KhatriRao product). Note that this objective function is with respect to row A _{ i:} not the whole patient factor matrix A. Gradient of f(A _{ i:}) is
and hessian of f(A _{ i:}) is
Using Newton’s gradient descent method, if i ∈ L, we update A _{ i:} as
If i ∉ L, we update A _{ i:} as Eq. (6) with ω = 0, which is a traditional CP decomposition without any regularization. Time complexity of Eq. (6) is bounded by O(JKR ^{2}) for i ∈ L; total time complexity to update A is bounded by O(IJKR ^{2}) (Table S1). The supervised term had negligible effects on the total time complexity. This updating rule can be linearly scaled up to the size of A. Updating the logistic regression parameters θ followed a typical logistic regression modeling method. We added a ridge penalty to shrink the size of θ and avoid overfitting (c is a weighting constant)^{34} as
Similaritybased phenotyping for distinct power
To derive distinct phenotypes with less overlapping with each other, we made phenotypes only consist of similar elements. We first derived components’ similarities from contexts in EHRs, used the similarities to infer cluster structures, and let phenotypes reflect the cluster structures.
Deriving contextual similarity
We derived contextual similarities from EHRs. Farhan et al. generate a vector representation of medical events (or elements in phenotype)^{17}. Based on this work, we generated sequences that consist of diagnoses and prescriptions from EHRs in time order (Table 1). We applied Word2Vec, a twolayer neural network for natural language processing for numerical representation of discrete words^{35}. We input the timeordered EHRs sequences into Word2Vec and derived a set of vectors for each diagnosis or prescription. After several trials, we set cardinality of the vector as 500 and window size of the sequence (i.e., the number of diagnoses or prescriptions in a sequence to consider them contextually similar) as 30. We found that, as the cardinality increases, distribution of the pairwise similarities spreads widely (i.e., many similarity values are close to −1 or 1 other than 0), but computation time also increases rapidly. We also observed that most of the pairwise similarities become close to 0 as the window size decreases, and close to 1 as the window size increases.
We then computed cosine similarities between the vector representation of elements, and derived a pairwise similarity matrix (either J × J matrix S ^{B} for diagnosis or K × K matrix S ^{C} for prescription). For example, let us say the j _{1} th and j _{2} th diagnoses in our dataset refer to atrial fibrillation and congestive heart failure, respectively. The vector representation is atrial fibrillation = (0.1, 0.6, 0.2, 0.1) and congestive heart failure = (0.3, 0.7, 0.1, 0.2). The similarity between them is stored at (j _{1}, j _{2})entry of S ^{B}, and the value is \({{\bf{S}}}_{{j}_{1},{j}_{2}}^{B}=\frac{0.1\times 0.3+0.6\times 0.7+0.2\times 0.1+0.1\times 0.2}{\sqrt{0.42}\sqrt{0.63}}\approx 0.95\).
We made S sparse for efficiency by ignoring trivial values. Many similarities were close to zero, and their small variance did not provide useful information. Similarities less than zero refer to dissimilarity, which was not the focus of this work. Considering all the less useful similarity values can increase computational overhead. We only used the highest l similarities value for each element, and consider the others as 0. We choose \(l=\lfloor {\mathrm{log}}_{2}J\rfloor \) for diagnosis and \(\lfloor {{\rm{l}}{\rm{o}}{\rm{g}}}_{2}K\rfloor (l > 0)\) for prescription according to previous works^{36, 37}.
We converted S into a normalizedcut similarity matrix^{38}. Incorporating the normalized cut similarity helped our problem to increase both the total dissimilarity between the different phenotypes and the total similarity within the phenotypes, thus avoid overlapping between the phenotypes. Converting to the normalized cut similarity matrix is
where D is a diagonal matrix of \({\bf{D}}=diag({d}_{1},\ldots ,{d}_{J}),{d}_{j}={\sum }_{l=1}^{J}{{\bf{S}}}_{jl}\).
Incorporating cluster structure
With the similarity matrix, we inferred a cluster structure from the similarity and incorporated it to our NTF optimization. The cluster structure contained information on which elements should be in the same phenotype together. We introduced a regularization term for the spectral clustering. We increased the sum of pairwise similarity within a phenotype. Because how much the elements are involved in each phenotype is different, the pairwise similarity was weighted by the two elements’ membership values to the phenotype. That is, in terms of diagnosis similarity matrix S ^{B}, the sum of weighted pairwise similarity within a phenotype r is
and the sum of all the similarity in Eq. (9) throughout the R phenotypes is
Here, Tr(B ^{T} S ^{B} B) is the objective of spectral clustering in which B represent the clustering assignment of each element^{37}. Consequently, the phenotypes preserved the spectral clustering structure by incorporating sum of weighted similarity. Meanwhile, Tr(B ^{T} S ^{B} B) is also equivalent to symmetric nonnegative matrix factorization of similarity matrix S ^{B} ^{36, 39}, i.e.,
because
by relaxing a constraint on B ^{T} B = I ^{39}. This transformation is beneficial because it helps phenotypes to be more orthogonal (or distinct) by retaining B ^{T} orthogonality approximately^{39}. Thus, the objective function with the cluster structure is
with a weighting constant μ. By incorporating \({{\bf{S}}}^{{\bf{B}}}{\bf{B}}{{\bf{B}}}^{T}{}^{2}\), our phenotyping method can absorb the spectral clustering structure and improve the orthogonality at the same time. Although it is a fourthorder nonconvex function and it is difficult to find a global optimum, it can converge to a stationary point^{36}. To find an optimum value, we derived the gradient of g(B):
and hessian of g(B):
where a vec(B) of length JR is a vectorization of B by column i.e., \(vec({\bf{B}})=[{{\bf{B}}}_{\mathrm{:1}}^{T},\ldots ,{{\bf{B}}}_{:R}^{T}{]}^{T}\), and \(\otimes \) refers to Kronecker product. Using Newton’s gradient descent method, we updated B as
Time complexity of Eq. (16) is bounded by O(IJKR) + O(J ^{3} R ^{3}). The similarity term had negligible effects on the total time complexity (Table S2). The updating rule for B contained matrix inversion of \({\nabla }^{2}g(vec({\bf{B}}))\in {{\mathbb{R}}}^{JR\times JR}\), which may not be scaled up well with large J. In this case, we can use a constant learning rate instead of \({\nabla }^{2}g{(vec({\bf{B}}))}^{1}\) although sacrificing converging rate.
Similarly, the factor matrix C for prescriptions followed the same update procedure. We repeated the updating procedures for the factor matrices A, B and C and logistic regression parameter θ until convergence. We assumed convergence when \(fi{t}_{old}fit < 5\times {10}^{4}\) where fit is defined as \(fit=1\frac{{\mathscr{O}}{\mathscr{X}}}{{\mathscr{O}}}\), and fit _{ old } is the fit of the previous iteration. After normalizing, we removed trivial values less than threshold ε because those values are too small for meaningful membership value and worsen the conciseness. We summarized the entire updating procedures in Algorithm 1.
Input: \({\mathscr{O}},\omega ,\mu \) 
1: Randomly initialize A, B, C. 
2: repeat 
3: \({{\bf{A}}}_{i:}=\,\max (\mathrm{0,}\,{{\bf{A}}}_{i:}{\nabla }^{2}f{({{\bf{A}}}_{i:})}^{1}\nabla f({{\bf{A}}}_{i:}))\) for all i. 
4: Update θ for logistic regression 
5: \(vec({\bf{B}})=\,{\rm{\max }}(\mathrm{0,}\,vec({\bf{B}}){\nabla }^{2}g{(vec({\bf{B}}))}^{1}\nabla g(vec({\bf{B}})))\). 
6: \(vec({\bf{C}})=\,{\rm{\max }}(\mathrm{0,}\,vec({\bf{C}}){\nabla }^{2}g{(vec({\bf{C}}))}^{1}\nabla g(vec({\bf{C}})))\). 
7: until Converged 
8: \({\overline{{\bf{A}}}}_{:r}\leftarrow \frac{{{\bf{A}}}_{:r}}{{{\bf{A}}}_{:r}}\), \({\overline{{\bf{B}}}}_{:r}\leftarrow \frac{{{\bf{B}}}_{:r}}{{{\bf{B}}}_{:r}}\), \({\overline{{\bf{C}}}}_{:r}\leftarrow \frac{{{\bf{C}}}_{:r}}{{{\bf{C}}}_{:r}}\), \(\forall r\) 
9: \({\overline{{\bf{A}}}}_{ir}\leftarrow 0\,{\rm{if}}\,{\overline{{\bf{A}}}}_{ir} < {10}^{6}\), \({\overline{{\bf{B}}}}_{jr}\leftarrow 0\,{\rm{if}}\,{\overline{{\bf{B}}}}_{jr} < {10}^{3}\), \({\overline{{\bf{C}}}}_{kr}\leftarrow 0\,{\rm{if}}\,{\overline{{\bf{C}}}}_{kr} < {10}^{3}\,\forall i,j,k,r\) 
10: return \({\mathscr{X}}={\sum }_{r=1}^{R}{\lambda }_{r}{\overline{{\bf{A}}}}_{:r}{\overline{{\bf{B}}}}_{:r}{\overline{{\bf{C}}}}_{:r}\). 
Experiment design
Dataset and preprocessing
We used a large publicly available dataset MIMICIII (Medical Information Mart for Intensive Care III)^{40}. MIMICIII contains comprehensive deidentified data on around 46,520 patients in critical care units of the Beth Israel Deaconess Medical Center between 2001 and 2012, and it includes information such as demographics, prescription, diagnosis ICD codes, and clinical outcomes such as mortality. We selected 10,028 patients, including all 5,014 patients who died during admission and a random sample of 5,014 of patients who survived. If a patient who survived had multiple admission histories, we used the first admission. We used 202 diagnosis ICD9 codes that are appeared in the charts of at least 5% of the patients and 316 prescription codes that appeared in at least 10% of the patients. We excluded diagnosis ICD9 ‘V’ or ‘E’ codes that describe supplementary factors for health status. We excluded trivial base type prescriptions such as 0.9% sodium chloride, 5% dextrose, and sterile water. Most nonzero cooccurrence values are one, and skewed right (Fig. S1). To prevent smalldosage frequent medicines from having high cooccurrences, we truncated the cooccurrence values to 1% percentile, 10 (Fig. S1).
Evaluation measures
We evaluated our proposed method in terms of discrimination and distinction. We measured the discrimination by the area under the receiver operating characteristic curve (AUC), sensitivity, and specificity. We measured distinction by a relative length of phenotype and an average overlap. An absolute length of r th phenotype is the number of nonzero in membership vector B _{:r } and C _{:r }. The relative length of the phenotype is the absolute length divided by the maximum length J + K. We averaged the R relative lengths of phenotype. The average overlap^{41} measures the degree of overlapping between all phenotype pairs. It is defined as the average of cosine similarities between phenotype pairs:
Setting R = 50, we repeatedly ran our models ten times for 10fold cross validation. We used the training set to compute the regression parameter θ and the likelihood term in supervised phenotyping, and used the test set to measure the discrimination (Table S3). Because tensor factorization is not deterministic method, the factorized tensors are different in each trial; so, we computed mean and 95% confidence interval.
Baselines
We compared the discrimination and the distinction of our proposed methods with that of several baseline methods. The baselines are:

APACHE II, SAPS II, OASIS, APS III score: Disease severity scores for predicting mortality in intensive care unit (for comparing discrimination only)^{42,43,44,45}. These scores assess the severity of disease using variables from preexisting conditions, physiological measurements, biochemical/hematological indices, and source of admission. The weighted sum of individual values produces the severity scores^{46}.

CP: Basic NTF model^{47, 48}.

Rubik: A stateoftheart computational phenotyping method based on CP. Rubik generates a phenotype candidate using count of diagnoses and treatments. It incorporates the orthogonality between phenotypes to derive concise phenotypes^{41}. We assume no existing knowledge term and bias term.
Our proposed methods are:

The supervised phenotyping that incorporates the prediction term for discriminative phenotypes (ω ≠ 0, μ = 0).

The similaritybased phenotyping that incorporates the cluster structure term for distinct phenotypes (ω = 0, μ ≠ 0).

The final model that incorporates the both supervised and similaritybased approach (ω ≠ 0, μ ≠ 0).

When evaluating discrimination (AUC, sensitivity, specificity) of NTFbased models, we used the patient’s membership values (i.e., \({\overline{{\bf{A}}}}_{i:}\) of size 1 × R) as features to fit a binary logistic regression to predict mortality. Particularly, for the supervised model, we fitted a binary logistic regression (after normalization) other than θ that are used during updating procedures. To examine the performance of the supervised and similaritybased phenotyping respectively, we compared the discrimination of CP and the supervised phenotyping (regardless of similarity term), and also compared the distinction of Rubik and similaritybased phenotyping (regardless of supervised term). We then combined the supervised approach and similaritybased approach together to achieve both discrimination and distinction. The weighting constants ω and μ were selected as ω = 1 and μ = 1000 after several trials. Note that ω was comparably small because it sensitively applied to each row of A whereas μ applies to the l _{2} norm of the whole matrix B or C. We used a tensor Matlab Tensor Toolbox Version 2.5^{49} from Sandia National Laboratories to represent tensors and compute tensor operations.
Results
We present the experimental evaluation and phenotypes derived from our method.
Discriminative and distinction power comparison
We found that our methods outperformed other baselines in terms of discrimination and distinction. The supervised phenotyping method showed the highest AUC and sensitivity among the other methods including APACHE II and SAPS II (Table 2). The similaritybased phenotyping method showed the lowest relative length and average overlap among the other methods. Particularly when compared with Rubik^{25} that considers orthogonality for the distinction, the similaritybased method improved the distinction significantly (the relative length of 0.3934 vs 0.0714).
Phenotypes
We presented the phenotypes that are derived from the similaritybased phenotyping method for maximum conciseness. After the tensor decomposition procedures with R = 50, we selected 25 phenotypes by forward feature selection^{50} to remove phenotypes that are redundant and not statistically significant for predicting mortality (Table 3). Among them, we reported ten representative phenotypes in which coefficients from the feature selection were large enough (absolute value of coefficient >20) to discriminate mortality (Table 4): sepsis with acute kidney injury, cardiac surgery, anemia, respiratory failure, heart failure, cardiac arrest, metastatic cancer (requiring ICU), endstage dementia (requiring ICU – sepsis, aspiration, trauma – and transitioned to comfort care), intraabdominal conditions, and alcohol abuse/withdrawal.
We categorized the phenotypes into four groups according to frequency (common or rare) and risk (high or low). Common phenotypes were the top five with high λ and prevalence (and rare otherwise). Highrisk (lowrisk) phenotypes were ones with positive (negative) logistic regression coefficients (Table 3). As a result, common and highrisk phenotypes are sepsis with acute kidney injury, respiratory failure, and heart failure; rare but highrisk phenotypes are cardiac arrest, metastatic cancer requiring ICU, and endstage dementia requiring ICU; common but lowrisk phenotypes are anemia and cardiac surgery; and rare and lowrisk phenotypes are intraabdominal conditions and alcohol abuse/withdrawal (Fig. 4).
To examine the risk of each phenotype in detail, we computed mortality of patients who were highly involved to each phenotype (Table 5). We observed that the mortality of patients who have high membership to phenotypes that are denoted as highrisk in Fig. 4 tends to increase to 1.
Discussion
The objective of this study was to develop a phenotyping method that can generate discriminative and distinct phenotypes. As a result, we derived phenotypes that consist of interactions between related diagnoses and prescriptions, and patients had membership to each phenotype. The phenotypes from the supervised model were more discriminative than APACHE II, SAPS II scores and the phenotypes from CP model^{32, 33}; the phenotypes from the similaritybased model were more distinct than the phenotypes from Rubik^{25}. We also observed that the supervised phenotyping and the similaritybased phenotyping have an opposite effect on each other in terms of the discrimination and distinction. The distinct phenotypes from the similaritybased approach lost its discriminative power, and the discriminative phenotypes from the supervised approach lost distinction power. A possible explanation for this tradeoff is that the similaritybased model tends to ignore less relevant elements in a phenotype to achieve the best distinction, although the “less relevant elements“ can contribute to increasing the discriminative power overall. However, the combined phenotypes from both approaches achieved the high discrimination and distinction at the same time (Table 2). When combining the supervised and the similaritybased phenotyping, the discrimination increased (with the AUC of 0.8389) compared to the similarity model (with the AUC of 0.7796), and distinction improved (with the relative length of 0.3958 and average overlap of 0.1267) compared to the supervised model (with the relative length of 0.6828 and average overlap of 0.3787).
We also described the most representative phenotypes: sepsis with acute kidney injury, cardiac surgery, anemia, respiratory failure, heart failure, cardiac arrest, metastatic cancer (requiring ICU), endstage dementia (requiring ICU and transitioned to comfort care), intraabdominal conditions, and alcohol abuse/withdrawal. These conditions are fairly consistent with the list of conditions known to require ICU care in US hospitals^{51}.
Our study also had some limitations. One limitation is that our approach used the entire ICU stay to generate our predictive models. Other predictive models, such as SAPS II, use only the first 24 hours of data as prediction at that point of the hospitalization is more clinically useful. However, our objective was to demonstrate how our approach could be used with a clinically significant outcome. Future work could create additional phenotypes using only the first 24 hours of data to generate models. A second limitation is that some of the phenotypes generated are not obvious to clinicians. For example, the main medications in the “anemia” phenotype are diabetic medications. This is likely because nonpharmacologic therapy is the main treatment for anemia and diabetic patients were highly represented in the “anemia” population.
With refinement, future applications of our proposed computational phenotyping method include clinical decision support to quickly identify subgroups of patients at different levels of important clinical outcomes (e.g., mortality, clinical decompensation, hospital readmission, etc.). It could also be used in cohort identification for quality improvement or research projects to find those who share similar characteristics by representing patients’ heterogeneous medical records into membership of phenotypes. In addition, the phenotypes we derived can provide genomic scientists an insight into genotypephenotype mapping for precision medicine^{52, 53}. In conclusion, computational phenotyping using nonnegative tensor factorization shows promise as an objective method for identification of important cohorts with promise for clinical, quality improvement and research purposes.
References
Freimer, N. & Sabatti, C. The human phenome project. Nature genetics 34, 15–21, doi:10.1038/ng050315 (2003).
McCarty, C. A. et al. The emerge network: a consortium of biorepositories linked to electronic medical records data for conducting genomic studies. BMC medical genomics 4, 1, doi:10.1186/17558794413 (2011).
Overhage, J. M., Ryan, P. B., Reich, C. G., Hartzema, A. G. & Stang, P. E. Validation of a common data model for active safety surveillance research. Journal of the American Medical Informatics Association 19, 54–60, doi:10.1136/amiajnl2011000376 (2012).
Hripcsak, G. & Albers, D. J. Nextgeneration phenotyping of electronic health records. Journal of the American Medical Informatics Association 20, 117–121, doi:10.1136/amiajnl2012001145 (2013).
Kho, A. N. et al. Use of diverse electronic medical record systems to identify genetic risk for type 2 diabetes within a genomewide association study. Journal of the American Medical Informatics Association 19, 212–218, doi:10.1136/amiajnl2011000439 (2012).
Nguyen, A. N. et al. Symbolic rulebased classification of lung cancer stages from freetext pathology reports. Journal of the American Medical Informatics Association 17, 440–445, doi:10.1136/jamia.2010.003707 (2010).
Schmiedeskamp, M., Harpe, S., Polk, R., Oinonen, M. & Pakyz, A. Use of international classification of diseases, ninth revision clinical modification codes and medication use data to identify nosocomial clostridium difficile infection. Infection Control & Hospital Epidemiology 30, 1070–1076, doi:10.1086/606164 (2009).
Klompas, M. et al. Automated identification of acute hepatitis b using electronic medical record data to facilitate public health surveillance. PLOS one 3, e2626, doi:10.1371/journal.pone.0002626 (2008).
Pathak, J. et al. Mapping clinical phenotype data elements to standardized metadata repositories and controlled terminologies: the emerge network experience. Journal of the American Medical Informatics Association 18, 376–386, doi:10.1136/amiajnl2010000061 (2011).
Kim, Y. et al. Discovery of prostate specific antigen pattern to predict castration resistant prostate cancer of androgen deprivation therapy. BMC Medical Informatics and Decision Making 63, doi:10.1186/s1291101602970 (2016).
Moskovitch, R. & Shahar, Y. Medical temporalknowledge discovery via temporal abstraction. In AMIA (2009).
Yu, S. et al. Toward highthroughput phenotyping: unbiased automated feature extraction and selection from knowledge sources. Journal of the American Medical Informatics Association 22, 993–1000, doi:10.1093/jamia/ocv034 (2015).
Savova, G. K. et al. Mayo clinical text analysis and knowledge extraction system (ctakes): architecture, component evaluation and applications. Journal of the American Medical Informatics Association 17, 507–513, doi:10.1136/jamia.2009.001560 (2010).
Friedman, C., Shagina, L., Lussier, Y. & Hripcsak, G. Automated encoding of clinical documents based on natural language processing. Journal of the American Medical Informatics Association 11, 392–402, doi:10.1197/jamia.M1552 (2004).
Lasko, T. A., Denny, J. C. & Levy, M. A. Computational phenotype discovery using unsupervised feature learning over noisy, sparse, and irregular clinical data. PloS one 8, e66341, doi:10.1371/journal.pone.0066341 (2013).
Choi, E. et al. Multilayer representation learning for medical concepts. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 1495–1504 (ACM, 2016).
Farhan, W. et al. A predictive model for medical events based on contextual embedding of temporal sequences. Journal of medical Interenet Research (2016).
Winslow, R. L., Trayanova, N., Geman, D. & Miller, M. I. Computational medicine: translating models to clinical care. Science translational medicine 4, 158rv11–158rv11, doi:10.1126/scitranslmed.3003528 (2012).
White, N. et al. Probabilistic subgroup identification using bayesian finite mixture modelling: A case study in parkinson’s disease phenotype identification. Statistical methods in medical research 21, 563–583, doi:10.1177/0962280210391012 (2012).
Zhou, J., Wang, F., Hu, J. & Ye, J. From micro to macro: data driven phenotyping by densification of longitudinal electronic medical records. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining 135–144 (ACM, 2014).
Liu, C., Wang, F., Hu, J. & Xiong, H. Temporal phenotyping from longitudinal electronic health records: A graph based framework. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 705–714 (ACM, 2015).
Luo, Y., Xin, Y., Joshi, R., Celi, L. & Szolovits, P. Predicting icu mortality risk by grouping temporal trends from a multivariate panel of physiologic measurements. In AAAI, 42–50 (2016).
Ho, J. C. et al. Limestone: Highthroughput candidate phenotype generation via tensor factorization. Journal of biomedical informatics 52, 199–211 (2014).
Ho, J. C., Ghosh, J. & Sun, J. Marble: highthroughput phenotyping from electronic health records via sparse nonnegative tensor factorization. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining 115–124 (ACM, 2014).
Wang, Y. et al. Rubik: Knowledge guided tensor factorization and completion for health data analytics. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 1265–1274 (ACM, 2015).
Luo, Y. et al. Subgraph augmented nonnegative tensor factorization (santf) for modeling clinical narrative text. Journal of the American Medical Informatics Association ocv016 (2015).
Luo, Y., Wang, F. & Szolovits, P. Tensor factorization toward precision medicine. Briefings in bioinformatics bbw026 (2016).
Perros, I., Chen, R., Vuduc, R. & Sun, J. Sparse hierarchical tucker factorization and its application to healthcare. In Data Mining (ICDM), 2015 IEEE International Conference on 943–948 (IEEE, 2015).
Ho, J. C. et al. Limestone: Highthroughput candidate phenotype generation via tensor factorization. Journal of biomedical informatics 52, 199–211 (2014).
Newton, K. M. et al. Validation of electronic medical recordbased phenotyping algorithms: results and lessons learned from the emerge network. Journal of the American Medical Informatics Association 20, e147–e154, doi:10.1136/amiajnl2012000896 (2013).
Richesson, R. L. et al. A comparison of phenotype definitions for diabetes mellitus. Journal of the American Medical Informatics Association 20, e319–e326, doi:10.1136/amiajnl2013001952 (2013).
Carroll, J. D. & Chang, J.J. Analysis of individual differences in multidimensional scaling via an nway generalization of “eckartyoung” decomposition. Psychometrika 35, 283–319, doi:10.1007/BF02310791 (1970).
Harshman, R. A. Foundations of the parafac procedure: Models and conditions for an “explanatory” multimodal factor analysis (1970).
Le Cessie, S. & Van Houwelingen, J. C. Ridge estimators in logistic regression. Applied statistics 41, 191–201, doi:10.2307/2347628 (1992).
Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S. & Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems 3111–3119 (2013).
Gegick, M. Symmetric nonnegative matrix factorization for graph clustering. In Proceedings of the 2012 SIAM International Conference on Data Mining (SIAM, 2012).
Von Luxburg, U. A tutorial on spectral clustering. Statistics and computing 17, 395–416, doi:10.1007/s112220079033z (2007).
Shi, J. & Malik, J. Normalized cuts and image segmentation. In Computer Vision and Pattern Recognition, 1997. Proceedings., 1997 IEEE Computer Society Conference on 731–737 (IEEE, 1997).
Ding, C. H., He, X. & Simon, H. D. On the equivalence of nonnegative matrix factorization and spectral clustering. In SDM vol. 5, 606–610 (SIAM, 2005).
Johnson, A. E. et al. Mimiciii, a freely accessible critical care database. Scientific data 3, 160035, doi:10.1038/sdata.2016.35 (2016).
Wang, Y. et al. Rubik: Knowledge guided tensor factorization and completion for health data analytics. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 1265–1274 (ACM, 2015).
Knaus, W. A., Draper, E. A., Wagner, D. P. & Zimmerman, J. E. Apache ii: a severity of disease classification system. Critical care medicine 13, 818–829, doi:10.1097/0000324619851000000009 (1985).
Le Gall, J.R., Lemeshow, S. & Saulnier, F. A new simplified acute physiology score (saps ii) based on a european/north american multicenter study. Jama 270, 2957–2963, doi:10.1001/jama.1993.03510240069035 (1993).
Johnson, A. E., Kramer, A. A. & Clifford, G. D. A new severity of illness scale using a subset of acute physiology and chronic health evaluation data elements shows comparable predictive accuracy. Critical care medicine 41, 1711–1718, doi:10.1097/CCM.0b013e31828a24fe (2013).
Pollack, M. M., Patel, K. M. & Ruttimann, U. E. et al. The pediatric risk of mortality iii—acute physiology score (prism iiiaps): a method of assessing physiologic instability for pediatric intensive care unit patients. The Journal of pediatrics 131, 575–581, doi:10.1016/S00223476(97)700659 (1997).
Bouch, D. C. & Thompson, J. P. Severity scoring systems in the critically ill. Continuing Education in Anaesthesia, Critical Care & Pain 8, 181–185 (2008).
Carroll, J. D. & Chang, J.J. Analysis of individual differences in multidimensional scaling via an nway generalization of “eckartyoung” decomposition. Psychometrika 35, 283–319, doi:10.1007/BF02310791 (1970).
Harshman, R. A. Foundations of the parafac procedure: Models and conditions for an “explanatory” multimodal factor analysis. UCLA Working Papers in Phonetics 16, 184 (1970).
Bader, B. W. & Kolda, T. G. Matlab tensor toolbox version 2.5. Available online, January 7 (2012).
Jain, A. & Zongker, D. Feature selection: Evaluation, application, and small sample performance. IEEE transactions on pattern analysis and machine intelligence 19, 153–158, doi:10.1109/34.574797 (1997).
Barrett, M. L., Smith, M. W., Elixhauser, A., Honigman, L. S. & Pines, J. M. Utilization of intensive care services  statistical brief 185. Healthcare Cost and Utilization Project (HCUP) Statistical Briefs (2014).
Robinson, P. N. Deep phenotyping for precision medicine. Human mutation 33, 777–780, doi:10.1002/humu.22080 (2012).
Zemojtel, T. et al. Effective diagnosis of genetic disease by computational phenotype analysis of the diseaseassociated genome. Science translational medicine 6, 252ra123–252ra123, doi:10.1126/scitranslmed.3009262 (2014).
Acknowledgements
This research was partly supported by NHGRI (R00HG008175, R01HG008802), NLM (R00LM011392, R21LM012060), NIPA (NIPA2014H0201141001), NRF (2012M3C4A7033344), MSIP/IITP (B0101150307), and MOTIE (10049079).
Author information
Authors and Affiliations
Contributions
Y.K., R.E. and X.J. participated in writing draft; J.S. and X.J. provided important motivations of this study; Y.K. performed the experiments; R.E. and X.J. analyzed data and results; J.S., H.Y. and X.J. provided administrative and supervisory support.
Corresponding authors
Ethics declarations
Competing Interests
The authors declare that they have no competing interests.
Additional information
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kim, Y., ElKareh, R., Sun, J. et al. Discriminative and Distinct Phenotyping by Constrained Tensor Factorization. Sci Rep 7, 1114 (2017). https://doi.org/10.1038/s4159801701139y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s4159801701139y
This article is cited by
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.