Clustering Overly-Specific Features in Electronic Medical Records Jeremy Weiss1, Bess Berg1, Peggy L. Peissig2, Catherine A. McCarty2, David Page1 1 2 Computer Science Department Marshfield Clinic Research Foundation University of Wisconsin-Madison Marshfield Clinic Madison, WI Marshfield, WI {jcweiss,bess}@cs.wisc.edu, {peissig.peggy, mccarty.catherine} [email protected] @mcrf.mfldclin.edu

Abstract The recent widespread growth of electronic medical records (EMRs) provides us a first opportunity to predict future occurrences of disease using full patient medical records. The data available about each patient are relational, high-dimensional, and highly specific. These data characteristics create a challenge for many machine learning algorithms, which may not scale well with dimensionality or handle complex relations. We tackle the former problem by introducing an overlapping clustering technique, producing lower granularity features as extra inputs to rule learners, which in turn address the latter problem. Our clustering technique accounts for specific characteristics of medical data, which standard clustering methods using feature correlation or distance fail to address, by relying on similar patterns of usage. We evaluate the approach by the degree to which it empowers learning to predict disease susceptibility and show that it improves model accuracy on five distinct medical diagnosis tasks.

1

I n t ro d u c t i o n

Electronic medical records provide a rich resource of data about patients, including diagnoses, drug prescriptions, laboratory results, genetic markers and family history, all in the form of a relational database (Figure 1). With this new wealth of information, researchers may now investigate important machine learning tasks, such as prediction of future susceptibility of particular diseases using complete medical records. Because EMRs contain many relational tables, it is natural to use relational learning to predict disease susceptibility from medical history. One challenge in applying machine learning techniques to this problem task is the fine granularity of health records data. For example, EMR diagnoses are encoded to highly specific disease subtypes. Suppose we want to identify from patient history that any psychiatric diagnosis increases the risk of a given disease. Because there are roughly 100 detailed psychiatric diagnoses, the number of examples needed to make this discovery from the EMR is greater than if the diagnoses had all been encoded as the one broader diagnosis. The problem is exacerbated if the effect occurs only within a particular context (e.g., other prior diagnoses or exposure to certain drug classes).

PatientID Gender Birthdate

PatientID Date

M

3/22/65

P1 P1

PatientID Date

Lab Test

Result

P1

P1 P1

1/1/01 blood glucose 1/9/01 blood glucose

Physician Symptoms

1/1/01 2/1/03

Smith Jones

Diagnosis

palpitations hypoglycemic fever, aches influenza

PatientID SNP1 SNP2 … SNP500K

42 45

P1 P2

AA AB

AB BB

BB AA

PatientID Date Prescribed Date Filled Physician Medication Dose Duration P1

5/17/98

5/18/98

Jones

prilosec

10mg 3 months

Figure 1: Sample EMR. This EMR contains demographic data, diagnoses, labs, genetic markers and drug prescriptions. Some fields have been omitted from the tables, such as patient name, zip code, etc. Most EMRs at present do not have coded symptoms (they are in text fields) or the genetics table, but these are likely to become widespread in the future. If the granularity problem is that the features are too general, improving prediction will be difficult. But if the features are too specific, as is the case for EMRs, we can address the problem if we appropriately cluster more detailed feature values into coarser feature values, for example clustering the multiple psychiatric diagnoses into one. In this paper, we examine this approach and examine shortcomings of current clustering techniques for this task. We propose a clustering method and test it on this application of predicting disease susceptibility from an EMR using relational machine learning.

2

Motivation

Given overly-specific features, we want to construct features that represent larger-grained concepts. Medically this could mean clustering drugs into drug classes—selective serotonin reuptake inhibitors (SSRIs) into an SSRI drug class, for example. Our desired clustering needs to find such relationships given, for example, (patient, drug) pairs. The problem is that if patients are taking SSRIi (e.g. Prozac), they are unlikely to be taking any other SSRIs (e.g. Zoloft) because physicians usually avoid prescribing multiple drugs from the same drug class (Figure 2). Therefore, the cross-coverage over patients for each SSRI will be minimal. In other words, the correlation between SSRIs, defining each SSRI by a binary feature vector over patients, will not be indicative of their membership to the same drug class.

1 2

1110000010000000 0011110001000000 1111111100000000

Figure 2. A constructed example of drugs taken by patients. Each column represents a patient, and a ‘1’ indicates that the patient takes the drug in that row. Note that the correlation between SSRIs is small. Our algorithm will cluster the SSRIs together without zolpidem, while direct correlationbased algorithms cluster all three drugs together, or do not include the SSRIs in the same cluster at all. Our desired clustering should also allow features to belong to multiple, overlapping clusters. Take the example of timolol, a beta-blocker. Indications for timolol include hypertension, migraines, open-angle glaucoma, and regulation of cardiac stress. We want timolol to belong to the beta-blocker drug class, but because it has indications not shared by all beta blockers, we want it to also belong to drug classes defined by the medical conditions they treat, and to drug classes defined by tendencies of physicians to prescribe groups of drugs. Each feature may belong to any number of clusters, or to none at all. Partition-based and hierarchical clustering algorithms do not produce such clusters. Soft clustering, the most similar

clustering method to overlapping clustering, is used most often as an attempt to model uncertainty in belonging to clusters, producing multinomials over clusters indicating probability of cluster membership. For our problem task, membership to one cluster should not affect the probability of membership to another, as this "normalizing" effect is not useful when the clusters will be used in feature construction to empower rule learning. When might such overlapping clusters be useful? One answer is, precisely when the current feature representation is fine-grained and high-dimensional. As such, overlapping clusters as an additional feature representation have applications in text classification, computer vision, bioinformatics, and prediction from relational databases. Overlapping clustering over sparse data is relevant to relational learning algorithms because database tables can be mapped to sparse tabular representations, where clustering can then produce new binary relations. Without clustering, relational learning methods may become burdened with details that are useful but do not individually have statistical power to become incorporated during rule creation. In related work, Craven and Slattery used a related technique called predicate invention to form collections of related documents from hypertext documents and links, improving a decision-tree learner [1,2]. Davis et al devised a technique called view learning to learn new predicates that formed relations between existing literals and applied them in a statistical relational learning model to improve prediction accuracy [3,4]. Kok and Domingos proposed a second-order Markov logic algorithm to produce informative unary predicates [5]. Aggregation techniques maybe also be useful in this context and are reviewed by Provost and Perlich [6]. Several works have specifically addressed overlapping clustering [7-9]. The most closely related work comes from Kandylas et al [7], who proposed a streaming algorithm that showed improvement over sparse k-means clustering for partial clusterings in terms of weighted average entropy and normalized mutual information. Here, we present an overlapping clustering algorithm that is capable of identifying overlapping clusters, based on co-occurring, significant information gain. We apply our clustering algorithm to a diagnosis list from an electronic medical record for use in relational prediction of five diseases. The extensive detail available in EMRs holds promise for effective delivery of personalized care, but such detail also means that algorithms in place to achieve this goal are underpowered. Our method counteracts the dilution of signal from finegrained detail classes and constructs meaningful, data-driven clusters. We assess the quality of the clusters by comparing their utility in prediction of five diseases against prediction using baseline clinical details, the streaming algorithm in Kandylas et al [7], and the established ICD9 hierarchy of diagnosis clusters.

3

E q u i v a l e n c e c l u s t e r i n g a l gor i t h m

3.1

Summary

We want to form clusters over a feature space A, for example over diagnoses. To do so we iterate through a set of 'fingerprint' features B, for example, drug prescriptions, or drug prescriptions plus diagnoses. Each feature in A or B is described by a binary indicator vector of length equal to the number of data points, e.g. the number of patients. To build a cluster, all candidate features are chosen whose information gain I(a,b) for a particular b∈B exceeds a threshold β; cluster coverage over data points must fall within a desired size range. A newly found cluster is merged with other pre-existing clusters if their overlap surpasses threshold δ. The union of cluster members becomes a new feature in A. For convenience in this paper we set B=A and include the fingerprint feature b in the cluster only if the a's are sufficiently correlated with b (Figure 3). 3.2

Algorithm

We choose two sets of features: A, the set of features to cluster, and B the set of fingerprint features. Each set of features is described by a common set of examples, defined by a binary indicator vector. Then for each feature , we select all such that 1| 1 1 , and , , where β is a threshold parameter for cluster inclusion strictness, and I is the information gain of over b and a.

The a that satisfy these criteria are ordered by decreasing I(a,b). Starting with an empty cluster Ci, we iteratively add each feature if the coverage over examples is less than max(εm,|b=1|), where b is the fingerprint feature used for cluster formation, m is the number of patients, and ε is a small constant. Next, we check for cluster overlap and merge clusters if they are too similar: i.e. when the intersection-to-union ratio across diagnoses is greater than δ. The new cluster Ci is mapped to a binary feature vector by the union of its members and appended to both A and B. We iteratively test fingerprint features b to see if they form new clusters or modify existing ones until no new acceptable cluster is formed based on any b. Finally, we map the relation (example i, feature j) to (example i, cluster k) if the feature j is a member of the cluster k, producing a new binary relation for relational learners (Figure 3). One advantage of our clustering algorithm is its ability to identify functionally equivalent members, as highlighted in the example in Section 2. By setting B=A we provide the option of adding b to the cluster of features C, with ai∈C. This is desirable when the ai are correlated, but undesirable when they are not. This is because if the ai are correlated and each is correlated with b, then we cannot easily distinguish b from the ai, and therefore should include b in the cluster. We would like to add b to the cluster when the intersection of elements in C is greater than if they were all independent (i.e. not correlated): ,…,



,

,…,

.

Of course when the size of the cluster grows large, the intersection goes to the empty set, so as a heuristic we compare S, the log probability that an example is described by γ|C| or more cluster elements, with the sum of log probabilities multiplied by γ/|C|, where 0 < γ ≤ 1. We set γ = 2/3. By allowing b to be a cluster member, our algorithm produces both specific equivalence membership clusters and broader clusters based on co-occurring information gain. Input: binary matrix M: n diagnoses (rows), m patients (columns); constants β, δ, ε repeat for each diagnosis b (B, fingerprint, row) Find set Dig of all diagnoses (A, rows) with I > β, order Dig by decreasing I Form new cluster Ci = {} for each dig in Dig, dig≠b if |union(dig,dj,…dk)=1| < max(εm,|b=1|), where {dj,…,dk}=Ci Add dig to Ci end if end for Remove b from Ci if Ci is not sufficiently intercorrelated Add Ci to set of clusters C while |intersection(Ci,Cj)|/|union(Ci,Cj)| > δ, any Cj in C Merge Ci with Cj, i.e. Ci = union(Ci,Cj) Remove Cj from C, corresponding row in M end while Append to M: cluster diagnosis (row) = union(all dj in Ci) end for until |C| did not change given each di in M Figure 3: Equivalence clustering algorithm with A=B assumption Another advantage of our algorithm is that, compared to the algorithm presented in Kandylas et al, it is not susceptible to many of the limitations of streaming clustering noted in Sarmento et al [10]. Compared to streaming, our algorithm iterates over features as partitions (information gain acts as a partition function) to form new clusters, not as candidates to join or create a cluster. These new clusters then are merged with other clusters if their intersection to union ratio is greater than δ. Using a strict β threshold for membership to new clusters and then merging these clusters allows large clusters to form only by merging small clusters, meaning that for a large cluster to exist,

many of its members must match across multiple fingerprint features. This tractably produces overlapping clusters that are formed based on a larger “collective fingerprint.” Figure 4 shows an example of the overlapping clustering our algorithm might produce. Nodes (small ovals) represent features, arcs indicate significant information gain and positive correlation between the features, and the large shapes show the resulting clusters. Information gain arcs from large shapes (as new features) are not depicted for simplicity. In the top left, we see A— C—B, producing the cluster {A,B}. This highlights how the SSRIs in our previous example might cluster together despite their lack of correlation or information gain with one another. In the bottom left, we see the D—E—F—D triangle. Using D as our fingerprint feature, we will cluster all three features together because E and F share information gain and positive correlation. The right side of Figure 3 highlights how the clustering algorithm is partial (O does not belong to any cluster) and overlapping (G belongs to two clusters). Note that even though G is used as a fingerprint feature, we do not have the cluster {H,I,L}. Node L exemplifies a common diagnosis, such that it shares significant information gain with many other diagnoses simply due to the distribution of diagnoses over patients following Zipf’s law. Because it is common, it will not meet the patient coverage limit criterion, and the cluster that will form instead is {H,I}, which will merge with cluster {G,H,I}, formed when H or I is the fingerprint feature.

Figure 4. Sample equivalence clustering.

4

Experiments

We apply our equivalence clustering to a nested case-control population for prediction of five diseases: depression, migraine, rheumatoid arthritis, cirrhosis and osteoporosis. Our data set comes from the Marshfield Clinic, one of five academic institutions that are part of the Electronic Medical Records & Genomics (eMERGE) Network (www.gwas.net), sponsored by the National Human Genome Research Institute, National Institute of General Medical Sciences, and National Institutes of Health. eMERGE was formed to develop, disseminate, and apply approaches to research that combine DNA biorepositories with EMR systems for genetic research. As part of the deliverables for this network, Marshfield developed a nested age-related cataract case/control cohort defined by ophthalmic exams, cataract diagnoses and cataract extraction procedures in subjects aged 50 years and older. This cohort is a subset of the Personalized Medicine Research Project (PMRP) cohort [11,12]. The PMRP cohort contains approximately 20,000 subjects aged 18 years and older who reside in the Marshfield Epidemiology Study Area (MESA), a geographically-defined cohort located in Central Wisconsin. The majority of subjects in this cohort receive most of their medical care through the Marshfield Clinic integrated health care system. Marshfield has one of the oldest internally developed Electronic Medical Records (Cattails MD) in the US, with coded diagnoses dating back to the early 1960’s. Data collected in clinical care using the EMR is transferred daily into the Marshfield Clinic Data Warehouse (CDW) where it is integrated into the enterprise repository. The CDW has a variety of data including billing, practice management, clinical, insurance and ancillary system

observations, medications, and clinical documents. In addition, there are hundreds of reference tables that provide meaning to facts collected. The CDW is the source of data for this study. Programs were developed to select, de-identify by removing direct identifiers, and then transfer the data to a collaboration server where scientists conduct experiments. For this investigation the specific files used were: ICD9 diagnoses, observations (including lab results and other items such as weight, blood pressure, and height), two sources of medication information, patient demographics (gender and date of birth), and smoking history collected from a survey completed by PMRP subjects at time of enrollment. 4.1

Methodology

We approach the medical diagnosis prediction task by identifying 1:1 case-control populations matching for age, gender, cataracts and diabetes status. The last two control for their overrepresentation in the eMERGE cohort. Cases were identified from the EMR by the target diagnosis ICD9 code. We then eliminate EMR data from our analysis after the timepoint t, one month prior to the age of the target diagnosis in both case and control. Each clustering algorithm was run over all EMR diagnoses for the eMERGE cohort, without knowledge of the prediction task. The case-control population for each target diagnosis was split into 10 folds for cross-validation. We evaluate the performance of our clustering algorithm based on the improvement in predictive accuracy in comparison with predictive accuracy at baseline, using the streaming algorithm in Kandylas et al [7], and using the established ICD9 hierarchy of diagnosis groups. Because the Kandylas algorithm supports partial clustering but not overlapping clustering, we chose two settings of θ to produce different clusters and overlaid the results (parameters: θ=0.05, 0.2; K=100, 1000; similarity=cosine). The ICD9 is a World Health Organization hierarchical categorization of diagnoses, providing broad categories and subcategories. We also include prediction using both the ICD9 hierarchy and our clustering algorithm to observe if the inclusion of clusters from multiple sources improves prediction beyond either alone. We add the clusters to the original EMR data for prediction using two relational learning approaches: inductive logic programming by Aleph [13] and the statistical relational learning algorithm SAYU-TAN [3,4]. Aleph prediction uses theories defined as a disjunction of rules, which themselves are conjunctions of predicates. To find acceptable rules, we perform a top down search over the lattice extending from each case’s bottom clause, or saturation [2], using branch and bound search with minimum accuracy of 0.85, and minimum cases of 16, maximizing information gain. SAYU-TAN is a statistical relational learning algorithm that constructs the maximum likelihood tree-augmented naïve Bayes net with logical rules as nodes. A node is added to the tree if it improves area under the precision-recall curve (AUCPR) of the Bayes net on a cross-validated tune set by at least 1.03 times the previous AUCPR in 7 of the 9 folds. For Aleph runs we report accuracy and the one-sided Fisher’s exact test; for SAYU-TAN we report the AUCPR and maximum accuracy along the AUCPR curve. While accuracy is a standard measure for algorithm comparison, it weighs false positives and false negatives equally, and provides no information about the significance of its result. In medical domains, significant results that either maximize sensitivity for screening tests or specificity for diagnostic tests are desired. We also report the Fisher’s exact test as another measure for comparison that incorporates differences in proportions and their variances. We also use the sign test to test for significant differences in prediction accuracy across all five diagnoses, for each ten folds. We choose this nonparametric test because the accuracies are not drawn from a single distribution. 4.2

Results

Our algorithm resulted in 1175 potentially-overlapping clusters from 11,067 diagnoses, with 11% of diagnoses included in two or more clusters (range 0-12). Most clusters showed meaningful relationships among its members (Figure 4). Clusters members tended to be variants of each other, descriptive of similar patient types, or linked through standard medical guidelines of treatment and care. Large clusters often comprised diagnoses that were not interchangeable, but instead represented a broader abstraction of medical concepts, such as diagnoses all related to pregnancy (e.g. antepartum hemorrhage, chromosomal abnormality screening, delivery, Caesarian section).

Table 1. Test set accuracy and Fisher’s exact p-value by cluster method for ILP predictors. Depression (n=2160) acc P

Osteoporosis (n=1908) acc p

Migraine (n=1262) acc p

Rheumatoid arthritis (n=720) acc p

Cirrhosis (n=422) acc

p

Baseline

0.63

4e-35

0.50

0.43

0.59

1e-11

0.64

2e-14

0.63

1e-07

Streaming Equivalence clusters ICD9 Equivalence clusters + ICD9

0.63

7e-33

0.50

0.42

0.58

4e-09

0.65

3e-16

0.64

4e-09

0.65

2e-47

0.52

0.03

0.59

8e-12

0.66

1e-17

0.62

5e-07

0.65

1e-52

0.50

0.36

0.61

2e-15

0.65

8e-16

0.64

2e-08

0.65

7e-45

0.52

0.03

0.60

3e-12

0.64

3e-14

0.62

5e-07

Table 2. Test set accuracy and area under the precision-recall curve by cluster method for SAYUTAN predictors. AUCPR is reported for recalls greater than 0.5. Depression (n=2160)

Osteoporosis (n=1908)

Migraine (n=1262)

Rheumatoid arthritis (n=720)

Cirrhosis (n=422)

acc

AUCPR

acc

AUCPR

acc

AUCPR

acc

AUCPR

acc

Baseline

0.56

0.27

0.56

0.28

0.62

0.29

0.63

0.29

0.58

0.28

Streaming

0.56

0.27

0.57

0.28

0.63

0.30

0.59

0.28

0.61

0.29

Equivalence Clusters

0.64

0.29

0.58

0.28

0.63

0.29

0.63

0.29

0.66

0.31

0.64

0.30

0.56

0.28

0.63

0.29

0.66

0.31

0.62

0.31

0.63

0.29

0.57

0.28

0.62

0.28

0.64

0.31

0.62

0.31

ICD9 Equivalence clusters + ICD9

AUCPR

---Diagnosis cluster 310 based on fingerprint “Severe sepsis with acute organ dysfunction”--Gram negative septicemia, NOS Septic shock ---Diagnosis cluster 33 based on fingerprint “Schizophrenic, paranoid, chronic”--Paranoid personality Schizophrenic, paranoid, chronic Schizophrenia, paranoid, unspecified Paranoid state, NOS Schizoaffective disorder, chronic Schizoaffective disorder, unspecified Schizoid personality, NOS Schizophrenia, NOS, in remission Schizophrenia, NOS, unspecified Schizophrenic disorder, residual, chronic Schizophrenic disorder, residual, unspecified Schizophrenia, simple, unspecified Figure 5. Two clusters of 1175 total clusters. The first is an equivalence cluster, the second is a correlation cluster. The cluster title is a fingerprint feature and is not necessarily included in the equivalence cluster. The accuracy and one-sided Fisher’s exact test from cross-validation on each prediction task are shown in Table 1. Prediction using our clustering algorithm outperforms prediction from baseline and prediction using streaming clustering in four of the five data sets. Inclusion of both the ICD9 hierarchy and equivalence clusters showed no consistent improvement in predictive accuracy over either one alone. Likewise, prediction using equivalence clusters in SAYU-TAN outperformed prediction at baseline and using streaming clusters over all five diagnoses, and performed comparably to the ICD9 hierarchy (Table 2).

Predictive accuracy varied by diagnosis between SAYU-TAN and Aleph, but in each case the use of our algorithm in prediction outperforms both prediction using baseline EMR data and EMR data with streaming clusters. We show this difference is significant using the sign test (Table 3). The difference between using our clusters and ICD9 clusters was not significant. Table 3. Sign test and p-value of accuracy differences between prediction using equivalence clustering and comparison group1 Comparison group Baseline Streaming ICD9

5

Aleph (+/-/=) 32/16/2 31/16/3 16/27/7

SAYU-TAN p 0.03 0.04 0.07

(+/-/=) 36/10/4 36/10/4 26/19/5

p 2e-04 2e-04 0.37

D i s c u s s i on

Our results suggest that the equivalence clustering algorithm improves predictive accuracy of relation-based prediction algorithms and rivals the improvement seen by using the preestablished ICD9 hierarchy. Tables 1 and 2 show the trend towards improvement in prediction using our algorithm regardless of the target diagnosis or choice of predictive model. This suggests that our clustering algorithm is robust to use in different relation-based prediction algorithms and for different prediction tasks. Aleph’s disjunction of rules is an intrinsically different model than SAYU-TAN; regardless of model choice, informative features, like the clusters our algorithm produces, result in better prediction. Another advantage of our algorithm is that we can extend it to cluster over other relations beyond diagnoses. This means we can produce clusters over diagnoses, drugs and other clinical details simultaneously, inferring clusters across clinical properties. Clusters may then take on semantic meaning that relate different segments of the medical history. As we have demonstrated in table 3, clustering is a valuable first step for improving prediction algorithms dealing with overly-specific features. Because information is present in the relatedness of these features, it helps to cluster them prior to prediction, to enhance the power of their common signal. Given that features are frequently related over more than one domain, greater attention should be devoted to overlapping clustering, avoiding the restriction of assignment of features to single clusters characteristic of partition-based clustering. Whether clustering features is helpful for tabular classification algorithms remains to be tested. These additional cluster features may then be considered as logical interaction terms that may enhance the feature representation while moderately expanding the hypothesis space.

6

Conclusion

In this paper, we have presented a clustering algorithm using features as partitions on which to cluster other features. We highlight several advantages in the way our algorithm forms clusters. We show that, in conjunction with relational learning algorithms, our algorithm improves prediction compared to prediction using baseline information and streaming clustering and is comparable to established ICD9 hierarchies in empowering learning. We emphasize the utility of overlapping clustering as a preprocessing step for relational learners and offer several considerations for further improvement. The redirection of medical care towards including personalized medicine can benefit greatly from identification of disease susceptibility. With the expansion of EMRs, we now have the data to succeed. New algorithms, such as the overlapping clustering algorithm presented here, can help us achieve this goal.

1

Per diagnosis per fold. +: accuracy(equivalence clustering) > accuracy(comparison group). -: the opposite. =: the accuracies were equal.

Acknowledgments We gratefully acknowledge our support from the Institute of Clinical and Translational Research (Clinical and Translational Science Award (CTSA) program of the National Center for Research Resources, NIH 1UL1RR025011), the Marshfield Clinic Research Foundation (NGHRI 1U01HG004608-01), and the Medical Scientist Training Program. We also acknowledge the State of Wisconsin for their support through the Wisconsin Genomics Initiative. References [1] Craven, M., & Slattery, S. (2001). Relational learning with statistical predicate invention: Better models for hypertext. Machine Learning, 43(1), 97-119. [2] Muggleton, S., & Buntine, W. (1988). Machine invention of first-order predicates by inverting resolution. Proceedings of the Fifth International Conference on Machine Learning, 339-351. [3] Davis, J., Burnside, E., Page, D., Dutra, I., & Costa, V. S. (2005). Learning bayesian networks of rules with SAYU. Proceedings of the 4th International Workshop on Multi-Relational Mining, 13. [4] Davis, J., Ong, I., Struyf, J., Burnside, E., Page, D., & Costa, V. S. (2007). Change of representation for statistical relational learning. Proceedings of the 20th International Joint Conference on Artificial Intelligence, 2719–2726. [5] Kok, S., & Domingos, P. (2007). Statistical predicate invention. Proceedings of the 24th International Conference on Machine Learning, 440. [6] Perlich, C & Provost F. (2003). Aggregation-based feature invention and relational concept classes. Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining, 167-176. [7] Kandylas, V., Upham, S. P., & Ungar, L. H. (2008). Finding cohesive clusters for analyzing knowledge communities. Knowledge and Information Systems, 17(3), 335-354. [8] Segal, E., Battle, A., & Koller, D. (2002). Decomposing gene expression into cellular processes. Biocomputing 2003: Proceedings of the Pacific Symposium Hawaii, USA 3-7 January 2003, 89. [9] Banerjee, A., Krumpelman, C., Ghosh, J., Basu, S., & Mooney, R. J. (2005). Model-based overlapping clustering. Proceedings of the Eleventh ACM SIGKDD International Conference on Knowledge Discovery in Data Mining, 537. [10] Sarmento, L., Kehlenbeck, A., Oliveira, E., & Ungar, L. (2009). Efficient clustering of webderived data sets. Proceedings of the 6th International Conference on Machine Learning and Data Mining in Pattern Recognition, 412. [11] McCarty, C.A., Peissig, P., Caldwell, M. D., & Wilke, R. A. (2008). The Marshfield Clinic Personalized Medicine Research Project (PMRP): 2008 scientific update and lessons learned in the first six years. Personalized Medicine, 5: 529-541. [12] McCarty, C.A., Wilke, R. A., Giampietro, P. F., Wesbrook, S., & Caldwell, M. D (2005). Marshfield Clinic Personalized Medicine Research Project (PMRP): design, methods and recruitment for a large, population-based biobank. Personalized Medicine, 2: 49-79. [13] Srinivasan, A. (1999). The aleph manual. Computing Laboratory, Oxford University.

Clustering Overly-Specific Features in Electronic ...

{jcweiss,bess}@cs.wisc.edu, ... PatientID Date Prescribed Date Filled Physician Medication Dose Duration. P1. 5/17/98. 5/18/98 Jones .... Of course when the size of the cluster grows large, the intersection goes to the empty set, so as a.

196KB Sizes 0 Downloads 114 Views

Recommend Documents

CLUSTERING of TEXTURE FEATURES for CONTENT ... - CiteSeerX
storage devices, scanning, networking, image compression, and desktop ... The typical application areas of such systems are medical image databases, photo ...

Interacting with Features in GPlates
See ​www.earthbyte.org/Resources/earthbyte_gplates.html​ for EarthByte data sets. Background. GPlates ... Feature Type box at the top of the window). .... Indonesian Gateway and associated back-arc basins, Earth-Sci. Rev., vol 83, p.

Interacting with Features in GPlates
... the new changes. See the GPlates online manual for further information: ... In this exercise we will learn how to query features. To start with we .... 85-138, 1995. Gaina, C., and R. D. Müller, Cenozoic tectonic and depth/age evolution of the.

Measuring Volatility Clustering in Stock Markets
Sep 15, 2007 - Division of Business Administration, ... apply it to a high-frequency data of the financial markets. ... In the next section, we describe the data sets and methods used in this paper. .... istry of Education through the program BK 21.

New Features in SAS/OR® 13.1
SAS/OR® software for operations research includes mathematical .... and 2( ), are being minimized and 10 points, each corresponding to a solution, are ..... Other brand and product names are trademarks of their respective companies.

Features in Concert: Discriminative Feature Selection meets ...
... classifiers (shown as sample images. 1. arXiv:1411.7714v1 [cs.CV] 27 Nov 2014 ...... ImageNet: A large-scale hierarchical im- age database. In CVPR, 2009. 5.

Search features
Search Features: A collection of “shortcuts” that get you to the answer quickly. Page 2. Search Features. [ capital of Mongolia ]. [ weather Knoxville, TN ]. [ weather 90712 ]. [ time in Singapore ]. [ Hawaiian Airlines 24 ]. To get the master li

Clustering in Data Streams
Small(er)-Space Algorithm (cont'd). • Application in data stream model. − Input m (a multiple of 2k) points at a time. − Reduce the first m points to 2k medians. − Maintain at most m level-i medians. − On seeing m, generate 2k level-(i+1) m

Distributed Electronic Rights in JavaScript
any powerful references by default; any references it has implicit access to, such as ..... approach solves the open access problem by restricting access to members and regu- ..... Journal of Applied Corporate Finance 8(2), 4–18 (1995). 15.

Program features - MCShield
Feb 26, 2012 - Hard disk drives – enables initial scan of all hard drives ..... C:\Documents and Settings\All Users\Application Data\MCShield (for Windows XP).

Web page clustering using Query Directed Clustering ...
IJRIT International Journal of Research in Information Technology, Volume 2, ... Ms. Priya S.Yadav1, Ms. Pranali G. Wadighare2,Ms.Sneha L. Pise3 , Ms. ... cluster quality guide, and a new method of improving clusters by ranking the pages by.

C++98 features? - GitHub
Software Architect at Intel's Open Source Technology. Center (OTC). • Maintainer of two modules in ... Apple Clang: 4.0. Official: 3.0. 12.0. 2008. C++11 support.

data clustering
Clustering is one of the most important techniques in data mining. ..... of data and more complex data, such as multimedia data, semi-structured/unstructured.

Fuzzy Clustering
2.1 Fuzzy C-Means . ... It means we can discriminate clearly whether an object belongs to .... Sonali A., P.R.Deshmukh, Categorization of Unstructured Web Data.

Spectral Clustering - Semantic Scholar
Jan 23, 2009 - 5. 3 Strengths and weaknesses. 6. 3.1 Spherical, well separated clusters . ..... Step into the extracted folder “xvdm spectral” by typing.

Modeling Timing Features in Broadcast News Video ...
However, learning a classifier using ... While it is tempting to apply machine learning tech- nique to .... broadcast news programs, including ABC, CNN, and.

NEW FEATURES IN THE CU-HTK SYSTEM FOR TRANSCRIPTION ...
2000 evaluation of English conversational telephone speech tran- ... The triphone models were phone posi- ... smoothed with a class-based trigram model.

Strong Positions and Laryngeal Features in Yukatek ...
Nov 7, 2008 - Yukatek Maya has a five vowel system with the expected vowel qualities: [a], [e], [i], [o], [u]. • Additionally, it has ... b. sıinikh 'ant' (*sıinik, *sıinik^).

Neighborhood Features Help Detecting Non-Technical Losses in Big ...
Neighborhood Features Help Detecting Non-Technical. Losses in Big Data Sets. Patrick Glauner. Interdisciplinary Centre for. Security, Reliability and Trust,.