Predicting the Risk of Type 2 Diabetes Using Insurance Claims Data Hari Iyer∗ Department of Statistics Colorado State University 1877 Campus Delivery Fort Collins, Colorado 80523-1877 [email protected]

Todd Iverson Department of Mathematics and Statistics Saint Mary’s University of Minnesota 700 Terrace Heights #1511 Winona, Minnesota 55987-1399 [email protected]

Asa Ben-Hur Computer Science Department Colorado State University 1873 Campus Delivery Fort Collins, Colorado 80523-1873 [email protected]

Abstract Type 2 diabetes is a preventable disease that is affecting millions of people in the U.S. and elsewhere. Therefore an easy-to-apply method for early detection of the disease can have significant impact. In this paper we show that early detection of the disease is possible using only insurance claims data. We model the problem in the framework of kernel methods and use a support vector machine as a classifier. We propose kernels that represent a sequence of insurance claims data using ideas from text categorization and bioinformatics sequence analysis. On a dataset of R close to 60,000 enrollees extracted from the MarketScan database our method is able to predict a diabetes diagnosis up to 18 months before a doctor’s diagnosis with an area under the ROC curve of 0.85.

1

Introduction

Diabetes is a major source of this nation’s medical expense [1]. It differs from many health issues in two ways. First, type 2 diabetes is preceded by a pre-diabetes phase that allows for early detection of those at risk of type 2 diabetes [2]. Second, the onset of the disease can be delayed or completely prevented through life-style changes in the form of diet and exercise [2]. On the other hand, a failure to detect and treat diabetes can lead to a progression of the disease. A recent study showed that more than 25% of untreated patients with type 2 diabetes experienced worsening glycemic control [3]. This points to the importance of early detection of individuals at risk for type 2 diabetes, illustrated by the American Diabetes Association call for research on an effective method for identifying patients at risk for pre-diabetes and diabetes [2]. We feel the following work addresses this need by focusing on predicting risk of type 2 diabetes using only insurance claims. It has been shown that individuals at risk for type 2 diabetes can be accurately detected (0.859 area under the ROC curve) using logistic regression on a collection of features that include age, sex, ethnicity, blood pressure, body mass index, parental or sibling history of diabetes, and various measurements on the patient’s blood [4]. The tests used to make these blood measurements are time consuming, costly, and inconvenient [4]. Another study looked at predicting the onset of type 2 ∗

Currently at Caterpillar Inc

1

diabetes in women of the Pima Indian tribe using eight variables that include body mass index (BMI) and various measurements on the patient’s blood [13]. These data were made public through the UCI Machine Learning Repository and many papers were published in attempts to improve classification performance [14]. This study is based on a small sample and is relevant to a particular population. Administrative patient information such as insurance claims is an alternative to clinical data that has been shown to be useful in predicting patient mortality after a variety of hospital procedures [5, 6, 7, 8], as well as the resources required for a patient’s care [7, 9], and evaluation of quality of care by different providers [8, 9]. For diabetes claims data have been used to measure the adherence to medications [10] and identify existing cases of the disease (see, for example, [11] and [12]). No work has been done to our knowledge on the use of administrative data for the prediction of disease risk. Employers who subsidize health insurance premiums for their employees are seriously interested in identifying at risk employees (or dependents) and encouraging them to adopt simple life style changes. However, due to governmental regulations they are unable to collect and use health related data on their employees. Realizing that it may be possible to use claims data to identify at risk employees, Caterpillar Inc initiated this study and provided partial funding. We have developed methods that allow accurate prediction of persons at risk of type 2 diabetes based on claims data alone. These methods can be used as an initial screening, cutting down on the number of individuals that need the costly test and helping identify persons that might not have otherwise been screened. In this paper we show that the risk of type 2 diabetes can be predicted from insurance claims alone. Initially we try to make predictions using all claims preceding the first diagnosis of diabetes. While this is interesting, the real objective is to predict diabetes before a doctor’s diagnosis. To test our methods performance relative to this objective, we remove up to 18 months from an enrollee’s claims history, and show that prediction accuracy shows very little degradation.

2

Insurance claims data

Health insurance claims belong to one of several types: inpatient claims, outpatient claims, prescription drug claims, and enrollment information. The inpatient claims include services associated with a hospital admission. The outpatient claims are related to services that occur at a doctor’s office, emergency room, or other outpatient facility. The inpatient and outpatient claims contain diagnosis and procedure codes that describe the nature of the medical procedure and the condition that it addresses. The diagnosis codes come in the form of ICD-9-CM coded variables. The International Classification of Diseases, Ninth Revision, Clinical Modification (ICD-9-CM) is the recognized method for coding diagnoses and procedures in United States hospitals. The procedure codes are coded either using the ICD-9-CM, CPT-4, or HCPCA protocols. The Current Procedural Terminology, 4th Edition (CPT-4) is used most frequently, with ICD-9-CM procedure codes the second most common and HCFA Common Procedural Coding System (HCPCA) the least prevalent. The research discussed in this paper was inspired by and conducted on Thomson Medstat’s R MarketScan database of health insurance claims from 1999 to 2003 [15]. The database consists of private sector health claims collected from over 100 employers and includes over 10 million employees and dependents. The data contains the medical service uses and medical expenditures for insured employees, early retirees, COBRA continuations, and dependents covered under the employees’ health plan. Partial or complete drug information is available for a large number of enrollees. These include both mail-order and card program prescription drug claims. These claims contain all relevant information about a particular prescription, including drug brand, dosage, number of days covered by the prescription, and cost information. In addition to the claims themselves the database contains general demographic information about each enrollee. This includes age, gender, region, and employment information. 2

3

Methodology

Insurance claims data is interesting and complex. Each enrollee has a variable number of claims in their claim history. These claims also contain a variety of types of information, including diagnosis, procedure and drug codes, cost information, and demographic information. We have chosen to model this problem in the framework of kernel methods [16], and to use the Support Vector Machine (SVM) as a classifier [17]. SVMs were successfully applied in a wide variety of problems that involve data that has the form of a sequence of objects: words, letters etc. This includes text categorization [18], content-based audio classification and retrieval [19], and a wide variety of problems in computational biology which involve biological sequences—proteins or DNA [20]. The flexibility of kernel methods provides us with a convenient framework in which to model claims data. Our strategy is to split the claims data into separate components, each representing one aspect of the data: demographics, cost, diagnosis/procedure codes, and drug information, and model each aspect separately using an appropriate kernel. The kernels are then added into a single master-kernel, which is equivalent to modeling the data in a feature space which is the product of the feature spaces of the individual kernels [16]. This is a standard way of integrating different kinds of data with kernel methods [21]. Specifics on each kernel follow. 3.1

Kernels

In this section we give details of the kernel constructed for each type of data: diagnosis/procedure codes, drug codes, cost, and demographic information. 3.1.1

Code and drug kernels: bag of codes

Each diagnosis, procedure, or drug prescription is associated with a code, and an enrollee’s history can be described as a sequence of such codes. Guided by the way text is often represented in information retrieval as a “bag of words” [22], we consider a Bag of Codes (BoC) kernel. In this kernel each code in the database is associated with a feature whose value is either an indicator of the presence of that code in an enrollee’s claim history (BoCI features), or the count of the number of times the code appears in a claim history (BoC features). We then use a Gaussian kernel over the resulting feature vectors after normalizing them to unit norm. The BoC and BoCI kernels completely ignore the time at which a code is observed. To introduce some time information into the feature set, we implement the following change: We first divide each enrollee’s claim history into n time segments of a fixed number of time units w. For enrollees that have a claim history longer than n ∗ w time units, the most recent n ∗ w units are used. We then generate either the BoC or BoCI features for each segment and concatenate them into one combined feature vector. A Gaussian kernel over these features give the SBoC or SBoCI kernels, respectively. In our experiments we used 6 segments of 6 months each. All the variants of the BoC kernel showed very similar performance. 3.1.2

Time warping kernel

The various BoC kernels ignore the sequence nature of claims data. When using kernel methods in computational biology it is crucial to model biological sequences as such [20]. In comparing DNA and protein sequences the notion of an alignment is particularly useful: align two sequences against each other by inserting gaps in one sequence or the other such that a cost function that captures the similarity of aligned positions is maximized (see e.g., [23]). We consider a more flexible notion of an alignment called a time warping alignment, that allows multiple events to be aligned against each other. In order to express time warping alignments as kernels Cuturi and co-authors considered the space of all possible time warping alignments and summed the scores of all possible alignments between two sequences [24]. We have recently extended their kernel to consider local alignments that match segments of the two sequences, and more importantly, incorporate time information into the scoring of the time warping alignments. Full definition of this kernel is rather involved, and because of space constraints and the fact that the simpler BoC kernels had comparable performance we refer the reader to [reference omitted for anonymization] for details. Furthermore, the complexity 3

Table 1: Dimensionality of the data for each type of kernel. BoCI is the Bag of Codes kernel with indicator variables; SBoCI is the stratified BoC kernel with indicator variables. The summed BoCI/SBoCI refers to the kernel which is the sum of all kernels, using the BoCI/SBoCI kernel. Kernel Number of features Cost 48 Demographics 42 Procedure/Diagnosis BoCI 8644 Procedure/Diagnosis SBoWI 8644*6 = 51864 Drug BoCI 183 Drug SBoCI 183*6 = 1098 Summed BoCI 8917 Summed SBoCI 53052

of computing the time warping kernel is proportional to the product of the lengths of the sequences, whereas computation of the BoC kernel is linear in the sum of the length of the sequences. 3.1.3

Cost kernel

For each claim, we have the total cost of the claim as part of our data. We then compute the total cost over an enrollee’s claim history and construct a bin kernel as follows. Suppose that we have a continuous variable x. We select b0 < b1 < . . . < bn as the boundaries of n bins as shown below. We construct a feature vector with one component per bin. The value of the ith component is 1 provided that the variable x is larger than the bin’s upper limit bi , and 0 otherwise. When we take the inner product of such vectors, we get a count of how many bins the two values have in common. To construct our cost kernel we divided the claims into into six month segments. Within each segment we computed a bin kernel with bin boundaries chosen as the 10th, 20th, ..., 90th percentiles of the total cost. The overall cost kernel is a Gaussian kernel over the resulting features. 3.1.4

Demographics kernel

This kernel was used to incorporate an enrollee’s demographic information that includes age group, geographical region, employment classification of the primary beneficiary, employment status of the primary beneficiary, relationship to the primary beneficiary, industry classification of the employee responsible for payment of the claim, and gender. All of these variables are categorical in nature; each variable was associated with a set of indicator features, one for each possible value. 3.1.5

Combined kernel

Our overall kernel is a sum of kernels, one for each aspect of the data: Procedure/diagnosis codes, drug codes, cost, and demographics. To the summed kernel we apply an overall Gaussian kernel defined by: k(x, x0 ) = exp(−γ||x − x0 ||2 ) with a width parameter, γ. Before applying the gaussian kernel, the data for each kernel were normalized to unit vectors by dividing each example by its norm. The appropriate value of γ is chosen on a separate tuning set as described below. Linear and polynomial kernels gave slightly lower levels of accuracy, and the SVM convergence time was longer, so they were not used in the testing phase of our experiments. Because of the high dimensionality of the data (see Table 1), normalization and selection of a good value for the width parameter of the Gaussian kernel are crucial.

4

Experimental setup

This section contains details on the experimental setup: the choice of examples and their division of examples positive and negative examples, the different datasets we constructed, and our method for model selection. 4

Table 2: Dataset sizes. The number of positive examples and negative examples in data used for selecting classifier parameters (tuning-set) and data used for training and testing the final classifier (train-test). The extra set contains examples that did not satisfy our criteria for choosing positive and negative examples. Dataset Train-Test Set Tuning Set Extra Set

4.1 4.1.1

Positive examples 14,746 14,777 178,851

Negative 14,754 14,723 178,851

Total 29,500 29,500 357,702

Data set Selection of examples

A positive example is defined as an enrollee that has at least one observed diabetes diagnosis, with no insulin or other diabetes related prescriptions prior to the first observed diabetes diagnosis. Diabetes diagnosis was identified using the ICD-9 codes associated with the claims data. We make the assumption that the first observed diagnosis is the actual first diabetes diagnosis. A negative example is defined as an enrollee with no observed diabetes diagnosis and no observed insulin or other diabetes related prescriptions. Some enrollees did not have complete drug information, so we added the restriction that all positive and negative examples have full drug information. For each example we imposed the constraint of having data for an enrollee covering a period of at least 3 years, and up to 5 years. For enrollees identified as positive examples we required having at least 3 years prior to the diabetes diagnosis. 4.1.2

Truncated histories

Since our ultimate goal is to predict diabetes well ahead of a doctor’s diagnosis we also constructed datasets with truncated enrollee history by removing a section of an enrollee’s claims history before the first diabetes diagnosis. When constructing the truncated data set, we were concerned that removing the same portion of the histories from all examples would leave them aligned with respect to the first diabetes diagnosis. That is, they all ended at the same spot relative to the event that was being predicted. To make sure that we weren’t gaining a performance advantage through this alignment, we shortened the windows by truncating enrollee histories by a random amount. We constructed two data sets differing in the average amount of claims data that was removed. The first set had an average of 12 months removed and the second set had an average of 18 months removed. 4.2

Tuning, model selection, and testing

To avoid the need for nested cross-validation for parameter tuning, we divided the positive examples with a claim history of three to five years into two equally sized sets, one for classifier tuning and the other for training/testing the classifier. An equal number of randomly selected negative examples with a 3 to 5 year claim window was added to complete each set. We refer to these sets as the tuning set and the train-test set, respectively. The total size of each set is given in Table 2. Classifier accuracy was estimated by five-fold cross-validation on the train-test set using parameters selected on the tuning set. The SVM soft margin constant (C) and kernel width (γ) were chosen using five-fold crossvalidation on the tuning set on a grid of parameters using the values [1, 10, 100, 1000] for C, and [0.001, 0.01, 0.1, 1] for γ. The final performance is measured using the mean area under the ROC curve (AUC), averaged over five runs of five-fold cross-validation on the train-test set. Using the AUC provides an evaluation metric that does not depend on the (unknown) prevalence of diabetes in the population. Note that the test set was used in the final testing phase only and never used for kernel construction, tuning, or model selection. All the experiments were performed using the PyML machine learning environment which is available at http://pyml.sf.net. 5

Table 3: Results for individual data types/kernels. The value given in the table is the average of five independent runs of five-fold cross-validation, with the same set of five partitions used for each kernel. The standard error obtained in 5 runs of cross-validation was less than 0.001. AUC is the area under the ROC curve. Data Type/Kernel AUC Cost 0.686 Demographics 0.807 Diagnosis/Proc (BoCI) 0.821 Diagnosis/Proc (SBoCI) 0.818 Drug (BoCI) 0.759 Drug (SBoCI) 0.764

Table 4: ROC scores for the summed kernels using complete and truncated histories (see ROC curves in Figure 1). The 12 month and 18 month truncated histories had a random amount of their most recent history removed that averaged 12 months and 18 months respectively. Average Amount Removed None 12 Month 18 Month

5 5.1

AUC 0.867 0.848 0.845

Results Performance using complete claim histories

Table 3 gives the results for the individual kernels on the four types of data. There was no difference in performance between the kernels that stratify claims data into 6 month windows (SBoCI) and the kernels that lump all the claims together, for both diagnosis/procedure kernels and drug information kernels. Both the SBoCI and BoCI use indicator variables to count the presence of a code. The lower dimensionaity of the BoCI kernel and simpler calculation makes it a more attractive choice for usage in practice. We also note that there was little difference in performance between the binary BoC kernels and ones that used counts (data not shown, see [anonymous PhD thesis]). The kernel which uses the sum of a kernel from each data type has an ROC curve the dominates all the other ROC curves (Figure 1). Its AUC score was 0.867 compared to 0.821 for the best individual kernel. 5.2

Truncated claim histories

To determine whether diabetes risk can be predicted ahead of a doctor’s diagnosis we compared the ROC scores of a classifier that is trained using complete claim histories and a classifier provided with truncated claims histories, as described in Section 4.1. For this comparison we used the BoCI kernel rather than the SBoCI kernel for diagnosis/procedure and drug data. Because of their equivalent levels of performance we preferred the simpler method; we were also concerned that since we remove a random amount from an enrollee’s claim history the first segment or segments of claims histories would not yield meaningful comparisons. Recall that we considered two versions of truncated claim histories, one with an average of 12 month removed and another with an average of 18 months removed. The resulting classifier accuracies are as follows: A classifier which used the entire claims history had an AUC of 0.867; 12 month truncation gave 0.859, and 18 months truncation gave 0.845 (see Table 4, and the corresponding ROC curves in Figure 1). This is our most important result: Even after removing an average of 18 months of an enrollee’s claim history, we are still able to classify enrollees nearly as well. These results are promising in that our ability to successfully classify enrollees is nearly as good over a year before the first diabetes diagnosis as it is the day before the diagnosis. 6

1.0 0.0

0.2

0.4

0.6

0.8

0.6 0.2

0.4

True Positive Rate

0.8

1.0 0.8 0.6 0.4

Complete Codes 12 Months Removed 18 Months Removed

0.0

True Positive Rate

0.2 0.0

Combined Diag./Proc. Drug Demographics Cost 1.0

0.0

False Positive Rate

0.2

0.4

0.6

0.8

1.0

False Positive Rate

Figure 1: The left-hand graph shows ROC curves for various kernels. The kernel that combines all the data sources (denoted as Combined in the legend) performs better than any individual kernel. The diagnosis/procedure and drug kernels are BoCI kernels. The right-hand graph shows the ROC curves for the summed kernels using complete and truncated histories.

5.3

Feature selection

There are thousands of diagnosis, drug, and procedure codes in the database, leading to very high dimensionality for the corresponding kernels. This makes the resulting classifier difficult to interpret. In order to reduce the number of codes we considered a combination of two simple criteria to rank features. The first is the absolute value of the difference in the number of occurrences of a code in the two classes. Useful codes will tend to occur more in one class than the other. This criterion can be easily adjusted for class imbalance if it occurs. It can fail to capture features that occur a relatively small number of times but are highly predictive of one of the two classes when they do occur. Therefore we considered a second criterion based on the positive predictive value (ppv) of a code: the fraction of times it occurs in the set of positive examples out of the total number of its occurrences. The negative predictive value (npv) is defined analogously, and our second criterion is the maximum of the ppv and npv of a feature. We also put a threshold on the number of occurrences of a feature since very rare codes are not useful. Using cross-validation we searched a grid for a combination of the thresholds of the two criteria that yielded around 300 features and provided best performance over the tuning set. Using the features selected on the tuning set the performance of the combined kernel showed only a very slight drop for the AUC from 0.865 to 0.864. In future work we will perform an analysis of the codes that were selected. 5.4

Time warping

We compared the local time warping and BoC kernels on a small sample of our tuning set containing 1,000 examples, 500 from each class. The local time warping kernel provided performance that is very close to that of our BoC kernel on this data, with an AUC of 0.76, compared to an AUC of 0.77 for the stratified bag of codes kernel. But in view of its higher computational complexity we did not test it on the full dataset. This result along with the observation that the kernels that bin the procedure/diagnosis or drug codes into 6 month time segments did not perform better than those that lump the data across an enrollee’s complete history suggest that the time aspect of a claims history is not contributing to classifier performance. We believe that this property is what allows our classifier to perform nearly as well when provided with truncated claims histories. 7

6

Conclusion

We have shown that medical claims data can be used to provide good results for predicting the risk of a type 2 diabetes diagnosis. In this study, claims data was the only form of information used; to the best of our knowledge, this has not been tried before. We were able to reduce the number of diagnosis, drug, and procedure codes to a manageable number while retaining performance that was comparable to the best kernel using the full set of codes. Most importantly, we showed that these data allow us to identify enrollees that will have ICD-9-CM codes indicative of a type 2 diabetes diagnosis in the next 12 to 18 months with reasonable success. In this study we have used ICD-9CM codes to predict a diabetes diagnosis. It is well-known that identification of diabetes patients from these codes suffers from inaccuracy due to various reasons (see e.g., [12]), so verification of our approach using “gold standard” disease diagnosis should be performed. Diabetes is a costly and prevalent disease. It is somewhat unique in that it has a pre-disease state that can be detected before the onset of the disease. Overall, we see these methods as constituting a first step in an intervention program. A diabetes intervention program would give incentives to participants to lower their risk of type 2 diabetes through moderate lifestyle changes. Such programs have shown success in delaying or completely preventing the onset of type 2 diabetes. The one thing that remained was a cost effective method for identifying candidates for an intervention. The methods presented here could be used as a very cheap method of identifying possible candidates for intervention, provided that the person’s insurance claims data is available. Persons identified as “positive” could then be tested for diabetes or pre-diabetes using more expensive medical tests. The power of such a program is that we will be able to detect people at risk of type 2 diabetes, most of whom won’t get tested until it is too late. The coming era of electronic patient records promises even more data that should further improve the prediction of diabetes risk. The modularity of our method will allow one to easily extend our method when such data becomes available. Acknowledgements The authors thank Caterpillar Inc for partial funding for this project and also Dr. Syamala Srinivasan from Caterpillar Inc. for sponsoring this project and for providing us with the Marketscan database.

References [1] Centers for Disease Control and Prevention. National diabetes fact sheet: general information and national estimates on diabetes in the United States, 2005. [2] American Diabetes Association and National Institute of Diabetes, Digestive, and Kidney Diseases. The prevention or delay of type 2 diabetes. Diabetes Care, 25(4):742–749, April 2002. [3] L.N. Pani, D.N. Nathan, and R.W. Grant. Clinical predictors of disease progression and medication initiation in untreated patients with type 2 diabetes and A1C less than 7%. Diabetes Care, 31(3):386– 390, March 2008. [4] M.P. Stern, K. Williams, and S.M. Haffner. Identification of individuals at high risk of type 2 diabetes: do we need the oral glucose tolerance test? Ann Intern Med., 136(8):575–581, April 2002. [5] J.M. Geraci, M.L. Johnson, H.S. Gordon, N.J. Petersen, A.L. Shroyer, F.L. Grover, and N.P. Wray. Mortality after cardiac bypass surgery: prediction from administrative versus clinical data. Medical care, 43(2):149, 2005. [6] Y.P. Tabak, R.S. Johannes, and J.H. Silber. Using automated clinical data for risk adjustment: development and validation of six disease-specific mortality predictive models for pay-for-performance. Medical care, 45(8):789, 2007. [7] S. Schneeweiss, J.D. Seeger, M. Maclure, P.S. Wang, J. Avorn, and R.J. Glynn. Performance of comorbidity scores to control for confounding in epidemiologic studies using claims data. American Journal of Epidemiology, 154(9):854, 2001. [8] E.L. Hannan, M.J. Racz, J.G. Jollis, and E.D. Peterson. Using Medicare claims data to assess provider quality for CABG surgery: does it work well enough? Health Services Research, 31(6):659, 1997. [9] A.P. Legorreta, J.-F. Ricci, M. Markowitz, and P. Jhingran. Patients diagnosed with irritable bowel syndrome: Medical record validation of a claims-based identification algorithm. Disease Management & Health Outcomes, 10(11):715–722, 2002.

8

[10] M. Pladevall, L.K. Williams, L.A. Potts, G. Divine, H. Xi, and J.E. Lafata. Clinical outcomes and adherence to medications measured by claims data in patients with diabetes. Diabetes Care, 27(12):2800–2805, Dec 2004. [11] D.R. Miller, M.M. Safford, and L.M. Pogach. Who has diabetes? Best estimates of diabetes prevalence in the Department of Veterans Affairs based on computerized patient data. Diabetes Care, 27:B10–B21, 2004. [12] P.L. Hebert, L.S. Geiss, E.F. Tierney, M.M. Engelgau, B.P. Yawn, and A.M. McBean. Identifying persons with diabetes using Medicare claims data. American Journal of Medical Quality, 14(6):270–277, NovDec 1999. [13] J.W. Smith, J.E. Everhart, W.C. Dickson, W.C. Knowler, and R.S. Johannes. Using the ADAP learning algorithm to forecast the onset of diabetes mellitus. In 12th Annual Symposium on Computer Applications in Medical Care, pages 261 – 265, Washington DC(USA), Nov 1988. [14] A. Frank and A. Asuncion. UCI machine learning repository, 2010. R [15] Thomson Medstat. MarketScan database. [16] J. Shawe-Taylor and N. Cristianini. Kernel Methods for Pattern Analysis. Cambridge UP, Cambridge, UK, 2004. [17] B. E. Boser, I. M. Guyon, and V. N. Vapnik. A training algorithm for optimal margin classifiers. In D. Haussler, editor, 5th Annual ACM Workshop on COLT, pages 144–152, Pittsburgh, PA, 1992. ACM Press. [18] T. Joachims. Text categorization with support vector machines: learning with many relevant features. In C Nedellec and C Rouveirol, editors, Proceedings of ECML-98, 10th European Conference on Machine Learning, number 1398, pages 137–142, Chemnitz, DE, 1998. Springer Verlag, Heidelberg, DE. [19] G. Li and A. Khokhar. Content-based indexing and retrieval of audio data using wavelets. In IEEE International Conference on Multimedia and Expo (II), pages 885–888, 2000. [20] A. Ben-Hur, C.S. Ong, S. Sonnenburg, B. Sch¨olkopf, and G. R¨atsch. Support vector machines and kernels for computational biology. PLoS Computational Biology, 4(10), 2008. [21] W. S. Noble and A. Ben-Hur. Integrating information for protein function prediction. In Bioinformatics From Genomes to Therapies, volume 3, pages 1297–1314. Thomas Lengauer, 2007. [22] G. Salton, A. Wong, and C. S. Yang. A vector space model for automatic indexing. Commun. ACM, 18(11):613–620, November 1975. [23] T.F. Smith and M.S. Waterman. Identification of common molecular subsequences. Journal of Molecular Biology, 147:195–197, 1981. [24] M. Cuturi, J.P. Vert, O. Birkenes, and T. Matsui. A kernel for time series based on global alignments. In IEEE International Conference on Acoustics, Speech and Signal Processing, 2007. ICASSP 2007, volume 2, 2007.

9

Predicting the Risk of Type 2 Diabetes Using Insurance ...

need by focusing on predicting risk of type 2 diabetes using only insurance claims. .... We have chosen to model this problem in the framework of kernel methods ...

171KB Sizes 1 Downloads 141 Views

Recommend Documents

Predicting the Risk of Type 2 Diabetes Using Insurance ...
sex, ethnicity, blood pressure, body mass index, parental or sibling history of ... by different providers [8, 9]. ... The data contains the medical service uses and medical expenditures ... These include both mail-order and card program prescription

pdf diabetes type 2
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. pdf diabetes ...

Management of Type 2 Diabetes in Youth: An ... - Semantic Scholar
Sep 1, 2007 - Academy of Pediatrics does not recommend population-based screen- ... Diabetes Association recommends screening high-risk youths every two years ... KEVIN PETERSON, MD, MPH, University of Minnesota Medical School, Minneapolis, Minnesota

Management of Hyperglycemia in Type 2 Diabetes: A ...
Apr 19, 2012 - fact that patients with type 2 diabetes are at ... work is properly cited, the use is educational and not for profit, and the work is not ... diovascular disorders, blindness, end-stage ...... meaningful glycemic reduction (i.e., “no

Genomics and Proteomics of Type 2 Diabetes in ... - MyScienceWork
development of disease not only shows exponential growth but also ... associated with type 2 diabetes in Asian Indian subjects and the XA ... Among various applications, clinical .... of view, the cost of applyingthese technologies and the.

Increased glutathionylated hemoglobin (HbSSG) in type 2 diabetes ...
Jul 26, 2005 - liquid chromatography/electrospray ionization-mass spectrometry (LC/ESI-MS) for HbSSG h-globin chains. Levels of GSH and ... E-mail address: [email protected] (M. Balasubramanyam). ... into both eyes and drops were repeated until the be

Predicting the Density of Algae Communities using Local Regression ...
... the combination of features of different data analysis techniques can be useful to obtain higher predictive accuracy. ... Download as a PDF; Download as a PS ...

Perceptual Reasoning Using Interval Type-2 Fuzzy ...
[7] for word modeling [in which interval end-point data are collected from a group of ... Two fuzzy reasoning models that fit the concept of rational description are ...

Predicting Prime Numbers Using Cartesian Genetic Programming
that can map quite long sequences of natural numbers into a sequence of dis- ..... and assigned a fitness value based on the hamming distance from the perfect.

Title Predicting Unroll Factors Using Supervised ... - CiteSeerX
SPECfp benchmarks, loop unrolling has a large impact on performance. The SVM is able to ... tool for clarifying and discerning complex decision bound- aries. In this work ... with applying machine learning techniques to the problem of heuristic tunin