Workshop track - ICLR 2016

A M INIMALISTIC A PPROACH TO S UM -P RODUCT N ETWORK L EARNING FOR R EAL A PPLICATIONS Moshe Looks Google [email protected]

Viktoriya Krakovna Department of Statistics, Harvard University [email protected]

arXiv:1602.04259v3 [cs.AI] 24 Apr 2016

A BSTRACT Sum-Product Networks (SPNs) are a class of expressive yet tractable hierarchical graphical models. LearnSPN is a structure learning algorithm for SPNs that uses hierarchical co-clustering to simultaneously identifying similar entities and similar features. The original LearnSPN algorithm assumes that all the variables are discrete and there is no missing data. We introduce a practical, simplified version of LearnSPN, MiniSPN, that runs faster and can handle missing data and heterogeneous features common in real applications. We demonstrate the performance of MiniSPN on standard benchmark datasets and on two datasets from Google’s Knowledge Graph exhibiting high missingness rates and a mix of discrete and continuous features.

1

I NTRODUCTION

The Sum-Product Network (SPN) [Poon & Domingos, 2011] is a tractable and interpretable deep model. An advantage of SPNs over other graphical models such as Bayesian Networks is that they allow efficient exact inference in linear time with network size. An SPN represents a multivariate probability distribution with a directed acyclic graph consisting of sum nodes (clusters over instances), product nodes (partitions over features), and leaf nodes (univariate distributions over features), as shown in Figure 1. The standard algorithms for learning SPN structure assume discrete data with no missingness, and mostly test on the same set of benchmark data sets that satisfy these criteria [Rooshenas & Lowd, 2014]. This is not a reasonable assumption when dealing with messy data sets in real applications. The Google Knowledge Graph (KG) is a semantic network of facts, based on Freebase [Bollacker et al., 2008], used to generate Knowledge Panels in Google Search. KG data is quite heterogeneous, with a lot of it missing, since much more is known about some entities in the graph than others. High missingness rates can also worsen the impact of discretizing continuous variables before doing structure learning, which results in losing more of the already scarce covariance information. Applications like the KG are common, and there is a need for an SPN learning algorithm that can handle this kind of data. In this paper, we present MiniSPN, a simplification of a state-of-the-art SPN learning algorithm that improves its applicability, performance and speed. We demonstrate the performance of MiniSPN on KG data and on standard benchmark data sets.

2

VARIATION ON THE L EARN SPN

ALGORITHM

LearnSPN [Gens & Domingos, 2013] is a greedy algorithm that performs co-clustering by recursively partitioning variables into approximately independent sets and partitioning the training data into clusters of similar instances, as shown in Figure 2. The variable and instance partitioning steps are applied to data slices (subsets of instances and variables) produced by previous steps. The variable partition step uses pairwise independence tests on the variables, and the approximately independent sets are the connected components in the resulting dependency graph. The instance clustering step uses a naive Bayes mixture model for the clusters, where the variables in each cluster are assumed independent. The clusters are learned using hard EM with restarts, avoiding overfitting using an exponential prior on the number of clusters. The splitting process continues until the data 1

Workshop track - ICLR 2016

Figure 1: Example of an SPN structure (figure from Zhao et al. [2015])

Figure 2: Recursive partitioning process in the LearnSPN algorithm (figure from Gens & Domingos [2013])

slice has too few instances to test for independence, at which point all the variables in that slice are considered independent. The end result is a tree-structured SPN. The standard LearnSPN algorithm assumes that all the variables are discrete and there is no missing data. Hyperparameter values for the cluster penalty and the independence test critical value are determined using grid search. The clustering step seems unnecessarily complex, involving a penalty prior, EM restarts, and hyperparameter tuning. It is by far the most complicated part of the algorithm in a way that seems difficult to justify, and likely the most time-consuming due to the restarts and hyperparameter tuning. We propose a variation on LearnSPN called MiniSPN that handles missing data, performs lazy discretization of continuous data in variable partition step, simplifies the model in the instance clustering step, and does not require hyperparameter search. We simplify the naive Bayes mixture model in the instance clustering step by attempting a split into two clusters at any given point, and eliminating the cluster penalty prior, which results in a more greedy approach than in LearnSPN that does not require restarts or hyperparameter tuning. This seems like a natural choice of simplification - an extension of the greedy approach used at the top level of the LearnSPN algorithm. We compare a partition into univariate leaves to a mixture of two partitions into univariate leaves (generated using hard EM), and the split succeeds if the two-cluster version has higher validation set likelihood. If the split succeeds, we apply it to each of the two resulting data slices, and only move on to a variable partition step after the clustering step fails. The greedy approach is similar to the one used in the SPN-B method [Vergari et al., 2015], which however alternates between variable and instance splits by default, thus building even deeper SPNs. In the variable partition step, we perform an independence test using the subset of rows where both variables are not missing, and conclude independence if the number of such rows is below threshold. We apply binary binning to each continuous variable, using its median in the data slice as a cutoff. We compare to the “Pareto” algorithm, previously used for learning SPN models in KG, inspired by the work of Grosse et al. [2012]. It produces a Pareto-optimal set of models, trading off between degrees of freedom and validation set log likelihood score. At each iteration, production rules are randomly applied to add partition and mixture splits to the models in the current model set, and the new models are added to the model set. If a model in the model set has both lower degrees of freedom and higher log likelihood score than another model, the inferior model is removed from the set. The algorithm returns the model from the Pareto model set with the highest validation log likelihood. We also compare to a hybrid method, with the Pareto algorithm initialized by MiniSPN.

3

S UMMARY OF EXPERIMENTS

We use two data sets from the Knowledge Graph People collection. In the KG Professions data set, most of the variables are boolean indicators of whether each person belongs to a particular profession. There are 83 boolean variables and 4 continuous variables. In the KG Dates data set, there are 14 continuous variables representing dates of life events for each person and their spouse(s), 2

Workshop track - ICLR 2016

Table 1: Average log likelihood and runtime comparison on KG data sets (best performing methods are shown in bold). Data set Professions-10K Professions-100K Dates-10K Dates-100K

Test set log likelihood Pareto Hybrid MiniSPN -10.2 -6.2 -6.09 -6.61 -6.53 -6.44 -8.66 -8.53 -8.68 -17.1 -16.7 -16.5

Runtime (seconds) Pareto Hybrid MiniSPN 5.3 3.7 0.4 72 131 7.2 1.7 2.4 0.26 29 566 5.4

Table 2: Average log likelihood and runtime comparison on literature data sets (best performing methods are shown in bold). Data set NLTCS MSNBC KDDCup Plants Audio Jester Netflix Accidents Retail Pumsb-star DNA Kosarek MSWeb Book EachMovie WebKB Reuters-52 Newsgroup BBC Ad

Pareto -6.33 -6.54 -2.17 -17.3 -41.9 -54.6 -59.5 -40.4 -11.1 -40.8 -98.1 -11.2 -10.7 -35.1 -55 -161 -92 -156 -258 -52.3

Test set log likelihood Hybrid MiniSPN LearnSPN -6.03 -6.12 -6.1 -6.4 -6.61 -6.11 -2.13 -2.14 -2.21 -13.1 -13.2 -13 -39.9 -40 -40.5 -52.9 -53 -53.4 -56.7 -56.8 -57.3 -32.5 -32.6 -30.3 -11 -11.1 -11.09 -28.4 -28.3 -25 -91.5 -93.9 -89 -10.8 -10.9 -11 -9.94 -10.1 -10.26 -34.7 -34.7 -36.4 -52.3 -52.2 -52.5 -155 -155 -162 -85.2 -84.7 -86.5 -152 -152 -160.5 -250 -249 -250 -49.5 -49.2 -22

Pareto 4.8 61 152 28 28 13 27 31 25 47 6.3 90 75 83 62 37 76 181 33 58

Runtime (seconds) Hybrid MiniSPN 35 1.4 212 5.6 2080 23 780 11 556 12 193 6.7 766 14 1140 18 63 7.3 1100 22 45 3 537 22 572 34 181 32 218 22 169 38 656 95 1190 139 123 42 92 50

LearnSPN 60 2400 400 160 955 1190 1230 330 100 350 300 200 260 350 220 900 2900 28000 900 300

with around 95% of the data missing. We use subsets of 10000 and 100000 instances from each of these data sets, and randomly split the data sets into a training and test set. On the KG data sets, we compare MiniSPN, Pareto and Hybrid algorithms. We were not able to apply the standard LearnSPN algorithm on these data sets, since they contain missing data. Table 1 shows log likelihood performance on the test set and runtime performance. MiniSPN does better than Pareto, both in terms of log likelihood and runtime. Hybrid performs comparably to MiniSPN, but is usually the slowest of the three. We use 20 benchmark data sets from the literature (exactly the same ones used in the LearnSPN paper [Gens & Domingos, 2013]) to compare the performance of MiniSPN with the standard LearnSPN algorithm. We are particularly interested in the effect of MiniSPN’s simple two-cluster instance split relative to the more complex instance split with the exponential prior and EM restarts used in the standard LearnSPN. Table 2 shows log likelihood performance on the test set and runtime performance. Like on the KG data, we find that MiniSPN uniformly outperforms Pareto, and performs similarly to Hybrid and LearnSPN but runs much faster (on the most time-intensive data set, Newsgroup, MiniSPN takes 2 minutes while LearnSPN takes 8 hours).

4

C ONCLUSION

Sum-product networks have been receiving increasing attention from researchers due to their expressiveness, efficient inference and interpretability, and many learning algorithms have been developed in the past few years. While recent developments have mostly focused on improving performance on benchmark data sets, our variation on a classical learning algorithm is simple yet has a large impact on usability, by improving speed and making it possible to apply to messy real data sets. 3

Workshop track - ICLR 2016

R EFERENCES Kurt Bollacker, Colin Evans, Praveen Paritosh, Tim Sturge, and Jamie Taylor. Freebase: a collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD international conference on Management of data, pp. 1247–1250. ACM, 2008. Robert Gens and Pedro M. Domingos. Learning the Structure of Sum-Product Networks. In Proceedings of the 30th International Conference on Machine Learning, ICML 2013, Atlanta, GA, USA, 16-21 June 2013, pp. 873–880, 2013. Roger B. Grosse, Ruslan Salakhutdinov, William. T. Freeman, and Joshua B. Tenenbaum. Exploiting compositionality to explore a large space of model structures. In Proceedings of the 28th Conference on Uncertainty in AI (UAI), 2012. Hoifung Poon and Pedro M. Domingos. Sum-Product Networks: A New Deep Architecture. In UAI 2011, Proceedings of the Twenty-Seventh Conference on Uncertainty in Artificial Intelligence, Barcelona, Spain, July 14-17, 2011, pp. 337–346, 2011. Amirmohammad Rooshenas and Daniel Lowd. Learning Sum-Product Networks with Direct and Indirect Variable Interactions. In Tony Jebara and Eric P. Xing (eds.), Proceedings of the 31st International Conference on Machine Learning (ICML-14), pp. 710–718. JMLR Workshop and Conference Proceedings, 2014. Antonio Vergari, Nicola Di Mauro, and Floriana Esposito. Simplifying, Regularizing and Strengthening Sum-Product Network Structure Learning. In Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2015, Porto, Portugal, September 7-11, 2015, Proceedings, Part II, pp. 343–358, 2015. Han Zhao, Mazen Melibari, and Pascal Poupart. On the Relationship between Sum-Product Networks and Bayesian Networks. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015, pp. 116–124, 2015.

4

arXiv:1602.04259v3 [cs.AI] 24 Apr 2016 - Research at Google

using an exponential prior on the number of clusters. The splitting process continues until the data. 1. arXiv:1602.04259v3 [cs.AI] 24 Apr 2016 ...

179KB Sizes 2 Downloads 156 Views

Recommend Documents

arXiv:1502.05678v2 [cs.CV] 17 Apr 2015 - Research at Google
which image depicts the person in the most important role? We introduce a .... Social networking sites .... pair, ten AMT workers were asked to pick the more im-.

arXiv:0704.3250v1 [nlin.CD] 24 Apr 2007
Apr 24, 2007 - assumed similar to the cyclotron type motion in constant magnetic fields, (Van. Allen and ... done using a mixture of analytical and numerical techniques. Here, we are in- terested in surveying the exact results on the Störmer problem

2016 Apr Minutes.pdf
Page 1 of 3. Tanque Verde Elementary School Parent Teacher Group. April 14. th. , 2017. Call to order, 8:30 by Andrea Johnson. Roll Call. Board Members: ...

expires apr. 30, 2016 - nccdn.net
Page 1. EXPIRES APR. 30, 2016.

Mathematics at - Research at Google
Index. 1. How Google started. 2. PageRank. 3. Gallery of Mathematics. 4. Questions ... http://www.google.es/intl/es/about/corporate/company/history.html. ○.

APR 33-2016.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. APR 33-2016.

BPH4Sem Apr 2016 Supp.pdf
Page 1 of 6. (BPH 40112). B.Pharmacy DEGREE I~XAMINATlON,ArHIL 20] G. :3. (a) Classify Antimalariala, discuss their SAR,. therapeutic uses and write the synthesis of. pyrimethamine. Or. (Supplementary) (b) What do you know of. Fourth Semester (i) Ant

arXiv:1611.04482v1 [cs.CR] 14 Nov 2016 - Research at Google
Nov 14, 2016 - across user-held training data on mobile devices, using Secure Aggregation to .... case and a more formal security analysis to the full version.

BPH6Sem Apr 2016 Supp.pdf
(c) BCG vaccine. (d) Mantoux test. 4. (a) Write the principle and procedure for the. microbiological assay of penicillin. Add note. on importance of microbial limit ...

arXiv:1606.05978v1 [cs.IR] 20 Jun 2016 - Research at Google
Jun 20, 2016 - mitted independently and (2) to follow a constant rate λ, which re- sults in a .... blogs and comments on the Youtube2[20]. We remind its .... For brevity, we show only the top 12 most prolific users, but ...... monitoring in news.

Faucet - Research at Google
infrastructure, allowing new network services and bug fixes to be rapidly and safely .... as shown in figure 1, realizing the benefits of SDN in that network without ...

BeyondCorp - Research at Google
41, NO. 1 www.usenix.org. BeyondCorp. Design to Deployment at Google ... internal networks and external networks to be completely untrusted, and ... the Trust Inferer, Device Inventory Service, Access Control Engine, Access Policy, Gate-.

VP8 - Research at Google
coding and parallel processing friendly data partitioning; section 8 .... 4. REFERENCE FRAMES. VP8 uses three types of reference frames for inter prediction: ...

JSWhiz - Research at Google
Feb 27, 2013 - and delete memory allocation API requiring matching calls. This situation is further ... process to find memory leaks in Section 3. In this section we ... bile devices, such as Chromebooks or mobile tablets, which typically have less .

Yiddish - Research at Google
translation system for these language pairs, although online dictionaries exist. ..... http://www.unesco.org/culture/ich/index.php?pg=00206. Haifeng Wang, Hua ...

traits.js - Research at Google
on the first page. To copy otherwise, to republish, to post on servers or to redistribute ..... quite pleasant to use as a library without dedicated syntax. Nevertheless ...

sysadmin - Research at Google
On-call/pager response is critical to the immediate health of the service, and ... Resolving each on-call incident takes between minutes ..... The conference has.

Introduction - Research at Google
Although most state-of-the-art approaches to speech recognition are based on the use of. HMMs and .... Figure 1.1 Illustration of the notion of margin. additional ...

References - Research at Google
A. Blum and J. Hartline. Near-Optimal Online Auctions. ... Sponsored search auctions via machine learning. ... Envy-Free Auction for Digital Goods. In Proc. of 4th ...

BeyondCorp - Research at Google
Dec 6, 2014 - Rather, one should assume that an internal network is as fraught with danger as .... service-level authorization to enterprise applications on a.

Browse - Research at Google
tion rates, including website popularity (top web- .... Several of the Internet's most popular web- sites .... can't capture search, e-mail, or social media when they ..... 10%. N/A. Table 2: HTTPS support among each set of websites, February 2017.

Continuous Pipelines at Google - Research at Google
May 12, 2015 - Origin of the Pipeline Design Pattern. Initial Effect of Big Data on the Simple Pipeline Pattern. Challenges to the Periodic Pipeline Pattern.

Accuracy at the Top - Research at Google
We define an algorithm optimizing a convex surrogate of the ... as search engines or recommendation systems, since most users of these systems browse or ...