t-Logistic Regression Nan Ding2 , S.V. N. Vishwanathan1,2 Departments of 1 Statistics and 2 Computer Science Purdue University [email protected], [email protected]

Abstract We extend logistic regression by using t-exponential families which were introduced recently in statistical physics. This gives rise to a regularized risk minimization problem with a non-convex loss function. An efficient block coordinate descent optimization scheme can be derived for estimating the parameters. Because of the nature of the loss function, our algorithm is tolerant to label noise. Furthermore, unlike other algorithms which employ non-convex loss functions, our algorithm is fairly robust to the choice of initial values. We verify both these observations empirically on a number of synthetic and real datasets.

1

Introduction

Many machine learning algorithms minimize a regularized risk [1]: m

J(θ) = Ω(θ) + Remp (θ), where Remp (θ) =

1 X l(xi , yi , θ). m i=1

(1)

Here, Ω is a regularizer which penalizes complex θ; and Remp , the empirical risk, is obtained by averaging the loss l over the training dataset {(x1 , y1 ), . . . , (xm , ym )}. In this paper our focus is on binary classification, wherein features of a data point x are extracted via a feature map φ and the label is usually predicted via sign(hφ(x), θi). If we define the margin of a training example (x, y) as u(x, y, θ) := y hφ(x), θi, then many popular loss functions for binary classification can be written as functions of the margin. Examples include1 l(u) = 0 if u > 0 and 1 otherwise . l(u) = max(0, 1 − u) l(u) = exp(−u) l(u) = log(1 + exp(−u))

(0 − 1 loss) (Hinge Loss) (Exponential Loss) (Logistic Loss).

(2) (3) (4) (5)

The 0 − 1 loss is non-convex and difficult to handle; it has been shown that it is NP-hard to even approximately minimize the regularized risk with the 0 − 1 loss [2]. Therefore, other loss functions can be viewed as convex proxies of the 0 − 1 loss. Hinge loss leads to support vector machines (SVMs), exponential loss is used in Adaboost, and logistic regression uses the logistic loss. Convexity is a very attractive property because it ensures that the regularized risk minimization problem has a unique global optimum [3]. However, as was recently shown by Long and Servedio [4], learning algorithms based on convex loss functions are not robust to noise2 . Intuitively, the convex loss functions grows at least linearly with slope |l0 (0)| as u ∈ (−∞, 0), which introduces the overwhelming impact from the data with u  0. There has been some recent and some notso-recent work on using non-convex loss functions to alleviate the above problem. For instance, a recent manuscript by [5] uses the cdf of the Guassian distribution to define a non-convex loss. 1

We slightly abuse notation and use l(u) to denote l(u(x, y, θ)). Although, the analysis of [4] is carried out in the context of boosting, we believe, the results hold for a larger class of algorithms which minimize a regularized risk with a convex loss function. 2

1

In this paper, we continue this line of inquiry and propose a non-convex loss function which is firmly grounded in probability theory. By loss extending logistic regression from the exLogistic exp ponential family to the t-exponential fam6 ily, a natural extension of exponential family Hinge of distributions studied in statistical physics [6–10], we obtain the t-logistic regression 4 algorithm. Furthermore, we show that a simple block coordinate descent scheme can be used to solve the resultant regularized 2 0-1 loss risk minimization problem. Analysis of this procedure also intuitively explains why tmargin logistic regression is able to handle label -4 -2 0 2 4 noise.

Figure 1: Some commonly used loss functions for binary

Our paper is structured as follows: In sec- classification. The 0-1 loss is non-convex. The hinge, expotion 2 we briefly review logistic regression nential, and logistic losses are convex upper bounds of the especially in the context of exponential fam- 0-1 loss. ilies. In section 3, we review t-exponential families, which form the basis for our proposed t-logistic regression algorithm introduced in section 4. In section 5 we utilize ideas from convex multiplicative programming to design an optimization strategy. Experiments that compare our new approach to existing algorithms on a number of publicly available datasets are reported in section 6, and the paper concludes with a discussion and outlook in section 7. Some technical details as well as extra experimental results can be found in the supplementary material.

2

Logistic Regression

Since we build upon the probabilistic underpinnings of logistic regression, we briefly review some salient concepts. Details can be found in any standard textbook such as [11] or [12]. Assume we are given a labeled dataset (X, Y) = {(x1 , y1 ), . . . , (xm , ym )} with the xi ’s drawn from some domain X and the labels yi ∈ {±1}. Given a family of conditional distributions parameterized by θ, using Bayes rule, and making a standard iid assumption about the data allows us to write p(θ | X, Y) = p(θ)

m Y

p(yi | xi ; θ)/p(Y | X) ∝ p(θ)

i=1

m Y

p(yi | xi ; θ)

(6)

i=1

where p(Y | X) is clearly independent of θ. To model p(yi | xi ; θ), consider the conditional exponential family of distributions p(y| x; θ) = exp (hφ(x, y), θi − g(θ | x)) ,

(7)

with the log-partition function g(θ | x) given by g(θ | x) = log (exp (hφ(x, +1), θi) + exp (hφ(x, −1), θi)) . If we choose the feature map φ(x, y) = that p(y| x; θ) is the logistic function p(y| x; θ) =

y 2 φ(x),

(8)

and denote u = y hφ(x), θi then it is easy to see

exp(u/2) 1 = . exp(u/2) + exp(−u/2) 1 + exp(−u)

(9)

By assuming a zero mean isotropic Gaussian prior N (0, √1λ I) for θ, plugging in (9), and taking logarithms, we can rewrite (6) as m X λ 2 − log p(θ | X, Y) = kθk + log (1 + exp (−yi hφ(xi ), θi)) + const. . 2 i=1

(10)

Logistic regression computes a maximum a-posteriori (MAP) estimate for θ by minimizing (10) as a function of θ. Comparing (1) and (10) it is easy to see that the regularizer employed in logistic 2 regression is λ2 kθk , while the loss function is the negative log-likelihood − log p(y| x; θ), which thanks to (9) can be identified with the logistic loss (5). 2

3

t-Exponential family of Distributions

In this section we will look at generalizations of the log and exp functions which were first introduced in statistical physics [6–9]. Some extensions and machine learning applications were presented in [13]. In fact, a more general class of functions was studied in these publications, but for our purposes we will restrict our attention to the so-called t-exponential and t-logarithm functions. The t-exponential function expt for (0 < t < 2) is defined as follows: ( exp(x) if t = 1 expt (x) := 1/(1−t) [1 + (1 − t)x]+ otherwise.

(11)

where (·)+ = max(·, 0). Some examples are shown in Figure 2. Clearly, expt generalizes the usual exp function, which is recovered in the limit as t → 1. Furthermore, many familiar properties of exp are preserved: expt functions are convex, non-decreasing, non-negative and satisfy expt (0) = 1 [9]. But expt does not preserve one very important property of exp, namely expt (a + b) 6= expt (a) · expt (b). One can also define the inverse of expt namely logt as  log(x) if t = 1  logt (x) := (12) x1−t − 1 /(1 − t) otherwise. Similarly, logt (ab) 6= logt (a) + logt (b). From Figure 2, it is clear that expt decays towards 0 more slowly than the exp function for 1 < t < 2. This important property leads to a family of heavy tailed distributions which we will later exploit. t = 1 (logistic) expt t = 1.5 exp(x) 7 logt t→0 6 t = 0.5 t = 0.5 5 t = 1.3 log(x) 2 4 t = 1.6 t→0 t = 1.5 1 t = 1.9 3 x 0 2 -1 1 2 3 4 5 6 7 0-1 loss 1 -2 -3 -2 -1 0 1 2 x -3 -4 -2

loss 6 4 2 margin 0

2

4

Figure 2: Left: expt and Middle: logt for various values of t indicated. The right figure depicts the t-logistic loss functions for different values of t. When t = 1, we recover the logistic loss Analogous to the exponential family of distributions, the t-exponential family of distributions is defined as [9, 13]: p(x; θ) := expt (hφ(x), θi − gt (θ)) .

(13)

A prominent member of the t-exponential family is the Student’s-t distribution [14]. Just like in the exponential family case, gt the log-partition function ensures that p(x; θ) is normalized. However, no closed form solution exists for computing gt exactly in general. A closely related distribution, which often appears when working with t-exponential families is the so-called escort distribution [9, 13]: qt (x; θ) := p(x; θ)t /Z(θ) where Z(θ) = integrates to 1.

R

(14)

t

p(x; θ) dx is the normalizing constant which ensures that the escort distribution

Although gt (θ) is not the cumulant function of the t-exponential family, it still preserves convexity. In addition, it is very close to being a moment generating function ∇θ gt (θ) = Eqt (x;θ) [φ(x)] .

(15)

The proof is provided in the supplementary material. A general version of this result appears as Lemma 3.8 in Sears [13] and a version specialized to the generalized exponential families appears as Proposition 5.2 in [9]. The main difference from ∇θ g(θ) of the normal exponential family is that now ∇θ gt (θ) is equal to the expectation of its escort distribution qt (x; θ) instead of p(x; θ). 3

4

Binary Classification with the t-exponential Family

In t-logistic regression we model p(y| x; θ) via a conditional t-exponential family distribution p(y| x; θ) = expt (hφ(x, y), θi − gt (θ | x)) , (16) where 1 < t < 2, and compute the log-partition function gt by noting that expt (hφ(x, +1), θi − gt (θ | x)) + expt (hφ(x, −1), θi − gt (θ | x)) = 1. (17) Even though no closed form solution exists, one can compute gt given θ and x using numerical techniques efficiently. The Student’s-t distribution can be regarded as a counterpart of the isotropic Gaussian prior in the t-exponential family [14]. Recall that a one dimensional Student’s-t distribution is given by  −(v+1)/2 Γ((v + 1)/2) (x − µ)2 St(x|µ, σ, v) = √ , (18) 1 + vσ vπΓ(v/2)σ 1/2 where Γ(·) denotes the usual Gamma function and v > 1 so that the mean is finite. If we select t satisfying −(v + 1)/2 = 1/(1 − t) and denote,  −2/(v+1) Γ((v + 1)/2) Ψ= √ , vπΓ(v/2)σ 1/2 then by some simple but tedious calculation (included in the supplementary material) ˜ − µ)2 /2 − g˜t ) St(x|µ, σ, v) = exp (−λ(x (19) t

2Ψ ˜= where λ (t − 1)vσ

and

Ψ−1 . t−1

g˜t =

Therefore, we work with the Student’s-t prior in our setting: p(θ) =

d Y

p(θj ) =

d Y

St(θj |0, 2/λ, (3 − t)/(t − 1)).

(20)

j=1

j=1

Here, the degree of freedom for Student’s-t distribution is chosen such that it also belongs to the expt family, which in turn yields v = (3 − t)/(t − 1). The Student’s-t prior is usually preferred to the Gaussian prior when the underlying distribution is heavy-tailed. In practice, it is known to be a robust3 alternative to the Gaussian distribution [16, 17]. As before, if we let φ(x, y) = y2 φ(x) and plot the negative log-likelihood − log p(y| x; θ), then we no longer obtain a convex loss function (see Figure 2). Similarly, − log p(θ) is no longer convex when we use the Student’s-t prior. This makes optimizing the regularized risk challenging, therefore we employ a different strategy. Since logt is also a monotonically increasing function, instead of working with log, we can equivalently work with the logt function (12) and minimize the following objective function: m Y ˆ J(θ) = − logt p(θ) p(yi | xi ; θ)/p(Y | X) i=1

1 = t−1

p(θ)

m Y

!1−t p(yi | xi ; θ)/p(Y | X)

+

i=1

1 , 1−t

(21)

where p(Y | X) is independent of θ. Using (13), (18), and (11), we can further write d  m  Y Dy E  Y i ˜ 2 /2 − g˜t ) ˆ J(θ) ∝ 1 + (1 − t)(−λθ 1 + (1 − t)( φ(xi ), θ − gt (θ | xi )) +const. . j 2 {z {z } i=1 | } j=1 | rj (θ)

=

d Y j=1

rj (θ)

m Y

li (θ)

li (θ) + const.

(22)

i=1

3 There is no unique definition of robustness. For example, one of the definitions is through the outlierproneness [15]: p(θ | X, Y, xn+1 , yn+1 ) → p(θ | X, Y) as xn+1 → ∞.

4

Since t > 1, it is easy to see that rj (θ) > 0 is a convex function of θ. On the other hand, since gt ˆ is convex and t > 1 it follows that li (θ) > 0 is also a convex function of θ. In summary, J(θ) is a product of positive convex functions. In the next section we will present an efficient optimization strategy for dealing with such problems.

5

Convex Multiplicative Programming

In convex multiplicative programming [18] we are interested in the following optimization problem: N Y min P(θ) , zn (θ) s.t. θ ∈ Rd , (23) θ

n=1

where zn (θ) are positive convex functions. Clearly, (22) can be identified with (23) by setting N = d+m and identifying zn (θ) = rn (θ) for n = 1, . . . , d and zn+d (θ) = ln (θ) for n = 1, . . . , m. The optimal solutions to the problem (23) can be obtained by solving the following parametric problem (see Theorem 2.1 of Kuno et al. [18]): N N X Y min min MP(θ, ξ) , ξn zn (θ) s.t. θ ∈ Rd , ξ > 0, ξn ≥ 1. (24) ξ

θ

n=1

n=1

The optimization problem In logistic regression,

in (24) is very reminiscent of logistic regression.

 ln (θ) = − y2n φ(xn ), θ + g(θ | xn ), while here ln (θ) = 1 + (1 − t) y2n φ(xn ), θ − gt (θ | xn ) . The key difference is that in t-logistic regression each data point xn has a weight (or influence) ξn associated with it. Exact algorithms have been proposed for solving (24) (for instance, [18]). However, the computational cost of these algorithms grows exponentially with respect to N which makes them impractical for our purposes. Instead, we apply a block coordinate descent based method. The main idea is to minimize (24) with respect to θ and ξ separately. ξ-Step: Assume that θ is fixed, and denote z˜n = zn (θ) to rewrite (24) as: min ξ

N X

ξn z˜n s.t.

ξ > 0,

N Y

ξn ≥ 1.

(25)

n=1

n=1

Since the objective function is linear in ξ and the feasible region is a convex set, (25) is a convex optimization problem. By introducing a non-negative Lagrange multiplier γ ≥ 0, the partial Lagrangian and its gradient with respect to ξn0 can be written as ! N N X Y L(ξ, γ) = ξn z˜n + γ · 1 − ξn (26) n=1

n=1

Y ∂ L(ξ, γ) = z˜n0 − γ ξn . 0 ∂ξn 0

(27)

n6=n

Setting the gradient to 0 obtains γ =

Q z˜n0

ξn . Since QN that n=1 ξn

n6=n0

K.K.T. conditions [3], we can conclude

z˜n0 > 0, it follows that γ cannot be 0. By the = 1. This in turn implies that γ = z˜n0 ξn0 or

(ξ1 , . . . , ξN ) = (γ/˜ z1 , . . . , γ/˜ zN ), with γ =

N Y

1

z˜nN .

(28)

n=1

Recall that ξn in (24) is the weight (or influence) of each term zn (θ). The above analysis shows that γ = z˜n (θ)ξn remains constant for all n. If z˜n (θ) becomes very large then its influence ξn is reduced. Therefore, points with very large loss have their influence capped and this makes the algorithm robust to outliers. θ-Step: In this step we fix ξ > 0 and solve for the optimal θ. This step is essentially the same as logistic regression, except that each component has a weight ξ here. N X min ξn zn (θ) s.t. θ ∈ Rd . (29) θ

n=1

5

This is a standard unconstrained convex optimization problem which can be solved by any off the shelf solver. In our case we use the L-BFGS Quasi-Newton method. This requires us to compute the gradient ∇θ zn (θ): ˜ n · en ∇θ zn (θ) = ∇θ rn (θ) = (t − 1)λθ y  n for n = 1, . . . , m ∇θ zn+d (θ) = ∇θ ln (θ) = (1 − t) φ(xn ) − ∇θ gt (θ | xn ) hy i  y2 n n φ(xn ) − Eqt (yn | xn ;θ) φ(xn ) , = (1 − t) 2 2 where en denotes the d dimensional vector with one at the n-th coordinate and zeros elsewhere (n-th unit vector). qt (y| x; θ) is the escort distribution of p(y| x; θ) (16): for n = 1, . . . , d

qt (y| x; θ) =

p(y| x; θ)t . p(+1| x; θ)t + p(−1| x; θ)t

(30)

The objective function is monotonically decreasing and is guaranteed to converge to a stable point of P(θ). We include the proof in the supplementary material.

6

Experimental Evaluation

Our experimental evaluation is designed to answer four natural questions: 1) How does the generalization capability (measured in terms of test error) of t-logistic regression compare with existing algorithms such as logistic regression and support vector machines (SVMs) both in the presence and absence of label noise? 2) Do the ξ variables we introduced in the previous section have a natural interpretation? 3) How much overhead does t-logistic regression incur as compared to logistic regression? 4) How sensitive is the algorithm to initialization? The last question is particularly important given that the algorithm is minimizing a non-convex loss. To answer the above questions empirically we use six datasets, two of which are synthetic. The Long-Servedio dataset is an artificially constructed dataset to show that algorithms which minimize a differentiable convex loss are not tolerant to label noise Long and Servedio [4]. The examples have 21 dimensions and play one of three possible roles: large margin examples (25%, x1,2,...,21 = y); pullers (25%, x1,...,11 = y, x12,...,21 = −y); and penalizers (50%, Randomly select and set 5 of the first 11 coordinates and 6 out of the last 10 coordinates to y, and set the remaining coordinates to −y). The Mease-Wyner is another synthetic dataset to test the effect of label noise. The input x is a 20-dimensional vector where each coordinate is uniformly distributed on [0, 1]. The label y is P5 +1 if j=1 xj ≥ 2.5 and −1 otherwise [19]. In addition, we also test on Mushroom, USPS-N (9 vs. others), Adult, and Web datasets, which are often used to evaluate machine learning algorithms (see Table 1 in supplementary material for details). For simplicity, we use the identity feature map φ(x) = x in all our experiments, and set t ∈ {1.3, 1.6, 1.9} for t-logistic regression. Our comparators are logistic regression, linear SVMs4 , and an algorithm (the probit) which employs the probit loss, L(u) = 1 − erf (2u), used in BrownBoost/RobustBoost [5]. We use the L-BFGS algorithm [21] for the θ-step in t-logistic regression. L-BFGS is also used to train logistic regression and the probit loss based algorithms. Label noise is added by randomly choosing 10% of the labels in the training set and flipping them; each dataset is tested with and without label noise. We randomly select and hold out 30% of each dataset as a validation set and use the rest of the 70% for 10-fold cross validation. The optimal parameters namely λ for t-logistic and logistic regression and C for SVMs is chosen by performing a grid search over the parameter space 2−7,−6,...,7 and observing the prediction accuracy over the validation set. The convergence criterion is to stop when the change in the objective function value is less than 10−4 . All code is written in Matlab, and for the linear SVM we use the Matlab interface of LibSVM [22]. Experiments were performed on a Qual-core machine with Dual 2.5 Ghz processor and 32 Gb RAM. In Figure 3, we plot the test error with and without label noise. In the latter case, the test error of t-logistic regression is very similar to logistic regression and Linear SVM (with 0% test error in 4 We also experimented with RampSVM [20], however, the results are worser than the other algorithms. We therefore report these results in the supplementary material.

6

6.0

TestError(%)

32

1.2 4.5

24

0.9 3.0

16 8

1.5

0.3

0

0.0

0.0

16.8

3.2

16.0

2.4

6.0

TestError(%)

0.6

4.5 3.0

1.6

15.2 1.5

0.8 14.4

0.0

logis. t=1.3 t=1.6 t=1.9 probit SVM

logis. t=1.3 t=1.6 t=1.9 probit SVM

0.0

logis. t=1.3 t=1.6 t=1.9 probit SVM

Figure 3: The test error rate of various algorithms on six datasets (left to right, top: Long-Servedio, Mease-Wyner, Mushroom; bottom: USPS-N, Adult, Web) with and without 10% label noise. All algorithms are initialized with θ = 0. The blue (light) bar denotes a clean dataset while the magenta (dark) bar are the results with label noise added. Also see Table 3 in the supplementary material.

Long-Servedio and Mushroom datasets), with a slight edge on some datasets such as Mease-Wyner. When label noise is added, t-logistic regression (especially with t = 1.9) shows significantly5 better performance than all the other algorithms on all datasets except the USPS-N, where it is marginally outperformed by the probit. To obtain Figure 4 we used the noisy version of the datasets, chose one of the 10 folds used in the previous experiment, and plotted the distribution of the 1/z ∝ ξ obtained after training with t = 1.9. To distinguish the points with noisy labels we plot them in cyan while the other points are plotted in red. Analogous plots for other values of t can be found in the supplementary material. Recall that ξ denotes the influence of a point. One can clearly observe that the ξ of the noisy data is much smaller than that of the clean data, which indicates that the algorithm is able to effectively identify these points and cap their influence. In particular, on the Long-Servedio dataset observe the 4 distinct spikes. From left to right, the first spike corresponds to the noisy large margin examples, the second spike represents the noisy pullers, the third spike denotes the clean pullers, while the rightmost spike corresponds to the clean large margin examples. Clearly, the noisy large margin examples and the noisy pullers are assigned a low value of ξ thus capping their influence and leading to the perfect classification of the test set. On the other hand, logistic regression is unable to discriminate between clean and noisy training samples which leads to bad performance on noisy datasets. Detailed timing experiments can be found in Table 4 in the supplementary material. In a nutshell, t-logistic regression takes longer to train than either logistic regression or the probit. The reasons are not difficult to see. First, there is no closed form expression for gt (θ | x). We therefore resort to pre-computing it at some fixed locations and using a spline method to interpolate values at other locations. Second, since the objective function is not convex several iterations of the ξ and θ steps might be needed. Surprisingly, the L-BFGS algorithm, which is not designed to optimize nonconvex functions, is able to minimize (22) directly in many cases. When it does converge, it is often faster than the convex multiplicative programming algorithm. However, on some cases (as expected) it fails to find a direction of descent and exits. A common remedy for this is the bundle L-BFGS with a trust-region approach. [21] Given that the t-logistic objective function is non-convex, one naturally worries about how different initial values affect the quality of the final solution. To answer this question, we initialized the algorithm with 50 different randomly chosen θ ∈ [−0.5, 0.5]d , and report test performances of the various solutions obtained in Figure 5. Just like logistic regression which uses a convex loss and hence converges to the same solution independent of the initialization, the solution obtained 5

We provide the significance test results in Table 2 of supplementary material.

7

300

1000 60

Frequency

240

800

45

180

600

120

30

400

60

15

200

0 0.0

0.2

0.4

0.6

0.8

0 0.0

1.0

0.2

0.4

0.6

0.8

1.0

0 0.0

0.2

0.4

0.2

0.4

0.6

0.8

1.0

0.6

0.8

1.0

600 1200

8000

900

6000

600

4000

300

2000

Frequency

450 300 150 0 0.0

0.2

0.4

0.6

0.8

1.0

0 0.0

0.2

0.4

ξ

0.6

0.8

1.0

0 0.0

ξ

ξ

Figure 4: The distribution of ξ obtained after training t-logistic regression with t = 1.9 on datasets with 10% label noise. Left to right, top: Long-Servedio, Mease-Wyner, Mushroom; bottom: USPSN, Adult, Web. The red (dark) bars (resp. cyan (light) bars) indicate the frequency of ξ assigned to points without (resp. with) label noise. by t-logistic regression seems fairly independent of the initial value of θ. On the other hand, the performance of the probit fluctuates widely with different initial values of θ. probit t = 1.9 t = 1.6 t = 1.3 logistic 0

10

20

30

0

10

20

30

40

0.00

0.15

0.30

0.45

probit t = 1.9 t = 1.6 t = 1.3 logistic 3.0

4.5

6.0 7.5 TestError(%)

9.0

15

18 21 TestError(%)

24

1.5

2.0 2.5 TestError(%)

3.0

3.5

Figure 5: The Error rate by different initialization. Left to right, top: Long-Servedio, Mease-Wyner, Mushroom; bottom: USPS-N, Adult, Web.

7

Discussion and Outlook

In this paper, we generalize logistic regression to t-logistic regression by using the t-exponential family. The new algorithm has a probabilistic interpretation and is more robust to label noise. Even though the resulting objective function is non-convex, empirically it appears to be insensitive to initialization. There are a number of avenues for future work. On Long-Servedio experiment, if the label noise is increased significantly beyond 10%, the performance of t-logistic regression may degrade (see Fig. 6 in supplementary materials). Understanding and explaining this issue theoretically and empirically remains an open problem. It will be interesting to investigate if t-logistic regression can be married with graphical models to yield t-conditional random fields. We will also focus on better numerical techniques to accelerate the θ-step, especially a faster way to compute gt . 8

References [1] Choon Hui Teo, S. V. N. Vishwanthan, Alex J. Smola, and Quoc V. Le. Bundle methods for regularized risk minimization. J. Mach. Learn. Res., 11:311–365, January 2010. [2] S. Ben-David, N. Eiron, and P.M. Long. On the difficulty of approximately maximizing agreements. J. Comput. System Sci., 66(3):496–514, 2003. [3] S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University Press, Cambridge, England, 2004. [4] Phil Long and Rocco Servedio. Random classification noise defeats all convex potential boosters. Machine Learning Journal, 78(3):287–304, 2010. [5] Yoav Freund. A more robust boosting algorithm. Technical Report Arxiv/0905.2138, Arxiv, May 2009. [6] J. Naudts. Deformed exponentials and logarithms in generalized thermostatistics. Physica A, 316:323–334, 2002. URL http://arxiv.org/pdf/cond-mat/0203489. [7] J. Naudts. Generalized thermostatistics based on deformed exponential and logarithmic functions. Physica A, 340:32–40, 2004. [8] J. Naudts. Generalized thermostatistics and mean-field theory. Physica A, 332:279–300, 2004. [9] J. Naudts. Estimators, escort proabilities, and φ-exponential families in statistical physics. Journal of Inequalities in Pure and Applied Mathematics, 5(4), 2004. [10] C. Tsallis. Possible generalization of boltzmann-gibbs statistics. J. Stat. Phys., 52, 1988. [11] Christopher Bishop. Pattern Recognition and Machine Learning. Springer, 2006. [12] Trevor Hastie, Robert Tibshirani, and Jerome Friedman. The Elements of Statistical Learning. Springer, New York, 2 edition, 2009. [13] Timothy D. Sears. Generalized Maximum Entropy, Convexity, and Machine Learning. PhD thesis, Australian National University, 2008. [14] Andre Sousa and Constantino Tsallis. Student’s t- and r-distributions: Unified derivation from an entropic variational principle. Physica A, 236:52–57, 1994. [15] A O’hagan. On outlier rejection phenomena in bayes inference. Royal Statistical Society, 41 (3):358–367, 1979. [16] Kenneth L. Lange, Roderick J. A. Little, and Jeremy M. G. Taylor. Robust statistical modeling using the t distribution. Journal of the American Statistical Association, 84(408):881–896, 1989. [17] J. Vanhatalo, P. Jylanki, and A. Vehtari. Gaussian process regression with student-t likelihood. In Neural Information Processing System, 2009. [18] Takahito Kuno, Yasutoshi Yajima, and Hiroshi Konno. An outer approximation method for minimizing the product of several convex functions on a convex set. Journal of Global Optimization, 3(3):325–335, September 1993. [19] David Mease and Abraham Wyner. Evidence contrary to the statistical view of boosting. J. Mach. Learn. Res., 9:131–156, February 2008. [20] R. Collobert, F.H. Sinz, J. Weston, and L. Bottou. Trading convexity for scalability. In W.W. Cohen and A. Moore, editors, Machine Learning, Proceedings of the Twenty-Third International Conference (ICML 2006), pages 201–208. ACM, 2006. [21] J. Nocedal and S. J. Wright. Numerical Optimization. Springer Series in Operations Research. Springer, 1999. [22] C.C. Chang and C.J. Lin. LIBSVM: a library for support vector machines, 2001. Software available at http://www.csie.ntu.edu.tw/˜cjlin/libsvm. [23] Fabian Sinz. UniverSVM: Support Vector Machine with Large Scale CCCP Functionality, 2006. Software available at http://www.kyb.mpg.de/bs/people/fabee/universvm. html.

9

A

Student’s-t Distribution

Recall that a k-dimensional Student’s-t distribution St(x|µ, Σ, v) with 1 < v < +∞ degrees of freedom has the following probability density function: Γ ((v + k)/2)

St(x|µ, Σ, v) =

k/2

(πv)

Γ(v/2)|Σ|1/2

−(v+k)/2 1 + (x − µ)> (vΣ)−1 (x − µ) .

(31)

Here Γ(·) denotes the usual Gamma function. In fact, the Student’s-t distribution is a member of the t-exponential family. To see this we first set −(v + k)/2 = 1/(1 − t) and Γ ((v + k)/2)

Ψ=

(πv)

k/2

!−2/(v+k)

Γ(v/2)|Σ|1/2

to rewrite (31) as 1/(1−t) St(x|µ, Σ, v) = Ψ + Ψ · (x − µ)> (vΣ)−1 (x − µ) .

(32)

Next we set φ(x) = [x; xx> ], θ = [θ1 , θ2 ], where K = (vΣ)−1 and θ1 = −2ΨKµ/(1 − t) while θ2 = ΨK/(1 − t). Then we define    Ψ hφ(x), θi = x> Kx − 2µ> Kx and 1−t    Ψ 1 gt (θ) = − µ> Kµ + 1 + 1−t 1−t to rewrite (32) as 1/(1−t)

St(x|µ, Σ, v) = (1 + (1 − t) (hφ(x), θi − gt (θ)))

.

Comparing with (11) clearly shows that St(x|µ, Σ, v) = pt (x; θ) = expt (hφ(x), θi − gt (θ)) . Furthermore, using this fact and some simple algebra yields the escort distribution of Student’s-t distribution: qt (x; θ) = St(x|µ, vΣ/(v + 2), v + 2) Interestingly, the mean of the Student’s-t pdf is µ, and its variance is vΣ/(v − 2) while the mean and variance of the escort are µ and Σ respectively.

B

Properties of gt (θ)

Although gt (θ) is not the cumulant function of the t-exponential family, it still preserves convexity. As the following theorem asserts, its first derivative can still be written as an expectation of φ(x) but now with respect to the escort distribution. Note that the theorem and proof here is only a special case of a more general one appeared in Sears [13] and [9]. Theorem 1 The function gt (θ) is convex. Moreover, if the following regularity condition Z Z ∇θ p(x; θ)dx = ∇θ p(x; θ)dx

(33)

holds, then ∇θ gt (θ) = Eqt (x;θ) [φ(x)] , where qt (x; θ) is the escort distribution (14). 10

(34)

Proof To prove convexity, we rely on the elementary arguments. Recall that expt is an increasing and strictly convex function. Choose θ1 and θ2 such that gt (θi ) < ∞ for i = 1, 2, and let α ∈ (0, 1). Set θα = αθ1 + (1 − α)θ2 , and observe that Z expt (hφ(x), θα i − αgt (θ1 ) − (1 − α)gt (θ2 )) dx Z Z < α expt (hφ(x), θ1 i − gt (θ1 ))dx + (1 − α) expt (hφ(x), θ2 i − gt (θ2 ))dx = 1. On the other hand, we also have Z expt (hφ(x), θα i − gt (θα )) dx = 1. Again, using the fact that expt is an increasing function, we can conclude from the above two equations that gt (θα ) < αgt (θ1 ) + (1 − α)gt (θ2 ). This shows that gt is a strictly convex function. To show (34) use (33) and observe that Z Z ∇θ p(x; θ)dx = ∇θ p(x; θ)dx = ∇θ 1 = 0. d Combining with the fact that dx expt (x) = exptt (x), use (14) and the chain rule to write Z Z ∇θ p(x; θ)dx = ∇θ expt (hφ(x), θi − gt (θ)) dx Z = exptt (hφ(x), θi − gt (θ)) (φ(x) − ∇θ gt (θ))dx Z ∝ qt (x; θ)(φ(x) − ∇θ gt (θ))dx = 0.

Rearranging terms and using

C

R

qt (x; θ)dx = 1 directly yields (34).

Convergence of Convex Multiplicative Programming

In the Convex Multiplicative Programming, we convert the problem: argmin P(θ) , θ

N Y

zn (θ)

n=1

into the problem: argmin MP(θ, ξ) , θ,ξ

N X

N Y

ξn zn (θ) s.t.

n=1

ξn = 1 and ξ > 0

n=1

by introducing the latent variable ξ. In the kth ξ-step, assuming the current variables are θ (k−1) and ξ (k−1) , we fix θ (k−1) , denote z˜ = z(θ (k−1) ), and minimize over ξ. It turns out that: ξn(k) =

N 1 Y N1 z˜n z˜n n=1

Therefore, MP(θ (k−1) , ξ (k) ) = min MP(θ (k−1) , ξ) = N P(θ (k−1) )1/N ≤ MP(θ (k−1) , ξ (k−1) ) ξ

11

The θ-step is to fix ξ (k) and minimize θ, the result is MP(θ (k) , ξ (k) ) = min MP(θ, ξ (k) ) ≤ MP(θ (k−1) , ξ (k) ) = N P(θ (k−1) )1/N θ

The above two equalities hold if and only if ξ k = ξ k−1 and θ k = θ k−1 , which follows the convergence of the algorithm at the kth iteration. Therefore, before convergence we have MP(θ (k) , ξ (k) ) < N P(θ (k−1) )1/N < MP(θ (k−1) , ξ (k−1) ). But since P(θ) > 0, the algorithm must converge at some point. ˜ after convergence is a stable point of the P(θ). Assume that θ ˜ and ξ˜ Next, we want to show that θ is the convergence point, then the θ-step is: ! ! N N Y N N X Y X 1 dzn (θ)| ˜ dz (θ)| 1 1 1 ˜ n θ=θ θ=θ ˜ N ˜ N 0= zn (θ) = zn (θ) ˜ ˜ dθ dθ zn (θ) n=1 n=1 n=1 n=1 zn (θ) which implies that, N X

N QN ˜ X dzn (θ)|θ=θ˜ dP(θ)|θ=θ˜ 1 ˜ n=1 zn (θ) dzn (θ)|θ=θ 0= = = ˜ z (θ)|θ=θ˜ dθ dθ dθ zn (θ) n=1 n n=1

˜ is a stable point of P(θ). Therefore, θ

D

Gradient Based method

It is also possible to directly use the gradient based method such as L-BFGS to solve (23). To do this, it is convenient to take log of (23). log P(θ) =

N X

log zn (θ) =

n=1 d X

d X j=1

log rj (θ) +

m X

log li (θ)

(35)

i=1

m  X  Dy E  i 2 ˜ = log 1 + (1 − t)(−λθj /2 − g˜t ) + log 1 + (1 − t)( φ(xi ), θ − gt (θ | xi )) 2 j=1 i=1



(36) Take the derivative,  for n = 1, . . . , d for n = 1, . . . , m

∇θ log zn (θ) = ∇θ log rn (θ) =

t−1 rn (θ)



˜ n en · λθ

(37)

∇θ log zn+d (θ) = ∇θ log ln (θ)    hy i 1−t yn n = · φ(xn ) − Eqt (yn | xn ;θ) φ(xn ) (38) ln (θ) 2 2

where en denotes the d dimensional vector with one at the n-th coordinate and zeros elsewhere (n-th unit vector). There is an obvious relation of (37) (38), and the previous routine, given that ξn = 1/zn (θ) here and ξ˜n ∝ 1/˜ zn in (29). We report the performance of the t-logistic regression by directly using L-BFGS as the optimizer in Table 5. The algorithms use the same parameters as in Table 1. Since L-BFGS is not designed to optimize non-convex functions, it may fail sometimes, in which case we randomly restart with a different initialization.

E

Higher Label Noise

Our algorithm appears to be more robust than logistic regression (especially when t = 1.9) against the label noise (10%). A natural question to ask then is how well it performs when the label noise is larger than 10%. In Figure 6, we compare the t-logistic regression with logistic regression and the probit in the cases when 20% and 30% label noise is added. We also report the test error in Table 6. 12

TestError(%)

30

12

20

8

10

4

0.9

0 0%

10%

20%

30%

0

0.6 0.3 0.0 0%

10%

20%

30%

0%

10%

20%

30%

0%

10%

20%

30%

7.5 3.0

TestError(%)

15.75 6.0

2.5

15.60 4.5 3.0 0%

10%

20%

30%

15.45

2.0

15.30

1.5 0%

10%

20%

30%

Figure 6: The test performance with the change of the label noise (left to right, top: Long-Servedio, Mease-Wyner, Mushroom; bottom: USPS-N, Adult, Web). The magenta dash line with uppertriangles is the logistic regression; the green line with circles is t = 1.3; the cyan line with squares is t = 1.6; the red line with diamonds is t = 1.9; and the blue dash line with lower-triangles is the probit.

F

Significance Test

We performed the paired T-test of the test error rates for each dataset obtained by the t-logistic regression with t = 1.9 and by the other algorithms. To do this, we take the difference of the error rate for each split in each dataset by any two algorithms. The hypothesis is that the difference of the two algorithms for each split is drawn from a zero-mean normal distribution with unknown variance in the same dataset. We report the significance test results in Table 2.

G

Selected Results from RampSVM

Unfortunately, we do not obtain good results by using RampSVM [20] on our datasets. We used the UniverSVM package [23], and performed a grid search of the parameters C and s. It appears that the optimal parameter C is consistent with that used by the Linear SVM, since the UniverSVM uses the solution of Linear SVM to initialize. We report the test performance for s = 0, −0.1, −1, −10, −100 in Table 7. Usually the results are significantly worse than the other algorithms, except for the LongServedio dataset where RampSVM performs as well as other non-convex losses with label noise. We therefore do not report results the RampSVM in the main body of the paper.

13

Table 1: Datasets used in our experiments.(λl denotes λ for the logistic regression. λ1.3 , λ1.6 , λ1.9 are the λ values used for t-logistic regression with t = 1.3, 1.6, and 1.9 respectively. λp is λ for the probit algorithm. C is the parameter C for the C-SVC which is equivalent to 1/λ. ) Name Noise Dimensions Num. of examples λl λ1.3 λ1.6 λ1.9 λp Long-Servedio 0% 21 2000 2−7 2−7 2−7 2−7 2−7 Mease-Wyner 0% 20 2000 2−7 26 24 23 2−7 −7 −7 −2 −5 Mushroom 0% 112 8124 2 2 2 2 2−2 4 2 1 0 USPS-N 0% 256 11000 2 2 2 2 24 5 5 4 3 Adult 0% 123 48842 2 2 2 2 22 −5 −7 0 −1 Web 0% 300 64700 2 2 2 2 2−7 −7 −7 −1 0 Long-Servedio 10% 21 2000 2 2 2 2 2−2 −2 7 7 6 Mease-Wyner 10% 20 2000 2 2 2 2 2−7 1 3 3 3 Mushroom 10% 112 8124 2 2 2 2 21 7 7 7 6 USPS-N 10% 256 11000 2 2 2 2 25 4 5 4 3 Adult 10% 123 48842 2 2 2 2 23 −7 −7 −7 −7 Web 10% 300 64700 2 2 2 2 2−7

Frequency

Table 2: Significance Test of the test error rates by the t-logistic regression with t = 1.9 and the other algorithms. The significance factor α is set as 0.05. ’Y’ means that the difference is significant. ’N’ means the difference is not significant. Dataset Noise Logistic t=1.3 t=1.6 Probit SVM Long-Servedio 0% N N N N N Mease-Wyner 0% Y N N Y Y Mushroom 0% N N N Y N USPS-N 0% N N N N N Adult 0% N N N N N Web 0% N Y N Y N Long-Servedio 10% Y N N N Y Mease-Wyner 10% Y Y N N Y Mushroom 10% N N N Y N USPS-N 10% Y Y N N Y Adult 10% Y N N N Y Web 10% Y Y N Y Y 300

40

1000

240

32

800

180

24

600

120

16

400

60

8

200

0 0.0

0.2

0.4

0.6

0.8

1.0

600

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

2500

0 0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

16000

2000

450 Frequency

0 0.2

12000

1500 300

8000 1000

150 0 0.3

4000

500 0.4

0.5

0.6

0.7

0.8

0.9

1.0

0 0.3

0.4

0.5

0.6

ξ

0.7 ξ

0.8

0.9

1.0

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 ξ

Figure 7: The ξ distribution with 10% label noise added. t = 1.3. Left to right, top: Long-Servedio, Mease-Wyner, Mushroom; bottom: USPS-N, Adult, Web. The red bars (resp. cyan bars) indicate the ξ assigned to points without (resp. with) label noise. 14

C 2−5 27 2−2 2−5 2−3 27 23 23 24 2−3 2−1 21

15

Noise 0% 0% 0% 0% 0% 0% 10% 10% 10% 10% 10% 10%

Logistic 0.00 ± 0.00 0.71 ± 0.71 0.00 ± 0.00 2.40 ± 0.62 15.23 ± 0.62 1.43 ± 0.18 25.50 ± 4.26 4.29 ± 1.58 0.09 ± 0.09 3.92 ± 0.96 15.48 ± 0.53 1.61 ± 0.21

t=1.3 0.00 ± 0.00 0.36 ± 0.51 0.00 ± 0.00 2.47 ± 0.58 15.27 ± 0.62 1.34 ± 0.16 2.00 ± 6.32 2.00 ± 1.00 0.12 ± 0.17 3.21 ± 0.95 15.36 ± 0.52 1.51 ± 0.18

t=1.6 0.00 ± 0.00 0.29 ± 0.37 0.00 ± 0.00 2.48 ± 0.67 15.30 ± 0.59 1.41 ± 0.17 0.00 ± 0.00 1.00 ± 0.90 0.11 ± 0.12 2.91 ± 0.84 15.35 ± 0.49 1.47 ± 0.17

t=1.9 0.00 ± 0.00 0.21 ± 0.35 0.00 ± 0.00 2.48 ± 0.67 15.31 ± 0.62 1.42 ± 0.16 0.00 ± 0.00 0.86 ± 0.74 0.09 ± 0.09 2.82 ± 0.76 15.30 ± 0.58 1.43 ± 0.15

Probit 0.00 ± 0.00 0.71 ± 0.71 0.95 ± 0.43 2.43 ± 0.74 15.37 ± 0.62 3.02 ± 0.23 0.00 ± 0.00 1.79 ± 1.55 0.93 ± 0.45 2.58 ± 0.84 15.44 ± 0.56 3.02 ± 0.23

SVM 0.00 ± 0.00 1.90 ± 1.09 0.00 ± 0.00 2.26 ± 0.64 15.38 ± 0.64 1.38 ± 0.17 23.36 ± 8.90 4.71 ± 1.18 0.11 ± 0.09 4.10 ± 1.02 16.19 ± 0.48 1.77 ± 0.12

Table 3: Test Error Rate in % for various algorithms. The noisy version is obtained by flipping the labels of randomly chosen fixed fraction of the training data.

Dataset Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web

16

Dataset Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web

Noise 0% 0% 0% 0% 0% 0% 10% 10% 10% 10% 10% 10%

Logistic 0.02 ± 0.00 0.02 ± 0.01 0.33 ± 0.03 0.87 ± 0.11 1.03 ± 0.06 4.92 ± 0.36 0.01 ± 0.00 0.01 ± 0.00 0.33 ± 0.02 0.78 ± 0.12 1.74 ± 0.18 8.45 ± 0.60

t=1.6 0.15 ± 0.03 1.70 ± 0.46 1.89 ± 0.28 12.65 ± 3.05 27.73 ± 13.10 31.40 ± 11.44 0.27 ± 0.02 0.42 ± 0.05 2.57 ± 0.40 4.77 ± 0.77 13.49 ± 2.92 51.02 ± 6.96

t=1.9 0.15 ± 0.03 0.98 ± 0.20 0.94 ± 0.18 17.81 ± 3.01 36.29 ± 8.21 40.27 ± 10.26 0.33 ± 0.01 0.59 ± 0.08 3.46 ± 1.07 6.36 ± 1.50 16.99 ± 5.24 48.21 ± 6.14

Table 4: CPU Time (in seconds)

t=1.3 0.17 ± 0.12 1.12 ± 0.30 0.77 ± 0.10 7.09 ± 2.20 11.11 ± 4.70 41.52 ± 9.01 0.33 ± 0.09 0.34 ± 0.08 2.47 ± 0.54 3.13 ± 0.82 7.83 ± 1.76 43.79 ± 9.39

Probit 0.07 ± 0.04 0.06 ± 0.06 0.25 ± 0.08 0.75 ± 0.07 0.96 ± 0.10 1.86 ± 0.28 0.04 ± 0.01 0.04 ± 0.03 0.19 ± 0.03 0.66 ± 0.08 1.05 ± 0.17 0.46 ± 0.07

SVM 0.59 ± 0.17 0.27 ± 0.14 0.40 ± 0.04 8.25 ± 0.66 111.33 ± 7.92 488.30 ± 127.42 6.20 ± 4.12 0.32 ± 0.07 8.38 ± 1.63 61.92 ± 2.95 265.70 ± 24.55 3312.38 ± 5455.18

17

Noise 0% 0% 0% 0% 0% 0% 10% 10% 10% 10% 10% 10%

Dataset

Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web

Test Error (%) t=1.6 t=1.9 0.00 ± 0.00 0.00 ± 0.00 0.29 ± 0.50 0.43 ± 0.69 0.00 ± 0.00 0.02 ± 0.06 2.48 ± 0.62 2.55 ± 0.65 15.30 ± 0.60 15.32 ± 0.60 1.40 ± 0.16 1.44 ± 0.18 0.00 ± 0.00 0.00 ± 0.00 1.00 ± 0.90 0.86 ± 0.74 0.11 ± 0.12 0.09 ± 0.09 2.92 ± 0.85 2.87 ± 0.78 15.34 ± 0.49 15.31 ± 0.58 1.47 ± 0.17 1.43 ± 0.16 t=1.3 0.09 ± 0.05 0.07 ± 0.02 0.75 ± 0.07 3.48 ± 0.46 5.13 ± 0.69 11.79 ± 1.52 0.06 ± 0.04 0.06 ± 0.03 0.72 ± 0.09 1.53 ± 0.21 4.23 ± 0.18 27.04 ± 2.00

Time (s) t=1.6 0.42 ± 0.72 0.08 ± 0.02 0.97 ± 0.10 5.69 ± 0.81 7.96 ± 0.99 8.30 ± 3.59 0.04 ± 0.01 0.07 ± 0.01 0.63 ± 0.06 1.72 ± 0.19 5.35 ± 0.65 23.18 ± 1.77

Table 5: Results of t-logistic regression by using L-BFGS directly.

t=1.3 0.00 ± 0.00 0.36 ± 0.51 0.00 ± 0.00 2.45 ± 0.59 15.27 ± 0.62 1.34 ± 0.15 0.00 ± 0.00 2.00 ± 1.00 0.12 ± 0.17 3.22 ± 0.93 15.36 ± 0.53 1.50 ± 0.18

t=1.9 0.09 ± 0.03 0.10 ± 0.11 0.98 ± 0.76 6.27 ± 2.12 9.21 ± 0.97 8.48 ± 7.76 0.05 ± 0.01 0.08 ± 0.01 0.71 ± 0.10 2.11 ± 0.33 6.54 ± 0.64 13.37 ± 6.15

300

60

Frequency

240

1000 800

45

180

600 30

120

400 15

60 0 0.0

0.2

0.4

0.6

0.8

1.0

0 0.0

200 0.2

0.4

0.6

0.8

1.0

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0

600 1600

10000

Frequency

450 1200 300 150 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 ξ

7500

800

5000

400

2500

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 ξ

0 0.0

0.2

0.4

0.6

0.8

1.0

ξ

Figure 8: The ξ distribution with 10% label noise added. t = 1.6. Left to right, top: Long-Servedio, Mease-Wyner, Mushroom; bottom: USPS-N, Adult, Web. The red bars (resp. cyan bars) indicate the ξ assigned to points without (resp. with) label noise.

Dataset Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web Long-Servedio Mease-Wyner Mushroom USPS-N Adult Web

Noise 20% 20% 20% 20% 20% 20% 30% 30% 30% 30% 30% 30%

Logistic 26.14 ± 5.39 7.86 ± 2.36 0.21 ± 0.27 4.82 ± 0.84 15.59 ± 0.49 1.77 ± 0.21 26.71 ± 4.34 11.07 ± 2.46 0.49 ± 0.40 7.27 ± 1.09 15.79 ± 0.49 1.98 ± 0.18

t=1.3 25.43 ± 5.23 5.86 ± 1.46 0.19 ± 0.24 4.73 ± 0.82 15.56 ± 0.50 1.69 ± 0.21 26.57 ± 4.43 10.50 ± 3.03 0.49 ± 0.41 7.29 ± 1.19 15.76 ± 0.46 1.97 ± 0.18

t=1.6 23.64 ± 5.49 4.07 ± 1.91 0.19 ± 0.24 4.21 ± 0.82 15.51 ± 0.50 1.64 ± 0.21 26.43 ± 4.48 6.64 ± 2.48 0.44 ± 0.39 6.62 ± 0.97 15.72 ± 0.48 1.96 ± 0.17

t=1.9 9.93 ± 12.97 3.07 ± 1.47 0.19 ± 0.24 3.77 ± 0.68 15.47 ± 0.45 1.59 ± 0.19 26.07 ± 4.55 5.29 ± 2.41 0.48 ± 0.29 6.01 ± 0.84 15.73 ± 0.47 1.95 ± 0.17

Probit 0.00 ± 0.00 3.00 ± 1.38 0.95 ± 0.45 3.23 ± 0.79 15.65 ± 0.62 3.02 ± 0.23 2.86 ± 2.54 9.50 ± 3.40 1.02 ± 0.52 4.35 ± 1.97 15.87 ± 0.62 3.02 ± 0.23

Table 6: Test Error Rate in % for various algorithms with higher label noise.

Dataset Long-Servedio Mease-Wyner Mushroom Long-Servedio Mease-Wyner Mushroom

Noise 0% 0% 0% 10% 10% 10%

C 2−5 27 2−2 23 23 24

s=0 0.00 ± 0.00 50.5 ± 4.10 0.04 ± 0.08 0.00 ± 0.00 50.5 ± 4.10 5.25 ± 8.87

s=-0.1 0.00 ± 0.00 50.5 ± 4.10 0.04 ± 0.08 0.00 ± 0.00 50.5 ± 4.10 5.74 ± 8.68

s=-1 0.00 ± 0.00 50.5 ± 4.10 0.04 ± 0.08 0.00 ± 0.00 50.5 ± 4.10 9.08 ± 10.47

Table 7: The test error (%) of RampSVM on selected datasets.

18

s=-10 0.00 ± 0.00 50.5 ± 4.10 0.04 ± 0.08 23.29 ± 8.84 50.5 ± 4.10 47.19 ± 1.66

s=-100 0.00 ± 0.00 50.5 ± 4.10 0.04 ± 0.08 23.29 ± 8.84 50.5 ± 4.10 47.19 ± 1.66

t-Logistic Regression

All code is written in Matlab, and for the linear SVM we use the Matlab .... The red (dark) bars (resp. cyan (light) bars) indicate the frequency of ξ assigned to .... Software available at http://www.kyb.mpg.de/bs/people/fabee/universvm. html. 9 ...

750KB Sizes 0 Downloads 183 Views

Recommend Documents

regression testing
iterative, parallel development cycle. Because the software ... The parallel nature .... prior to delivery implies that different code is delivered than was tested. The.

REGRESSION: Concept of regression, Simple linear ...
... Different smoothing techniques, General linear process, Autoregressive Processes AR(P),. Moving average Process Ma(q): Autocorrelation,. Partial autocorrelation, Spectral analysis,. Identification in time domain, Forecasting,. Estimation of Param

Regression models in R Bivariate Linear Regression in R ... - GitHub
cuny.edu/Statistics/R/simpleR/ (the page still exists, but the PDF is not available as of Sept. ... 114 Verzani demonstrates an application of polynomial regression.

Quantile Regression
The prenatal medical care of the mother is also divided into four categories: those ..... Among commercial programs in common use in econometrics, Stata and TSP ... a public domain package of quantile regression software designed for the S.

Linear regression
Linear regression attempts to model the relationship between two variables by fitting a linear equation to the a sample of n observations (i=1,...,n) as such. Yi = β0 + β1Xi + ui. Yi dependent variable, regressand. Xi independent variable, regresso

Logistic Regression - nicolo' marchi
These scripts set up the dataset for the problems and make calls to functions that you will write. .... 1.2.3 Learning parameters using fminunc. In the previous ...

regression testing
finding new errors in previously tested software—is an absolutely crit- ... As more companies embrace the object paradigm, more managers will reg ognize that .... 3 10. 6 \ / 12. ERROR. 13 we. FIGURE 2. Automated regression testing using ...

structural threshold regression
Meeting of the Canadian Econometrics Study Group in Vancouver, and 23rd (EC) 2 ... Our finding is similar to the result of Perron and Qu (2006) who consider ...... 24. ANDROS KOURTELLOS ET AL. 5. Conditional on γ , estimation in each ...

Structural Threshold Regression
Feb 26, 2015 - terms is required to restore the conditional mean zero assumption of the .... Assumption 1.3 assumes conditional mean independence between ...

(EBMAL) for Regression
†Human Research and Engineering Directorate, U.S. Army Research Laboratory, Aberdeen Proving Ground, MD USA ... ¶Brain Research Center, National Chiao-Tung University, Hsinchu, Taiwan ... Administration (NHTSA) [36], 2.5% of fatal motor vehicle ..

Data enriched linear regression - arXiv
big data set is thought to have similar but not identical statistical characteris- tics to the small one. ... using the big data set at the risk of introducing some bias. Our goal is to glean ...... Stanford University Press,. Stanford, CA. Stein, C.

Least Angle Regression
tions, and interactions between variables), to survival and other additional regression models, and to ..... The advantage over stepwise procedures is that the best set of two predictors need not include the predictor that was best in isolation. The

[PDF] Applied Logistic Regression
of correlated outcome data A wealth of additional material for topics ranging from Bayesian methods to assessing model fit Rich data sets from real-world studies ...

Quadratic Modeling and Regression Homework.pdf
C) Determine the equivalent temperature in Celsius degrees for a body temperature of 98.6 degrees Farenheit. 3. A study compared the speed s (in miles per ...

Boosting Methodology for Regression Problems
Decision-support Systems, with an Application in. Gastroenterology” (with discussion), Journal of the Royal. Statistical Society (Series A), 147, 35-77. Zheng, Z. and G.I. Webb [1998]. “Lazy Bayesian Rules,”. Technical Report TR C98/17, School

Regression Discontinuity Design with Measurement ...
Nov 20, 2011 - All errors are my own. †Industrial Relations Section, Princeton University, Firestone Library, Princeton, NJ 08544-2098. E-mail: zpei@princeton.