Learning Features by Contrasting Natural Images with Noise Michael Gutmann University of Helsinki

Aapo Hyvärinen University of Helsinki

[email protected]

[email protected]

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 1/17

Introduction

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 2/17

Natural images? A natural scene Natural image patches

Introduction

3

● Preliminaries ● Nat. image vs. noise ● Classifier Contrastive feature learning Simulations

5

4



Most prob. models are models of natural image patches.



Difficult enough as the data is high dimensional.



Can serve as building block for models of entire scenes.

2 1 6

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 3/17

Why model natural images?

Introduction ● Preliminaries ● Nat. image vs. noise ● Classifier Contrastive feature learning Simulations

Possible motivations for building statistical models of natural images: ■ You can use the model as prior in tasks which involve Bayesian inference (not the topic of this presentation). ■ You can use the model to generate artificial natural images (not the topic of this presentation). ■ There are some connections to visual neuroscience (not the topic of this presentation). ■ Finding interesting features of natural images: ◆ What kind of features appear in natural images? ◆ What kind of structure is characteristic for natural images? ◆ How do natural images differ from other, artificial, image data (noise)?

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 4/17

Natural images vs. noise

Introduction ● Preliminaries

In what aspects do the two datasets differ from each other?

● Nat. image vs. noise ● Classifier Contrastive feature learning Simulations

natural images

Michael Gutmann – University of Helsinki

noise images

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 5/17

Learning features by classification

Introduction ● Preliminaries ● Nat. image vs. noise ● Classifier Contrastive feature learning Simulations

Key idea: ■ Train a classifier to discriminate between natural images and some artificial noise. ■ To succeed in the discrimination task, the classifier must “discover structure” in the data, i.e. identify features of natural images. Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 6/17

Learning features by classification

Introduction ● Preliminaries ● Nat. image vs. noise ● Classifier Contrastive feature learning Simulations

Key idea: ■ Train a classifier to discriminate between natural images and some artificial noise. ■ To succeed in the discrimination task, the classifier must “discover structure” in the data, i.e. identify features of natural images. Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 7/17

Contrastive feature learning

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 8/17

The elements of contrastive feature learning

Introduction Contrastive feature learning ● Elements ● More details ● Nonlinearities Simulations

1. Classifier: Assign C = 1 if input x is a natural image, and C = 0 if input is noise. 2. Estimation method: Fit the parameters in the classifier to the data (supervised learning!) 3. Noise: The reference data that is used for comparison with the natural images.

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 9/17

More on the classifier and its estimation ■

Use a classification approach based on logistic regression

Introduction

1 P (C = 1|x) = 1 + exp(−y(x))

Contrastive feature learning ● Elements ● More details ● Nonlinearities Simulations





y(x) =

M X

T g(wm x + bm ) + γ

m=1

Parameters in the model are the features wm , the bias terms bm , the offset γ, as well as possibly the function g(u). The parameters can be estimated by maximum (conditional) likelihood. This is the same as minimization of the cross-entropy error J T 1X −Ct log [P (Ct = 1|xt )]−(1−Ct ) log [1 − P (Ct = 1|xt )] J= T t=1



Michael Gutmann – University of Helsinki

Reference data: Use noise with the same covariance structure as natural images.

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 10/17

Choice of the nonlinearity in the discriminant y(x) Simple choices

10

Introduction Contrastive feature learning ● Elements ● More details

Classical choices

● Nonlinearities

"Next" simplest choices

Simulations

40

Linear

Quadratic

5

30

0

20

−5

10

−10 −6 1

−4

−2

0

2

4

0 −6 1

6

−4

−2

Tanh

0

2

4

6

2

4

6

Logistic

0.5 0

0.5

−0.5 −1 −6

−4

−2

0

2

1

4

0 −6

6

−4

5

Symm logistic

0.8

3

0.4

2

0.2

1

0

20

Lin thresh

4

0.6

−2

Squared thresh 15 10

0

−10

−5

0

5

10

0 −6

5

−4

−2

0

2

4

6

0 −6

−4

−2

0

2

4

6

Parameterized nonlinearity:

g(u) = α1 [max(0, u − β1 )]η1 + α2 [max(0, −(u − β2 ))]η2 Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 11/17

Simulations

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 12/17

The questions addressed

Introduction Contrastive feature learning Simulations ● Questions addressed ● Performance ● Features

1. Which nonlinearity g(u) gives the best performance? 2. How do the features wm look like? 3. Which principle does the classifier use to solve the discrimination task?

● Classification principle

Michael Gutmann – University of Helsinki

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 13/17

Classification performance Cross entropy error

False classification rate

0.34

0.45 0.32 0.3

0.4

0.28

0.35

0.26 Tanh

Logistic

Symm. Logistic Lin thresh

Squared thresh

Tanh

Logistic

Symm. Logistic Lin thresh

Squared thresh

0.4 0.3 0.38 0.29 0.36 0.28 0.34 0.27 0.32 0.26 0.3 0.25 0.28 Optimized nonlinearity

Michael Gutmann – University of Helsinki

Symm. Logistic

Optimized nonlinearity

Symm. Logistic

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 14/17

The learned features Symm. logistic

Michael Gutmann – University of Helsinki

Optimized nonlinearity

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 15/17

Classification principle The discriminant y(x) = Introduction

PM

T g(w x) m m=1

y(x) > 0 ⇒ x is a natural image.



+ γ/M rules:

y(x) < 0 ⇒ x is noise.

Contrastive feature learning Simulations

Symm. logistic

● Questions addressed

Optimized nonlinearity

● Performance

15

1.2

● Features ● Classification principle

1 10 g(u)+γ/M

g(u) + γ/M

0.8 0.6 0.4 0.2

5

0

0 −0.2 −15

−10

−5

0 u

5

10

15

Thresholding of each feature output

Michael Gutmann – University of Helsinki

−5 −6

−4

−2

0 u

2

4

6

Sign consistency across feature outputs

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 16/17

Summary ■ Introduction Contrastive feature learning Simulations Summary



Michael Gutmann – University of Helsinki

The minimum to retain: 1. The talk was about learning features in data (here: natural images). 2. Features are learned by training a classifier to distinguish between the data and some artificial noise. 3. We used nonlinear logistic regression to do the classification. Some more details: 1. Classification by thresholding outputs of gabor-like feature detectors. 2. Optimizing the nonlinearity gives an asymmetric solution. Classification performance improves. 3. An alternative classification principle: the outputs of some gabor-like feature detectors need to have the same sign for natural images.

ICANN2009: Learning Features by Contrasting Natural Images with Noise - p. 17/17

Learning Features by Contrasting Natural Images with ...

Michael Gutmann – University of Helsinki. ICANN2009: Learning ... Estimation method: Fit the parameters in the classifier to the data (supervised learning!) 3.

448KB Sizes 0 Downloads 228 Views

Recommend Documents

Learning features by contrasting natural images with ...
1 Dept. of Computer Science and HIIT, University of Helsinki,. P.O. Box 68, FIN-00014 University of Helsinki, Finland. 2 Dept. of Mathematics and Statistics, University of ... rameterized family of probability distributions. In non-overcomplete ICA,

Reading Digits in Natural Images with Unsupervised Feature Learning
Reliably recognizing characters in more complex scenes like ... As a result systems based on hand-engineered representations perform far worse on this task ...

Reading Digits in Natural Images with Unsupervised ... - Deep Learning
ture learning schemes applied to a large corpus of digit data and ... The SVHN dataset was obtained from a large number of Street View images using a ...

Learning with Augmented Features for Supervised and ...
... applications clearly demonstrate that our SHFA and HFA outperform the existing HDA methods. Index Terms—Heterogeneous domain adaptation, domain adaptation, transfer learning, augmented features. 1 INTRODUCTION. IN real-world applications, it is

Interacting with Features in GPlates
See ​www.earthbyte.org/Resources/earthbyte_gplates.html​ for EarthByte data sets. Background. GPlates ... Feature Type box at the top of the window). .... Indonesian Gateway and associated back-arc basins, Earth-Sci. Rev., vol 83, p.

Interacting with Features in GPlates
... the new changes. See the GPlates online manual for further information: ... In this exercise we will learn how to query features. To start with we .... 85-138, 1995. Gaina, C., and R. D. Müller, Cenozoic tectonic and depth/age evolution of the.

Learning Natural Image Structure with a Horizontal ...
2 Helsinki Institute for Information Technology. 3 Department of Mathematics ... assumed to be a known supergaussian probability density function (pdf). Due to.

Learning features to compare distributions
Goal of this talk. Have: Two collections of samples X Y from unknown distributions. P and Q. Goal: Learn distinguishing features that indicate how P and Q differ. 2/28 ...

key features for math teachers - Proven Learning
and analyze data, and then automatically transfer grades into any electronic gradebook. GRIDDED ANSWERS. Customize vertical columns with numbers and.

Contrasting effects of bromocriptine on learning of a ... - Springer Link
Materials and methods Adult male Wistar rats were subjected to restraint stress for 21 days (6 h/day) followed by bromocriptine treatment, and learning was ...

Contrasting effects of bromocriptine on learning of a ...
neurochemistry experiments, Bhagya and Veena for their help in data entry. ..... of dopaminergic agonists could produce the behavioural recovery by acting on .... Conrad CD, Galea LA, Kuroda Y, McEwen BS (1996) Chronic stress impairs rat ...

Sparse Coding of Natural Images Using an ...
Computer Science Department. Carnegie Mellon University ... represent input in such a way as to reduce the high degree of redun- dancy. Given a noisy neural ...

Text Detection from Natural Scene Images: Towards a ...
When a visually impaired person is walking around, it is important to get text information which is present in the scene. For example, a 'stop' sign at a crossing ...

Image Compression of Natural Images Using Artificial ...
frequency domain using a two-dimensional discrete cosine transform (DCT). The fifth step achieves the actual data reduction (quantisation), by dividing each of ...

Contrasting vulnerability.pdf
implementing them in a Geographic Information System (GIS) with. a 10-m spatial resolution. We used the average ratio of precipita- tion (P) to potential ...

Provision of contrasting ecosystem services by soil ... - Toby Kiers
maize roots was determined using molecular tools. Detailed information on the .... correlated when all data were included in the analysis. (Fig. 2a), and likewise when ..... (1999) Visualization of ribosomal DNA loci in spore interphasic nuclei of ..

Contrasting styles of volcanic activity as observed by ...
eruptive column that extended up to 20 km and pyroclastic flows that extended up to 7.5 km of the summit. (Gardeweg & Medina, 1994). The more active volcanoes in Southern Volcanic Zone (SVZ) are Villarrica and Llaima volcanoes (Moreno &. Fuentealba,

Learning Invariant Features Using Inertial Priors ... - Semantic Scholar
Nov 30, 2006 - Department of Computer Science .... Figure 1: The simple three-level hierarchical Bayesian network (a) employed as a running example in the ...

VSEPR Lab Images with Notes.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. VSEPR Lab ...

Explain Images with Multimodal Recurrent Neural Networks
Oct 4, 2014 - In this paper, we present a multimodal Recurrent Neural Network (m-RNN) model for generating .... It needs a fixed length of context (i.e. five words), whereas in our model, ..... The perplexity of MLBL-F and LBL now are 9.90.