Deep Learning Master Class

Tel Aviv University | Wednesday-Thursday, November 5-6, 2014

Wednesday Nov. 5th

9:10-9:30: Opening

9:30-10:30: Yann LeCun (Facebook, NYU)
TBD

10:30-11:30: Tomaso Poggio (Massachusetts Institute of Technology)
Learning of invariant representations in visual cortex: m-theory
The theory shows how a hierarchical architecture of simple-complex cells modules can learn in an unsupervised way to be automatically invariant to transformations of a new object, achieving the goal of recognition with very few labeled examples. M-theory makes specific predictions about the architecture of the ventral stream, including the dependence on eccentricity of the magnification factor in various areas, and on the tuning properties of its neurons from early generic, Gabor-like tuning to class-specific tuning in AIT. This approach is an example of what could be the next phase in the theory of learning: how to learn in an unsupervised way good representations that allow a supervised classifier to learn from very few labeled examples, similar to how children learn.

11:30-12:00: Break

12:00-13:00: Yaniv Taigman (Facebook)
Web-Scale Training for Face Identification
Scaling machine learning methods to massive datasets has attracted considerable attention in recent years, thanks to easy access to ubiquitous sensing and data from the web. Face recognition is a task of great practical interest for which (i) very large labeled datasets exist, containing billions of images; (ii) the number of classes can reach billions; and (iii) complex features are necessary in order to encode subtle differences between subjects, while maintaining invariance to factors such as pose, illumination, and aging. In this talk I will present an elaborate pipeline and several customized deep architectures, that learn representations that generalize well on the tasks of face verification and identification.

13:00-14:00: lunch break

14:00-15:00 Amnon Shashua (HUJI)
SimNets: A Generalization of Convolutional Networks
We present a deep layered architecture that generalizes classical convolutional neural networks (ConvNets). The architecture, called SimNets, is driven by two operators, one being a similarity function whose family contains the convolution operator used in ConvNets, and the other is a new "soft max-min-mean" operator called MEX that realizes classical operators like ReLU and max-pooling, but has additional capabilities that make SimNets a powerful generalization of ConvNets. Two interesting properties that emerge from the architecture are: (i) the basic input to hidden-units to output-nodes machinery contains as special case a kernel machine, and (ii) initializing networks using unsupervised learning is natural. Experiments demonstrate the capability of achieving state of the art accuracy with networks that are 1/8 the size of comparable ConvNets.

15:00-15:30: break

15:30-18:00: Open pannel with students