School of Computer Science

Learning and Neural Computation

0368-4034-01

Semester I, 1999-0,     Wednesday 9-12,   Dan David 211

The course will be given in English!

Submission of the project till Thursday, Feb 24 8pm
Presentation of the project in Mid March.

Dr. Nathan Intrator,  nin@math.tau.ac.il
Office hours: Wednesday, 13:30-14:30, Schreiber 216
Prof. Manfred Joswig,   manfred@ndc.soreq.gov.il
Office hours: ***, Schreiber ***

Updated April 10, 2000     This page will be updated throughout the course.

Download the test mat file for all signals (click on right button)

Please send me an email with a list of 19 numbers between zero and one which represent your prediction regarding each pattern. Please indicate the architecture and parameters used to get these values.

The course is intended for third-year and Master students. The first part of the course will deal with classical topics in computational learning theory and neural network learning. The second part will be geared more towards applications. In this part, students will apply the acquired methods on some Seismic Data Classification and few other real-world problems. Part of the course will therefore be devoted to topics related to detection in time-series and acoustic signals. This will be done by Dr. Manfred Joswig who is visiting this year and is a world expert on Seismic Data Analysis.

Suggested Reading

Course Outline  

Description of the final   

The purpose of the final this year is to develop expertise in using regularization constraints during training to reduce the chance of over-fitting. We shall thus train with different bias constraints using the general framework of Intrator (1993). In addition we shall master the use of object preprocessing and ensemble training and prediction confidence as essential tools for a practical neural network expert system.
We shall be using the following preprocessing methods:
  1. Principal components
  2. Dimensionality reduction via entropy constraints
  3. K-means Clustering
  4. Your choice.
The following training constraints:
  1. Reconstruction constraint
  2. Kurtosis constraint
  3. BCM constraint
  4. Minimal entropy constraint
  5. Maximal entropy constraint
  6. Mixture of Gaussians constraint
  7. Your choice
The following network architectures and training algorithms
  1. Orr's Radial Basis Function code.
  2. Bishop's Radial Basis Function code.
  3. Neural networks toolbox back-prop code with Levenberg-Marquardt training.
  4. Neural networks toolbox RBF code.
  5. Your choice.
The following prediction-confidence methods:
  1. Variance between different experts weighted by their past performance.
  2. Weighted reconstruction error.
  3. Your choice.

Description of the data, and past projects (from which code can be taken) can be found here.

You will need to supply three functions:

  1. [preproc, netarch] = findarch(train_data, train_label, options)
    This function finds an optimal preprocessing and a set of architectures with their corresponding parameters which are best for the given training data. This function calls internally to findpreproc

  2. [processed, preproc] = findpreproc(train_data, train_label, procoptsions, preproc)
    This function has two tasks. First it finds an optimal preprocessing and second, it generates the pre-processed data. If preproc is supplied, the program runs the data through the preprocessing methodology which should be completely kept in preproc.

  3. [label_pred, confidence] = netpred(data, preproc, netarch, options)
    This function performs the preprocessing on the given data based on the preprocessing scheme that was chosen by findarch and is stored in preproc and then prediction of the class labels based on the architectures that were found by findarch and are stored in netarch. Note that if a collection of several experts is found by findarch, then the algorithm for fusing the experts should also be stored in netarch to be used by netpred.
    You should suggest a way to calculate the confidence and justify in your written report why this confidence method is useful.

Each function should print once on the screen the names of the students and a short description of the methods that are used, the type of preprocessing and the architectures that are found. In the prediction code, again, the names should be displayed as well as the type of expert fusion and the method to calculate the confidence. There should also be a clear description of the possible options to the functions and default values (which were found optimal) in case options are not chosen.
All other functions should be internal to the above three functions.

Note:   You are allowed to use code written by last year's projects or code taken from the web as long as you clearly state it at the beggining of the code, insert it into your own code, make sure that you understand what the code is doing and the code has good documentation.

Project submission

There should be an HTML page describing the project and having a link to the functions. It should include the following:

  1. Clear description of the type of algorithms that were used and the way optimal architectures were chosen.
  2. Clear justification based on what was learned in the course for your decisions. (This can include graphs.)
  3. Clear description and justification of how the confidence is calculated.
In the code itself, there should be a help file similar to those found in matlab code with clear description of the structure of your variables (the net architectures, the options and the preprocessing methods and parameters).

Project presentation is scheduled to Tuesday February 8, 2000 at 9am.

Specific-group tasks

  1. Project:  Akavia  Gindi  Erez 
    • Preprocessing:   clustering and your choice.
    • Training constraints:   Kurtosis and your choice.
    • Architecture and code:   RBF with Orr and Back-prop with LM.
    • Confidence:   Reconstruction and your choice.
    • Data:   Sonl and Sono.
  2. Project:  Lifchitz  Weisman  Arnold 
    • Preprocessing:   Minimal entropy and clustering.
    • Training constraints:   BCM and Kurtosis.
    • Architecture and code:   Back-prop with LM.
    • Confidence:   Variance and your choice.
    • Data:   Sono and wvlet.
  3. Project   Hoffman  Stav 
    • Preprocessing:   Maximal entropy and clustering.
    • Training constraints:   Reconstruction and minimal entropy.
    • Architecture and code:   RBF with Orr and Bishop.
    • Confidence:   Variance and your choice.
    • Data:   Sonl and psd.
  4. Project   Tobias  
    • Preprocessing:   Clustering and your choice.
    • Training constraints: Reconstruction and Kurtosis 
    • Architecture and code:   RBF with Bishop.
    • Confidence:   Reconstruction and your choice.
    • Data:   Sonl and wvlet
  5. Project   Sidi,   Bogodlov
    • Preprocessing:   clustering and your choice.
    • Training constraints:   BCM and Kurtosis.
    • Architecture and code:   RBF with Orr.
    • Confidence:   Reconstruction and your choice.
    • Data:   Sonl and Sono.
Data Sets    Sonl    Sono    Psd    WVlet    Analog Data    Download using right mouse botton

Copyright © 1997-1999 Nathan Intrator. All rights reserved.       http://www.math.tau.ac.il/~nin/learn99/index.html