Copy Link
Add to Bookmark
Report

Neuron Digest Volume 04 Number 17

eZine's profile picture
Published in 
Neuron Digest
 · 1 year ago

Neuron Digest	Tuesday, 25 Oct 1988		Volume 4 : Issue 17 

Today's Topics:

Administrivia
Congress on Cybernetics and Systems
Report: Markov Models and Multilayer Perceptrons
AAAIC '88
3rd Intl. Conference on Genetic Algorithms
Abstract: ANNS and Radial Basis Functions
Abstract: A Dynamic Connectionist Model For Phoneme Recognition
tech report: Laerning Algorithm for Fully Recurrent ANNS
Paper from nEuro'88 in Paris
Cary Kornfeld to speak on neural networks and bitmapped graphics
Neural Network Symposium Announcement
NIPS Student Travel Awards


Send submissions, questions, address maintenance and requests for old issues to
"neuron-request@hplabs.hp.com" or "{any backbone,uunet}!hplabs!neuron-request"

------------------------------------------------------------

Subject: Administrivia
From: "Neuron-Digest Moderator -- Peter Marvit" <neuron@hplms2>
Date: Tue, 25 Oct 88 15:00:08 -0700

[[ As you all noticed, nearly everyone received a duplicate of the last issue
of the Digest. I've traced it to a machine which has a too few CPU cycles
available. We're waiting for a faster one, but in the mean time I'm sending
this through a different route. Thanks you all who sent me headers so I
could trace the problem.

This issue contains all paper and conference announcements. I'll send the
next one of discussions and requests. I'm saving the issue of the
discussion of Consciousness till the one after that.

Keep those cards and letters coming. -PM ]]

------------------------------

Subject: Congress on Cybernetics and Systems
From: SPNHC@CUNYVM.CUNY.EDU (Spyros Antoniou)
Organization: The City University of New York - New York, NY
Date: 08 Oct 88 03:28:19 +0000


WORLD ORGANIZATION OF SYSTEMS AND CYBERNETICS

8 T H I N T E R N A T I O N A L C O N G R E S S

O F C Y B E R N E T I C S A N D S Y S T E M S

JUNE 11-15, 1990 at Hunter College, City University of New York, USA

This triennial conference is supported by many international
groups concerned with management, the sciences, computers, and
technology systems.

The 1990 Congress is the eighth in a series, previous events
having been held in London (1969), Oxford (1972), Bucharest (1975),
Amsterdam (1978), Mexico City (1981), Paris (1984) and London (1987).

The Congress will provide a forum for the presentation
and discussion of current research. Several specialized sections
will focus on computer science, artificial intelligence, cognitive
science, biocybernetics, psychocybernetics and sociocybernetics.
Suggestions for other relevant topics are welcome.

Participants who wish to organize a symposium or a section,
are requested to submit a proposal ( sponsor, subject, potential
participants, very short abstracts ) as soon as possible, but not
later than September 1989. All submissions and correspondence
regarding this conference should be addressd to:

Prof. Constantin V. Negoita
Congress Chairman
Department of Computer Science
Hunter College
City University of New York
695 Park Avenue, New York, N.Y. 10021 U.S.A.

=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=
| Spyros D. Antoniou SPNHC@CUNYVM.BITNET SDAHC@HUNTER.BITNET |
| |
| Hunter College of the City University of New York U.S.A. |
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=

------------------------------

Subject: Report: Markov Models and Multilayer Perceptrons
From: prlb2!welleken@uunet.UU.NET (Wellekens)
Date: Sat, 08 Oct 88 18:00:24 +0100

The following report is available free of charge from

Chris.J.Wellekens, Philips Research Laboratory Brussels,
2 Avenue van Becelaere, B-1170 Brussels,Belgium.
Email wlk@prlb2.uucp

LINKS BETWEEN MARKOV MODELS AND MULTILAYER PERCEPTRONS
H.Bourlard and C.J.Wellekens
Philips Research Laboratory Brussels

ABSTRACT

Hidden Markov models are widely used for automatic speech recognition. They
inherently incorporate the sequential character of the speech signal and are
statistically trained. However, the a priori choice of a model topology
limits the flexibility of the HMM's. Another drawback of these models is
their weak discriminating power.

Multilayer perceptrons are now promising tools in the connectionist approach
for classification problems and have already been successfully tested on
speech recognition problems. However, the sequential nature of the speech
signal remains difficult to handle in that kind of machine.

In this paper, a discriminant hidden Markov model is defined and it is shown
how a particular multilayer perceptron with contextual and extra feedback
input units can be considered as a general form of such Markov models.
Relations with other recurrent networks commonly used in speech recognition
are also pointed out.

Chris

------------------------------

Subject: AAAIC '88
From: wilsonjb%avlab.dnet@AFWAL-AAA.ARPA
Organization: The Internet
Date: 08 Oct 88 18:20:00 +0000

Aerospace Applications of Artificial Intelligence (AAAIC) '88

Special Emphasis
On
Neural Network Applications



LOCATION: Stouffer Dayton Plaza Hotel
Dayton, OH

DATES: Monday, 24 Oct - Friday, 28 Oct 88


PLENARY SESSION Tuesday Morning

Lt General John M. Loh,
Commander, USAF Aeronautical Systems Division

Dr. Stephen Grossberg,
President, Association of Neural Networks


TECHNICAL SESSIONS Tuesday - Thursday (in paralell)

I. Neural Network Aerospace Applications
Integrating Neural Netorks and Expert Systems
Neural Networks and Signal Processing
Neural Networks and Man-Machine Interface Issues
Parallel Processing and Neural Networks
Optical Neural Networks
Back Propogation with Momentum, Shared Weights and Recurrency
Cybernetics

II. AI Aerospace Applications
Developmental Tools and Operational and Maintenance Issues
Using Expert Systems
Real Time Expert Systems
Automatic Target Recognition
Data Fusion/Sensor Fusion
Combinatorial Optimaztion for Scheduling and Resource Control
Machine Learining, Cognition, and Avionics Applications
Advanced Problem Solving Techniques
Cooperative and Competitive Network Dynamics in Aerospace

Tutorials

I. Introduction to Neural Nets Mon 8:30 - 11:30
II. Natural LAnguage Processing 8:30 - 11:30
III. Conditioned Response in Neural Nets 1:30 - 4:30
IV. Verification and Validation of Knowledge 1:30 - 4:30
Based Systems

Workshops

I. Robotics, Vision, and Speech Fri 8:30 - 11:30
II. AI and Human Engineering Issues 8:30 - 11:30
III. Synthesis of Intelligence 1:30 - 4:30
IV. A Futurists View of AI 1:30 - 4:30


REGISTRATION INFORMATION
(after 30 Sept)

Conference $225
Individual Tech Session (ea) $ 50
Tutorials (ea) $ 50
Workshops (ea) $ 50


Conference Reistration includes: Plenary Session
Tuesday Luncheon
Wednesday Banquet
All Technical Sessions
Proceedings

Tutorials and Workshops are extra.

For more information, contact:

AAAIC '88
Dayton SIGART
P.O. Box 31434
Dayton, OH 45431

Darrel Vidrine
(513) 255-2446

Hotel information:

Stouffer Dayton Plaza Hotel
(513) 224-0800

Rates: Govt Non-Govt

Single $55 $75

Double $60 $80

------------------------------

Subject: 3rd Intl. Conference on Genetic Algorithms
From: gref@AIC.NRL.NAVY.MIL
Organization: The Internet
Date: 08 Oct 88 18:20:00 +0000


Call for Papers

The Third International Conference on Genetic Algorithms
(ICGA-89)


The Third International Conference on Genetic Algorithms (ICGA-
89), will be held on June 4-7, 1989 at George Mason University
near Washington, D.C. Authors are invited to submit papers on
all aspects of Genetic Algorithms, including: foundations of
genetic algorithms, search, optimization, machine learning using
genetic algorithms, classifier systems, apportionment of credit
algorithms, relationships to other search and learning paradigms.
Papers discussing specific applications (e.g., OR, engineering,
science, etc.) are encouraged.


Important Dates:

10 Feb 89: Submissions must be received by program chair
10 Mar 89: Notification of acceptance or rejection
10 Apr 89: Camera ready revised versions due
4-7 Jun 89: Conference Dates


Authors are requested to send four copies (hard copy only) of a
full paper by February 10, 1989 to the program chair:


Dr. J. David Schaffer
Philips Laboratories
345 Scarborough Road
Briarcliff Manor, NY 10510
ds1@philabs.philips.com
(914) 945-6168


Conference Committee:

Conference Chair: Kenneth A. De Jong, George Mason University
Local Arrangements: Lashon B. Booker, Naval Research Lab
Program Chair: J. David Schaffer, Philips Laboratories
Program Committee: Lashon B. Booker
Lawrence Davis, Bolt, Beranek and Newman, Inc.
Kenneth A. De Jong
David E. Goldberg, University of Alabama
John J. Grefenstette, Naval Research Lab
John H. Holland, University of Michigan
George G. Robertson, Xerox PARC
J. David Schaffer
Stephen F. Smith, Carnegie-Melon University
Stewart W. Wilson, Rowland Institute for Science

------------------------------

Subject: Abstract: ANNS and Radial Basis Functions
From: "M. Niranjan" <niranjan%digsys.engineering.cambridge.ac.uk@NSS.Cs.Ucl.AC.UK>
Date: Mon, 10 Oct 88 11:59:27 -0000

Here is an extended summary of a Tech report now available. Apologies for
the incomplete de-TeXing.

niranjan

PS: Remember, reprint requests should be sent to
"niranjan@dsl.eng.cam.ac.uk"

=============================================================================


NEURAL NETWORKS AND RADIAL BASIS FUNCTIONS
IN CLASSIFYING STATIC SPEECH PATTERNS

Mahesan Niranjan & Frank Fallside

CUED/F-INFENG/TR 22

University Engineering Department
Cambridge, CB2 1PZ, England
Email: niranjan@dsl.eng.cam.ac.uk

SUMMARY

This report compares the performances of three non-linear pattern classifiers
in the recognition of static speech patterns. Two of these classifiers are
neural networks (Multi-layered perceptron and the modified Kanerva model
(Prager & Fallside, 1988)). The third is the method of radial basis functions
(Broomhead & Lowe, 1988).

The high performance of neural-network based pattern classifiers shows
that simple linear classifiers are inadequate to deal with complex patterns
such as speech. The Multi-layered perceptron (MLP) gives a mechanism to
approximate an arbitrary classification boundary (in the feature space) to a
desired precision. Due to this power and the existence of a simple learning
algorithm (error back-propagation), this technique is in very wide use
nowadays.

The modified Kanerva model (MKM) for pattern classification is derived from
a model of human memory (Kanerva, 1984). It attempts to take advantage of
certain mathematical properties of binary spaces of large dimensionality.
The modified Kanerva model works with real valued inputs. It compares an
input feature vector with a large number of randomly populated `location
cells' in the input feature space; associated with every cell is a `radius'.
Upon comparison, the cell outputs value 1 if the input vector lies within
a volume defined by the radius; its output is zero otherwise. The discrimi-
nant function of the Modified Kanerva classifier is a weighted sum of these
location-cell outputs. It is trained by a gradient descent algorithm.

The method of radial basis functions (RBF) is a technique for non-linear
discrimination. RBFs have been used by Powell (1985) in multi-variable
interpolation. The non-linear discriminant function in this method is of the
form,

g( x) = sum_j=1^m lambda_j phi (||x - x_j||)

Here, x is the feature vector. lambda_j are weights associated with each of
the given training examples x_j. phi is a kernel function defining the
range of influence of each data point on the class boundary. For a particular
choice of the phi function, and a set of training data {x_j,f_j}, j=1,...,N,
the solution for the lambda_j s is closed-form. Thus this technique is
computationally simpler than most neural networks. When used as a non-
parametric technique, each computation at classification stage involves the
use of all the training examples. This, however, is not a disadvantage since
much of this computing can be done in parallel.

In this report, we compare the performance of these classifiers on speech
signals. Several techniques similar to the method of radial basis functions
are reviewed. The properties of the class boundaries generated by the MLP,
MKM and RBF are derived on simple two dimensional examples and an experimental
comparison with speech data is given.

============================================================================

------------------------------

Subject: Abstract: A Dynamic Connectionist Model For Phoneme Recognition
From: Tony Robinson <ajr@DSL.ENG.CAM.AC.UK>
Date: Wed, 12 Oct 88 11:29:55 -0000

For those people who did not attend the nEuro'88 connectionists conference
in Paris, our contribution is now available, abstract included below.

Tony Robinson

PS: Remember, reprint requests should be sent to
"ajr@dsl.eng.cam.ac.uk"

==============================================================================

A DYNAMIC CONNECTIONIST MODEL FOR PHONEME RECOGNITION

A J Robinson, F Fallside
Cambridge University Engineering Department
Trumpington Street, Cambridge, England
ajr@dsl.eng.cam.ac.uk

ABSTRACT

This paper describes the use of two forms of error propagation net trained
to ascribe phoneme labels to successive frames of speech from multiple
speakers. The first form places a fixed length window over the speech and
labels the central portion of the window. The second form uses a dynamic
structure in which the successive frames of speech and state vector
containing context information are used to generate the output label. The
paper concludes that the dynamic structure gives a higher recognition rate
both in comparison with the fixed context structure and with the
established k nearest neighbour technique.

============================================================================

------------------------------

Subject: tech report: Laerning Algorithm for Fully Recurrent ANNS
From: farrelly%ics@ucsd.edu (Kathy Farrelly)
Date: Wed, 12 Oct 88 14:58:00 -0700

If you'd like a copy of the following tech report, please write, call,
or send e-mail to:

Kathy Farrelly
Cognitive Science, C-015
University of California, San Diego
La Jolla, CA 92093-0115
(619) 534-6773
farrelly%ics@ucsd.edu


Report Info:

A LEARNING ALGORITHM FOR CONTINUALLY RUNNING
FULLY RECURRENT NEURAL NETWORKS

Ronald J. Williams, Northeastern University
David Zipser, University of California, San Diego

The exact form of a gradient-following learning algorithm for
completely recurrent networks running in continually sampled time
is derived. Practical learning algorithms based on this result
are shown to learn complex tasks requiring recurrent connections.
In the recurrent networks studied here, any unit can be connected
to any other, and any unit can receive external input. These
networks run continually in the sense that they sample their
inputs on every update cycle, and any unit can have a training
target on any cycle. The storage required and computation time on
each step are independent of time and are completely determined
by the size of the network, so no prior knowledge of the temporal
structure of the task being learned is required. The algorithm is
nonlocal in the sense that each unit must have knowledge of the
complete recurrent weight matrix and error vector. The algorithm
is computationally intensive in sequential computers, requiring a
storage capacity of order the 3rd power of the number of units
and computation time on each cycle of order the 4th power the
number of units. The simulations include examples in which
networks are taught tasks not possible with tapped delay lines;
that is, tasks that require the preservation of state. The most
complex example of this kind is learning to emulate a Turing
machine that does a parenthesis balancing problem. Examples are
also given of networks that do feedforward computations with
unknown delays, requiring them to organize into the correct
number of layers. Finally, examples are given in which networks
are trained to oscillate in various ways, including sinusoidal
oscillation.


------------------------------

Subject: Paper from nEuro'88 in Paris
From: Orjan Ekeberg <mcvax!bion.kth.se!orjan@uunet.UU.NET>
Date: Thu, 13 Oct 88 09:48:35 +0100

The following paper, presented at the nEuro'88 conference in Paris,
has been sent for publication in the proceedings. Reprint requests
can be sent to orjan@bion.kth.se

===============

AUTOMATIC GENERATION OF INTERNAL REPRESENTATIONS IN A
PROBABILISTIC ARTIFICIAL NEURAL NETWORK

Orjan Ekeberg, Anders Lansner

Department of Numerical Analysis and Computing Science
The Royal Institute of Technology, S-100 44 Stockholm, Sweden

ABSTRACT

In a one layer feedback perceptron type network, the connections can be
viewed as coding the pairwise correlations between activity in the
corresponding units. This can then be used to make statistical inference
by means of a relaxation technique based on bayesian inferences.

When such a network fails, it might be because the regularities are not
visible as pairwise correlations. One cure would then be to use a different
internal coding where selected higher order correlations are explicitly
represented. A method for generating this representation automatically is
presented with a special focus on the networks ability to generalize properly.

------------------------------

Subject: Cary Kornfeld to speak on neural networks and bitmapped graphics
From: pratt@zztop.rutgers.edu (Lorien Y. Pratt)
Organization: Rutgers Univ., New Brunswick, N.J.
Date: 13 Oct 88 19:22:14 +0000


Fall, 1988
Neural Networks Colloquium Series
at Rutgers

Bitmap Graphics and Neural Networks
-----------------------------------

Cary Kornfeld
AT&T Bell Laboratories

Room 705 Hill center, Busch Campus
Monday October 31, 1988 at 11:00 AM
NOTE DAY AND TIME ARE DIFFERENT FROM USUAL
Refreshments served before the talk


From the perspective of system architecture and hardware
design, bitmap graphics and neural networks are surprisingly
alike.

I will describe two key components of a graphics
processor, designed and fabricated at Xerox PARC, this engine is
based on Leo Guiba's Bitmap Calculus. While implementing that
machine I got interested in building tiny, experimental flat
panel displays. In the second part of this talk, I will
describe a few of the early prototypes and (if facilities per-
mit), will show a short video clip of their operation.
When I arrived at Bell Labs three years ago I began
building larger display panels using amorphous silicon,
thin film transistors on glass substrates. It was this display
work that gave birth to the idea of fabricating large neural
networks using light sensitive synaptic elements. In May of this
year we demonstrated working prototypes of these arrays in an ex-
perimental neuro-computer at the Atlanta COMDEX show.

This is one of the first neuro-computers built and is
among the largest. Each of its 14,000 synapses is independently
programmable over a continuous range of connection strength that
can theoretically span more than five orders of magnitude
(we've measured about three in our first-generation arrays).
The computer has an animated, graphical user interface that en-
ables the operator to both monitor and control its operation.
This machine is "programmed" to solve a pattern reconstruction
problem. (Again, facilities permitting) I will show a video tape
of its operation and will demonstrate the user interface on a
color SUN 3.
- --
- -------------------------------------------------------------------
Lorien Y. Pratt Computer Science Department
pratt@paul.rutgers.edu Rutgers University
Busch Campus
(201) 932-4634 Piscataway, NJ 08854

------------------------------

Subject: Neural Network Symposium Announcement
From: RCE1@APLVM.BITNET (RUSS EBERHART)
Organization: The Internet
Date: 15 Oct 88 17:10:39 +0000


ANNOUNCEMENT AND CALL FOR ABSTRACTS

SYMPOSIUM ON THE BIOMEDICAL APPLICATIONS OF NEURAL NETWORKS
***********************************************************
Saturday, April 22, 1989
Parsons Auditorium
The Johns Hopkins University Applied Physics Laboratory
Laurel, Maryland

The study and application of neural networks has increased significantly
in the past few years. This applications-oriented symposium focuses on
the use of neural networks to solve biomedical tasks such as the
classification of biopotential signals.

Abstracts of not more than 300 words may be submitted prior to January
31, 1989. Accepted abstracts will be allotted 20 minutes for oral
presentation.

Registration fee is $20.00 (U.S.); $10.00 for full-time students.
Registration fee includes lunch. For more information and/or to
register, contact Russ Eberhart (RCE1 @ APLVM), JHU Applied Physics
Lab., Johns Hopkins Road, Laurel, MD 20707.

The Symposium is sponsored by the Baltimore Chapter of the IEEE Engineering
in Medicine and Biology Society. Make check for registration fee payable
to "EMB Baltimore Chapter".

------------------------------

Subject: NIPS Student Travel Awards
From: terry@cs.jhu.edu (Terry Sejnowski <terry@cs.jhu.edu>)
Date: Tue, 18 Oct 88 18:00:21 -0400

We have around 80 student applications for travel awards for
the NIPS meeting in November. All students who are presenting
an oral paper or poster will receive $250-500 depending on
their expenses. Other students that have applied will very
likely receive at least $250 --- but this depends on what
the registration looks like at the end of the month.
Official letters will go out on November 1.

The deadline for 30 day supersaver fares is coming up soon.
There is a $200+ savings for staying over Saturday night,
so students who want to go to the workshop can actually
save travel money by doing so.

Terry

- -----

------------------------------

End of Neurons Digest
*********************

← previous
next →
loading
sending ...
New to Neperos ? Sign Up for free
download Neperos App from Google Play
install Neperos as PWA

Let's discover also

Recent Articles

Recent Comments

Neperos cookies
This website uses cookies to store your preferences and improve the service. Cookies authorization will allow me and / or my partners to process personal data such as browsing behaviour.

By pressing OK you agree to the Terms of Service and acknowledge the Privacy Policy

By pressing REJECT you will be able to continue to use Neperos (like read articles or write comments) but some important cookies will not be set. This may affect certain features and functions of the platform.
OK
REJECT