Copy Link
Add to Bookmark
Report
Neuron Digest Volume 10 Number 18
Neuron Digest Thursday, 12 Nov 1992 Volume 10 : Issue 18
Today's Topics:
Neural Nets Based Systems
Re: Neural Nets Based Systems
Re: Neural Nets Based Systems
Stock Market
Stock Market
Help on hybrid systems
Postdoc Position in Lund
Free Neural Network Simulation and Analysis SW (am6.0)
Send submissions, questions, address maintenance, and requests for old
issues to "neuron-request@cattell.psych.upenn.edu". The ftp archives are
available from cattell.psych.upenn.edu (130.91.68.31). Back issues
requested by mail will eventually be sent, but may take a while.
----------------------------------------------------------------------
Subject: Neural Nets Based Systems
From: ehf@philabs.philips.com (Eberhard Fisch)
Organization: Philips Laboratories, Briarcliff, NY 10510
Date: 09 Nov 92 15:09:59 +0000
[[ Editor's Note: Given the recent post about a "stock market game," I
gleaned this message and the following two from a mailing list devoted to
general investing. Just For Your Information, of course. -PM ]]
I am interested in investigating the potential of neural networks for
forcasting markets. Im looking for articles or books dealing with the
application of neural nets for forcasting market moves or finding
patterns that occur in markets. I am not looking for, nor do I expect,
neural nets or any other model for market analysis to be a magic
solution. I just want to see whether neural network based systems have
merit. Any opinions from people who have had some experience in using
some of the standard neural network software or have developed their own
software are welcome.
Eberhard Fisch
Philips Labs, NY
------------------------------
Subject: Re: Neural Nets Based Systems
From: venky@thelonious.bellcore.com (G A Venkatesh)
Organization: Bellcore, Morristown NJ
Date: 09 Nov 92 18:56:52 +0000
Re: article by ehf@philabs.philips.com (Eberhard Fisch) writes:
I don't know where you could find more information about them but there are
at least two mutual funds (Fidelity Disciplined Equity and Fidelity Stock
Selector) that use a neural net program to help with their buy/sell decisions.
They seem to be doing well but I don't know how much the programs help.
venky
------------------------------
Subject: Re: Neural Nets Based Systems
From: mincy@think.com (Jeffrey Mincy)
Organization: Thinking Machines Corporation, Cambridge MA, USA
Date: 09 Nov 92 19:13:21 +0000
Re: article by ehf@philabs.philips.com (Eberhard Fisch) writes:
The WSJ did the following article a few weeks ago:
"Fidelity's Bradford Lewis Takes Aim at Indexes With His 'neural Network'
Computer Program, by Robert Mcgough
...
This fund is an index killer, he says of his fidelity disciplined equity fund.
Mr Lewis leaves the stock-picking to a "neural network," a computer program
that tries to mimic the intricate structure of the human brain.
...
Disciplined equity has beaten the S&P 500 by 2.3 to 5.6 percentage points in
each of three years to 1991. It's doint the same in the ytd with 5.8% return
...Only three other stock funds tracked by lipper did better than the market in
all four of these periods.
...
Anyway, perhaps this gives you some information.
-- jeff
mincy@think.com
------------------------------
Subject: Stock Market
From: D.Navin.Chandra@ISL1.RI.CMU.EDU
Date: Sun, 08 Nov 92 20:26:40 -0500
[[ Editor's Note: See previous and following note about the use of Neural
Networks in the stock market. It should be noted also that various "dart
board" investment strategoes regularly outperform the profession money
managers. See also Vol 10 #11 and #13 for related bibliographies. -PM ]]
Gary Bradsky
I agree the announcement was wrong, but please dont belittle the things
that other people try to work on. The Stock Market game is a legitimate
game entered into by 1000's of students from across the country. Many
people use NNets and GA's to "crack" the market trend. The area has also
been the most successful application of NN to date.
navin
------------------------------
Subject: Stock Market
From: bradski@cns.bu.edu
Date: Mon, 09 Nov 92 21:01:49 -0500
[[ Re: previous message from D.Navin.Chandra@ISL1.RI.CMU.EDU ]]
If the stock market game had not been a money making scheme, but simply
an "educational" contest as advertised, I would not have complained at
all. I also don't personally care if some people make this type of
contest a money making business -- just let them advertise in magazines,
or TV, not in a news group (I get enough ads as it is).
Now, to add increasing amounts of content:
(1) APPLICATIONS OF NN:
Stock market price prediction is not "the most successful application
of NN to date", the most successful application of neural networks is
clearly in adaptive filtering -- particularly for echo cancellation.
Adalines or their variants (lattice filters etc...) are used in nearly
every long distance phone line, modems, in EKG, EEG, ultra sound and
the list goes on.
(2) PREDICTION, STOCKS, AND SELECTIVE INFORMATION:
Neuron readers ought to be aware, that even if someone "cracks" the
market with their NN application and wins a stock market contest, it
means absolutely nothing more than if someone won the contest by throwing
darts at the Wall St Journal and looking for the holes on the stock
quotation pages. It means nothing more because we know nothing about the
statistical sample out of which we hear about the "winner" (and never the
losers). If 5000 people try to predict stocks with backprop and one of
them wins big, and the rest fall around the mean of stock performance,
should you put your *future* money with the one winner? Should you use
backprop? Or, did the winner just happen to hit on a lucky fit of the
now, past data, that tells you nothing about future performance?
(3) PREDICTING MARKET TRENDS MAY BE IN NEED OF BELITTLEMENT:
Trying to "crack the market trend" may in fact be hopeless, if by "trend"
one means the future expected value of stock prices. If you are going to
use NNs with financial data, get an intro into financial theory (but be
prepared for a little math). I suggest: Easy intro: John Hull, "Options,
Futures, and Other Derivative Securities", Prentice-Hall, 1989, More
cutting edge, but still decipherable: "Continuous-Time Finance", R.C.
Merton, Basil Blackwell, 1992 (get the cheaper paperback version).
In Continuous-Time Finance, read chapter 3 for an easy development of
stochastic calculus and Ito's lemma. In chapter 3, Merton uses "order
statistics" to develop his equations (order statistics just tell you how
fast a term blows up or goes to zero). On page 68-9 he develops a model
of stock market price changes based on an Ito process:
X(t) - X(t-1) = mean(t)*h + std_dev(t)*noise*h^(1/2) (A)
where h is a small time increment which goes to zero in the limit, and
noise can be taken as zero-mean Gaussian. The point of all this is that
in a small increment of time, the h^(1/2) "noise" or variance term
dominates the mean. This is observed in practice, eg. the variability of
a stock's price swamps out the level, or expected value of its price.
Moreover to estimate the mean of (A) over n observations over a total
time period T (T = h*n), we'd use:
est_mean = Sum_{k=1 to n} (X(k) - X(k-1))/T = mean(t)
In other words, the estimate of the mean is not affected by choosing
finer observation intervals, only in the length of time "T" that we
measure over. For estimating the variance of (A):
est_var = (Sum_{k=1 to n} (X(k) - X(k-1))^2/T = std_dev^2 + h*mean^2
= std_dev^2 + (T/n)*mean^2
becomes better and better with finer measurements (larger n).
THE CONCLUSION: variances (or covariances) are easier to estimate from
stochastic time series than are the means. This is a fundamental fact,
which no algorithm, no matter how neural or not, can overcome.
WHAT IT "MEANS": The stock and option models developed by Black-Scholes,
Merton and others have shown themselves to be useful (so useful in fact
that they are essentially what are keeping our large banks profitable
now). These models don't depend on having to know the mean, but do depend
on having to estimate the variance of stocks. THUS, if you want to come
up with a useful NN financial application, try using NNs to estimate
stock variance, not stock prices. You might start by using adaptive
filters as per (1) above to do this.
--Gary Bradski (bradski@cns.bu.edu)
------------------------------
Subject: Help on hybrid systems
From: Fernando Passold <EEL3FPS%BRUFSC.bitnet@UICVM.UIC.EDU>
Organization: Universidade Federal de Santa Catarina/BRASIL
Date: Thu, 12 Nov 92 17:37:23 -0300
[[ Editor's Note: I assume this person wanted his message to be
published. If you, loyal reader, feel *your* reply to this fellow
might be of more general interest, please send a copy to
neuron-request@cattell.psych.upenn.edu -PM ]]
I would like to know the intend of this list and if I
could obtain some replies.
I am a researcher tacking part of Biomedical Engineering
Group Laboratory of Electrical Engineering Department at the
Federal University of Santa Catarina. We develop little
biomedical equipments and Expest Systems - such a Hospital
Infection Control Aid and a Hybrid System (rule-based and
neural network based) for Proposal and Evaluation of
Anesthesian Plan.
I am a master degree student, now, grapple in develop a
Expert Network (also called, hybrid system: those whom
combine neural networks with rule-based methods) for
Planning and Evaluation of Plans of Anesthesia, continuing
a PHD thesis, for Critical Patients (their who need
critical cares) or Problem Patients (exceptional cases of
patientes whom evaluate to critical patients). I have some
troubles choosing the most suitable approach to develop
this system. I do not know if neural networks could be the
key to solve the main part of the problem, because we are
dealing with exceptions that probably could best solved
throught a rule-based method. At lately, there are a few
shells systems applying object-orienthed technics with
rule-based or frames methods such the new Kappa PC
Application Development Systems for Windows environment,
from IntellCorp. Inc.. So, I am interested in
implementations of hybrid systems using object-oriented
programming. Maybe, there will be a way to link neural
networks simulators using object-oriented approach to
another heuristical languages such as Turbo-Prolog object-
oriented.
I would be glad if someone could make a comment about
it, as well as indicate if there exists a _similar research_.
Please, reply directly to me as I am not subcribed to
this list.
Thanks a lot respect to this matter,
best regards,
Fernando Passold
Biomedical Engineering Group Lab.
UFSC/BRAZIL
E-mail: ee3fps@brufsc.bitnet
P.S.: I am interested in subcribing in your list if not a
large amount of material is normaly posted to it.
------------------------------
Subject: Postdoc Position in Lund
From: carsten@thep.lu.se
Date: Tue, 10 Nov 92 15:01:40 +0100
A two year postdoc position will be available within the Complex Systems
group at the Department of Theoretical Physics, University of Lund,
Sweden, starting September 1st 1993. The major research area of the group
is Artificial Neural Networks with tails into chaos and difficult
computational problems in general. Although some application studies
occur, algorithmic development is the focus in particular within the
following areas:
* Using Feed-back ANN for finding good solutions to combinatorial
optimization problems; knapsacks, scheduling, track-finding.
* Time-series prediction.
* Robust multi-layer perceptron updating procedures including noise.
* Deformable template methods -- robust statistics.
* Configurational Chemistry -- Polymers, Proteins ...
* Application work within the domain of experimental physics, in particular
in connection with the upcoming SSC/LHC experiments.
Lund University is the largest campus in Scandinavia located in a
picturesque 1000 year old city (100k inhabitants). Lund is strategically
well located in the south of Sweden with 1.5 hrs commuting distance to
Copenhagen (Denmark).
The candidate should have a PhD in a relevant field, which need not be
Physics/Theoretical Physics.
Applications and three letters of recommendation should be sent to (not
later than December 15):
Carsten Peterson
Department of Theoretical Physics
University of Lund
Solvegatan 14A
S-223 62 Lund
Sweden
or
Bo S\"{o}derberg
Department of Theoretical Physics
University of Lund
Solvegatan 14A
S-223 62 Lund
Sweden
------------------------------
Subject: Free Neural Network Simulation and Analysis SW (am6.0)
From: Russell R Leighton <taylor@world.std.com>
Date: Fri, 30 Oct 92 09:09:54 -0500
*************************************************************************
**** delete all prerelease versions!!!!!!! (they are not up to date) ****
*************************************************************************
The following describes a neural network simulation environment made
available free from the MITRE Corporation. The software contains a neural
network simulation code generator which generates high performance ANSI C
code implementations for modular backpropagation neural networks. Also
included is an interface to visualization tools.
FREE NEURAL NETWORK SIMULATOR
AVAILABLE
Aspirin/MIGRAINES
Version 6.0
The Mitre Corporation is making available free to the public a neural
network simulation environment called Aspirin/MIGRAINES. The software
consists of a code generator that builds neural network simulations by
reading a network description (written in a language called "Aspirin")
and generates an ANSI C simulation. An interface (called "MIGRAINES") is
provided to export data from the neural network to visualization tools.
The previous version (Version 5.0) has over 600 registered installation
sites world wide.
The system has been ported to a number of platforms:
Host platforms:
convex_c2 /* Convex C2 */
convex_c3 /* Convex C3 */
cray_xmp /* Cray XMP */
cray_ymp /* Cray YMP */
cray_c90 /* Cray C90 */
dga_88k /* Data General Aviion w/88XXX */
ds_r3k /* Dec Station w/r3000 */
ds_alpha /* Dec Station w/alpha */
hp_parisc /* HP w/parisc */
pc_iX86_sysvr4 /* IBM pc 386/486 Unix SysVR4 */
pc_iX86_sysvr3 /* IBM pc 386/486 Interactive Unix SysVR3 */
ibm_rs6k /* IBM w/rs6000 */
news_68k /* News w/68XXX */
news_r3k /* News w/r3000 */
next_68k /* NeXT w/68XXX */
sgi_r3k /* Silicon Graphics w/r3000 */
sgi_r4k /* Silicon Graphics w/r4000 */
sun_sparc /* Sun w/sparc */
sun_68k /* Sun w/68XXX */
Coprocessors:
mc_i860 /* Mercury w/i860 */
meiko_i860 /* Meiko w/i860 Computing Surface */
Included with the software are "config" files for these platforms.
Porting to other platforms may be done by choosing the "closest" platform
currently supported and adapting the config files.
New Features
- ------------
- ANSI C ( ANSI C compiler required! If you do not
have an ANSI C compiler, a free (and very good)
compiler called gcc is available by anonymous ftp
from prep.ai.mit.edu (18.71.0.38). )
Gcc is what was used to develop am6 on Suns.
- Autoregressive backprop has better stability
constraints (see examples: ringing and sequence),
very good for sequence recognition
- File reader supports "caching" so you can
use HUGE data files (larger than physical/virtual
memory).
- The "analyze" utility which aids the analysis
of hidden unit behavior (see examples: sonar and
characters)
- More examples
- More portable system configuration
for easy installation on systems
without a "config" file in distribution
Aspirin 6.0
- ------------
The software that we are releasing now is for creating, and evaluating,
feed-forward networks such as those used with the backpropagation
learning algorithm. The software is aimed both at the expert
programmer/neural network researcher who may wish to tailor significant
portions of the system to his/her precise needs, as well as at casual
users who will wish to use the system with an absolute minimum of effort.
Aspirin was originally conceived as ``a way of dealing with MIGRAINES.''
Our goal was to create an underlying system that would exist behind the
graphics and provide the network modeling facilities. The system had to
be flexible enough to allow research, that is, make it easy for a user to
make frequent, possibly substantial, changes to network designs and
learning algorithms. At the same time it had to be efficient enough to
allow large ``real-world'' neural network systems to be developed.
Aspirin uses a front-end parser and code generators to realize this goal.
A high level declarative language has been developed to describe a
network. This language was designed to make commonly used network
constructs simple to describe, but to allow any network to be described.
The Aspirin file defines the type of network, the size and topology of
the network, and descriptions of the network's input and output. This
file may also include information such as initial values of weights,
names of user defined functions.
The Aspirin language is based around the concept of a "black box". A
black box is a module that (optionally) receives input and (necessarily)
produces output. Black boxes are autonomous units that are used to
construct neural network systems. Black boxes may be connected
arbitrarily to create large possibly heterogeneous network systems. As a
simple example, pre or post-processing stages of a neural network can be
considered black boxes that do not learn.
The output of the Aspirin parser is sent to the appropriate code
generator that implements the desired neural network paradigm. The goal
of Aspirin is to provide a common extendible front-end language and
parser for different network paradigms. The publicly available software
will include a backpropagation code generator that supports several
variations of the backpropagation learning algorithm. For
backpropagation networks and their variations, Aspirin supports a wide
variety of capabilities:
1. feed-forward layered networks with arbitrary connections
2. ``skip level'' connections
3. one and two-dimensional weight tessellations
4. a few node transfer functions (as well as user defined)
5. connections to layers/inputs at arbitrary delays,
also "Waibel style" time-delay neural networks
6. autoregressive nodes.
7. line search and conjugate gradient optimization
The file describing a network is processed by the Aspirin parser and
files containing C functions to implement that network are generated.
This code can then be linked with an application which uses these
routines to control the network. Optionally, a complete simulation may be
automatically generated which is integrated with the MIGRAINES interface
and can read data in a variety of file formats. Currently supported file
formats are:
Ascii
Type1, Type2, Type3 Type4 Type5 (simple floating point file formats)
ProMatlab
Examples
- --------
A set of examples comes with the distribution:
xor: from RumelHart and McClelland, et al, "Parallel Distributed
Processing, Vol 1: Foundations", MIT Press, 1986, pp. 330-334.
encode: from RumelHart and McClelland, et al, "Parallel Distributed
Processing, Vol 1: Foundations", MIT Press, 1986, pp. 335-339.
bayes: Approximating the optimal bayes decision surface for a gauss-gauss
problem.
detect: Detecting a sine wave in noise.
iris: The classic iris database.
characters: Learing to recognize 4 characters independent of rotation.
ring: Autoregressive network learns a decaying sinusoid impulse response.
sequence: Autoregressive network learns to recognize a short sequence of
orthonormal vectors.
sonar: from Gorman, R. P., and Sejnowski, T. J. (1988). "Analysis of
Hidden Units in a Layered Network Trained to Classify Sonar Targets" in
Neural Networks, Vol. 1, pp. 75-89.
spiral: from Kevin J. Lang and Michael J, Witbrock, "Learning to Tell Two
Spirals Apart", in Proceedings of the 1988 Connectionist Models Summer
School, Morgan Kaufmann, 1988.
ntalk: from Sejnowski, T.J., and Rosenberg, C.R. (1987). "Parallel
networks that learn to pronounce English text" in Complex Systems, 1,
145-168.
perf: a large network used only for performance testing.
monk: The backprop part of the monk paper. The MONK's problem were the
basis of a first international comparison of learning algorithms. The
result of this comparison is summarized in "The MONK's Problems - A
Performance Comparison of Different Learning algorithms" by S.B. Thrun,
J. Bala, E. Bloedorn, I. Bratko, B. Cestnik, J. Cheng, K. De Jong, S.
Dzeroski, S.E. Fahlman, D. Fisher, R. Hamann, K. Kaufman, S. Keller, I.
Kononenko, J. Kreuziger, R.S. Michalski, T. Mitchell, P. Pachowicz, Y.
Reich H. Vafaie, W. Van de Welde, W. Wenzel, J. Wnek, and J. Zhang has
been published as Technical Report CS-CMU-91-197, Carnegie Mellon
University in Dec. 1991.
wine: From the ``UCI Repository Of Machine Learning Databases and Domain
Theories'' (ics.uci.edu: pub/machine-learning-databases).
Performance of Aspirin simulations
- ----------------------------------
The backpropagation code generator produces simulations that run very
efficiently. Aspirin simulations do best on vector machines when the
networks are large, as exemplified by the Cray's performance. All
simulations were done using the Unix "time" function and include all
simulation overhead. The connections per second rating was calculated by
multiplying the number of iterations by the total number of connections
in the network and dividing by the "user" time provided by the Unix time
function. Two tests were performed. In the first, the network was simply
run "forward" 100,000 times and timed. In the second, the network was
timed in learning mode and run until convergence. Under both tests the
"user" time included the time to read in the data and initialize the
network.
Sonar:
This network is a two layer fully connected network
with 60 inputs: 2-34-60.
Millions of Connections per Second
Forward:
SparcStation1: 1
IBM RS/6000 320: 2.8
HP9000/720: 4.0
Meiko i860 (40MHz) : 4.4
Mercury i860 (40MHz) : 5.6
Cray YMP: 21.9
Cray C90: 33.2
Forward/Backward:
SparcStation1: 0.3
IBM RS/6000 320: 0.8
Meiko i860 (40MHz) : 0.9
HP9000/720: 1.1
Mercury i860 (40MHz) : 1.3
Cray YMP: 7.6
Cray C90: 13.5
Gorman, R. P., and Sejnowski, T. J. (1988). "Analysis of Hidden Units in
a Layered Network Trained to Classify Sonar Targets" in Neural Networks,
Vol. 1, pp. 75-89.
Nettalk:
This network is a two layer fully connected network
with [29 x 7] inputs: 26-[15 x 8]-[29 x 7]
Millions of Connections per Second
Forward:
SparcStation1: 1
IBM RS/6000 320: 3.5
HP9000/720: 4.5
Mercury i860 (40MHz) : 12.4
Meiko i860 (40MHz) : 12.6
Cray YMP: 113.5
Cray C90: 220.3
Forward/Backward:
SparcStation1: 0.4
IBM RS/6000 320: 1.3
HP9000/720: 1.7
Meiko i860 (40MHz) : 2.5
Mercury i860 (40MHz) : 3.7
Cray YMP: 40
Cray C90: 65.6
Sejnowski, T.J., and Rosenberg, C.R. (1987). "Parallel networks that
learn to pronounce English text" in Complex Systems, 1, 145-168.
Perf:
This network was only run on a few systems. It is very large with very
long vectors. The performance on this network is in some sense a peak
performance for a machine.
This network is a two layer fully connected network
with 2000 inputs: 100-500-2000
Millions of Connections per Second
Forward:
Cray YMP 103.00
Cray C90 220
Forward/Backward:
Cray YMP 25.46
Cray C90 59.3
MIGRAINES
- ------------
The MIGRAINES interface is a terminal based interface that allows you to
open Unix pipes to data in the neural network. This replaces the NeWS1.1
graphical interface in version 4.0 of the Aspirin/MIGRAINES software. The
new interface is not a simple to use as the version 4.0 interface but is
much more portable and flexible. The MIGRAINES interface allows users to
output neural network weight and node vectors to disk or to other Unix
processes. Users can display the data using either public or commercial
graphics/analysis tools. Example filters are included that convert data
exported through MIGRAINES to formats readable by:
- Gnuplot 3
- Matlab
- Mathematica
- Xgobi
Most of the examples (see above) use the MIGRAINES interface to dump data
to disk and display it using a public software package called Gnuplot3.
Gnuplot3 can be obtained via anonymous ftp from:
>>>> In general, Gnuplot 3 is available as the file gnuplot3.?.tar.Z
>>>> Please obtain gnuplot from the site nearest you. Many of the major ftp
>>>> archives world-wide have already picked up the latest version, so if
>>>> you found the old version elsewhere, you might check there.
>>>>
>>>> NORTH AMERICA:
>>>>
>>>> Anonymous ftp to dartmouth.edu (129.170.16.4)
>>>> Fetch
>>>> pub/gnuplot/gnuplot3.?.tar.Z
>>>> in binary mode.
>>>>>>>> A special hack for NeXTStep may be found on 'sonata.cc.purdue.edu'
>>>>>>>> in the directory /pub/next/submissions. The gnuplot3.0 distribution
>>>>>>>> is also there (in that directory).
>>>>>>>>
>>>>>>>> There is a problem to be aware of--you will need to recompile.
>>>>>>>> gnuplot has a minor bug, so you will need to compile the command.c
>>>>>>>> file separately with the HELPFILE defined as the entire path name
>>>>>>>> (including the help file name.) If you don't, the Makefile will over
>>>>>>>> ride the def and help won't work (in fact it will bomb the program.)
NetTools
- -----------
We have include a simple set of analysis tools by Simon Dennis and Steven
Phillips. They are used in some of the examples to illustrate the use of
the MIGRAINES interface with analysis tools. The package contains three
tools for network analysis:
gea - Group Error Analysis
pca - Principal Components Analysis
cda - Canonical Discriminants Analysis
Analyze
- -------
"analyze" is a program inspired by Denis and Phillips' Nettools. The
"analyze" program does PCA, CDA, projections, and histograms. It can read
the same data file formats as are supported by "bpmake" simulations and
output data in a variety of formats. Associated with this utility are
shell scripts that implement data reduction and feature extraction.
"analyze" can be used to understand how the hidden layers separate the
data in order to optimize the network architecture.
How to get Aspirin/MIGRAINES
- -----------------------
The software is available from two FTP sites, CMU's simulator collection
and UCLA's cognitive science machines. The compressed tar file is a
little less than 2 megabytes. Most of this space is taken up by the
documentation and examples. The software is currently only available via
anonymous FTP.
> To get the software from CMU's simulator collection:
1. Create an FTP connection from wherever you are to machine "pt.cs.cmu.edu"
(128.2.254.155).
2. Log in as user "anonymous" with password your username.
3. Change remote directory to "/afs/cs/project/connect/code". Any
subdirectories of this one should also be accessible. Parent directories
should not be. ****You must do this in a single operation****:
cd /afs/cs/project/connect/code
4. At this point FTP should be able to get a listing of files in this
directory and fetch the ones you want.
Problems? - contact us at "connectionists-request@cs.cmu.edu".
5. Set binary mode by typing the command "binary" ** THIS IS IMPORTANT **
6. Get the file "am6.tar.Z"
> To get the software from UCLA's cognitive science machines:
1. Create an FTP connection to "ftp.cognet.ucla.edu" (128.97.50.19)
(typically with the command "ftp ftp.cognet.ucla.edu")
2. Log in as user "anonymous" with password your username.
3. Change remote directory to "alexis", by typing the command "cd alexis"
4. Set binary mode by typing the command "binary" ** THIS IS IMPORTANT **
5. Get the file by typing the command "get am6.tar.Z"
Other sites
- -----------
If these sites do not work well for you, then try the archie
internet mail server. Send email:
To: archie@cs.mcgill.ca
Subject: prog am6.tar.Z
Archie will reply with a list of internet ftp sites
that you can get the software from.
How to unpack the software
- --------------------------
After ftp'ing the file make the directory you
wish to install the software. Go to that
directory and type:
zcat am6.tar.Z | tar xvf -
-or-
uncompress am6.tar.Z ; tar xvf am6.tar
How to print the manual
- -----------------------
The user documentation is located in ./doc in a
few compressed PostScript files. To print
each file on a PostScript printer type:
uncompress *.Z
lpr -s *.ps
Why?
- ----
I have been asked why MITRE is giving away this software. MITRE is a
non-profit organization funded by the U.S. federal government. MITRE does
research and development into various technical areas. Our research into
neural network algorithms and applications has resulted in this software.
Since MITRE is a publically funded organization, it seems appropriate
that the product of the neural network research be turned back into the
technical community at large.
Thanks
- ------
Thanks to the beta sites for helping me get the bugs out and make this
portable.
Thanks to the folks at CMU and UCLA for the ftp sites.
Copyright and license agreement
- -------------------------------
Since the Aspirin/MIGRAINES system is licensed free of charge, the MITRE
Corporation provides absolutely no warranty. Should the Aspirin/MIGRAINES
system prove defective, you must assume the cost of all necessary
servicing, repair or correction. In no way will the MITRE Corporation be
liable to you for damages, including any lost profits, lost monies, or
other special, incidental or consequential damages arising out of the use
or in ability to use the Aspirin/MIGRAINES system.
This software is the copyright of The MITRE Corporation. It may be
freely used and modified for research and development purposes. We
require a brief acknowledgement in any research paper or other
publication where this software has made a significant contribution. If
you wish to use it for commercial gain you must contact The MITRE
Corporation for conditions of use. The MITRE Corporation provides
absolutely NO WARRANTY for this software.
October, 1992
Russell Leighton * *
MITRE Signal Processing Center *** *** *** ***
7525 Colshire Dr. ****** *** *** ******
McLean, Va. 22102, USA *****************************************
***** *** *** ******
INTERNET: taylor@world.std.com, ** *** *** ***
leighton@mitre.org * *
------------------------------
End of Neuron Digest [Volume 10 Issue 18]
*****************************************