DADiSP / NeuralNet

Neural Network Module

neural networkDADiSP NeuralNet is an add-on module to DADiSP that provides direct and easy access to the demonstrated predictive power and pattern recognition capability of neural networking technology. With DADiSP/NeuralNet users can build their own artificial neural networks (ANNs) and apply them to achieve more accurate predictions and pattern classifications.
 

 

 

Key features

  • Menu-driven Network design.
  • Automatic normalisation of data.
  • Choice of the number of hidden layers.
  • Unlimited input and output variables.
  • Unlimited rumber of runs.
  • Cross-validation training to verify output results simultaneously.
  • Built-in protection against local minima distorting output results.
  • User selectable desired mean square error, minimum gradient norm and desired absolute error.
  • Digital Error, Analogue Error, Maximum Error and Gradient Values Post-Training Error Graph Types.
  • Extract Random Seeds gives the values used to start a network and enables building another network with the same weights.
  • Extract network weights returns the weights and biases that define the network.

Neural network training

Neural networks resemble the human brain because they can learn. A back-propagation neural network develops its predictive capabilities by being trained on a set of historical inputs and known resulting outputs. The neural net applies random weights to each designated input variable. It then adjusts the weights depending on how closely the actual output values match the desired output values in the training set of historical data. Once the appropriate variable weights have been set that minimise the difference between expected and actual output from the neural net, the neural net can then be applied to new data for classification.

Back-propagation learning algorithm

DADiSP/NeuralNet employs the back-propagation learning algorithm. Back-propagation has become the most widely used neural network paradigm for modelling, forecasting, and classification. To minimise the error in the network DADiSP/NeuralNet uses a rapid-descent algorithm derived from the Vogl method of locating the global minimum. Since results depend on the initial conditions the neural net module allows you to train a lot of neural networks on the same data with different initial configurations and pick the best one.

Powerful preprocessing functions

Preprocessing of the data is one of the largest problems in using neural network tools. DADiSP/NeuralNet is fully integrated with DADiSP so hundreds of analysis functions are available to pre- and post-process neural network data. DADiSP has mathematical and statistical functions to scale, filter and process the data to identify features to be learned by the neural network. A typical Worksheet will contain the preprocessing steps, the neural network and the output results. Simply change the input data or initial conditions and each dependent Window is automatically recalculated. You immediately see the effects of your changes on the neural network.

NeuralNet functions

DADiSP/NeuralNet includes several functions to create, apply and analyse neural networks.

applynet Apply a neural network to data.
getrundata Extract a particular set of data for a run.
getseeds Extract neural network random seed values.
getweights Extract neural network weights.
makenet Create a neural network.
normalise Normalise target data to +-1 range.

Hier geht es weiter

Sie möchten kaufen?

Was sagen unsere Kunden über uns?

I have had a very helpful response, and have passed it on to my Oxford colleagues with success. I am impressed with the way Adept seems to maintain the same staff for many years, which speaks very well for the organisation.

HG, Oxford, UK

Best technical support for a software program that I’ve come across – painless for once!

KH, Tunbridge Wells, UK

You have set a benchmark standard that many other companies should aspire to.

JS, Chippenham, UK

Bob’s great, he should be franchised.