DADiSP / NeuralNet
Neural Network Module
DADiSP NeuralNet is an add-on module to DADiSP that provides direct and easy access to the demonstrated predictive power and pattern recognition capability of neural networking technology. With DADiSP/NeuralNet users can build their own artificial neural networks (ANNs) and apply them to achieve more accurate predictions and pattern classifications.
- Menu-driven Network design.
- Automatic normalisation of data.
- Choice of the number of hidden layers.
- Unlimited input and output variables.
- Unlimited rumber of runs.
- Cross-validation training to verify output results simultaneously.
- Built-in protection against local minima distorting output results.
- User selectable desired mean square error, minimum gradient norm and desired absolute error.
- Digital Error, Analogue Error, Maximum Error and Gradient Values Post-Training Error Graph Types.
- Extract Random Seeds gives the values used to start a network and enables building another network with the same weights.
- Extract network weights returns the weights and biases that define the network.
Neural network training
Neural networks resemble the human brain because they can learn. A back-propagation neural network develops its predictive capabilities by being trained on a set of historical inputs and known resulting outputs. The neural net applies random weights to each designated input variable. It then adjusts the weights depending on how closely the actual output values match the desired output values in the training set of historical data. Once the appropriate variable weights have been set that minimise the difference between expected and actual output from the neural net, the neural net can then be applied to new data for classification.
Back-propagation learning algorithm
DADiSP/NeuralNet employs the back-propagation learning algorithm. Back-propagation has become the most widely used neural network paradigm for modelling, forecasting, and classification. To minimise the error in the network DADiSP/NeuralNet uses a rapid-descent algorithm derived from the Vogl method of locating the global minimum. Since results depend on the initial conditions the neural net module allows you to train a lot of neural networks on the same data with different initial configurations and pick the best one.
Powerful preprocessing functions
Preprocessing of the data is one of the largest problems in using neural network tools. DADiSP/NeuralNet is fully integrated with DADiSP so hundreds of analysis functions are available to pre- and post-process neural network data. DADiSP has mathematical and statistical functions to scale, filter and process the data to identify features to be learned by the neural network. A typical Worksheet will contain the preprocessing steps, the neural network and the output results. Simply change the input data or initial conditions and each dependent Window is automatically recalculated. You immediately see the effects of your changes on the neural network.
DADiSP/NeuralNet includes several functions to create, apply and analyse neural networks.
|applynet||Apply a neural network to data.|
|getrundata||Extract a particular set of data for a run.|
|getseeds||Extract neural network random seed values.|
|getweights||Extract neural network weights.|
|makenet||Create a neural network.|
|normalise||Normalise target data to +-1 range.|
Hier geht es weiter
Sie möchten kaufen?
Was sagen unsere Kunden über uns?
Here is to the best software support team – ever! Many, many thanks. You have definitely made my day! yay!!HE
You have set a benchmark standard that many other companies should aspire to.JS, Chippenham, UK
Amazing, wish I had learned about Endnote 5 years ago
Ref: Mathtype 6.6a. I love this product, thanks for decent support on this softwareJT, Derbyshire, UK
Zur Zeit ist es uns leider nicht möglich die folgenden Produkte anzubieten. Wir arbeiten jedoch daran Ihnen ein möglichst vielfältiges Produktsortiment zur Verfügung zu stellen. Bitte kontaktieren Sie uns wenn Sie Fragen oder Anregungen haben.