**The static adjustment of a neural network previous training of a nuclear neural network is considered. The choice of synoptic scales and displacement for the first and the latent layers of a neural network is carried out.** The basic mathematical parities which are used at performance of algorithm of training of a nuclear neural network are resulted.

At designing of information systems on the basis of neural networks the training of nuclear neural networks is played the important role.

Before training it is necessary to set initial values of weight factors in the neural network in this or other way. The weight factors are usually initialized randomly. Statistical adjustment is intended to improve the initialization algorithm on the basis of additional information about the data [1,2]. The purpose of static adjustment is:

- reduction of input symbols real levels to the operating range of the neural network by the expedient choice of synoptic weights and displacement in the first layer;
- reduction of the network output signals to the range of true values by the expedient choice of synoptic weights and displacement in the last layer;
- setting the initial state of the network into the area of general position, that is, maximum sensitivity to parameter variation by optimal choice of initial values of displacements and synoptic weights for the neurons of hidden layers.

Let us mark the weight factors of synoptic maps for layer m as:

*wi ^{m}*(

*a*,

*b*) (1)

where *i* is the number of the nucleus, and *m* is number of the layer.

_{A}m

Data processing for the nucleus * ^{Ai}* is defined by the expression

*Sm* (*V*) = ∑ *X (u)W' (u, v)*

* ^{i} u ^{i i}* (2)

where *u* = 1,2, ..., p.

In the training algorithm along with synoptic chart adjustment the nonlinear functions selection by changing their arguments displacement is performed. Formally, displacement is realized by adding a fictitious (dummy) ^{m}

^{x}i^{(*)} coordinate, which usually has a constant value equal to +1, and, besides, for each neuron nucleus an adjusted synoptic weight value ^{w}i^{(*,}^{b}^{)} is added. This technique allows you not to change the form of the expression (7), assuming that the number of summands in this sum is increased by one.

Static adjustment of the first layer Source information:

1) For each input variable ^{x}^{(}^{u}^{)} the following is considered to be known - the average value of the variable: * ^{x}* - change range: ʌ

^{х}*Optimality principles:*

For each input neuron the input variables change in the authentic levels range should have about the same change of the neuron output signal.

The level of displacement for each neuron must be set so that to provide at the average the maximum value of the derivative for the activation function in the training sampling.

Design equations:

29

30