The document discusses key concepts in neural networks including:
1) Neural networks consist of simple processing units (neurons) connected by weighted connections. The strength of the connection between neurons i and j is represented by the weight wi,j.
2) Data is transferred between neurons via connections. Neurons process this input data using a propagation function, activation function and optional output function to determine the neuron's output.
3) Common network topologies include feedforward networks with layers of neurons and shortcut connections, as well as recurrent networks where neurons can influence themselves. Learning strategies are used to adjust the network weights to fit the needed task.
The document discusses key concepts in neural networks including:
1) Neural networks consist of simple processing units (neurons) connected by weighted connections. The strength of the connection between neurons i and j is represented by the weight wi,j.
2) Data is transferred between neurons via connections. Neurons process this input data using a propagation function, activation function and optional output function to determine the neuron's output.
3) Common network topologies include feedforward networks with layers of neurons and shortcut connections, as well as recurrent networks where neurons can influence themselves. Learning strategies are used to adjust the network weights to fit the needed task.
The document discusses key concepts in neural networks including:
1) Neural networks consist of simple processing units (neurons) connected by weighted connections. The strength of the connection between neurons i and j is represented by the weight wi,j.
2) Data is transferred between neurons via connections. Neurons process this input data using a propagation function, activation function and optional output function to determine the neuron's output.
3) Common network topologies include feedforward networks with layers of neurons and shortcut connections, as well as recurrent networks where neurons can influence themselves. Learning strategies are used to adjust the network weights to fit the needed task.
The concept of time in neural networks • The term time or the number of cycles of the neural network is used in some definition's theory. • Time is divided into discrete time steps: • The concept of time • The current time (present time) is referred to as (t), the next time step as (t+1), the preceding one as (t−1). All other time steps are referred to analogously. • Several mathematical variables (e.g. netj or oi) refer to a certain point in time, the notation will be, for example, netj (t − 1) or oi(t).
Components of neural networks • A technical neural network consists of simple processing units, the neurons, and directed, weighted connections between those neurons. • The strength of a connection (or the connecting weight) between two neurons i and j is referred to as wi,j
Neural network • A neural network is a sorted triple (N, V, w) with two sets N, V and a function w, where N is the set of neurons and V a set {(i,j)|i,j N} whose elements are called connections between neuron i and neuron j. • The function w : V → R defines the weights, where w((i, j)), the weight of the connection between neuron i and neuron j, is shortened to wi,j . • Depending on the point of view it is either undefined or 0 for connections that do not exist in the network.
Neural network • The weights can be implemented in a square weight matrix W or, optionally, in a weight vector W • The row number of the matrix indicating where the connection begins, and the column number of the matrix indicating, which neuron is the target. • Indeed, in this case the numeric 0 marks a non- existing connection. • This matrix representation is also called Hinton diagram
Connections carry information that is processed by neurons • Data are transferred between neurons via connections with the connecting weight being either excitatory or inhibitory. • The definition of connections has already been included in the definition of the neural network.
The propagation function converts vector inputs to scalar network inputs • At a neuron , there are a lot of neurons with a connection to , i.e. which transfer their output to j. • For a neuron j the propagation function receives the outputs 𝑖1, . . . , 𝑖𝑛 of other neurons 1 2 (which are connected to ), and transforms them in consideration of the connecting weights 𝑖𝑗 into the network input 𝑗 that can be further processed by the activation function. • Thus, the network input is the result of the propagation function
Propagation function and network input • Let I = {i1, i2, . . . , in } be the set of neurons, such that z { 1 , . . . , n} : . Then the network input of j , called net j , is calculated by the propagation function fprop as follows:
• Here the weighted sum is very popular: The
multiplication of the output of each neuron i by wi,j , and the summation of the results:
The activation is the "switching status" of a neuron • The reactions of the neurons to the input values depend on this activation state . • The activation state indicates the extent of a neuron’s activation and is often shortly referred to as activation. • Definition: Let j be a neuron. The activation state , in short activation, is explicitly assigned to j , indicates the extent of the neuron’s activity and results from the activation function.
The activation function determines the activation of a neuron dependent on network input and threshold value • Definition of Activation function and Activation: • Let j be a neuron. The activation function is defined as
• It transforms the network input , as well as the
previous activation state ( t − 1) into a new activation state ( t ), with the threshold value Θ playing an important role • the activation function is often defined globally for all neurons or at least for a set of neurons and only the threshold values are different for each neuron
Common activation functions • The simplest activation function is the binary threshold function which can only take on two values (also referred to as Heaviside function). • If the input is above a certain threshold, the function changes from one value to another, but otherwise remains constant. • This implies that the function is not differentiable at the threshold and for the rest the derivative is 0
An output function may be used to process the activation once again • The output function of a neuron j calculates the values which are transferred to the other neurons connected to j . More formally: • (Output function) . Let j be a neuron. The output function
• calculates the output value o j of the neuron j from its
activation state a j . • Generally, the output function is defined globally, too. Often this function is the identity , i.e. the activation is directly output
Learning strategies adjust a network to fit our needs • General learning rule: . • The learning strategy is an algorithm that can be used to change and thereby train the neural network, so that the network produces a desired output for a given input.
Network topologies • an overview of the usual topologies (= designs) of neural networks, i.e. to construct net • Every topology is illustrated by a map and its Hinton diagram works consisting of these elements.
Feedforward networks consist of layers and connections towards each following layer
• Feedforward are the networks that consist of the
neurons that are grouped in the following layers: One input layer, n hidden processing layers (invisible from the outside, that’s why the neurons are also referred to as hidden neurons ) and one output layer. • In a feedforward network each neuron in one layer has only directed connections to the neurons of the next layer (towards the output layer) • Feedforward networks in which every neuron i is connected to all neurons of the next layer (these layers are called completely linked).
Feedforward networks • Definition Feedforward network • The neuron layers of a feedforward network are clearly separated: One input layer, one output layer and one or more processing layers which are invisible from the outside (also called hidden layers). • Connections are only permitted to neurons of the following layer.
Shortcut connections skip layers • Definition Feedforward network with shortcut connections: • Similar to the feedforward network, but the connections may not only be directed towards the next layer but also towards any other subsequent layer.
COURSE : ARTIFICIAL NEURAL NETWORK dprini@unsri.ac.id Recurrent networks have influence on themselves • Recurrence is defined as the process of a neuron influencing itself by any means or by any connection. • Recurrent networks do not always have explicitly defined input or output neurons.
Direct recurrences start and end at the same neuron • Some networks allow for neurons to be connected to themselves, which is called direct recurrence (or sometimes self-recurrence • Definition of Direct recurrence: • The feedforward network by connecting a neuron j to itself, with the weights of these connections being referred to as . In other words: the diagonal of the weight matrix W may be different from 0.
COURSE : ARTIFICIAL NEURAL NETWORK dprini@unsri.ac.id Indirect recurrences can influence their starting neuron only by making detours • If connections are allowed towards the input layer, they will be called indirect recurrences. Then a neuron j can use indirect forwards connections to influence itself, • for example, by influencing the neurons of the next layer and the neurons of this next layer influencing j • Definition Indirect recurrence • brain our network is based on a feedforward network, now with additional connections between neurons and their preceding layerbeing allowed. Therefore, below the diagonal of W is different from 0
COURSE : ARTIFICIAL NEURAL NETWORK dprini@unsri.ac.id Lateral recurrences connect neurons within one layer • Connections between neurons within one layer are called lateral recurrences • each neuron often inhibits the other neurons of the layer and strengthens itself. As a result only the strongest neuron becomes active (winner-takes-all scheme). • Definition Lateral recurrence • A laterally recurrent network permits connections within one layer.
COURSE : ARTIFICIAL NEURAL NETWORK dprini@unsri.ac.id Completely linked networks allow any possible connection • Completely linked networks permit connections between all neurons, except for direct recurrences. Furthermore, the connections must be symmetric. • A popular example are the self-organizing maps • Definition Complete interconnection • In this case, every neuron is always allowed to be connected to every other neuron – but as a result every neuron can become an input neuron. Therefore, direct recurrences normally cannot be applied • Thus, the matrix W may be unequal to 0 everywhere, except along its diagonal
COURSE : ARTIFICIAL NEURAL NETWORK dprini@unsri.ac.id The bias neuron is a technical trick to consider threshold values as connection weights • In many network paradigms neurons have a threshold value that indicates when a neuron becomes active. Thus, the threshold value is an activation function parameter of a neuron. • From the biological point of view this sounds most plausible, but it is complicated to access the activation function at runtime in order to train the threshold value. • But Threshold values Θ j1 , . . . , Θ jn for neurons j1, j2, . . . , jn can also be realized as connecting weight of a continuously firing neuron • For this purpose an additional bias neuron whose output value is always 1 is integrated in the network and connected to the neurons j1, j2, . . . , j n . These new connections get the weights − Θ j1 , . . . , − Θ jn , i.e. they get the negative threshold values
• Definition A bias neuron is a neuron whose output value is always 1 and which is represented by
• It is used to represent neuron biases as connection weights,
which enables any weight training algorithm to train the biases at the same time. • Then the threshold value of the neurons j1, j2, . . . , jn is set to 0. Now the threshold values are implemented as connection weights and can directly be trained together with the connection weights, which considerably facilitates the learning process.
COURSE : ARTIFICIAL NEURAL NETWORK dprini@unsri.ac.id Representing neurons • Another useful representation, which will be use several times is to illustrate neurons according to their type of data processing.
Neuron activations : Synchronous activation • All neurons change their values synchronously, i.e. they simultaneously calculate network inputs, activation and output, and pass them on. • Synchronous activation corresponds closest to its biological counterpart, but it is – if to be implemented in hardware – only useful on certain parallel computers and especially not for feedforward networks • Definition Synchronous activation : • All neurons of a network calculate network inputs at the same time by means of the propagation function, activation by means of the activation function and output by means of the output function. After that the activation cycle is complete.
Asynchronous activation • The neurons do not change their values simultaneously but at different points of time. • There exist different orders: Random order, Random permutation, Topological order, Fixed orders of activation during implementation
Topological order • This procedure can only be considered for non- cyclic, i.e. non-recurrent, networks, since otherwise there is no order of activation
Fixed orders of activation during implementation • Feedforward networks is very popular to determine the order of activation once according to the topology and to use this order without further verification at runtime. • But this is not necessarily useful for networks that are capable to change their topology