当前位置:网站首页>Introduction of neural network (BP) in Intelligent Computing

Introduction of neural network (BP) in Intelligent Computing

2022-06-22 20:25:00 Foreign rookie

Catalog

1. Reasons for the popularity of neural networks

2. Definition of artificial neural network

3. Learning ability of artificial neural network

4.  The basic principle of artificial neural network

5. Research progress of neural networks

6.  Typical structure of neural network

6.1 Single layer perceptron network

6.2 Feedforward network

6.3 Feedforward inner interconnection network

6.4 Feedback network

6.5 Full Internet

7. Learning algorithm of neural network

7.1 Learning methods

7.2 Learning rule

8. BP neural network

8.1 The basic idea

8.2 Algorithm flow

8.3 give an example


1. Reasons for the popularity of neural networks

         The reason why artificial neural networks are so popular , From the characteristics and advantages of artificial neural network , Mainly in the following three aspects :
        (1) With self-learning function . for example , When realizing image recognition , Just input many different image templates and corresponding recognition results into the artificial neural network , The network will use the self-learning function , Learn to recognize similar images slowly . The self-learning function is particularly important for prediction , It is expected that the future artificial neural network computer will provide economic forecasts for mankind , market prediction , Benefit forecast , Its application prospect is very great .
        (2) With associative storage function . This association can be realized by using the feedback network of artificial neural network .(3) It has the ability to find the optimal solution at high speed . Find the optimal solution of a complex problem , It often requires a lot of computation , Using a feedback artificial neural network designed for a certain problem , Give full play to the computer's high-speed computing ability , The optimization solution may be found soon .

2. Definition of artificial neural network

         A mathematical model of artificial neural network simulating human brain and its activities , It consists of a large number of processing units (process element) Interconnected in an appropriate manner to form , It is a large-scale nonlinear adaptive system .1988 year ,Hecht-Nielsen The following definitions have been given to artificial neural networks :
        “ Artificial neural network is a parallel network 、 Distributed processing structure , It consists of a processing unit and an undirected signal channel called a connection . These processing units have local memory , And can complete local operations . Each processing unit has a single output connection , This output can be branched into multiple parallel joins as needed , And these parallel connections all output the same signal , That is, the signal of the corresponding processing unit and its signal are too small to change with the number of branches . The output signal of the processing unit can be any desired mathematical model , The operations performed in each processing unit must be completely local . in other words , It must only depend on the current values of all input signals reaching the processing unit through the input connection and the values stored in the local range of the processing unit .”

This definition mainly emphasizes the following aspects of neural networks 4 All aspects :

        (1) parallel 、 Distributed processing structure ;
        (2) The output of a processing unit can be arbitrarily branched and of constant size ;
        (3) The output signal can be any mathematical model ;
        (4) Local operation of processing unit
         The processing unit here is the artificial neuron (artificial neuron ,AN). 

according to Rumellhart ,MeClelland , Hinton And so on (paralleldistributed processing ,PDP) Model , The artificial neural network consists of 8 It is composed of three elements :

        (1) A set of processing units (PE or AN);
        (2) Activation status of the processing unit (a);
        (3) Output function of each processing unit (f);
        (4) The connection mode between processing units ;
        (5) Transfer rules ;
        (6) An activation rule that combines the input of the processing unit with the current state to generate an activation value (F);
        (7) Learning rules for modifying connection strength through experience ;
        (8) The environment in which the system operates ( Sample set ). 

1987 year .Simpson Starting from the topological structure of artificial neural network , The following definition is given, which is not too strict but concise :
“ Artificial neural network is a nonlinear directed graph . The graph contains weighted edges that can store patterns by changing the weight size , And you can find patterns from incomplete or unknown inputs .”
For general applications , Artificial neural networks can be called PDP outside , It can also be called an artificial nervous system (ANS)、 neural network (NN)、 Adaptive systems (adaptive systems)、 Adaptive network (adap-tive networks), Join model (connectionism) And neural computers (neuron computer) etc. .

The artificial neural network not only simulates the biological neural system in form , It does have some of the brain in two ways basic feature :
        (1) Neurons and their connections . From the form of system composition , Because the artificial neural network is inspired by the biological neural system , From the neuron itself to the connection pattern , They basically work in a way similar to the biological nervous system . there AN And biological neurons (BN) Corresponding , Connections that can change strength correspond to synapses .
        (2) Storage and processing of information . In terms of performance characteristics , Artificial neural network also tries to simulate the basic operation mode of biological neural system . for example , Through the corresponding learning / Training algorithm , Abstract the data relation contained in a large data set , Just like people can constantly explore the law , Sum up the experience , A general rule or framework can be found from the previous examples , And then generate new instances as required .

3. Learning ability of artificial neural network

         Artificial neural network can change its behavior according to its environment . in other words , Artificial neural network can accept the sample set submitted by users , According to the algorithm given by the system , Constantly modify the strength of the connections between neurons used to determine the behavior of the system , And after the basic composition of the network is determined , This change is made naturally according to the set of samples it receives . Generally speaking , Users do not need to adjust the learning algorithm of the network according to the sample set they encounter , in other words , Artificial neural network has good learning function .
         The neural network is in contrast to the learning problems of traditional artificial intelligence systems . In the research of traditional artificial intelligence system , Although people are not satisfied with “ machine learning ” Enough attention has been paid to the problem and great efforts have been put into it , however , The poor self-learning ability of the system is still the biggest obstacle to the wide application of these systems , The artificial neural network has a good learning function , It makes people have a great interest in it . This characteristic of artificial neural network is called “ Nature has a learning function ”.
         In the learning process , The artificial neural network continuously extracts the basic information contained in the set from the accepted sample set , It is stored in the system in the form of connection weights between neurons . for example , For a pattern , A network can be trained with its data containing different noises , On the premise that these data are properly selected , When the network encounters similar data with certain defects in the future , You can still get the complete schema corresponding to it . This shows that , An artificial neural network can learn to produce patterns on demand that it has never encountered . This function of artificial neural network is called “ abstract ”.

4.  The basic principle of artificial neural network

The structure of neurons is as follows :

  There are biological neurons in biology , What is the relationship between biological neurons and artificial neurons ?

From the figure, we can get a rough idea of how artificial neurons imitate biological neurons to work :

 Insert picture description here

5. Research progress of neural networks

6.  Typical structure of neural network

There are many different models of artificial neural networks , You can usually press the following 5 Principles categorize .·

  • According to the network structure , There are forward networks and feedback networks .
  • Distinguish according to the learning style , There are supervised learning networks and unsupervised learning networks .
  • Distinguish according to network performance , There are continuous and discrete networks , Stochastic and deterministic networks .· Distinguish according to synaptic properties , There are first-order linear correlation networks and higher-order nonlinear correlation networks .
  • According to the hierarchical simulation of biological nervous system , There is a neural hierarchy model , Combined model ﹑ Network hierarchy model ﹑ Hierarchical model and intelligent model of nervous system .

6.1 Single layer perceptron network

         The single-layer perceptron was first used , It is also the simplest neural network structure , It consists of one or more linear threshold units .

Because this network structure is relatively simple , Therefore, the ability is also very limited , Generally, it is less used .

6.2 Feedforward network

         The signal of feedforward network is transmitted unidirectionally from input layer to output layer .

         The neurons in each layer are only connected to the neurons in the previous layer , Only accept the information transmitted from the previous layer .

         It is a widely used neural network model , Because its structure is not too complex , Learning and adjusting programs are also easier to operate , And because of the multi-layer network structure , Its ability to solve problems has also been significantly strengthened , Basically, it can meet the use requirements .

6.3 Feedforward inner interconnection network

         This network structure is still a feedforward network from the outside , But there are some internal nodes interconnected in the layer .

6.4 Feedback network

         This network structure also establishes another relationship between input and output , The output layer of the network has a feedback loop to the input layer as an input of the input layer , The network itself is feedforward .

         The input layer of this neural network not only receives external input signals , At the same time, it receives the output signal of the network itself . The output feedback signal may be the original output signal , It can also be a converted output signal ; It can be the output signal at this time , It can also be an output signal after a certain delay .

         Such networks are often used for system control 、 Real time signal processing and other occasions that need to be adjusted according to the current state of the system .

6.5 Full Internet

Fully interconnected network is a network between all neurons all There are mutual connections .

  Such as Hopfiled and Boltgmann Networks are of this type .

7. Learning algorithm of neural network

7.1 Learning methods

  • Supervised learning Supervised learning : It has characteristics (feature) And labels (label) Of , Even if there is no label , The machine can also use the relationship between features and labels , Judge the label .
  • Unsupervised learning Unsupervised learning : Only features , No label .
  • Semi-Supervised learning Semi-supervised learning : Data used , Part of it is marked , And most of them are unmarked . Compared with supervised learning , The cost of semi supervised learning is lower , But it can achieve high accuracy .
  • Reinforcement learning Reinforcement learning : Reinforcement learning also uses unlabeled data , But there are some ways to know whether you are getting closer or farther away from the correct answer ( Reward and punishment function ).

7.2 Learning rule

Hebb Learning rule : The simple understanding is , If the processing unit receives an input from another processing unit , And if both units are highly active , At this time, the connection weight between the two units will be strengthened .
Hebb Learning rules is a learning method without guidance , It only changes the weight according to the activation level between neuronal connections , Therefore, this method is also called correlation learning or parallel learning .

Delta(δ \deltaδ) Learning rule :Delta Rules are the most commonly used learning rules , The key point is to change the connection weight between units to reduce the error between the actual output and the expected output of the system .
This rule is also called Widrow-Hoff Learning rule , First, in the Adaline Apply... In the model , It can also be called Minimum mean square deviation rule .

BP The learning algorithm of network is called BP Algorithm , Is in Delta Developed on the basis of rules , It can learn effectively on multi-layer network .

Gradient descent learning rule : The key point of gradient descent learning rule is in the learning process , Keep the gradient of the error curve down .
The error curve may have a local minimum , When learning online , Try to get rid of the local minimum of the error , And reach the real minimum error .
At present, many algorithms to reduce the value of loss function use the learning rule of gradient descent . There are many categories of gradient descent , Those who are interested can find out by themselves .

Kohonen Learning rule : The rule is made by Teuvo Kohonen Based on the study of biological system learning , Only for networks that do not train under guidance .

Backward propagation learning rules : Back propagation (Back Propagation,BP) Study , It is a widely used neural network learning rule .

Probabilistic learning rules : From statistical mechanics 、 Starting from the standard of steady-state energy in molecular thermodynamics and probability theory , The way of learning neural network is called probabilistic learning .
The error curve may have a local minimum , When learning online , Try to get rid of the local minimum of the error , And reach the real minimum error .

Competitive learning rules : Competitive learning belongs to unsupervised learning . This learning method uses excitatory connections between neurons in different layers and the same excitatory connections between neurons close to each other in the same layer , But the nerve which is far away has no inhibitory connection .
The essence of competitive learning rules is that high-level neurons in neural networks compete to identify the input patterns of low-level neurons .

8. BP neural network

8.1 The basic idea

        BP Neural network is also called : Feedforward neural network with backward propagation learning (Back Propagation Feed-forward Neural Network,BPFNN/BPNN), It is one of the most widely used neural networks .

         stay BPNN in , Backward propagation is a learning algorithm , Embodied in BPNN Training process of , This process needs teacher guidance ; Feedforward network is a kind of structure , Embodied in BPNN The network structure of .

         Back propagation algorithm through iterative processing , Constantly adjust the weight of the network connecting neurons , The error between the final output result and the expected result is minimized .

        BPNN It is a typical neural network , It is widely used in various classification systems , It also includes two stages of training and use . Because the training phase is BPNN The basis and premise for putting into use , The use phase itself is a very simple process , That is, give input ,BPNN Will calculate according to the trained parameters , Get the output .

8.2 Algorithm flow

  In fact, this is essentially similar to the principle of perceptron .

Steps are as follows :

8.3 give an example

That's all for Yangyang , It's my first time to learn , Please advise , If you have questions, just comment or chat privately , When you see it, you will reply , First of all , I don't know when I will go online , Sometimes I don't come here once a week . Ha ha ha ! 

 

原网站

版权声明
本文为[Foreign rookie]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/173/202206221852574617.html