CN112232494A - Method for constructing pulse neural network for feature extraction based on frequency induction - Google Patents
Method for constructing pulse neural network for feature extraction based on frequency induction Download PDFInfo
- Publication number
- CN112232494A CN112232494A CN202011246943.4A CN202011246943A CN112232494A CN 112232494 A CN112232494 A CN 112232494A CN 202011246943 A CN202011246943 A CN 202011246943A CN 112232494 A CN112232494 A CN 112232494A
- Authority
- CN
- China
- Prior art keywords
- neuron
- pulse
- feature extraction
- learning
- neurons
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000605 extraction Methods 0.000 title claims abstract description 48
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 46
- 238000000034 method Methods 0.000 title claims abstract description 38
- 230000006698 induction Effects 0.000 title claims abstract description 28
- 210000002569 neuron Anatomy 0.000 claims abstract description 119
- 210000000225 synapse Anatomy 0.000 claims abstract description 67
- 238000004088 simulation Methods 0.000 claims abstract description 32
- 238000001208 nuclear magnetic resonance pulse sequence Methods 0.000 claims abstract description 16
- 230000007246 mechanism Effects 0.000 claims abstract description 11
- 238000012549 training Methods 0.000 claims abstract description 11
- 240000002627 Cordeauxia edulis Species 0.000 claims abstract description 10
- 238000012545 processing Methods 0.000 claims abstract description 10
- 230000001242 postsynaptic effect Effects 0.000 claims description 24
- 238000012421 spiking Methods 0.000 claims description 18
- 230000000946 synaptic effect Effects 0.000 claims description 16
- 230000005284 excitation Effects 0.000 claims description 15
- 238000010304 firing Methods 0.000 claims description 14
- 210000005215 presynaptic neuron Anatomy 0.000 claims description 11
- 230000000694 effects Effects 0.000 claims description 9
- 230000005764 inhibitory process Effects 0.000 claims description 9
- 239000011159 matrix material Substances 0.000 claims description 9
- 230000007774 longterm Effects 0.000 claims description 7
- 230000008569 process Effects 0.000 claims description 7
- 230000003247 decreasing effect Effects 0.000 claims description 6
- 230000002238 attenuated effect Effects 0.000 claims description 3
- 230000002860 competitive effect Effects 0.000 claims description 3
- 239000000463 material Substances 0.000 claims description 3
- 238000012067 mathematical method Methods 0.000 claims description 3
- 210000005036 nerve Anatomy 0.000 claims description 3
- 238000003062 neural network model Methods 0.000 claims description 3
- 230000001629 suppression Effects 0.000 claims description 3
- 238000012163 sequencing technique Methods 0.000 claims description 2
- 230000003313 weakening effect Effects 0.000 claims description 2
- 239000000284 extract Substances 0.000 abstract description 7
- 208000012868 Overgrowth Diseases 0.000 abstract description 4
- 230000002708 enhancing effect Effects 0.000 description 3
- 230000003956 synaptic plasticity Effects 0.000 description 3
- 210000004556 brain Anatomy 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000002401 inhibitory effect Effects 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
- 230000003518 presynaptic effect Effects 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
A method for constructing a pulse neural network for feature extraction based on frequency induction comprises establishing an input layer, performing primary processing on input data, and converting the input data into a pulse sequence; establishing a feature extraction layer, introducing a frequency induction heuristic mechanism on a pulse neuron, performing structure learning by using an accumulative growth algorithm based on an HEBB rule, performing weight learning by using an STDP rule, inputting data needing feature extraction after learning is finished, wherein the output of the last layer of the feature extraction layer is the feature, and starting simulation for each piece of training data for a plurality of times for learning. The method can effectively generate memory, efficiently extract features, have good robustness through small sample learning, and lead the network capability to be stronger, the learning to be more effective and the operation efficiency to be higher due to the introduction of a heuristic mechanism based on frequency induction, thereby leading the network synapse growth speed to be kept in a reasonable range, leading the network scale to be in a dynamic balance state and preventing the overgrowth condition.
Description
Technical Field
The invention belongs to the technical field of artificial intelligence and neural networks, and particularly relates to a method for constructing a pulse neural network for feature extraction based on frequency induction.
Background
The impulse neural network is called as a third generation artificial neural network, a model which is closer to a real neuron is considered to have greater potential in the aspect of intelligence, but now the sample processing in the process of training the impulse neural network is also based on discrete numerical values, so that the information is lost by converting numerical values into impulses, some characteristics in the samples are lost due to the loss, namely, a great deal of noise is added, the learning effect of the impulse neural network is influenced, and the learning is insufficient. There is currently a lack of an excellent method of converting features into pulse trains. When a neural network is used for processing specific problems, the efficiency of a feature extraction stage is low, the system function and the overall processing capacity are seriously influenced, and the synaptic plasticity based on frequency induction is common in the brain, such as LTP and LTD mechanisms, but the synaptic plasticity is not widely discussed in a pulse neural network learning algorithm.
Disclosure of Invention
In order to overcome the disadvantages of the prior art, the present invention aims to provide a method for constructing a spiking neural network for feature extraction based on frequency induction, which enables the spiking neural network to learn features in data for a specific task based on frequency-induced synaptic plasticity, and finally outputs a pulse sequence as an extraction of the features in the data. The method can effectively generate memory, efficiently extract features, has good robustness through small sample learning, has stronger network capability, more effective learning and higher operation efficiency due to the introduction of a heuristic mechanism based on frequency induction, can keep the network synapse growth speed in a reasonable range, and can prevent overgrowth when the network scale is in a dynamic balance state.
In order to achieve the purpose, the invention adopts the technical scheme that:
a method for constructing a pulse neural network for feature extraction based on frequency induction comprises the following steps:
step 1: establishing an input layer, performing primary processing on input data, and converting the input data into a pulse sequence;
step 2: establishing a feature extraction layer, introducing a frequency induction heuristic mechanism on a pulse neuron, performing structure learning by using an accumulative growth algorithm based on an HEBB rule, performing weight learning by using an STDP rule, inputting data needing feature extraction after learning is finished, wherein the output of the last layer of the feature extraction layer is the feature, and starting simulation for each input data sample for a plurality of times for learning.
The step 1 specifically comprises the following steps:
step 1.1: constructing an input layer using p pulse generators;
step 1.2: performing primary processing on input data samples, weakening noise in the samples by a non-neural network method, and sequencing the samples according to categories;
step 1.3: converting the data processed in the step 1.2 into pulse excitation time, and exciting the data after the value is larger;
step 1.4: a simulation is started for each sample, and the pulse generator sends a pulse at a corresponding time to convert the data into a pulse sequence.
The step 2 specifically comprises the following steps:
step 2.1: constructing a feature extraction layer by using three layers of pulse neural networks, wherein each layer of network is provided with q normal pulse neurons, and a plurality of synapses are randomly generated from a previous layer to a next layer during initialization;
step 2.2: setting a time window and an excitation counter for each pulse neuron, judging whether the pulse neuron is under the influence of frequency-induced inhibition or enhancement effect for each simulation, and if so, applying the influence;
step 2.3: for each simulation, the pulse neurons between two adjacent layers carry out structure learning through an accumulative growth algorithm based on the HEBB rule, the limitation of material and energy is guaranteed when synapses grow, namely the synapses are a long-term process and a neuron cannot grow in large quantity, and a competitive relationship exists, namely as learning progresses, some growing synapses become unimportant and degenerate;
step 2.4: for each simulation, the synapses of the spiking neurons are weight learned by the STDP rule, i.e. if the pre-synaptic neuron fires before the post-synaptic neuron, the synaptic weight needs to be increased, the closer the interval, the more increased, and if the pre-synaptic neuron fires after the post-synaptic neuron, the synaptic weight is decreased, the closer the interval, the more decreased.
In step 2.1, the weight is initialized to one third of the upper limit of the synaptic weight.
The step 2.2 specifically comprises the following steps:
step 2.2.1: setting a firing counter cc and a time window with the size of L for each pulse neuron;
step 2.2.2: counting the value of cc in the time window, if satisfying formula 1, it indicates that the impulse neuron is in the enhanced state, and satisfying formula 2, it indicates that the impulse neuron is in the inhibited state, and theta in formulas 1 and 2PAnd thetaDJudgment thresholds representing enhancement and suppression, respectively;
cc∈[θP,L]equation 1
cc∈(0,θD]Equation 2
Step 2.2.3: if the neuron is in the enhanced state, establishing a weight between the neuron and all the post-synaptic neurons thereof, and fixing the weight to wPSynapse of (a), wherein wPShould be large enough to ensure excitation of the postsynaptic nerve group; if the neuron is in the inhibited state, the synapses between the neuron and all its postsynaptic neurons are disconnected and then fixed to w with a weightDSynapse replacement of wherein wDThe postsynaptic neurons are difficult to excite, meanwhile, disconnected synaptic information is stored, and a user can recover subsequently;
step 2.2.4: at the end of each simulation run, all neurons in the enhancement or inhibition state are traversed, and the enhancement or inhibition state is exited if it has been maintained for a time window.
For the neuron being enhanced, exiting the enhancing state disconnects the synapse temporarily established; for the inhibited neurons, additionally exerting a long-term attenuating effect, the probability of attenuation P is calculated according to equation 3:
for each synapse, the synapse is disconnected with the probability of P, if the synapse is not disconnected, the weight of the synapse is multiplied by an attenuation factor alpha with the probability of P, if the synapse is not disconnected or attenuated, the original value when the synapse does not enter the inhibited state is recovered, wherein rho is a probability constant in formula 3, rho is smaller than 1, and Dc is a counter for the neuron to continuously enter the inhibited state.
The step 2.3 specifically comprises the following steps:
step 2.3.1: for each two adjacent layers, a coefficient matrix A, A is establishedijIndicates the importance of growing synapses between neurons i in the previous layer and neurons j in the subsequent layer, if synapses already exist between ij, AijIs always 0;
step 2.3.2: after each simulation is finished, traversing the neuron pairs between two adjacent layers, and updating the coefficient matrix according to a formula 4:
wherein λHEBBIs the coefficient of attenuation, Δ AmaxIs the maximum value at which the coefficient can be increased each time,andrespectively the m-th firing time of j neuron and the n-th firing time of i neuron in the simulation, th is a statistical time threshold,it is shown that the firing of j-pulse neurons within the time of firing of i-pulse neurons th is counted,is the updated coefficient after the k-th simulation is finished,is the coefficient before update;
step 2.3.3: after updating the coefficient matrix, setting a growth threshold value thetaHEBBIf, ifGreater than thetaHEBBThen neuron i will generate a synapse to neuron j;
step 2.3.4: traversing presynaptic neurons which will grow, and selecting traversed neurons iThe largest K postsynaptic neurons j synapse for growth.
The step 2.4 specifically comprises the following steps:
2.4.1: after the first simulation is finished, scanning all synapses, and calculating delta t according to a formula 5 for each excitation of front and back neurons:
Δt=tpre-tpost+ delay formula 5
Wherein t ispreAnd tpostRespectively representing the excitation time of a presynaptic neuron and a postsynaptic neuron, and delay is synaptic delay;
2.4.2: the synaptic weight is modified according to equation 6:
wherein wmaxIs the upper limit of the weight, λSTDPIs the learning rate, μ+And mu-The weight determination coefficients, alpha, during weight increase and decay, respectivelySTDPIs the asymmetry factor, K+And K-Respectively, the time convergence coefficients of weight decay and growth, e is a natural constant, tau-and tau+Time scaling factors for weight increase and decay, respectively, and w' and w are weights after and before update, respectively.
The features are pulse sequence features, are directly used as input for training other pulse neural networks, or are converted into numerical features by using a mathematical method, and are used for learning other non-pulse neural network models.
Compared with the prior art, the invention has the beneficial effects that:
1) the invention can efficiently extract the characteristics of the data, can directly extract the characteristics into pulse sequence characteristics for training other pulse neural networks, and fills the blank of the neighborhood research of the existing pulse neural networks.
2) The invention introduces a heuristic mechanism based on frequency induction, improves a synapse growth algorithm based on an HEBB rule and a weight change algorithm based on an STDP rule, and has the advantages of stronger network capability, more effective learning, higher operation efficiency, dynamic balance state of network scale, prevention of overgrowth and good robustness through small sample learning.
3) The method can be used for extracting the characteristics of data such as pictures and voice, and for the pictures, the pictures can be converted into the pulse sequence through the method, the pulse sequence can represent the characteristics contained in the pictures and can be input into the pulse neural network for classified learning of the pictures, and the learning effect of the pulse neural network on the pictures can be improved. For voice data, the method can convert audio frequency in voice into pulse information, and the pulse information is input into a pulse neural network for voice recognition to train, so that the recognition rate of the voice and the robustness to abnormal pronunciation can be improved.
Drawings
Fig. 1 is an overall structural view of the present invention.
Detailed Description
The embodiments of the present invention will be described in detail below with reference to the drawings and examples.
As shown in FIG. 1, a method for constructing a spiking neural network for feature extraction based on frequency induction constructs an input layer and a feature extraction layer. The sample is input into the input layer after being coded, neurons in the input layer are excited at corresponding time, and then the input is input into the feature extraction layer for learning. A heuristic mechanism based on frequency induction is introduced to the pulse neuron of the feature extraction layer, under the mechanism, if the pulse neuron is excited at a high frequency within a period of time, the neuron is considered to have a high value on current data, the performance of the neuron is enhanced, if the pulse neuron is excited at a low frequency, the neuron is considered to be noise and is inhibited, and the enhancing or inhibiting effect lasts for a long time. Meanwhile, the synapse structure learning of the impulse neurons in the feature extraction layer uses an accumulative growth algorithm based on the HEBB rule, the growth of synapses in the algorithm is used for the reference of the situation of a real brain, and the energy, substance and speed limits are considered, namely, the growth of a synapse is a long-term process, and the synapse may become less important as the learning progresses, and the synapse needs to grow in other places, so the synapse stops growing. The learning of the weight of synapses is performed using the STDP rule. The output of the last layer after learning is the feature converted into the pulse sequence. The algorithm can enable the network to effectively generate memory, efficiently extract features, achieve good robustness through small sample learning, meanwhile, a heuristic mechanism induced by frequency has a positive effect on learning of the impulse neural network, enable the network to have stronger capacity, enable the learning to be more effective, enable the operation efficiency to be higher, enable the network synapse growth speed to be kept in a reasonable range, enable the network scale to be in a dynamic balance state, and prevent overgrowth.
Referring to fig. 1, taking feature extraction for picture recognition as an example, the method includes the following steps:
step 1: and establishing an input layer, performing primary processing on input data, and converting the input data into a pulse sequence.
Step 1.1: the samples are subjected to preliminary processing, noise in the samples is weakened through a non-neural network method such as denoising and enhancing (in the embodiment, one layer of convolution is selected to be performed on the pictures), and the samples are sorted according to categories.
Step 1.2: and (3) constructing an input layer by using p pulse generators, converting the data processed in the step 1.1 into pulse excitation time, and exciting the data after the convolution is larger.
Step 1.3: a simulation is started for each sample, and the pulse generator sends a pulse at a corresponding time to convert the data into a pulse sequence.
Step 2: establishing a feature extraction layer, introducing a frequency induction heuristic mechanism on a pulse neuron, performing structure learning by using an accumulative growth algorithm based on an HEBB rule, performing weight learning by using an STDP rule, inputting data needing feature extraction after learning is finished, wherein the output of the last layer of the feature extraction layer is the feature, and starting simulation for each input data sample for a plurality of times for learning.
Step 2.1: and constructing a feature extraction network by using three layers of pulse neural networks, wherein each layer of network has q normal pulse neurons, a previous layer generates a plurality of synapses to a next layer randomly during initialization, and weight initialization is preferably one third of the upper limit of the synapse weight according to experience.
Step 2.2: setting a time window and an excitation counter for each pulse neuron, judging whether the pulse neuron is under the influence of frequency-induced inhibition or enhancement effect for each simulation, and if so, applying the influence.
Step 2.2.1: setting a firing counter cc and a time window with the size of L for each pulse neuron;
step 2.2.2: counting the value of cc in the time window, if satisfying formula 1, it indicates that the neuron is in the enhanced state, and satisfying formula 2, it indicates that the neuron is in the inhibited state, and theta in formulas 1 and 2PAnd thetaDJudgment thresholds respectively representing enhancement and suppression states;
cc∈[θP,L]equation 1
cc∈(0,θD]Equation 2
Step 2.2.3: if the neuron is in the enhanced state, establishing a weight between the neuron and all the post-synaptic neurons thereof, and fixing the weight to wPSynapse of (a), wherein wPShould be large enough to ensure excitation of the postsynaptic nerve group.
Step 2.2.4: if the neuron is in the inhibited state, the synapses between the neuron and all its postsynaptic neurons are disconnected and then fixed to w with a weightDSynapse replacement of wherein wDThe postsynaptic neurons are difficult to excite, meanwhile, disconnected synaptic information is stored, and a user can recover subsequently;
step 2.2.5: at the end of each simulation run, all neurons in the enhancement or inhibition state are traversed, and the enhancement or inhibition state is exited if it has been maintained for a time window. For the neurons being enhanced, exiting the enhanced state disconnects the synapses that were temporarily established. For inhibited neurons, which require additional long-term attenuating effects, the probability of attenuation P will be calculated according to equation 3:
for each synapse, the synapse is disconnected with the probability of P, if the synapse is not disconnected, the weight of the synapse is multiplied by an attenuation factor alpha with the probability of P, if the synapse is not disconnected or attenuated, the original value when the synapse does not enter the inhibited state is recovered, wherein rho is a probability constant in formula 3, rho is smaller than 1, and Dc is a counter for the neuron to continuously enter the inhibited state.
Step 2.3: for each simulation, the pulse neurons between two adjacent layers are subjected to structure learning through an accumulative growth algorithm based on the HEBB rule, the limitation of material and energy in synapse growth is guaranteed, namely the synapse growth is a long-term process and a neuron cannot grow in large quantity, and a competitive relationship exists, namely as learning progresses, some growing synapses become unimportant and are degraded.
Step 2.3.1: for every two adjacentLayer, establishing a coefficient matrix A, AijIndicates the importance of growing synapses between neurons i in the previous layer and neurons j in the subsequent layer, if synapses already exist between ij, AijIs always 0;
step 2.3.2: after each simulation is finished, traversing the neuron pairs between two adjacent layers, and updating the coefficient matrix according to a formula 4:
wherein λHEBBIs the coefficient of attenuation, Δ AmaxIs the maximum value at which the coefficient can be increased each time,andrespectively the m-th firing time of j neuron and the n-th firing time of i neuron in the simulation, th is a statistical time threshold,it is shown that the firing of j-pulse neurons within the time of firing of i-pulse neurons th is counted,is the updated coefficient after the k-th simulation is finished,is the coefficient before update.
Step 2.3.3: after updating the coefficient matrix, setting a growth threshold value thetaHEBBIf, ifGreater than thetaHEBBThen neuron i will generate a synapse to neuron j;
step 2.3.4: traversing presynaptic neurons that will grow, pairSelecting the traversed neuron iThe largest K postsynaptic neurons j synapse for growth.
Step 2.4: for each simulation, the synapses of the spiking neurons are weight learned by the STDP rule, i.e. if the pre-synaptic neuron fires before the post-synaptic neuron, the synaptic weight needs to be increased, the closer the interval, the more increased, and if the pre-synaptic neuron fires after the post-synaptic neuron, the synaptic weight is decreased, the closer the interval, the more decreased.
2.4.1: after the first simulation is finished, scanning all synapses, and calculating delta t according to a formula 5 for each excitation of front and back neurons:
Δt=tpre-tpost+ delay formula 5
Wherein t ispreAnd tpostRespectively, pre-synaptic and post-synaptic neuron firing times, and delay is synaptic delay.
2.4.2: the synaptic weight is modified according to equation 6:
wherein wmaxIs the upper limit of the weight, λSTDPIs the learning rate, μ+And mu-The weight determination coefficients, alpha, during weight increase and decay, respectivelySTDPIs the asymmetry factor, K+And K-Respectively, the time convergence coefficient of weight attenuation and increase, e is a natural constant, tau-And τ+Time scaling factors for weight increase and decay, respectively, and w' and w are weights after and before update, respectively.
Step 2.5: after learning is finished, data needing feature extraction is input, the output of the last layer of the feature extraction layer is the feature, the feature is a pulse sequence feature, the pulse sequence feature can be directly used as input for training other pulse neural networks, and can also be converted into a numerical feature by using a mathematical method for learning other non-pulse neural network models.
The following is a specific application embodiment of the present invention in MNIST data set identification.
The steps and effects of the invention are specifically described by taking the establishment of a spiking neural network for feature extraction, namely MNIST data set identification, as an example based on frequency induction.
In this example, 100 corresponding pictures are randomly selected for each digit as a training set for training the method of the present invention. The convolution operation is first performed on the pictures in the selected data set, in this example 4 convolution kernels are selected, and the size of the convolved pictures is 12 x 12. The input layer in the inventive spiking neural network needs to have 12 × 4 pulse generators. All pictures in the training set are sorted by category next, i.e. pictures of 0 are put together, pictures of 1 are put together, and pictures of 2, 3, 4, 5, 6, 7, 8 and 9 are also put together individually. Next, a feature extraction layer is constructed using step 2.1 of the present invention, in this case 1000 x 3 in size. And inputting the pictures in sequence, wherein the pictures are input in sequence, and the picture process of inputting 0 is repeated three times after the input, all the pictures of 1 are input in sequence after the input is finished, the process of inputting in sequence is repeated 3 times, and then the pictures of 2, 3, 4, 5, 6, 7, 8 and 9 are input in the mode. The simulation was started 500ms each time a picture was entered. The network is then trained according to step 2.2, step 2.3 and step 2.4 as described in the present invention. And after training is finished, closing the weight change capability of the feature extraction layer to fix the network, namely completing the construction of the pulse neural network for feature extraction based on frequency induction. Then, an output layer with 10 pulse neurons can be established after the feature extraction layer, the feature extraction layer and the output layer are fully connected, and a HEBB algorithm can be used for supervised learning to achieve the aim of identifying the picture. Experiments show that the pulse neural network for feature extraction based on frequency induction can effectively extract the features of the pictures under the training of small data after being constructed, and different responses are generated to different types of pictures. The method extracts the features and then inputs the features into the pulse neural network for MNIST handwritten digit recognition to learn, thereby improving the accuracy of the network and improving the recognition accuracy of the illegible handwritten data. The method can be used for image entity recognition, the image is firstly subjected to feature extraction according to the method, then the image is input into an impulse neural network for learning, and then possible existing entities in the image are output.
Meanwhile, the invention can also be used for extracting the characteristics of other data, such as voice data. In the task of converting voice into characters, the method can convert the audio frequency in the voice into pulse information, and the pulse information is input into a pulse neural network for voice recognition to train, so that the recognition rate of the voice and the robustness to nonstandard pronunciation can be improved.
Claims (9)
1. A method for constructing a pulse neural network for feature extraction based on frequency induction is characterized by comprising the following steps:
step 1: establishing an input layer, performing primary processing on input data, and converting the input data into a pulse sequence;
step 2: establishing a feature extraction layer, introducing a frequency induction heuristic mechanism on a pulse neuron, performing structure learning by using an accumulative growth algorithm based on an HEBB rule, performing weight learning by using an STDP rule, inputting data needing feature extraction after learning is finished, wherein the output of the last layer of the feature extraction layer is the feature, and starting simulation for each input data sample for a plurality of times for learning.
2. The method for constructing a spiking neural network for feature extraction based on frequency induction according to claim 1, wherein the step 1 specifically comprises the following steps:
step 1.1: constructing an input layer using p pulse generators;
step 1.2: performing primary processing on input data samples, weakening noise in the samples by a non-neural network method, and sequencing the samples according to categories;
step 1.3: converting the data processed in the step 1.2 into pulse excitation time, and exciting the data after the value is larger;
step 1.4: a simulation is started for each sample, and the pulse generator sends a pulse at a corresponding time to convert the data into a pulse sequence.
3. The method for constructing a spiking neural network for feature extraction based on frequency induction according to claim 1, wherein the step 2 specifically comprises the following steps:
step 2.1: constructing a feature extraction layer by using three layers of pulse neural networks, wherein each layer of network is provided with q normal pulse neurons, and a plurality of synapses are randomly generated from a previous layer to a next layer during initialization;
step 2.2: setting a time window and an excitation counter for each pulse neuron, judging whether the pulse neuron is under the influence of frequency-induced inhibition or enhancement effect for each simulation, and if so, applying the influence;
step 2.3: for each simulation, the pulse neurons between two adjacent layers carry out structure learning through an accumulative growth algorithm based on the HEBB rule, the limitation of material and energy is guaranteed when synapses grow, namely the synapses are a long-term process and a neuron cannot grow in large quantity, and a competitive relationship exists, namely as learning progresses, some growing synapses become unimportant and degenerate;
step 2.4: for each simulation, the synapses of the spiking neurons are weight learned by the STDP rule, i.e. if the pre-synaptic neuron fires before the post-synaptic neuron, the synaptic weight needs to be increased, the closer the interval, the more increased, and if the pre-synaptic neuron fires after the post-synaptic neuron, the synaptic weight is decreased, the closer the interval, the more decreased.
4. The method for constructing a spiking neural network for feature extraction based on frequency induction as claimed in claim 3, wherein in the step 2.1, the weight is initialized to one third of the synapse weight upper limit.
5. The method for constructing a spiking neural network for feature extraction based on frequency induction according to claim 3, wherein the step 2.2 comprises the following steps:
step 2.2.1: setting a firing counter cc and a time window with the size of L for each pulse neuron;
step 2.2.2: counting the value of cc in the time window, if satisfying formula 1, it indicates that the impulse neuron is in the enhanced state, and satisfying formula 2, it indicates that the impulse neuron is in the inhibited state, and theta in formulas 1 and 2PAnd thetaDJudgment thresholds representing enhancement and suppression, respectively;
cc∈[θP,L]equation 1
cc∈(0,θD]Equation 2
Step 2.2.3: if the neuron is in the enhanced state, establishing a weight between the neuron and all the post-synaptic neurons thereof, and fixing the weight to wPSynapse of (a), wherein wPShould be large enough to ensure excitation of the postsynaptic nerve group; if the neuron is in the inhibited state, the synapses between the neuron and all its postsynaptic neurons are disconnected and then fixed to w with a weightDSynapse replacement of wherein wDThe postsynaptic neurons are difficult to excite, meanwhile, disconnected synaptic information is stored, and a user can recover subsequently;
step 2.2.4: at the end of each simulation run, all neurons in the enhancement or inhibition state are traversed, and the enhancement or inhibition state is exited if it has been maintained for a time window.
6. The method for constructing a spiking neural network for feature extraction based on frequency induction according to claim 5, wherein for the enhanced neuron, exiting the enhanced state disconnects the temporarily established synapse; for the inhibited neurons, additionally exerting a long-term attenuating effect, the probability of attenuation P is calculated according to equation 3:
for each synapse, the synapse is disconnected with the probability of P, if the synapse is not disconnected, the weight of the synapse is multiplied by an attenuation factor alpha with the probability of P, if the synapse is not disconnected or attenuated, the original value when the synapse does not enter the inhibited state is recovered, wherein rho is a probability constant in formula 3, rho is smaller than 1, and Dc is a counter for the neuron to continuously enter the inhibited state.
7. The method for constructing a spiking neural network for feature extraction based on frequency induction according to claim 3, wherein the step 2.3 comprises the following steps:
step 2.3.1: for each two adjacent layers, a coefficient matrix A, A is establishedijIndicates the importance of growing synapses between neurons i in the previous layer and neurons j in the subsequent layer, if synapses already exist between ij, AijIs always 0;
step 2.3.2: after each simulation is finished, traversing the neuron pairs between two adjacent layers, and updating the coefficient matrix according to a formula 4:
wherein λHEBBIs the coefficient of attenuation, Δ AmaxIs the maximum value at which the coefficient can be increased each time,andrespectively the m-th firing time of j neuron and the n-th firing time of i neuron in the simulation, th is a statistical time threshold,indicating that the j pulse neuron is excited before the i pulse neuron by thThe excitation conditions in the inner tube are counted,is the updated coefficient after the k-th simulation is finished,is the coefficient before update;
step 2.3.3: after updating the coefficient matrix, setting a growth threshold value thetaHEBBIf, ifGreater than thetaHEBBThen neuron i will generate a synapse to neuron j;
8. The method for constructing a spiking neural network for feature extraction based on frequency induction according to claim 3, wherein the step 2.4 comprises the following steps:
2.4.1: after the first simulation is finished, scanning all synapses, and calculating delta t according to a formula 5 for each excitation of front and back neurons:
Δt=tpre-tpost+ delay formula 5
Wherein t ispreAnd tpostRespectively representing the excitation time of a presynaptic neuron and a postsynaptic neuron, and delay is synaptic delay;
2.4.2: the synaptic weight is modified according to equation 6:
wherein wmaxIs the upper limit of the weight, λSTDPIs the learning rate, μ+And mu-The weight determination coefficients, alpha, during weight increase and decay, respectivelySTDPIs the asymmetry factor, K+And K-Respectively, the time convergence coefficient of weight attenuation and increase, e is a natural constant, tau-And τ+Time scaling factors for weight increase and decay, respectively, and w' and w are weights after and before update, respectively.
9. The method for constructing the spiking neural network for feature extraction based on frequency induction as claimed in claim 1 or 3, wherein the features are pulse sequence features, and are directly used as input for other spiking neural network training or are converted into numerical features by using a mathematical method for learning of other non-spiking neural network models.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011246943.4A CN112232494A (en) | 2020-11-10 | 2020-11-10 | Method for constructing pulse neural network for feature extraction based on frequency induction |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011246943.4A CN112232494A (en) | 2020-11-10 | 2020-11-10 | Method for constructing pulse neural network for feature extraction based on frequency induction |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112232494A true CN112232494A (en) | 2021-01-15 |
Family
ID=74122098
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011246943.4A Pending CN112232494A (en) | 2020-11-10 | 2020-11-10 | Method for constructing pulse neural network for feature extraction based on frequency induction |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112232494A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114220089A (en) * | 2021-11-29 | 2022-03-22 | 北京理工大学 | Method for carrying out pattern recognition based on segmented progressive pulse neural network |
-
2020
- 2020-11-10 CN CN202011246943.4A patent/CN112232494A/en active Pending
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114220089A (en) * | 2021-11-29 | 2022-03-22 | 北京理工大学 | Method for carrying out pattern recognition based on segmented progressive pulse neural network |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111858989B (en) | Pulse convolution neural network image classification method based on attention mechanism | |
Liang et al. | Stacked denoising autoencoder and dropout together to prevent overfitting in deep neural network | |
Hinton et al. | Improving neural networks by preventing co-adaptation of feature detectors | |
CN112633497A (en) | Convolutional pulse neural network training method based on reweighted membrane voltage | |
CN108304912B (en) | System and method for realizing pulse neural network supervised learning by using inhibition signal | |
CN113657561B (en) | Semi-supervised night image classification method based on multi-task decoupling learning | |
CN111639754A (en) | Neural network construction, training and recognition method and system, and storage medium | |
CN113094357A (en) | Traffic missing data completion method based on space-time attention mechanism | |
CN112906828A (en) | Image classification method based on time domain coding and impulse neural network | |
CN109635938B (en) | Weight quantization method for autonomous learning impulse neural network | |
CN114266351A (en) | Pulse neural network training method and system based on unsupervised learning time coding | |
Chandra et al. | Encoding subcomponents in cooperative co-evolutionary recurrent neural networks | |
CN116796207A (en) | Self-organizing mapping clustering method based on impulse neural network | |
CN112232494A (en) | Method for constructing pulse neural network for feature extraction based on frequency induction | |
CN113553918B (en) | Machine ticket issuing character recognition method based on pulse active learning | |
Dong et al. | Training generative adversarial networks with binary neurons by end-to-end backpropagation | |
CN109101984B (en) | Image identification method and device based on convolutional neural network | |
CN115984942A (en) | Facial expression recognition method based on impulse neural network, storage medium and equipment | |
CN113628615B (en) | Voice recognition method and device, electronic equipment and storage medium | |
CN112288078B (en) | Self-learning, small sample learning and migration learning method and system based on impulse neural network | |
CN115456173A (en) | Generalized artificial neural network unsupervised local learning method, system and application | |
CN113948067B (en) | Voice countercheck sample repairing method with hearing high fidelity characteristic | |
CN115546556A (en) | Training method of pulse neural network for image classification | |
CN114118378A (en) | Hardware-friendly STDP learning method and system based on threshold self-adaptive neurons | |
CN111340329B (en) | Actor evaluation method and device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210115 |