CN108345938A - A kind of neural network processor and its method including bits switch device - Google Patents

A kind of neural network processor and its method including bits switch device Download PDF

Info

Publication number
CN108345938A
CN108345938A CN201810170612.3A CN201810170612A CN108345938A CN 108345938 A CN108345938 A CN 108345938A CN 201810170612 A CN201810170612 A CN 201810170612A CN 108345938 A CN108345938 A CN 108345938A
Authority
CN
China
Prior art keywords
bits
bits switch
neural network
data
initial data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810170612.3A
Other languages
Chinese (zh)
Inventor
韩银和
闵丰
许浩博
王颖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Computing Technology of CAS
Original Assignee
Institute of Computing Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Computing Technology of CAS filed Critical Institute of Computing Technology of CAS
Priority to CN201810170612.3A priority Critical patent/CN108345938A/en
Priority to PCT/CN2018/082179 priority patent/WO2019165679A1/en
Publication of CN108345938A publication Critical patent/CN108345938A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Neurology (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of neural network processor, and the method for carrying out bits switch to the data of neural network using the neural network processor.The neural network processor includes bits switch device, which includes:Input interface, control unit, Date Conversion Unit and output interface;Wherein, described control unit is used to generate the control signal for the Date Conversion Unit;The input interface is for receiving initial data;The Date Conversion Unit is used to carry out bits switch to the initial data according to the control signal, and the initial data is converted to the bits switch result expressed using less number of bits;The output interface is used to the bits switch result exporting the bits switch device.Number of bits used by expression data can be reduced through the invention, reduced hardware cost and energy consumption needed for calculating, improved calculating speed.

Description

A kind of neural network processor and its method including bits switch device
Technical field
The present invention relates to artificial intelligence fields, more particularly to the improvement to neural network processor.
Background technology
The depth learning technology of artificial intelligence was obtaining development at full speed in recent years, was solving high-level abstractions cognitive question On, such as image recognition, speech recognition, natural language understanding, weather forecasting, gene expression, commending contents and intelligent robot Equal fields are widely applied and have been proved there is outstanding performance.This make exploitation for artificial intelligence technology and It is modified to the research hotspot for academia and industrial quarters.
Deep neural network is one of the sensor model that artificial intelligence field has highest development level, such network passes through Model is established to simulate the neural connection structure of human brain, data characteristics is described using the layering of multiple conversion stages, Task, which is handled, for large-scale datas such as image, video and audios brings breakthrough.The model of deep neural network belongs to A kind of operational model uses netted interconnection structure it includes a large amount of node between these nodes, is called depth nerve The neuron of network.Weighted value of the bonding strength representation signal between the two nodes between two nodes, i.e. weight, with It is corresponding with the memory in the neural network on biological significance.
For the application specific processor of neural computing, i.e. neural network processor has also obtained corresponding development.In reality In the neural computing processing procedure on border, need repeatedly to carry out the operations such as convolution, activation, pond to a large amount of data, this It needs to consume the extremely large amount of calculating time, has seriously affected the usage experience of user.This makes the meter for how reducing neural network Evaluation time becomes a kind of improvement strategy for neural network processor.
Invention content
Therefore, it is an object of the invention to overcome the defect of the above-mentioned prior art, a kind of neural network processor is provided, it should Neural network processor includes bits switch device, which includes:
Input interface, control unit, Date Conversion Unit and output interface;
Wherein,
Described control unit is used to generate the control signal for the Date Conversion Unit;
The input interface is for receiving initial data;
The Date Conversion Unit is used to carry out bits switch to the initial data according to the control signal, by institute It states initial data and is converted to the bits switch result expressed using less number of bits;
The output interface is used to the bits switch result exporting the bits switch device.
Preferably, according to the neural network processor, wherein described control unit be used for according to the parameter of setting or The parameter of input determines the rule for executing bits switch, to generate the control signal;
Wherein, the parameter includes the bit with the number of bits of the initial data and the bits switch result The relevant information of number.
Preferably, according to the neural network processor, wherein the Date Conversion Unit is used to be believed according to the control Number, it determines the reserved bit in the initial data and blocks position, and according to the reserved bit of the initial data and described The highest order of initial data blocked in position determines the bits switch result.
Preferably, according to the neural network processor, wherein the Date Conversion Unit is used to be believed according to the control Number, it determines the reserved bit in the initial data and blocks position, and using the reserved bit in the initial data as described in Bits switch result.
Preferably, according to the neural network processor, wherein the Date Conversion Unit is used to be believed according to the control Number bits switch is carried out to the initial data, is converted into initial data and to be expressed using the number of bits of script half Bits switch result.
A kind of neural network processor using described in above-mentioned any one carries out bits switch to the data of neural network Method, including:
1) described control unit generates the control signal for Date Conversion Unit;
2) input interface receives the original number for needing to execute bits switch outside the bits switch device According to;
3) Date Conversion Unit carries out bits switch according to the control signal to the initial data, will be described Initial data is converted to the bits switch result expressed using less number of bits;
4) the bits switch result is exported the bits switch device by the output interface.
Preferably, according to the method, wherein step 1) includes:
1-1) described control unit determines the rule for executing bits switch according to the parameter of setting or the parameter of input;
1-2) described control unit generates control signal corresponding with the rule;
Wherein, the parameter includes the bit with the number of bits of the initial data and the bits switch result The relevant information of number.
Preferably, according to the method, wherein step 3) includes:
The Date Conversion Unit is according to the control signal, reserved bit based on the initial data and described original The highest order of data blocked in position determines the bits switch result.
Preferably, according to the method, wherein step 3) includes:
The Date Conversion Unit is according to the control signal, using the reserved bit in the initial data as the bit Transformation result.
Preferably, according to the method, the caching to Neural Network Data is being completed and is not yet completing convolution algorithm When, the Neural Network Data of caching is inputted into the bits switch device to execute step 1) -4), or be completed to data Convolution algorithm and when not yet completing activation operation, the result of convolution algorithm is inputted into the bits switch device to execute Step 1) -4).
A kind of computer readable storage medium, wherein being stored with computer program, the computer program is when executed For realizing the method described in above-mentioned any one.
Compared with the prior art, the advantages of the present invention are as follows:
The present invention provides a kind of bits switch device for neural network processor, can be used in neural network Number of bits used by expression data is adjusted in various calculating process.Bit used by data is expressed by reduction Digit can reduce the hardware cost needed for calculating, improve calculating speed, reduce neural network processor to data space Needs and reduce execute neural computing energy consumption.
Description of the drawings
Embodiments of the present invention is further illustrated referring to the drawings, wherein:
Fig. 1 shows the module map of bits switch device according to an embodiment of the invention;
Fig. 2 is the annexation figure of each unit in bits switch device according to an embodiment of the invention;
Fig. 3 be it is according to an embodiment of the invention using bits switch device as illustrated in FIG. 1 to neural network Data carry out the method flow of bits switch;
Fig. 4 a are according to one embodiment of present invention in the Date Conversion Unit of bits switch device in " four houses Five enter pattern " under execute bits switch hardware structure diagram;
Fig. 4 b are according to one embodiment of present invention in the Date Conversion Unit of bits switch device for " directly The hardware structure diagram of bits switch is executed under truncated mode ".
Specific implementation mode
It elaborates with reference to the accompanying drawings and detailed description to the present invention.
As described in the text, when designing neural network processor, it is desirable to be able to reduce the calculating time of neural network.It is right This, inventor thinks can be by suitably reducing the number of bits of the data for participating in neural computing, for example, by using more Few bit come represent originally need more bit to indicate data, reduce operand to reduce the calculating of neural network Time.This is because, inventor has found in the research to the prior art, the intermediate result of the algorithm of neural network for calculating There are relatively high fault-tolerances, and the data for needing more bit to indicate originally are represented even with less bit Way can change the precision for the data for participating in calculating to influence the accuracy of obtained intermediate result, however this can't Large effect is caused to the result of neural network final output.
In the present invention, the mode of the bit of data used in this Reduction Computation is referred to as " cutting out to data Cut operation ".Also, the process being adjusted to the binary bits digit needed for expression numerical value is referred to as " bits switch ". For example, be directed to metric numerical value 0.5, use the result that Q7 fixed-point datas are indicated be 01000000 (here Q7 using 8 compare The bit of the leftmost side first in spy indicates fractional parts as sign bit, using remaining 7 bit, thus indicates -1 The decimal that precision between to 1 is 7), when carrying out bits switch, script can be used to the results modification that Q7 is indicated to use Q3 is indicated, and obtaining result 0100, (analogously with Q7, Q3 equally uses the bit of the leftmost side first as sign bit, no Same is that it uses 3 bits to indicate fractional parts, can indicate the decimal that the precision between -1 to 1 is 3).
Based on above-mentioned analysis, the present invention proposes a kind of bits switch device for neural network processor.Pass through institute Stating bits switch device can be according to setting or based on parameter input by user determination execution bits switch rule, with right Data execute bits switch.Conversion in this way, neural network processor can be handled relatively small amounts of data, And the energy consumption for thus promoting processing speed, reducing neural network processor.It has been recognised by the inventors that in logic combination circuit, data The speed of operation and the number of bits of numerical expression are inversely proportional;The energy consumption of data operation is directly proportional to the bit of numerical expression; Therefore after carrying out bits switch to data, the effect for accelerating to calculate with reduce power consumption can reach.
Fig. 1 shows bits switch device 101 according to an embodiment of the invention, including:As input interface Input bus unit 102, Date Conversion Unit 103, the output bus unit 104 as output interface, control unit 105.
Wherein, input bus unit 102 is carried for obtaining the Neural Network Data for needing to carry out bits switch It is supplied to Date Conversion Unit 103.In some embodiments, input bus unit 102 concurrently can be received and/or be transmitted and is multiple Data to be converted.
Date Conversion Unit 103 is used for according to such as setting or execution of determination based on parameter input by user The rule of bits switch executes bits switch to the Neural Network Data from input bus unit 102.
Output bus unit 104, for will be handled via Date Conversion Unit 103 result of obtained bits switch from It is exported in bits switch device 101, with the device being provided in Processing with Neural Network for executing subsequent processing.
Control unit 105, the rule for determining bits switch select corresponding bits switch pattern and turn to control data Change the operation that unit 103 executes bits switch.The parameter or defeated by user that described control unit 105 can be arranged by analysis The parameter entered executes the rule of bits switch to determine, to be selected from pre-set various translative mode.Here join Number may include the number of bits of data to be converted and transformed data bit digit, can also be that data to be converted are adopted The binary expression way used desired by data after binary expression way and conversion, such as Q7, Q3 etc.. For example, according to parameter input by user, determines that the Neural Network Data that Q7 will be used to indicate is converted to and indicated using Q3.It is reducing Used by expression when bit, the mode of " rounding up " may be used, such as 0110 is converted to by 01011000, it can also 0101 is converted to by the way of " directly blocking ", such as by 01011000.Here " rounding up " or " directly blocking " Equal conversion regimes can both have been inputted by user, can also be set to be fixed.
In some embodiments, input bus unit 102 and/or output bus unit 104 can concurrently receive and/or Transmit multiple data to be converted.
Fig. 2 is the annexation figure of each unit in bits switch device according to an embodiment of the invention.Wherein, The number of bits of input bus unit is 128bit, and the number of bits of output bus is 64bit.Control unit is filled from bits switch External reception parameter input by user is set, is used to be used for Date Conversion Unit according to determining bits switch rule to generate Mode switching signal so that Date Conversion Unit, which can be informed under the present situation, to be needed to use which kind of mode to execute bit Conversion.Also, control unit can also be generated to be started to receive data or pause reception data for controlling input bus unit Input control signal, and for control output bus unit start export or suspend output bits switch result output Control signal.
Below by by one embodiment introduction using bits switch device as illustrated in FIG. 1 to Neural Network Data into The procedure of row bits switch.With reference to figure 3, the method includes:
Conversion requirements parameter or input by user parameter of the step 1. based on setting, by bits switch device 101 Control unit 105 determine used in bits switch rule.The conversion requirements parameter of the setting described is inputted by user Parameter include relevant with the data bit digit after the number of bits of Neural Network Data and conversion that need to convert Information.The conversion requirements parameter of the setting, the parameter input by user can also be included in when carrying out bits switch Block rule, such as the rules such as " rounding up " or " directly blocking ".
Based on above-mentioned rule, can be selected from pre-set bits switch pattern by control unit 105.According to One embodiment of the present of invention, the bits switch pattern includes " round up pattern " and " direct truncated mode ", for institute Stating the processing mode of two kinds of different modes will be introduced in a subsequent step.
What the input bus unit 102 in step 2. bits switch device 101 was obtained it needs to execute bits switch Neural Network Data be provided to Date Conversion Unit 103.
Here input bus unit 102 may include multiple interfaces that can receive data parallel, concurrently to receive The Neural Network Data for needing to execute bits switch outside bits switch device 101.Similar, input bus unit 102 can also include multiple interfaces for capableing of parallel output data, to concurrently provide data to Date Conversion Unit 103, to carry out the processing of bits switch.
Rule of step 3. Date Conversion Unit 103 according to bits switch determined by control unit 105, to needing to execute The Neural Network Data of bits switch executes bits switch.
In this step, the control signal from control unit 105 can be received by Date Conversion Unit 103 with according to institute It states rule and executes bits switch.
Inventor has found, when reducing the number of bits for calculating used data, if the number of bits after reduction is big In the half of the number of bits equal to original data, then can make neural network processor hardware cost, processing speed and Reach compromise between accuracy rate.Therefore, in the present invention preferably, it would be desirable to execute the ratio of the Neural Network Data of bits switch Special abbreviation is the half of script, and bits switch is executed for example, by using fixed hardware configuration, and the data of 32bit are turned 16bit is turned to, convert the data of 16bit to 8bit, convert the data of 8bit to 4bit, converts the data of 4bit to 2bit and convert the data of 2bit to 1bit etc..
It, can be according to the rule, it would be desirable to execute the neural network of bits switch during executing bits switch Each bit of data is divided into reserved bit and blocks position, and wherein reserved bit is each bit of the Neural Network Data In higher one or more bits, block remaining bit in each bit that position is the Neural Network Data. For example, for the data 10101111 of 8bit, according to the mode for the half that its number of bits is reduced to script, then its Reserved bit is 1010, and it is 1111 to block position.
Fig. 4 a are shown according to one embodiment of present invention in Date Conversion Unit 103 for " round up mould The hardware device structure of bits switch is executed under formula ", wherein the Neural Network Data for needing to execute bits switch of 16 8bit It is concurrently input into Date Conversion Unit 103, is removed in the reserved bit of the 4bit in the Neural Network Data of each 8bit Remove bit (such as a other than sign bit1、a2、a3) and corresponding highest order (such as a blocked in position4) served as Two inputs of adder, the sign bit in the output of the adder and the Neural Network Data are used as needle jointly The result after bits switch is executed to the Neural Network Data of the 8bit.
It is illustrated with reference to figure 4a, under " round up pattern ", it is assumed that the nerve being input in converting unit 103 Network data is 10101111 (radix-minus-one complements), indicates its expression metric -0.6328125, and it is 1111 to block position, will block position Highest order 1 be added with 3 bits 010 except divided-by symbol position in reserved bit, based on the symbol in the Neural Network Data It is 1011 (radix-minus-one complements) that the result of number position and adder, which obtains the result after bits switch, indicates metric -0.625.
Fig. 4 b are shown according to one embodiment of present invention in Date Conversion Unit 103 for " directly blocking mould The hardware device structure of bits switch is executed under formula ", wherein the Neural Network Data for needing to execute bits switch of 16 8bit It is concurrently input into Date Conversion Unit 103, the reserved bit of the 4bit in the Neural Network Data of each 8bit (such as a0、a1、a2、a3) be used directly as for the result after the Neural Network Data of 8bit execution bits switch.
It is illustrated with reference to figure 4b, under " direct truncated mode ", it is assumed that the nerve being input in converting unit 103 Network data is 10101111 (radix-minus-one complements), then it is 1010 to execute the result after bits switch.
Step 4. will be handled the knot of obtained bits switch by output bus unit 104 via Date Conversion Unit 103 Fruit exports from bits switch device 101, with the device being provided in Processing with Neural Network for executing subsequent processing.
The bits switch device provided by the above embodiment of the present invention can as a part for neural network processor, It is used in the various calculating process for neural network.
For example, can be when the caching to Neural Network Data be completed and not yet completes convolution algorithm, using bit Conversion equipment carries out bits switch to the Neural Network Data of caching.Reason for doing so is that the heterogeneous networks of neural network There may be different requirements to number of bits used by data for layer, in order to be adapted to required calculating speed and phase The energy consumption of prestige can carry out bits switch to the Neural Network Data of caching by bits switch device, and will pass through bits switch The result obtained is provided to for executing the unit of convolution algorithm to execute convolution algorithm.
In another example can be turned using bit when the convolution algorithm to data is completed and not yet completes activation operation Changing device carries out bits switch to the result of convolution algorithm.Reason for doing so is that the accumulation operations of convolution algorithm unit are past Toward the number of bits for the result that can increase obtained convolution algorithm, in order to be adapted to requirement of the subsequent operation to number of bits (such as some activation arithmetic elements realized using hardware mode, used number of bits is often fixed ), it needs to carry out bits switch to the result of convolution algorithm.
Based on above-described embodiment, the present invention provides a kind of bits switch device for neural network processor, can quilt For being adjusted to number of bits used by expression data in the various calculating process of neural network.It is expressed by reducing Number of bits used by data can reduce the hardware cost needed for calculating, improve calculating speed, reduce Processing with Neural Network Device executes the needs of data space and reducing the energy consumption of neural computing.
It should be noted that each step introduced in above-described embodiment is all not necessary, those skilled in the art Can carry out according to actual needs it is appropriate accept or reject, replace, modification etc..
It should be noted last that the above examples are only used to illustrate the technical scheme of the present invention and are not limiting.On although Text is described the invention in detail with reference to embodiment, it will be understood by those of ordinary skill in the art that, to the skill of the present invention Art scheme is modified or replaced equivalently, and without departure from the spirit and scope of technical solution of the present invention, should all be covered at this In the right of invention.

Claims (11)

1. a kind of neural network processor, which includes bits switch device, the bits switch device packet It includes:
Input interface, control unit, Date Conversion Unit and output interface;
Wherein,
Described control unit is used to generate the control signal for the Date Conversion Unit;
The input interface is for receiving initial data;
The Date Conversion Unit is used to carry out bits switch to the initial data according to the control signal, by the original Beginning data are converted to the bits switch result expressed using less number of bits;
The output interface is used to the bits switch result exporting the bits switch device.
2. neural network processor according to claim 1, wherein described control unit be used for according to the parameter of setting or The parameter of person's input determines the rule for executing bits switch, to generate the control signal;
Wherein, the parameter includes the number of bits phase with the number of bits of the initial data and the bits switch result The information of pass.
3. neural network processor according to claim 2, wherein the Date Conversion Unit is used for according to the control Signal determines the reserved bit in the initial data and blocks position, and according to the reserved bit of the initial data and institute The highest order blocked in position for stating initial data determines the bits switch result.
4. neural network processor according to claim 2, wherein the Date Conversion Unit is used for according to the control Signal determines the reserved bit in the initial data and blocks position, and using the reserved bit in the initial data as institute State bits switch result.
5. neural network processor according to claim 1, wherein the Date Conversion Unit is used for according to the control Signal carries out bits switch to the initial data, is converted into initial data and is expressed using the number of bits of script half Bits switch result.
6. a kind of neural network processor using as described in any one of claim 1-5 carries out the data of neural network The method of bits switch, including:
1) described control unit generates the control signal for Date Conversion Unit;
2) input interface receives the initial data for needing to execute bits switch outside the bits switch device;
3) Date Conversion Unit carries out bits switch according to the control signal to the initial data, will be described original Data are converted to the bits switch result expressed using less number of bits;
4) the bits switch result is exported the bits switch device by the output interface.
7. according to the method described in claim 6, wherein step 1) includes:
1-1) described control unit determines the rule for executing bits switch according to the parameter of setting or the parameter of input;
1-2) described control unit generates control signal corresponding with the rule;
Wherein, the parameter includes the number of bits phase with the number of bits of the initial data and the bits switch result The information of pass.
8. according to the method described in claim 7, wherein step 3) includes:
The Date Conversion Unit is according to the control signal, the reserved bit based on the initial data and the initial data The highest order blocked in position determine the bits switch result.
9. according to the method described in claim 7, wherein step 3) includes:
The Date Conversion Unit is according to the control signal, using the reserved bit in the initial data as the bits switch As a result.
10. according to the method described in any one of claim 6-9, being completed to the caching of Neural Network Data and When not yet completing convolution algorithm, the Neural Network Data of caching is inputted into the bits switch device to execute step 1) -4), or The result of convolution algorithm is inputted the ratio by person when the convolution algorithm to data is completed and not yet completes activation operation Special conversion equipment is to execute step 1) -4).
11. a kind of computer readable storage medium, wherein being stored with computer program, the computer program is used when executed In method of the realization as described in any one of claim 6-10.
CN201810170612.3A 2018-03-01 2018-03-01 A kind of neural network processor and its method including bits switch device Pending CN108345938A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201810170612.3A CN108345938A (en) 2018-03-01 2018-03-01 A kind of neural network processor and its method including bits switch device
PCT/CN2018/082179 WO2019165679A1 (en) 2018-03-01 2018-04-08 Neural network processor comprising bit conversion device and method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810170612.3A CN108345938A (en) 2018-03-01 2018-03-01 A kind of neural network processor and its method including bits switch device

Publications (1)

Publication Number Publication Date
CN108345938A true CN108345938A (en) 2018-07-31

Family

ID=62959552

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810170612.3A Pending CN108345938A (en) 2018-03-01 2018-03-01 A kind of neural network processor and its method including bits switch device

Country Status (2)

Country Link
CN (1) CN108345938A (en)
WO (1) WO2019165679A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021180201A1 (en) * 2020-03-13 2021-09-16 华为技术有限公司 Data processing method and apparatus for terminal network model, terminal and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106796668A (en) * 2016-03-16 2017-05-31 香港应用科技研究院有限公司 For the method and system that bit-depth in artificial neural network is reduced
CN107203808A (en) * 2017-05-08 2017-09-26 中国科学院计算技术研究所 A kind of two-value Convole Unit and corresponding two-value convolutional neural networks processor
CN107292458A (en) * 2017-08-07 2017-10-24 北京中星微电子有限公司 A kind of Forecasting Methodology and prediction meanss applied to neural network chip
CN107340993A (en) * 2016-04-28 2017-11-10 北京中科寒武纪科技有限公司 A kind of apparatus and method for the neural network computing for supporting less digit floating number

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109934331B (en) * 2016-04-29 2020-06-19 中科寒武纪科技股份有限公司 Apparatus and method for performing artificial neural network forward operations
CN106447034B (en) * 2016-10-27 2019-07-30 中国科学院计算技术研究所 A kind of neural network processor based on data compression, design method, chip
CN107423816B (en) * 2017-03-24 2021-10-12 中国科学院计算技术研究所 Multi-calculation-precision neural network processing method and system
CN107145939B (en) * 2017-06-21 2020-11-24 北京图森智途科技有限公司 Computer vision processing method and device of low-computing-capacity processing equipment

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106796668A (en) * 2016-03-16 2017-05-31 香港应用科技研究院有限公司 For the method and system that bit-depth in artificial neural network is reduced
CN107340993A (en) * 2016-04-28 2017-11-10 北京中科寒武纪科技有限公司 A kind of apparatus and method for the neural network computing for supporting less digit floating number
CN107203808A (en) * 2017-05-08 2017-09-26 中国科学院计算技术研究所 A kind of two-value Convole Unit and corresponding two-value convolutional neural networks processor
CN107292458A (en) * 2017-08-07 2017-10-24 北京中星微电子有限公司 A kind of Forecasting Methodology and prediction meanss applied to neural network chip

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
郑南宁: "《数字信号处理简明教程》", 30 September 2015 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021180201A1 (en) * 2020-03-13 2021-09-16 华为技术有限公司 Data processing method and apparatus for terminal network model, terminal and storage medium

Also Published As

Publication number Publication date
WO2019165679A1 (en) 2019-09-06

Similar Documents

Publication Publication Date Title
CN106447034B (en) A kind of neural network processor based on data compression, design method, chip
CN113361680B (en) Neural network architecture searching method, device, equipment and medium
CN106610867B (en) On-chip network task scheduling method and device
CN108345934A (en) A kind of activation device and method for neural network processor
CN110334802A (en) A kind of construction method of neural network model, device, equipment and storage medium
CN110163354A (en) A kind of computing device and method
CN109389212A (en) A kind of restructural activation quantization pond system towards low-bit width convolutional neural networks
CN108304925A (en) A kind of pond computing device and method
CN112101517A (en) FPGA implementation method based on piecewise linear pulse neuron network
CN114358319B (en) Machine learning framework-based classification method and related device
CN105389772A (en) Data processing method and device based on figure processor
CN111831358B (en) Weight precision configuration method, device, equipment and storage medium
CN116187548A (en) Photovoltaic power generation power prediction method and device, storage medium and electronic device
Geng et al. CQNN: a CGRA-based QNN framework
CN108345938A (en) A kind of neural network processor and its method including bits switch device
Shu et al. High energy efficiency FPGA-based accelerator for convolutional neural networks using weight combination
Lee et al. Energy-efficient control of mobile processors based on long short-term memory
CN109002885A (en) A kind of convolutional neural networks pond unit and pond calculation method
CN115292390B (en) Load information generation method and device, electronic equipment and computer readable medium
CN111831356A (en) Weight precision configuration method, device, equipment and storage medium
CN111027669A (en) Method and device for realizing deep neural network on field programmable gate array
CN111104767B (en) Variable-precision random gradient descending structure and design method for FPGA
WO2020051918A1 (en) Neuronal circuit, chip, system and method therefor, and storage medium
CN117634577B (en) Vector processor, neural network accelerator, chip and electronic equipment
CN105824285B (en) The Programming method of programmable logic control system for microcontroller

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination