[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN105681628A - Convolution network arithmetic unit, reconfigurable convolution neural network processor and image de-noising method of reconfigurable convolution neural network processor - Google Patents

Convolution network arithmetic unit, reconfigurable convolution neural network processor and image de-noising method of reconfigurable convolution neural network processor Download PDF

Info

Publication number
CN105681628A
CN105681628A CN201610003960.2A CN201610003960A CN105681628A CN 105681628 A CN105681628 A CN 105681628A CN 201610003960 A CN201610003960 A CN 201610003960A CN 105681628 A CN105681628 A CN 105681628A
Authority
CN
China
Prior art keywords
input
convolution
restructural
output
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610003960.2A
Other languages
Chinese (zh)
Other versions
CN105681628B (en
Inventor
张斌
饶磊
李艳婷
杨宏伟
赵季中
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN201610003960.2A priority Critical patent/CN105681628B/en
Publication of CN105681628A publication Critical patent/CN105681628A/en
Application granted granted Critical
Publication of CN105681628B publication Critical patent/CN105681628B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/21Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/21Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
    • H04N5/213Circuitry for suppressing or minimising impulsive noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/64Circuits for processing colour signals
    • H04N9/73Colour balance circuits, e.g. white balance circuits or colour temperature control

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Processing (AREA)
  • Complex Calculations (AREA)

Abstract

The invention discloses a convolution network arithmetic unit, a reconfigurable convolution neural network processor and an image de-noising method of the reconfigurable convolution neural network processor. The reconfigurable convolution neural network processor comprises a bus interface, a preprocessing unit, a reconfigurable hardware controller, an SRAM, an SRAM control module, an input caching module, an output caching module, a memory, a data memory controller and the convolution network arithmetic unit. The processor is featured by few resources and rapid speed and can be applicable to common convolution neural network architecture. According to the unit, the processor and the method provided by the invention, convolution neural networks can be realized; the processing speed is rapid; transplanting is liable to be carried out; the resource consumption is little; an image or a video polluted by raindrops and dusts can be recovered; and raindrop and dust removing operations can be taken as preprocessing operations for providing help in follow-up image identification or classification.

Description

A kind of convolutional network arithmetic element and restructural convolutional neural networks processor and realize the method that image denoising processes
Technical field
The present invention relates to image processing field, particularly to a kind of convolutional network arithmetic element and restructural convolutional neural networks processor with realize the method that image denoising processes.
Background technology
The removal of image raindrop and dust is significant for image procossing application, particularly video monitoring and navigation system. It can be used for recovering by raindrop, the image of contamination by dust or video, is alternatively arranged as pre-treatment operation and offers help for follow-up image recognition or classification.
The current method removing picture noise mostly utilizes the modes such as gaussian filtering, medium filtering, bilateral filtering to complete, and these method treatment effects are bad, usually can not meet specific image and process the demand of application. It is thus desirable to the method for a better effects if removes picture noise, the method for convolutional neural networks becomes a good selection.
Current degree of depth learning network is mostly run on GPU, but GPU is expensive, and power consumption is high, is not appropriate for large-scale extensive use. And the speed of service is slow on CPU, run large-scale degree of depth learning network efficiency low, it is impossible to meet performance requirement.
Can be seen that current technology is for application convolutional neural networks, the problem being primarily present has: the problems such as processor area is big, and cost is high, and power consumption is big, poor performance. Therefore this is accomplished by that a low-power consumption, area be little, the restructural convolutional neural networks processor of high treating effect.
Summary of the invention
It is an object of the invention to provide a kind of convolutional network arithmetic element and restructural convolutional neural networks processor and realize the method that image denoising processes, hardware resource consumption is low, area is little, can recover by raindrop, the image of contamination by dust or video.
To achieve these goals, the present invention adopts the following technical scheme that
A kind of convolutional network arithmetic element, including 2 restructural separation convolution module, nonlinear activation function unit and multiply-accumulator unit;
First restructural separation convolution module is output as the input of nonlinear activation function unit, and nonlinear activation function unit is output as the input of multiply-accumulator unit, and multiply-accumulator unit is output as the input of second restructural separation convolution module;
Picture signal is input to first restructural with configuration network parameter signal and separates convolution module; First restructural separation convolution module completes 16 × 16 convolution algorithms; Nonlinear activation function unit completes the computing of activation primitive in convolutional neural networks; Multiply-accumulator unit completes the computing of the articulamentum in convolutional neural networks; Second restructural separation convolution module simultaneously completes 48 × 8 convolution algorithms;
Described multiply-accumulator unit includes some multiply-accumulator and some depositors; Wherein multiply-accumulator for calculate the output valve of last layer convolutional network and weight parameter product and; The result of last layer convolutional network is input in multiply-accumulator by depositor.
Further, described restructural separation convolution module includes 16 4 × the 4 one-dimensional convolution module of restructural and the first Parasites Fauna; Picture signal and convolutional network parameter are input to the one-dimensional convolution module of restructural by the first Parasites Fauna; Restructural separation convolution module can complete 1 16 × 16 convolution or simultaneously complete 48 × 8 convolution algorithms; 4 × 4 one-dimensional convolution module of restructural include 4 first selectors, 4 the one 2 input multipliers, the one 4 input summer, 4 the 2nd 2 input multiplier and the 2nd 4 input summers; The outfan of 4 first selectors connects the input of 4 the 1st corresponding input multipliers, the weight that another one input is neutral net of 4 the 1st input multipliers; The outfan of 4 the 1st input multipliers connects the input of the one 4 input summer; The output that input is the one 4 input summer of 4 the 2nd 2 input multipliers and the weight of neutral net; The input of the 2nd 4 input summer is the output of 4 the 2nd 2 input multipliers.
Further, described nonlinear activation function unit include QD generator and arithmetical unit group; Wherein the input of QD generator is the output of restructural separation convolution, arithmetical unit group the output that input is QD generator; QD generator is for producing the parameter needed for activation primitive; Arithmetical unit, group was for calculating the end value that activation primitive is final;
Described QD generator comprises first divider; Input signal is input to the first divider, the first divider output business Q and remainder D; Described arithmetical unit, group comprised shift register, 2 first adders and the second divider; Shift register output and be the input of 2 first adders; 2 first adders are output as the input of the second divider; Shift register, first adder and the second divider are sequentially connected with;
A kind of restructural convolutional neural networks processor, including the convolutional network arithmetic element according to any one of EBI, pretreatment unit, reconfigurable hardware controller, SRAM, SRAM control unit, input buffer module, output buffer module, memorizer, data storage controller and some claims 1 to 3; EBI connects pretreatment unit, data storage controller, reconfigurable hardware controller and input-buffer, output buffer memory; Memorizer connects data storage controller; Input-buffer connects reconfigurable hardware controller and SRAM control unit; Convolutional network arithmetic element connects input buffer module, output buffer module;
The input of described pretreatment unit is image or video signal; Complete the pre-treatment such as white balance, noise filtering operation;
Described input buffer module, output buffer module are respectively used to input and the output of buffer memory convolutional network arithmetic element;
Convolutional network computing module is configured by described reconfigurable hardware controller, controls its calculating process; What send in calculating process or when terminating that interrupt requests completes with external system is mutual;
Described SRAM control unit is for controlling the transmission of convolutional network weight parameter.
Further, including 512 convolutional network arithmetic elements, it is achieved the image denoising based on convolutional neural networks processes.
Further, described a kind of restructural convolutional neural networks processor realizes 3 layers of convolutional neural networks, for removing raindrop and the dust of attachment in image or video; Described convolutional neural networks ground floor is made up of 512 16 × 16 convolution, and the second layer is neutral net articulamentum, and third layer is made up of 512 8 × 8 convolution.
A kind of restructural convolutional neural networks processor realizes the method that image denoising processes, including:
In the process that image denoising processes, the random convolution number that reduces, the consumption of minimizing hardware resource, improves processing speed;
Or, 16 × 16 convolution algorithm unit and 8 × 8 convolution algorithm unit are respectively classified into the convolution mask of 16 and 44 × 4 by the process that image denoising processes, the convolution to each 4 × 4 adopts one-dimensional convolution.
Relative to prior art, the method have the advantages that convolutional network arithmetic element utilizes Reconfiguration Technologies, 16 × 16 convolution can be completed or simultaneously complete 48 × 8 convolution algorithms, improve hardware performance and motility. The present invention utilizes the method that the degree of depth learns, it is achieved that can removing the denoising of image raindrop and dust, treatment effect meets demand. The present invention is under the premise not affecting treatment effect, and the random template number reducing convolutional network, but also the method utilizing the one-dimensional convolution of piecemeal, hardware resource consumption greatly reduces, and processing speed is greatly improved. This processor can realize the convolutional neural networks of 3 layers, it is possible to provides feature for follow-up higher level image recognition, classification. Expensive relative to GPU, power consumption is high, and area is big. The CPU speed of service is slow, runs large-scale degree of depth learning network efficiency low. The present invention adopts Reconfiguration Technologies and the method for above-mentioned minimizing template number and the one-dimensional convolution of piecemeal, it is achieved restructural convolutional neural networks processor low in resources consumption, be prone to hardware and realize, it is possible to recover by raindrop, the image of contamination by dust or video.
Accompanying drawing explanation
Fig. 1 is the structural representation of convolutional network arithmetic element;
Fig. 2 is the structural representation of non-linear activation primitive unit;
Fig. 3 is the structural representation of the one 4 × 4th one-dimensional convolution module of restructural;
Fig. 4 is the structural representation of restructural separation convolution module;
Fig. 5 is the structural representation of restructural convolutional neural networks processor;
Detailed description of the invention
Illustrate with detailed description of the invention, the present invention to be elaborated below in conjunction with accompanying drawing.
With reference to Fig. 1, the convolutional network arithmetic element used in restructural convolutional neural networks processor in the present invention includes 2 restructural separation convolution module, nonlinear activation function unit and multiply-accumulator unit; First restructural separation convolution module is output as the input of nonlinear activation function unit, and nonlinear activation function unit is output as the input of multiply-accumulator unit, and multiply-accumulator unit is output as the input of second restructural separation convolution module;
Picture signal is input to first restructural with configuration network parameter signal and separates convolution module; First restructural separation convolution module completes 16 × 16 convolution algorithms; Nonlinear activation function unit completes the computing of activation primitive in convolutional neural networks; Multiply-accumulator unit completes the computing of the articulamentum in convolutional neural networks; Second restructural separation convolution module simultaneously completes 48 × 8 convolution algorithms;
Refer to shown in Fig. 2, nonlinear activation function unit include QD generator and arithmetical unit group; Wherein the input of QD generator is the output of restructural separation convolution, arithmetical unit group the output that input is QD generator; QD generator is for producing the parameter needed for activation primitive; Arithmetical unit, group was for calculating the result that activation primitive is final.
The activation primitive of neutral net of the present invention is hyperbolic tangent function
tanh ( x ) = e x - e - x e x + e - x
By defining territory extension and Taylor series expansion, obtain
tanh ( Q l n 2 + D ) = 1 + D - 2 - 2 Q ( 1 - D ) 1 + D + 2 - 2 Q ( 1 - D ) Wherein | D | < ln2
QD generator comprises first divider, and input signal is input to the first divider, and the first divider, divided by definite value 0.69, exports business Q and remainder D; Described arithmetical unit, group comprised shift register, 2 first adders and the second divider; Shift register output and be the input of 2 first adders; 2 first adders are output as the input of the second divider; Shift register, first adder and the second divider are sequentially connected with;
Referring to shown in Fig. 3,4 × 4 one-dimensional convolution module of restructural include 4 first selector MUX, 4 the 1st input multipliers, the one 4 input summer, 4 the 2nd 2 input multipliers, the 2nd 4 input summer. Two inputs of first selector are picture signal and previous stage result; The outfan of 4 first selectors connects an input of 4 the 1st corresponding input multipliers, the weight that another one input is neutral net of 4 the 1st input multipliers; The outfan of 4 the 1st input multipliers connects the input of the one 4 input summer; The output that input is the one 4 input summer of 4 the 2nd 2 input multipliers and the weight of neutral net; The input of the 2nd 4 input summer is the output of 4 the 2nd 2 input multipliers.
Referring to shown in Fig. 4, restructural separation convolution module includes the first Parasites Fauna, 16 4 × 4 one-dimensional convolution module of restructural, 44 input first adders and 14 input second adder. Utilize Reconfiguration Technologies, restructural separation convolution module can complete 16 × 16 or simultaneously complete 48 × 8 convolution algorithms. Picture signal and configuration signal are input to the first Parasites Fauna. The input of the one 4 × 4th convolution 1 is 1-4 row picture signal, and the input of the one 4 × 4th convolution 5 is 5-8 row picture signal.
When convolution mask is 16 × 16, the output that input is the one 4 × 4th convolution 2 of the one 4 × 4th convolution 3, the output that input is the one 4 × 4th convolution 6 of the one 4 × 4th convolution 7, the output that input is the one 4 × 4th convolution 10 of the one 4 × 4th convolution 11, the output that input is the one 4 × 4th convolution 14 of the one 4 × 4th convolution 15. The input of the one 4 × 4th convolution 9 is 9-12 row picture signal, and the input of the one 4 × 4th convolution 13 is 13-16 row picture signal. Restructural separation convolution module is output as the result of second adder.
When convolution module is 8 × 8, the input of the one 4 × 4th convolution 3 is 1-4 row picture signal, the input of the one 4 × 4th convolution 7 is 1-4 row picture signal, and the input of the one 4 × 4th convolution 11 is 1-4 row picture signal, and the input of the one 4 × 4th convolution 15 is 1-4 row picture signal. The input of the one 4 × 4th convolution 9 is 1-4 row picture signal, and the input of the one 4 × 4th convolution 13 is 5-8 row picture signal. Restructural separation convolution module is output as the result of 4 first adders. One restructural separation convolution module can simultaneously complete 48 × 8 convolution algorithms.
Referring to shown in Fig. 5, one restructural convolutional neural networks processor of the present invention includes EBI, pretreatment unit, reconfigurable hardware controller, SRAM, SRAM control unit, input-buffer, output buffer memory, memorizer, data storage controller and some convolutional network arithmetic elements; EBI connects pretreatment unit, data storage controller, reconfigurable hardware controller and input-buffer, output buffer memory; Memorizer connects data storage controller; Input-buffer connects reconfigurable hardware controller and SRAM control unit; Convolutional network arithmetic element connects input buffer module, output buffer module.
The input of pretreatment unit is image or video signal; Complete the pre-treatment such as white balance, noise filtering operation; Input buffer module, output buffer memory are respectively used to input and the output of buffer memory convolutional network arithmetic element. Convolutional network arithmetic element is configured by reconfigurable hardware controller, controls its calculating process; What send in calculating process or when terminating that interrupt requests completes with external system is mutual; SRAM control unit is for controlling the transmission of convolutional network weight parameter.
In one convolutional neural networks realizing removal image raindrop and dust, including 512 convolutional network arithmetic elements. In order to reduce resource, improve processing speed, the present invention have employed following two method in implementing process: (1) reduces the method for convolution number at random: reduce the number of convolutional network arithmetic element under the premise not affecting treatment effect, reduce the consumption of hardware resource, improve processing speed; (2) method of the one-dimensional convolution of piecemeal: the convolution mask by 16 × 16 and 8 × 8 is respectively classified into the convolution mask of 16 and 44 × 4, the convolution to each 4 × 4 adopts the mode of one-dimensional convolution.
With reference to Fig. 5, restructural 16 × 16 convolution algorithm unit include 16 4 × 4 one-dimensional convolution module of restructural (1,2,3 ..., 16), row storing module and depositor; The input of row storing module is image or video signal, the output that input is row storing module of Parasites Fauna, the output that input is Parasites Fauna of 4 × 4 one-dimensional convolution module of restructural; Row storing module is used for preserving image; The view data of serial input deposited by depositor for preserving row, and view data is input to 4 × 4 one-dimensional convolution module of restructural.
Restructural 8 × 8 convolution algorithm unit includes 4 the one-dimensional convolution module of 4 × 4 restructurals (1,2,3,4), row storing module and depositor; The input of row storing module is the output of multiply-accumulator, the output that input is row storing module of Parasites Fauna, the output that input is Parasites Fauna of 4 × 4 one-dimensional convolution module of restructural.

Claims (7)

1. a convolutional network arithmetic element, it is characterised in that: include 2 restructural separation convolution module, nonlinear activation function unit and multiply-accumulator unit;
First restructural separation convolution module is output as the input of nonlinear activation function unit, and nonlinear activation function unit is output as the input of multiply-accumulator unit, and multiply-accumulator unit is output as the input of second restructural separation convolution module;
Picture signal is input to first restructural with configuration network parameter signal and separates convolution module; First restructural separation convolution module completes 16 × 16 convolution algorithms; Nonlinear activation function unit completes the computing of activation primitive in convolutional neural networks; Multiply-accumulator unit completes the computing of the articulamentum in convolutional neural networks; Second restructural separation convolution module simultaneously completes 48 × 8 convolution algorithms;
Described multiply-accumulator unit includes some multiply-accumulator and some depositors; Wherein multiply-accumulator for calculate the output valve of last layer convolutional network and weight parameter product and; The result of last layer convolutional network is input in multiply-accumulator by depositor.
2. a kind of convolutional network arithmetic element according to claim 1, it is characterised in that: described restructural separation convolution module includes 16 4 × the 4 one-dimensional convolution module of restructural and the first Parasites Fauna; First Parasites Fauna is input to the one-dimensional convolution module of restructural for picture signal or previous stage being exported with convolutional network parameter; Restructural separation convolution module has been used for 1 16 × 16 convolution or has simultaneously completed 48 × 8 convolution algorithms;
4 × 4 one-dimensional convolution module of restructural include 4 first selectors, 4 the one 2 input multipliers, the one 4 input summer, 4 the 2nd 2 input multiplier and the 2nd 4 input summers; The outfan of 4 first selectors connects the input of 4 the 1st corresponding input multipliers, the weight that another one input is neutral net of 4 the 1st input multipliers; The outfan of 4 the 1st input multipliers connects the input of the one 4 input summer; The output that input is the one 4 input summer of 4 the 2nd 2 input multipliers and the weight of neutral net; The input of the 2nd 4 input summer is the output of 4 the 2nd 2 input multipliers.
3. a kind of convolutional network arithmetic element according to claim 1, it is characterised in that: described nonlinear activation function unit include QD generator and arithmetical unit group; Wherein the input of QD generator is the output of restructural separation convolution, arithmetical unit group the output that input is QD generator; QD generator is for producing the parameter needed for activation primitive; Arithmetical unit, group was for calculating the end value that activation primitive is final;
Described QD generator comprises first divider; Input signal is input to the first divider, the first divider output business Q and remainder D; Described arithmetical unit, group comprised shift register, 2 first adders and the second divider; Shift register output and be the input of 2 first adders; 2 first adders are output as the input of the second divider; Shift register, first adder and the second divider are sequentially connected with.
4. a restructural convolutional neural networks processor, it is characterised in that: include the convolutional network arithmetic element according to any one of EBI, pretreatment unit, reconfigurable hardware controller, SRAM, SRAM control unit, input buffer module, output buffer module, memorizer, data storage controller and some claims 1 to 3; EBI connects pretreatment unit, data storage controller, reconfigurable hardware controller and input-buffer, output buffer memory; Memorizer connects data storage controller; Input-buffer connects reconfigurable hardware controller and SRAM control unit; Convolutional network arithmetic element connects input buffer module, output buffer module;
The input of described pretreatment unit is image or video signal; Complete the pre-treatment such as white balance, noise filtering operation;
Described input buffer module, output buffer module are respectively used to input and the output of buffer memory convolutional network arithmetic element;
Convolutional network computing module is configured by described reconfigurable hardware controller, controls its calculating process; What send in calculating process or when terminating that interrupt requests completes with external system is mutual;
Described SRAM control unit is for controlling the transmission of convolutional network weight parameter.
5. a kind of restructural convolutional neural networks processor according to claim 4, it is characterised in that: include 512 convolutional network arithmetic elements, it is achieved the image denoising based on convolutional neural networks processes.
6. a kind of restructural convolutional neural networks processor according to claim 4, it is characterised in that: described a kind of restructural convolutional neural networks processor realizes 3 layers of convolutional neural networks, for removing raindrop and the dust of attachment in image or video; Described convolutional neural networks ground floor is made up of 512 16 × 16 convolution, and the second layer is neutral net articulamentum, and third layer is made up of 512 8 × 8 convolution.
7. a kind of restructural convolutional neural networks processor according to any one of claim 4 to 6 realizes the method that image denoising processes, it is characterised in that: including:
In the process that image denoising processes, the random convolution number that reduces, the consumption of minimizing hardware resource, improves processing speed;
Or, 16 × 16 convolution algorithm unit and 8 × 8 convolution algorithm unit are respectively classified into the convolution mask of 16 and 44 × 4 by the process that image denoising processes, the convolution to each 4 × 4 adopts one-dimensional convolution.
CN201610003960.2A 2016-01-05 2016-01-05 A kind of convolutional network arithmetic element and restructural convolutional neural networks processor and the method for realizing image denoising processing Active CN105681628B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610003960.2A CN105681628B (en) 2016-01-05 2016-01-05 A kind of convolutional network arithmetic element and restructural convolutional neural networks processor and the method for realizing image denoising processing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610003960.2A CN105681628B (en) 2016-01-05 2016-01-05 A kind of convolutional network arithmetic element and restructural convolutional neural networks processor and the method for realizing image denoising processing

Publications (2)

Publication Number Publication Date
CN105681628A true CN105681628A (en) 2016-06-15
CN105681628B CN105681628B (en) 2018-12-07

Family

ID=56298840

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610003960.2A Active CN105681628B (en) 2016-01-05 2016-01-05 A kind of convolutional network arithmetic element and restructural convolutional neural networks processor and the method for realizing image denoising processing

Country Status (1)

Country Link
CN (1) CN105681628B (en)

Cited By (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106203617A (en) * 2016-06-27 2016-12-07 哈尔滨工业大学深圳研究生院 A kind of acceleration processing unit based on convolutional neural networks and array structure
CN106203621A (en) * 2016-07-11 2016-12-07 姚颂 The processor calculated for convolutional neural networks
CN106250103A (en) * 2016-08-04 2016-12-21 东南大学 A kind of convolutional neural networks cyclic convolution calculates the system of data reusing
CN106331433A (en) * 2016-08-25 2017-01-11 上海交通大学 Video denoising method based on deep recursive neural network
CN106843809A (en) * 2017-01-25 2017-06-13 北京大学 A kind of convolution algorithm method based on NOR FLASH arrays
CN106909970A (en) * 2017-01-12 2017-06-30 南京大学 A kind of two-value weight convolutional neural networks hardware accelerator computing module based on approximate calculation
CN106940815A (en) * 2017-02-13 2017-07-11 西安交通大学 A kind of programmable convolutional neural networks Crypto Coprocessor IP Core
CN107169563A (en) * 2017-05-08 2017-09-15 中国科学院计算技术研究所 Processing system and method applied to two-value weight convolutional network
CN107229967A (en) * 2016-08-22 2017-10-03 北京深鉴智能科技有限公司 A kind of hardware accelerator and method that rarefaction GRU neutral nets are realized based on FPGA
CN107248144A (en) * 2017-04-27 2017-10-13 东南大学 A kind of image de-noising method based on compression-type convolutional neural networks
CN107256424A (en) * 2017-05-08 2017-10-17 中国科学院计算技术研究所 Three value weight convolutional network processing systems and method
CN107480782A (en) * 2017-08-14 2017-12-15 电子科技大学 Learn neural network processor on a kind of piece
CN107609641A (en) * 2017-08-30 2018-01-19 清华大学 Sparse neural network framework and its implementation
WO2018018470A1 (en) * 2016-07-27 2018-02-01 华为技术有限公司 Method, apparatus and device for eliminating image noise and convolutional neural network
CN107844826A (en) * 2017-10-30 2018-03-27 中国科学院计算技术研究所 Neural-network processing unit and the processing system comprising the processing unit
CN107862374A (en) * 2017-10-30 2018-03-30 中国科学院计算技术研究所 Processing with Neural Network system and processing method based on streamline
CN107871163A (en) * 2016-09-28 2018-04-03 爱思开海力士有限公司 Operation device and method for convolutional neural networks
CN107909148A (en) * 2017-12-12 2018-04-13 北京地平线信息技术有限公司 For performing the device of the convolution algorithm in convolutional neural networks
CN108038815A (en) * 2017-12-20 2018-05-15 深圳云天励飞技术有限公司 Integrated circuit
CN108073550A (en) * 2016-11-14 2018-05-25 耐能股份有限公司 Buffer unit and convolution algorithm apparatus and method
CN108073983A (en) * 2016-11-10 2018-05-25 谷歌有限责任公司 Core is performed within hardware to cross over
CN108073977A (en) * 2016-11-14 2018-05-25 耐能股份有限公司 Convolution algorithm device and convolution algorithm method
CN108133264A (en) * 2016-12-01 2018-06-08 上海兆芯集成电路有限公司 Perform the neural network unit of efficient 3 dimension convolution
CN108133262A (en) * 2016-12-01 2018-06-08 上海兆芯集成电路有限公司 With for perform it is efficient 3 dimension convolution memory layouts neural network unit
CN108154194A (en) * 2018-01-18 2018-06-12 北京工业大学 A kind of method with the convolutional network extraction high dimensional feature based on tensor
CN108241484A (en) * 2016-12-26 2018-07-03 上海寒武纪信息科技有限公司 Neural computing device and method based on high bandwidth memory
CN108256628A (en) * 2018-01-15 2018-07-06 合肥工业大学 Convolutional neural networks hardware accelerator and its working method based on multicast network-on-chip
CN108268932A (en) * 2016-12-31 2018-07-10 上海兆芯集成电路有限公司 Neural network unit
CN108304923A (en) * 2017-12-06 2018-07-20 腾讯科技(深圳)有限公司 Convolution algorithm processing method and Related product
CN108629406A (en) * 2017-03-24 2018-10-09 展讯通信(上海)有限公司 Arithmetic unit for convolutional neural networks
CN108764336A (en) * 2018-05-28 2018-11-06 北京陌上花科技有限公司 For the deep learning method and device of image recognition, client, server
CN108804973A (en) * 2017-04-27 2018-11-13 上海鲲云信息科技有限公司 The hardware structure and its execution method of algorithm of target detection based on deep learning
CN108804974A (en) * 2017-04-27 2018-11-13 上海鲲云信息科技有限公司 Calculation of natural resources, configuration method and the system of the hardware structure of algorithm of target detection
CN108846420A (en) * 2018-05-28 2018-11-20 北京陌上花科技有限公司 Network structure and client
CN109117945A (en) * 2017-06-22 2019-01-01 上海寒武纪信息科技有限公司 Processor and its processing method, chip, chip-packaging structure and electronic device
CN109343826A (en) * 2018-08-14 2019-02-15 西安交通大学 A kind of reconfigurable processor arithmetic element towards deep learning
CN109409512A (en) * 2018-09-27 2019-03-01 西安交通大学 A kind of neural computing unit, computing array and its construction method of flexibly configurable
CN109472356A (en) * 2018-12-29 2019-03-15 南京宁麒智能计算芯片研究院有限公司 A kind of accelerator and method of restructural neural network algorithm
CN109711533A (en) * 2018-12-20 2019-05-03 西安电子科技大学 Convolutional neural networks module based on FPGA
CN109784483A (en) * 2019-01-24 2019-05-21 电子科技大学 Computation accelerator in binaryzation convolutional neural networks memory based on FD-SOI technique
CN109844738A (en) * 2016-10-19 2019-06-04 索尼半导体解决方案公司 Arithmetic processing circuit and identifying system
CN110070178A (en) * 2019-04-25 2019-07-30 北京交通大学 A kind of convolutional neural networks computing device and method
CN110088773A (en) * 2016-10-06 2019-08-02 谷歌有限责任公司 Image procossing neural network with separable convolutional layer
CN110147873A (en) * 2018-05-18 2019-08-20 北京中科寒武纪科技有限公司 The processor and training method of convolutional neural networks
CN110363292A (en) * 2018-04-11 2019-10-22 深圳市九天睿芯科技有限公司 A kind of mixed signal binary system CNN processor
WO2020042770A1 (en) * 2018-08-31 2020-03-05 北京嘉楠捷思信息技术有限公司 Image recognition method and apparatus
CN111008697A (en) * 2019-11-06 2020-04-14 北京中科胜芯科技有限公司 Convolutional neural network accelerator implementation architecture
US10621486B2 (en) 2016-08-12 2020-04-14 Beijing Deephi Intelligent Technology Co., Ltd. Method for optimizing an artificial neural network (ANN)
US10643124B2 (en) 2016-08-12 2020-05-05 Beijing Deephi Intelligent Technology Co., Ltd. Method and device for quantizing complex artificial neural network
CN111291874A (en) * 2018-12-06 2020-06-16 黄朝宗 Convolution neural network processor and data processing method thereof
US10698657B2 (en) 2016-08-12 2020-06-30 Xilinx, Inc. Hardware accelerator for compressed RNN on FPGA
US10762426B2 (en) 2016-08-12 2020-09-01 Beijing Deephi Intelligent Technology Co., Ltd. Multi-iteration compression for deep neural networks
CN111626399A (en) * 2019-02-27 2020-09-04 中国科学院半导体研究所 Convolutional neural network calculation device and data calculation method
US10802992B2 (en) 2016-08-12 2020-10-13 Xilinx Technology Beijing Limited Combining CPU and special accelerator for implementing an artificial neural network
US10810484B2 (en) 2016-08-12 2020-10-20 Xilinx, Inc. Hardware accelerator for compressed GRU on FPGA
US10832123B2 (en) 2016-08-12 2020-11-10 Xilinx Technology Beijing Limited Compression of deep neural networks with proper use of mask
US10936941B2 (en) 2016-08-12 2021-03-02 Xilinx, Inc. Efficient data access control device for neural network hardware acceleration system
US10984308B2 (en) 2016-08-12 2021-04-20 Xilinx Technology Beijing Limited Compression method for deep neural networks with load balance
TWI734598B (en) * 2020-08-26 2021-07-21 元智大學 Removing method of rain streak in image
US11087203B2 (en) 2016-11-10 2021-08-10 Beijing Baidu Netcom Science and Technology, Co., Ltd Method and apparatus for processing data sequence
CN113591025A (en) * 2021-08-03 2021-11-02 深圳思谋信息科技有限公司 Feature map processing method and device, convolutional neural network accelerator and medium
CN115841416A (en) * 2022-11-29 2023-03-24 白盒子(上海)微电子科技有限公司 Reconfigurable intelligent image processor architecture for automatic driving field
US11948279B2 (en) 2020-11-23 2024-04-02 Samsung Electronics Co., Ltd. Method and device for joint denoising and demosaicing using neural network

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4644488A (en) * 1983-10-12 1987-02-17 California Institute Of Technology Pipeline active filter utilizing a booth type multiplier
US4937774A (en) * 1988-11-03 1990-06-26 Harris Corporation East image processing accelerator for real time image processing applications
US20110029471A1 (en) * 2009-07-30 2011-02-03 Nec Laboratories America, Inc. Dynamically configurable, multi-ported co-processor for convolutional neural networks

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4644488A (en) * 1983-10-12 1987-02-17 California Institute Of Technology Pipeline active filter utilizing a booth type multiplier
US4937774A (en) * 1988-11-03 1990-06-26 Harris Corporation East image processing accelerator for real time image processing applications
US20110029471A1 (en) * 2009-07-30 2011-02-03 Nec Laboratories America, Inc. Dynamically configurable, multi-ported co-processor for convolutional neural networks

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
HIROKI NAKAHARA ETAL: "A Deep Convolutional Neural Network Based on Nested Residue Number System", 《FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL)》 *
MURUGAN SANKARADAS ETAL: "A Massively Parallel Coprocessor for Conv-olutional Neural Networks", 《2009 20TH IEEE INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS》 *
SHEFA A. DAWWD: "A reconfigurable interconnected filter for face recognition based on convolution neural network", 《DESIGN AND TEST WORKSHOP (IDT)》 *
方睿等: "卷积神经网络的FPGA并行加速方案设计", 《计算机工程与应用》 *

Cited By (98)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106203617B (en) * 2016-06-27 2018-08-21 哈尔滨工业大学深圳研究生院 A kind of acceleration processing unit and array structure based on convolutional neural networks
CN106203617A (en) * 2016-06-27 2016-12-07 哈尔滨工业大学深圳研究生院 A kind of acceleration processing unit based on convolutional neural networks and array structure
CN106203621A (en) * 2016-07-11 2016-12-07 姚颂 The processor calculated for convolutional neural networks
CN106203621B (en) * 2016-07-11 2019-04-30 北京深鉴智能科技有限公司 The processor calculated for convolutional neural networks
WO2018018470A1 (en) * 2016-07-27 2018-02-01 华为技术有限公司 Method, apparatus and device for eliminating image noise and convolutional neural network
CN108604369A (en) * 2016-07-27 2018-09-28 华为技术有限公司 A kind of method, apparatus, equipment and the convolutional neural networks of removal picture noise
CN108604369B (en) * 2016-07-27 2020-10-27 华为技术有限公司 Method, device and equipment for removing image noise and convolutional neural network
CN106250103A (en) * 2016-08-04 2016-12-21 东南大学 A kind of convolutional neural networks cyclic convolution calculates the system of data reusing
US10810484B2 (en) 2016-08-12 2020-10-20 Xilinx, Inc. Hardware accelerator for compressed GRU on FPGA
US10832123B2 (en) 2016-08-12 2020-11-10 Xilinx Technology Beijing Limited Compression of deep neural networks with proper use of mask
US10621486B2 (en) 2016-08-12 2020-04-14 Beijing Deephi Intelligent Technology Co., Ltd. Method for optimizing an artificial neural network (ANN)
US10643124B2 (en) 2016-08-12 2020-05-05 Beijing Deephi Intelligent Technology Co., Ltd. Method and device for quantizing complex artificial neural network
US10698657B2 (en) 2016-08-12 2020-06-30 Xilinx, Inc. Hardware accelerator for compressed RNN on FPGA
US10762426B2 (en) 2016-08-12 2020-09-01 Beijing Deephi Intelligent Technology Co., Ltd. Multi-iteration compression for deep neural networks
US10984308B2 (en) 2016-08-12 2021-04-20 Xilinx Technology Beijing Limited Compression method for deep neural networks with load balance
US10936941B2 (en) 2016-08-12 2021-03-02 Xilinx, Inc. Efficient data access control device for neural network hardware acceleration system
US10802992B2 (en) 2016-08-12 2020-10-13 Xilinx Technology Beijing Limited Combining CPU and special accelerator for implementing an artificial neural network
CN107229967A (en) * 2016-08-22 2017-10-03 北京深鉴智能科技有限公司 A kind of hardware accelerator and method that rarefaction GRU neutral nets are realized based on FPGA
CN107229967B (en) * 2016-08-22 2021-06-15 赛灵思公司 Hardware accelerator and method for realizing sparse GRU neural network based on FPGA
CN106331433B (en) * 2016-08-25 2020-04-24 上海交通大学 Video denoising method based on deep recurrent neural network
CN106331433A (en) * 2016-08-25 2017-01-11 上海交通大学 Video denoising method based on deep recursive neural network
CN107871163A (en) * 2016-09-28 2018-04-03 爱思开海力士有限公司 Operation device and method for convolutional neural networks
US11449745B2 (en) 2016-09-28 2022-09-20 SK Hynix Inc. Operation apparatus and method for convolutional neural network
CN107871163B (en) * 2016-09-28 2022-05-24 爱思开海力士有限公司 Operation device and method for convolutional neural network
CN110088773A (en) * 2016-10-06 2019-08-02 谷歌有限责任公司 Image procossing neural network with separable convolutional layer
US11593614B2 (en) 2016-10-06 2023-02-28 Google Llc Image processing neural networks with separable convolutional layers
US11922288B2 (en) 2016-10-06 2024-03-05 Google Llc Image processing neural networks with separable convolutional layers
CN109844738A (en) * 2016-10-19 2019-06-04 索尼半导体解决方案公司 Arithmetic processing circuit and identifying system
US11087203B2 (en) 2016-11-10 2021-08-10 Beijing Baidu Netcom Science and Technology, Co., Ltd Method and apparatus for processing data sequence
US11816532B2 (en) 2016-11-10 2023-11-14 Google Llc Performing kernel striding in hardware
CN108073983A (en) * 2016-11-10 2018-05-25 谷歌有限责任公司 Core is performed within hardware to cross over
CN108073983B (en) * 2016-11-10 2022-04-26 谷歌有限责任公司 Performing core crossing in hardware
CN108073550A (en) * 2016-11-14 2018-05-25 耐能股份有限公司 Buffer unit and convolution algorithm apparatus and method
CN108073977A (en) * 2016-11-14 2018-05-25 耐能股份有限公司 Convolution algorithm device and convolution algorithm method
CN108133264B (en) * 2016-12-01 2022-02-25 上海兆芯集成电路有限公司 Neural network unit performing efficient 3-dimensional convolution
CN108133262B (en) * 2016-12-01 2022-02-25 上海兆芯集成电路有限公司 Neural network cell with memory layout for performing efficient 3-dimensional convolution
CN108133264A (en) * 2016-12-01 2018-06-08 上海兆芯集成电路有限公司 Perform the neural network unit of efficient 3 dimension convolution
CN108133262A (en) * 2016-12-01 2018-06-08 上海兆芯集成电路有限公司 With for perform it is efficient 3 dimension convolution memory layouts neural network unit
CN108241484A (en) * 2016-12-26 2018-07-03 上海寒武纪信息科技有限公司 Neural computing device and method based on high bandwidth memory
CN108268932B (en) * 2016-12-31 2021-04-16 上海兆芯集成电路有限公司 Neural network unit
CN108268932A (en) * 2016-12-31 2018-07-10 上海兆芯集成电路有限公司 Neural network unit
CN106909970A (en) * 2017-01-12 2017-06-30 南京大学 A kind of two-value weight convolutional neural networks hardware accelerator computing module based on approximate calculation
CN106909970B (en) * 2017-01-12 2020-04-21 南京风兴科技有限公司 Approximate calculation-based binary weight convolution neural network hardware accelerator calculation device
CN106843809B (en) * 2017-01-25 2019-04-30 北京大学 A kind of convolution algorithm method based on NOR FLASH array
CN106843809A (en) * 2017-01-25 2017-06-13 北京大学 A kind of convolution algorithm method based on NOR FLASH arrays
CN106940815A (en) * 2017-02-13 2017-07-11 西安交通大学 A kind of programmable convolutional neural networks Crypto Coprocessor IP Core
CN106940815B (en) * 2017-02-13 2020-07-28 西安交通大学 Programmable convolutional neural network coprocessor IP core
CN108629406B (en) * 2017-03-24 2020-12-18 展讯通信(上海)有限公司 Arithmetic device for convolutional neural network
CN108629406A (en) * 2017-03-24 2018-10-09 展讯通信(上海)有限公司 Arithmetic unit for convolutional neural networks
CN107248144A (en) * 2017-04-27 2017-10-13 东南大学 A kind of image de-noising method based on compression-type convolutional neural networks
CN108804973A (en) * 2017-04-27 2018-11-13 上海鲲云信息科技有限公司 The hardware structure and its execution method of algorithm of target detection based on deep learning
CN108804974A (en) * 2017-04-27 2018-11-13 上海鲲云信息科技有限公司 Calculation of natural resources, configuration method and the system of the hardware structure of algorithm of target detection
CN107169563A (en) * 2017-05-08 2017-09-15 中国科学院计算技术研究所 Processing system and method applied to two-value weight convolutional network
US11551068B2 (en) 2017-05-08 2023-01-10 Institute Of Computing Technology, Chinese Academy Of Sciences Processing system and method for binary weight convolutional neural network
CN107256424A (en) * 2017-05-08 2017-10-17 中国科学院计算技术研究所 Three value weight convolutional network processing systems and method
CN107256424B (en) * 2017-05-08 2020-03-31 中国科学院计算技术研究所 Three-value weight convolution network processing system and method
CN109117945A (en) * 2017-06-22 2019-01-01 上海寒武纪信息科技有限公司 Processor and its processing method, chip, chip-packaging structure and electronic device
CN109117945B (en) * 2017-06-22 2021-01-26 上海寒武纪信息科技有限公司 Processor and processing method thereof, chip packaging structure and electronic device
CN107480782A (en) * 2017-08-14 2017-12-15 电子科技大学 Learn neural network processor on a kind of piece
CN107609641B (en) * 2017-08-30 2020-07-03 清华大学 Sparse neural network architecture and implementation method thereof
CN107609641A (en) * 2017-08-30 2018-01-19 清华大学 Sparse neural network framework and its implementation
CN107844826A (en) * 2017-10-30 2018-03-27 中国科学院计算技术研究所 Neural-network processing unit and the processing system comprising the processing unit
CN107862374A (en) * 2017-10-30 2018-03-30 中国科学院计算技术研究所 Processing with Neural Network system and processing method based on streamline
CN107844826B (en) * 2017-10-30 2020-07-31 中国科学院计算技术研究所 Neural network processing unit and processing system comprising same
CN108304923A (en) * 2017-12-06 2018-07-20 腾讯科技(深圳)有限公司 Convolution algorithm processing method and Related product
US11449576B2 (en) 2017-12-06 2022-09-20 Tencent Technology (Shenzhen) Company Limited Convolution operation processing method and related product
CN107909148A (en) * 2017-12-12 2018-04-13 北京地平线信息技术有限公司 For performing the device of the convolution algorithm in convolutional neural networks
US11429836B2 (en) 2017-12-12 2022-08-30 Nanjing Horizon Robotics Technology Co., Ltd. Apparatus for performing convolution operations in a convolutional neural network
CN108038815A (en) * 2017-12-20 2018-05-15 深圳云天励飞技术有限公司 Integrated circuit
CN108256628A (en) * 2018-01-15 2018-07-06 合肥工业大学 Convolutional neural networks hardware accelerator and its working method based on multicast network-on-chip
CN108154194A (en) * 2018-01-18 2018-06-12 北京工业大学 A kind of method with the convolutional network extraction high dimensional feature based on tensor
CN110363292A (en) * 2018-04-11 2019-10-22 深圳市九天睿芯科技有限公司 A kind of mixed signal binary system CNN processor
CN110147873B (en) * 2018-05-18 2020-02-18 中科寒武纪科技股份有限公司 Convolutional neural network processor and training method
CN110147873A (en) * 2018-05-18 2019-08-20 北京中科寒武纪科技有限公司 The processor and training method of convolutional neural networks
CN108846420A (en) * 2018-05-28 2018-11-20 北京陌上花科技有限公司 Network structure and client
CN108764336A (en) * 2018-05-28 2018-11-06 北京陌上花科技有限公司 For the deep learning method and device of image recognition, client, server
CN109343826B (en) * 2018-08-14 2021-07-13 西安交通大学 Reconfigurable processor operation unit for deep learning
CN109343826A (en) * 2018-08-14 2019-02-15 西安交通大学 A kind of reconfigurable processor arithmetic element towards deep learning
WO2020042770A1 (en) * 2018-08-31 2020-03-05 北京嘉楠捷思信息技术有限公司 Image recognition method and apparatus
CN109409512B (en) * 2018-09-27 2021-02-19 西安交通大学 Flexibly configurable neural network computing unit, computing array and construction method thereof
CN109409512A (en) * 2018-09-27 2019-03-01 西安交通大学 A kind of neural computing unit, computing array and its construction method of flexibly configurable
CN111291874A (en) * 2018-12-06 2020-06-16 黄朝宗 Convolution neural network processor and data processing method thereof
CN111291874B (en) * 2018-12-06 2023-12-01 神盾股份有限公司 Convolutional neural network processor and data processing method thereof
CN109711533B (en) * 2018-12-20 2023-04-28 西安电子科技大学 Convolutional neural network acceleration system based on FPGA
CN109711533A (en) * 2018-12-20 2019-05-03 西安电子科技大学 Convolutional neural networks module based on FPGA
CN109472356A (en) * 2018-12-29 2019-03-15 南京宁麒智能计算芯片研究院有限公司 A kind of accelerator and method of restructural neural network algorithm
CN109784483B (en) * 2019-01-24 2022-09-09 电子科技大学 FD-SOI (field-programmable gate array-silicon on insulator) process-based binary convolution neural network in-memory computing accelerator
CN109784483A (en) * 2019-01-24 2019-05-21 电子科技大学 Computation accelerator in binaryzation convolutional neural networks memory based on FD-SOI technique
CN111626399B (en) * 2019-02-27 2023-07-28 中国科学院半导体研究所 Convolutional neural network computing device and data computing method
CN111626399A (en) * 2019-02-27 2020-09-04 中国科学院半导体研究所 Convolutional neural network calculation device and data calculation method
CN110070178A (en) * 2019-04-25 2019-07-30 北京交通大学 A kind of convolutional neural networks computing device and method
CN110070178B (en) * 2019-04-25 2021-05-14 北京交通大学 Convolutional neural network computing device and method
CN111008697A (en) * 2019-11-06 2020-04-14 北京中科胜芯科技有限公司 Convolutional neural network accelerator implementation architecture
TWI734598B (en) * 2020-08-26 2021-07-21 元智大學 Removing method of rain streak in image
US11948279B2 (en) 2020-11-23 2024-04-02 Samsung Electronics Co., Ltd. Method and device for joint denoising and demosaicing using neural network
CN113591025A (en) * 2021-08-03 2021-11-02 深圳思谋信息科技有限公司 Feature map processing method and device, convolutional neural network accelerator and medium
CN115841416A (en) * 2022-11-29 2023-03-24 白盒子(上海)微电子科技有限公司 Reconfigurable intelligent image processor architecture for automatic driving field
CN115841416B (en) * 2022-11-29 2024-03-19 白盒子(上海)微电子科技有限公司 Reconfigurable intelligent image processor architecture for automatic driving field

Also Published As

Publication number Publication date
CN105681628B (en) 2018-12-07

Similar Documents

Publication Publication Date Title
CN105681628A (en) Convolution network arithmetic unit, reconfigurable convolution neural network processor and image de-noising method of reconfigurable convolution neural network processor
JP6857286B2 (en) Improved performance of neural network arrays
CN108764466B (en) Convolution neural network hardware based on field programmable gate array and acceleration method thereof
CN109102065B (en) Convolutional neural network accelerator based on PSoC
CN105930902B (en) A kind of processing method of neural network, system
CN108416327B (en) Target detection method and device, computer equipment and readable storage medium
CN110288086B (en) Winograd-based configurable convolution array accelerator structure
US20190026626A1 (en) Neural network accelerator and operation method thereof
CN111414994B (en) FPGA-based Yolov3 network computing acceleration system and acceleration method thereof
CN106203617A (en) A kind of acceleration processing unit based on convolutional neural networks and array structure
CN110321997B (en) High-parallelism computing platform, system and computing implementation method
CN110383300A (en) A kind of computing device and method
CN103019656A (en) Dynamically reconfigurable multi-stage parallel single instruction multiple data array processing system
CN117933314A (en) Processing device, processing method, chip and electronic device
CN110276447A (en) A kind of computing device and method
CN108320018A (en) A kind of device and method of artificial neural network operation
CN109472734B (en) Target detection network based on FPGA and implementation method thereof
de Sousa et al. An FPGA distributed implementation model for embedded SOM with on-line learning
CN111931927B (en) Method and device for reducing occupation of computing resources in NPU
CN110009644B (en) Method and device for segmenting line pixels of feature map
CN109146069B (en) Arithmetic device, arithmetic method, and chip
CN108960420B (en) Processing method and acceleration device
CN116888605A (en) Operation method, training method and device of neural network model
Li et al. Fpga-based object detection acceleration architecture design
CN204440489U (en) A kind of configurable multi-functional data path structure for infrared image processing

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant