US20150070530A1 - Imaging apparatus - Google Patents
Imaging apparatus Download PDFInfo
- Publication number
- US20150070530A1 US20150070530A1 US14/477,188 US201414477188A US2015070530A1 US 20150070530 A1 US20150070530 A1 US 20150070530A1 US 201414477188 A US201414477188 A US 201414477188A US 2015070530 A1 US2015070530 A1 US 2015070530A1
- Authority
- US
- United States
- Prior art keywords
- image data
- image
- processing
- section
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 149
- 238000007781 pre-processing Methods 0.000 claims abstract description 198
- 238000000034 method Methods 0.000 claims description 100
- 230000008569 process Effects 0.000 claims description 93
- 230000005540 biological transmission Effects 0.000 claims description 61
- 230000015654 memory Effects 0.000 claims description 33
- 238000006243 chemical reaction Methods 0.000 claims description 11
- 238000001914 filtration Methods 0.000 description 44
- 238000010586 diagram Methods 0.000 description 21
- 238000007792 addition Methods 0.000 description 6
- 230000006870 function Effects 0.000 description 6
- 238000007906 compression Methods 0.000 description 5
- 230000006835 compression Effects 0.000 description 4
- 230000006837 decompression Effects 0.000 description 4
- 238000005520 cutting process Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000005401 electroluminescence Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000003705 background correction Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000011946 reduction process Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/907—Television signal recording using static stores, e.g. storage tubes or semiconductor memories
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
- H04N5/772—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera the recording apparatus and the television camera being placed in the same enclosure
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
Definitions
- the present invention relates to an imaging apparatus.
- a process such as a spatial filtering process is performed on a signal of each pixel (hereinafter, referred to as a “pixel signal”) output from a solid-state imaging device provided in the imaging apparatus.
- the solid-state imaging device performs a raster scan of pixels, and sequentially outputs pixel signals in a horizontal direction of the image.
- an image-processing device that is provided in the imaging apparatus and performs the spatial filtering process needs to include a line memory that temporarily stores pixel signals in a horizontal direction (lateral direction), which are output from the solid-state imaging device, by the number of rows required for the spatial filtering process (refer to FIGS. 9A and 9B ).
- FIG. 9A when performing the spatial filtering process on “3 ⁇ 3” pixels of an image, the image-processing device processes image data of “3 rows ⁇ 3 columns” simultaneously. Therefore, as shown in FIG. 9B , the image-processing device needs to include two line memories in which pixel signals in the horizontal direction of the image are stored by one row.
- FIGS. 9A and 9B show an example of the configuration including two line memories (each for 6000 pixels) corresponding to an image having a horizontal width (image size in the horizontal direction) of 6000 pixels.
- the storage capacity of each line memory increases according to the corresponding number of pixels in the horizontal direction in the spatial filtering process.
- the increase in the storage capacity of each line memory in the image-processing device causes an increase in the circuit size of the imaging apparatus.
- Japanese Patent (Granted) Publication No. 4179701 discloses a technique of an image-processing device that performs image processing on one image a multiple of times.
- pixel signals output from a solid-state imaging device are temporarily stored in a frame memory provided outside the image-processing device, and then image processing is performed multiple times while reading pixel signals required for processing from the frame memory, thereby generating one image. That is, in the technique disclosed in Japanese Patent (Granted) Publication No. 4179701, one image is divided into a plurality of blocks, and image processing is performed by the same number of times as the number of divided blocks, thereby generating one image.
- an imaging apparatus including: a solid-state imaging device which has a plurality of pixels and outputs a pixel signal corresponding to a formed image of a subject as an image data; an image-capturing processing unit which outputs a pre-processed image data that is obtained by performing pre-processing on a part of the image data input from the solid-state imaging device, and an un-processed image data except the part of the image data from the image data input from the solid-state imaging device; and a storage unit which stores the pre-processed image data and the un-processed image data.
- the image-capturing processing unit may include: a pre-processing section that has a line memory, which stores the image data whose amount is smaller than a data amount of the image data in a first direction of the image, and outputs the pre-processed image data obtained by performing pre-processing on the image data of the image divided within a range of the data amount of the image data that is capable of being stored in the line memory; a first data transmission section that transmits the pre-processed image data to the storage unit; and a second data transmission section that transmits the image data in the first direction of the image that has not been pre-processed by the pre-processing section, to the storage unit as the un-processed image data.
- the image-capturing processing unit may further include: a data acquisition section that acquires image data stored in the storage unit; and a selection section that outputs the image data input from the solid-state imaging device or the image data acquired by the data acquisition section to at least one of the pre-processing section and the second data transmission section.
- the first data transmission section may transmit first pre-processed image data, which is obtained after the pre-processing section performs pre-processing on the image data of a first image obtained by dividing the image within a range of the data amount of the image data that is capable of being stored in the line memory, to the storage unit, and the second data transmission section may transmit the image data of a second image obtained by dividing the image that has not been pre-processed by the pre-processing section, as first un-processing image data, to the storage unit.
- the data acquisition section may acquire the first un-processing image data stored in the storage unit
- the selection section may output the first un-processing image data to the pre-processing section
- the first data transmission section may transmit second pre-processed image data, which is obtained after the pre-processing section performs pre-processing on the first un-processing image data, to the storage unit.
- the image-capturing processing unit may further include a conversion section that performs conversion processing for converting the pre-processed image data and outputs converted image data after the conversion processing.
- the conversion section may include a resizing section that resizes an image corresponding to the pre-processed image data to an image, which has a size set in advance in the first direction and a second direction perpendicular to the first direction, and outputs the converted image data of the resized image.
- the first data transmission section may transmit second pre-processed image data, which is obtained after the pre-processing section performs pre-processing on the first un-processing image data, to the storage unit, and the second data transmission section may transmit first converted image data, which is obtained after the resizing section performs a resizing process on the second pre-processed image data, to the storage unit.
- the data acquisition section may acquire the first pre-processed image data stored in the storage unit
- the selection section may output the first pre-processed image data to the resizing section
- the second data transmission section may transmit second converted image data, which is obtained after the resizing section performs a resizing process on the first pre-processed image data, to the storage unit.
- FIG. 1 is a block diagram showing the schematic configuration of an imaging apparatus according to a first embodiment of the present invention.
- FIG. 2 is a diagram schematically showing an example of the first operation in a spatial filtering process by the imaging apparatus of the first embodiment.
- FIG. 3 is a diagram schematically showing an example of the second operation in the spatial filtering process by the imaging apparatus of the first embodiment.
- FIG. 4 is a diagram illustrating an example of an image on which a spatial filtering process is performed by the imaging apparatus of the first embodiment.
- FIG. 5 is a block diagram showing the schematic configuration of an imaging apparatus according to a second embodiment of the present invention.
- FIG. 6 is a diagram schematically showing an example of the first operation in a spatial filtering process by the imaging apparatus of the second embodiment.
- FIG. 7 is a diagram schematically showing an example of the second operation in the spatial filtering process by the imaging apparatus of the second embodiment.
- FIG. 8 is a diagram schematically showing an example of the third operation in the spatial filtering process by the imaging apparatus of the second embodiment.
- FIG. 9A is a diagram showing a spatial filtering process in a known imaging apparatus and the schematic configuration to execute the spatial filtering process.
- FIG. 9B is a diagram showing a spatial filtering process in a known imaging apparatus and the schematic configuration to execute the spatial filtering process.
- FIG. 1 is a block diagram showing the schematic configuration of an imaging apparatus according to a first embodiment.
- An imaging apparatus 10 shown in FIG. 1 includes an image sensor 100 , an image-capturing processing unit 200 , an image-processing unit 300 , a display-processing unit 400 , a display device 401 , a card interface (IF) unit 500 , a recording medium 501 , a dynamic random access memory (DRAM) controller 600 , a DRAM 601 , and a CPU 700 .
- the image-capturing processing unit 200 of the imaging apparatus 10 includes a selection section 201 , a pre-processing section 202 , a first data transmission section 203 , a second data transmission section 204 , and a data acquisition section 205 .
- the image-capturing processing unit 200 , the image-processing unit 300 , the display-processing unit 400 , the card IF unit 500 , the DRAM controller 600 , and the CPU 700 in the imaging apparatus 10 are connected to each other through a data bus 800 .
- reading of data from the DRAM 601 connected to the DRAM controller 600 and writing of data into the DRAM 601 are performed by direct memory access (DMA).
- DMA direct memory access
- the image sensor 100 is a solid-state imaging device represented by a charge coupled device (CCD) image sensor or a complementary metal oxide semiconductor (CMOS) image sensor that photoelectrically converts an optical image of a subject that is formed by a lens (not shown).
- CCD charge coupled device
- CMOS complementary metal oxide semiconductor
- the image sensor 100 performs a raster scan of pixels provided in the image sensor 100 itself, and sequentially outputs pixel signals corresponding to subject light to the image-capturing processing unit 200 , as image data, in a horizontal direction of the image.
- the image-capturing processing unit 200 acquires the image data input from the image sensor 100 , performs pre-processing including a spatial filtering process on the acquired image data, and transmits (writes) image data obtained as a result of the pre-processing to the DRAM 601 through the DRAM controller 600 .
- the amount of image data that can be pre-processed by the image-capturing processing unit 200 is smaller than the amount of image data that the image sensor 100 outputs as the amount of data of one row. Therefore, the image-capturing processing unit 200 transmits (writes) image data, which cannot be pre-processed, to the DRAM 601 through the DRAM controller 600 as it is, that is, without performing pre-processing. Then, the image-capturing processing unit 200 completes pre-processing on data, which can be pre-processed, and stores the pre-processed data in the DRAM 601 as it is.
- the image-capturing processing unit 200 acquires image data that has not been pre-processed, performs pre-processing on the acquired image data, and transmits (writes) the pre-processed data to the DRAM 601 again. That is, when image data having an amount in the horizontal direction that cannot be pre-processed is input, the image-capturing processing unit 200 performs pre-processing on all pieces of the input image data by performing the pre-processing separately on the input image data divided in a range of the amount of data that can be pre-processed.
- the selection section 201 selects a component of the output destination of the input image data. Specifically, as an output destination of the image data input from the image sensor 100 and the image data input from the data acquisition section 205 , the selection section 201 selects one or both of the pre-processing section 202 and the second data transmission section 204 . Then, the selection section 201 outputs the input image data to the selected output destination.
- the pre-processing section 202 performs various kinds of pre-processing including a spatial filtering process, such as defect correction, shading correction, and pixel defect correction, on the image data input from the selection section 201 , and outputs image data obtained as a result of the processing (hereinafter, referred to as “pre-processed image data”) to the first data transmission section 203 .
- the pre-processing section 202 is configured to include a plurality of line memories each of which stores image data, which corresponds to one row in the horizontal direction (lateral direction, first direction) of the image, of the image data input from the selection section 201 , as shown in FIG. 9B .
- the configuration of the pre-processing section 202 is the same as the configuration of a pre-processing section provided in an imaging apparatus in the related art.
- the pre-processing operation of the pre-processing section 202 is also the same as the operation of the pre-processing section provided in the imaging apparatus in the related art. Therefore, detailed explanation regarding the configuration of the pre-processing section 202 or the pre-processing operation of the pre-processing section 202 will be omitted.
- the first data transmission section 203 stores (writes) the pre-processed image data input from the pre-processing section 202 in the DRAM 601 connected to the DRAM controller 600 by the DMA, for example. Since the configuration of the first data transmission section 203 or the operation when storing (writing) the pre-processed image data in the DRAM 601 by the DMA is the same as the normal configuration or operation to store data by the DMA, detailed explanation thereof will be omitted.
- the method used when the first data transmission section 203 stores (writes) the pre-processed image data in the DRAM 601 may be a method other than the DMA.
- the second data transmission section 204 stores (writes) the image data input from the selection section 201 , that is, image data that has not been pre-processed by the pre-processing section 202 (hereinafter, referred to as “original image data”) in the DRAM 601 connected to the DRAM controller 600 by the DMA, for example.
- original image data image data that has not been pre-processed by the pre-processing section 202
- the configuration of the second data transmission section 204 or the operation when storing (writing) the original image data in the DRAM 601 by the DMA is the same as the normal configuration or operation to store data by the DMA. Accordingly, detailed explanation thereof will be omitted.
- the method used when the second data transmission section 204 stores (writes) the original image data in the DRAM 601 may be a method other than the DMA.
- the data acquisition section 205 acquires (reads) the image data stored in the DRAM 601 through the DRAM controller 600 by the DMA, for example, and outputs the acquired image data to the selection section 201 .
- the image data acquired by the data acquisition section 205 may not only be the pre-processed image data stored by the first data transmission section 203 or the original image data stored by the second data transmission section 204 but also image data stored in the DRAM 601 after image processing by the image-processing unit 300 , for example.
- the image-processing unit 300 acquires (reads) the pre-processed image data stored in the DRAM 601 , and generates image data for display or image data for recording by performing various kinds of image processing, such as noise removal processing, distortion aberration correction, YC conversion processing, resizing process, JPEG compression processing, and video compression processing (for example, MPEG compression processing or H.264 compression processing), on the acquired pre-processed image data. Then, the image-processing unit 300 stores (writes) the generated image data for display or image data for recording in the DRAM 601 again.
- image processing such as noise removal processing, distortion aberration correction, YC conversion processing, resizing process, JPEG compression processing, and video compression processing (for example, MPEG compression processing or H.264 compression processing
- the image-processing unit 300 acquires (reads) the image data for recording stored in the DRAM 601 , and generates image data for display by performing various kinds of image processing, such as JPEG decompression processing and video decompression processing (for example, MPEG decompression processing or H.264 decompression processing). Then, the image-processing unit 300 stores (writes) the generated image data for display in the DRAM 601 again.
- JPEG decompression processing and video decompression processing for example, MPEG decompression processing or H.264 decompression processing.
- the display-processing unit 400 acquires (reads) the image data for display stored in the DRAM 601 , and performs display processing, such as a process of superimposing data for on-screen display (OSD) on the acquired image data for display. Then, the image data after display processing is output to the display device 401 .
- display processing such as a process of superimposing data for on-screen display (OSD) on the acquired image data for display. Then, the image data after display processing is output to the display device 401 .
- OSD on-screen display
- the display device 401 is a display device, such as a thin film transistor (TFT) liquid crystal display (LCD) or an organic electroluminescence (EL) display, and displays an image corresponding to the image data after display processing that is output from the display-processing unit 400 .
- TFT thin film transistor
- LCD liquid crystal display
- EL organic electroluminescence
- the display device 401 may be a display device, such as an electronic view finder (EVF), or an external display, such as a television. Although the display device 401 is also a component of the imaging apparatus 10 in FIG. 1 , the display device 401 may be configured to be detachable from the imaging apparatus 10 .
- EVF electronic view finder
- the card IF unit 500 acquires (reads) the image data for recording that is recorded in the DRAM 601 , and records the acquired image data for recording on the recording medium 501 . In addition, the card IF unit 500 reads the image data recorded on the recording medium 501 , and transmits (writes) the read image data to the DRAM 601 .
- the recording medium 501 is a recording medium, such as an SD memory card or a compact flash (CF: registered trademark), and records the image data for recording that is output from the card IF unit 500 . In addition, the image data recorded by the card IF unit 500 is read.
- the recording medium 501 is also a component of the imaging apparatus 10 in FIG. 1 , the recording medium 501 may be configured to be detachable from the imaging apparatus 10 .
- the DRAM controller 600 controls the storage (writing) of data into the connected DRAM 601 and the acquisition (reading) of data from the DRAM 601 in response to an access request to the DRAM 601 from a plurality of components in the imaging apparatus 10 connected to the data bus 800 , for example, in response to the access request of DMA.
- the DRAM 601 is a memory (storage unit) whose access control is performed by the DRAM controller 600 .
- the DRAM 601 temporarily stores (memorizes) various kinds of data in the process of each component of the imaging apparatus 10 .
- the CPU 700 controls the components of the imaging apparatus 10 , that is, the entire imaging apparatus 10 .
- the CPU 700 controls the operation of each component in the imaging apparatus 10 according to an imaging operation or a reproduction operation in the imaging apparatus 10 .
- the CPU 700 controls the start of the output of the image data from the image sensor 100 or the start of the acquisition of the image data by the image-capturing processing unit 200 when the imaging apparatus 10 performs an imaging operation.
- the CPU 700 controls the selection of the output destination of image data by the selection section 201 of the image-capturing processing unit 200 , or performs a setting of transmission of image data by the first and second data transmission sections 203 and 204 and a setting of the data acquisition section 205 .
- the imaging apparatus 10 of the first embodiment it is possible to acquire the image data output from the image sensor 100 in real time while performing various kinds of pre-processing including a spatial filtering process directly on the acquired image data.
- various kinds of pre-processing including a spatial filtering process can be performed on the image data stored in the DRAM 601 by acquiring the image data stored in the DRAM 601 by the data acquisition section 205 . That is, in the imaging apparatus 10 of the first embodiment, various kinds of pre-processing including a spatial filtering process can also be performed on image data other than the image data that is input in real time.
- the pre-processing section 202 includes a line memory having a storage capacity of 6000 pixels in the horizontal direction.
- input image data when the amount of image data in the horizontal direction that the image sensor 100 outputs for each row (hereinafter, referred to as “input image data”) is larger than the amount of image data of one row that can be stored in a line memory provided in the pre-processing section 202 , and accordingly, all pieces of the input image data of one row cannot be stored in the line memory.
- the imaging apparatus 10 When the amount of input image data is larger than the amount of data that can be stored in a line memory provided in the pre-processing section 202 , that is, when the amount of input image data that can be pre-processed by the image-capturing processing unit 200 is smaller than the amount of input image data that the image sensor 100 outputs for each row, the imaging apparatus 10 performs pre-processing on all pieces of the input image data by performing the pre-processing separately on the input image data divided in a range of the amount of data that can be pre-processed (here, 6000 pixels) as described above.
- a range of the amount of data that can be pre-processed here, 6000 pixels
- FIGS. 2 and 3 are diagrams schematically showing an example of the pre-processing operation of the imaging apparatus 10 of the first embodiment.
- FIG. 2 shows an example of the first operation when pre-processing section 202 performs pre-processing (spatial filtering process) on the input image data of one half (left side) that is input from the image sensor 100 .
- FIG. 3 shows an example of the second operation when pre-processing section 202 performs pre-processing (spatial filtering process) on the input image data of the other half (right side) that is input from the image sensor 100 .
- FIG. 2 shows the path of each piece of data in the first operation (operation of the first pass) on the block diagram of the imaging apparatus 10 shown in FIG. 1 .
- pre-processed image data obtained by performing pre-processing on the input image data in real time is transmitted (written) to the DRAM 601 .
- the image sensor 100 captures an image P1 having a horizontal width of 10000 pixels, and outputs the input image data of the captured image P1 to the image-capturing processing unit 200 .
- the image-capturing processing unit 200 input image data in all regions of the image P1 input from the image sensor 100 is acquired, and the acquired input image data of the image P1 is sequentially input to the selection section 201 of the image-capturing processing unit 200 .
- the selection section 201 outputs the input image data in all regions of the input image P1 sequentially to both the pre-processing section 202 and the second data transmission section 204 .
- the pre-processing section 202 performs pre-processing on input image data of an image P2, which is a region of the left half, of the input image data of the image P1 that is sequentially input from the selection section 201 , and sequentially outputs pre-processed image data of the image P2 after pre-processing to the first data transmission section 203 . Then, the first data transmission section 203 transmits (writes) the pre-processed image data of the image P2, which is sequentially input from the pre-processing section 202 , to the DRAM 601 through the DRAM controller 600 . As a result, the pre-processed image data of the image P2 after pre-processing is stored in the DRAM 601 .
- the second data transmission section 204 transmits (writes) input image data of an image P3 that is a region of the right half, of the input image data of the image P1 that is sequentially input from the selection section 201 , to the DRAM 601 through the DRAM controller 600 as it is as original image data of the image P3.
- the input image data of the image P3 of the right half of the image P1 output from the image sensor 100 is stored in the DRAM 601 as it is.
- pre-processed image data of the image P2 obtained by performing pre-processing on the region of the left half of the image P1 output from the image sensor 100 and the original image data of the image P3, which is a region of the right half of the image P1 output from the image sensor 100 , are stored in the DRAM 601 .
- the imaging apparatus 10 transmits (writes) to the DRAM 601 pre-processed image data obtained by performing pre-processing on the original image data of the image P3 that is the input image data of the right half of the image P1 input from the image sensor 100 , which could not be pre-processed in the first operation (operation of the first pass).
- FIG. 3 shows the path of data in the second operation (operation of the second pass) on the block diagram of the imaging apparatus 10 shown in FIG. 1 .
- pre-processed image data obtained by performing pre-processing on the original image data of the image P3 stored in the DRAM 601 is transmitted (written) to the DRAM 601 .
- the data acquisition section 205 sequentially acquires (reads) the original image data of the image P3 stored in the DRAM 601 , and sequentially outputs the acquired original image data of the image P3 to the selection section 201 .
- the selection section 201 sequentially outputs the input original image data of the image P3 to the pre-processing section 202 .
- the pre-processing section 202 performs pre-processing on the original image data of the image P3 that is sequentially input from the selection section 201 , and sequentially outputs pre-processed image data of an image P4 after pre-processing to the first data transmission section 203 . Then, the first data transmission section 203 transmits (writes) the pre-processed image data of the image P4, which is sequentially input from the pre-processing section 202 , to the DRAM 601 through the DRAM controller 600 . As a result, the pre-processed image data of the image P4 after pre-processing is stored in the DRAM 601 .
- pre-processed image data of the image P4 obtained by performing pre-processing on the image P3 of the right half of the image P1 output from the image sensor 100 , which is stored in the DRAM 601 , is stored in the DRAM 601 .
- pre-processed image data obtained by performing pre-processing on the input image data in all regions of one image P1 having a horizontal width of 10000 pixels, which is captured by the image sensor 100 , is stored in the DRAM 601 .
- pre-processing on all pieces of the input image data of one row can be performed by two operations of the first and second operations.
- the imaging apparatus 10 transmits (writes) pre-processed image data, which is obtained by performing pre-processing in real time on the input image data of the left half of the image P1 input from the image sensor 100 , and also transmits (writes) the input image data of the right half, which could not be pre-processed, to the DRAM 601 as it is.
- the imaging apparatus 10 performs pre-processing on all pieces of the input image data of one image P1 by transmitting (writing) to the DRAM 601 pre-processed image data obtained by performing pre-processing on the input image data of the right half of the image P1 input from the image sensor 100 , which could not be pre-processed in the first operation (operation of the first pass).
- the imaging apparatus 10 of the first embodiment even when the amount of image data of one row that can be stored in a line memory provided in the pre-processing section 202 of the image-capturing processing unit 200 is smaller than the amount of image data that the image sensor 100 outputs for each row, it is possible to perform pre-processing on all pieces of image data output from the image sensor 100 . That is, it is possible to perform pre-processing corresponding to the image sensor 100 in which the number of pixels has increased, without increasing the circuit size of the imaging apparatus 10 , especially, the storage capacity of the line memory. In other words, it is possible to meet an increase in the number of pixels of the image sensor with a small circuit size.
- FIG. 4 is a diagram illustrating an example of an image on which a spatial filtering process is performed by the imaging apparatus 10 of the first embodiment.
- FIG. 4 shows an example of the case where an overlapping region is provided in a region of the left half, which is processed in the first operation (operation of the first pass), and a region of the right half, which is processed in the second operation (operation of the second pass), at a position where the entire region of the image P1 is divided, when dividing the input image data of the entire region of the image P1 input from the image sensor 100 in the pre-processing operation of the imaging apparatus 10 shown in FIGS. 2 and 3 .
- an overlapping region when performing the spatial filtering process multiple times, it is possible to eliminate the inconvenience of having the image in a boundary where the process is divided.
- pre-processing on the input image data of the region of the right half of the image P1 may be performed in the first operation (operation of the first pass), and pre-processing on the input image data of the region of the left half of the image P1 may be performed in the second operation (operation of the second pass).
- the number of division times of pre-processing (spatial filtering process) performed on the input image data of one image is not limited to twice as described above, and may be 3 times or more.
- the operation of the pre-processing in the imaging apparatus 10 of the first embodiment does not necessarily need to be performed twice or more.
- the pre-processing section 202 can perform pre-processing only once. Accordingly, the pre-processing does not necessarily need to be performed twice or more.
- the number of times by which the pre-processing section 202 performs pre-processing can be changed according to the number of pixels in the horizontal direction of the image sensor mounted in the imaging apparatus.
- an image sensor can be driven in a plurality of driving modes, such as a driving mode in which all pixel signals are output (hereinafter, referred to as an “all pixel output mode”) and a driving mode in which pixel signals are added and output (hereinafter, referred to as a “pixel addition mode”).
- the all pixel output mode is set when capturing a high-resolution image, such as a still image, and the pixel addition mode is set when performing video imaging or imaging of a so-called live view function of displaying a video for checking a subject to be imaged on a display device.
- the amount of image data that the image sensor 100 outputs for one row is smaller than the amount of image data (6000 pixels) of one row that can be stored in the line memory provided in the pre-processing section 202 of the image-capturing processing unit 200 .
- the pre-processing section 202 can perform pre-processing only once. Therefore, the pre-processing does not necessarily need to be performed twice or more.
- pre-processing is performed twice when the image sensor is driven in the all pixel output mode, that is, in the mode in which a still image is captured, and pre-processing is performed once in real time when the image sensor is driven in the pixel addition mode, that is, in the mode in which video imaging or live view imaging is performed, for example. Therefore, the number of times by which the pre-processing section 202 performs pre-processing can be changed according to an image to be captured.
- the case has been described in which the image sensor 100 outputs a pixel signal, which is obtained by performing a raster scan in the horizontal direction of the image, as image data.
- the concept of the present invention can be similarly applied by regarding the horizontal direction in the first embodiment as a vertical direction, that is, by regarding the horizontal direction and the vertical direction of the image reversely.
- the case has been described in which pre-processing is performed multiple times by horizontally dividing the image input from the image sensor 100 .
- pre-processing is performed multiple times by vertically dividing the image input from the image sensor 100
- the concept of the present invention can be similarly applied by regarding the horizontal direction and the vertical direction of the image reversely.
- the image-capturing processing unit may perform various processes other than the pre-processing.
- a conversion section that generates a converted image by converting a normal image may be provided in addition to the configuration to record a normal image, which has the number of pixels that is almost the same as the number of effective pixels of the image sensor, as a still image.
- a resizing section that generates a reduced image having the same angle of view as a converted image with the number of pixels smaller than the normal number of pixels.
- the reduced image generated by the resizing section may be displayed on a display device so that the captured image can be immediately checked, or may be used as an image recorded together with a normal still image (so-called thumbnail) or an image for list display of captured images. That is, various applications can be considered.
- an imaging apparatus of a second embodiment will be described.
- the image-capturing processing unit 200 provided in the imaging apparatus 10 of the first embodiment shown in FIG. 1 is different. Therefore, in the explanation of the imaging apparatus of the second embodiment, the same reference numerals are given to the same components as in the imaging apparatus 10 of the first embodiment. In addition, only components and operations different from those in the imaging apparatus 10 of the first embodiment will be described, and detailed explanation regarding the same components and operations as in the imaging apparatus 10 of the first embodiment will be omitted.
- FIG. 5 is a block diagram showing the schematic configuration of the imaging apparatus according to the second embodiment.
- An imaging apparatus 11 shown in FIG. 5 includes an image sensor 100 , an image-capturing processing unit 210 , an image-processing unit 300 , a display-processing unit 400 , a display device 401 , a card IF unit 500 , a recording medium 501 , a DRAM controller 600 , a DRAM 601 , and a CPU 700 .
- the image-capturing processing unit 210 of the imaging apparatus 11 includes a selection section 201 , a pre-processing section 202 , a selection section 216 , a first data transmission section 203 , a resizing section 217 , a second data transmission section 204 , and a data acquisition section 205 .
- the image-capturing processing unit 210 , the image-processing unit 300 , the display-processing unit 400 , the card IF unit 500 , the DRAM controller 600 , and the CPU 700 in the imaging apparatus 11 are connected to each other through a data bus 800 , in the same manner as in the imaging apparatus 10 of the first embodiment shown in FIG. 1 .
- reading of data from the DRAM 601 connected to the DRAM controller 600 and writing of data into the DRAM 601 are performed by the DMA.
- the image sensor 100 performs a raster scan of pixels provided in the image sensor 100 itself, and sequentially outputs pixel signals corresponding to subject light to the image-capturing processing unit 210 , as image data, in a horizontal direction of the image.
- the image-capturing processing unit 210 acquires input image data that is input from the image sensor 100 , and transmits (writes) pre-processed image data, which is obtained by performing pre-processing including a spatial filtering process on the acquired input image data, and input image data, which cannot be pre-processed, to the DRAM 601 through the DRAM controller 600 .
- the amount of image data that can be pre-processed by the image-capturing processing unit 210 is smaller than the amount of input image data that the image sensor 100 outputs as the amount of data of one row. Therefore, when input image data having an amount in the horizontal direction that cannot be pre-processed is input, the image-capturing processing unit 210 also performs pre-processing on all pieces of the input image data by performing the pre-processing separately on the input image data divided in a range of the amount of data that can be pre-processed, in the same manner as in the image-capturing processing unit 200 in the first embodiment.
- the image-capturing processing unit 210 generates a reduced image by performing a reduction process, such as a resizing process or a cutting process, based on the pre-processed image data, and transmits (writes) the generated reduced image to the DRAM 601 through the DRAM controller 600 .
- a reduction process such as a resizing process or a cutting process
- the image-capturing processing unit 210 functions as a processing unit that generates a reduced image by performing a resizing process.
- the selection section 201 selects one or both of the pre-processing section 202 and the selection section 216 as an output destination of the input image data, which is input from the image sensor 100 , and the image data input from the data acquisition section 205 , and outputs the input image data to the selected output destination.
- the pre-processing section 202 outputs pre-processed image data, which is obtained by performing various kinds of pre-processing including a spatial filtering process on the image data input from the selection section 201 , to the selection section 216 .
- the selection section 216 selects a component of the output destination of the input image data. More specifically, the selection section 216 performs conversion processing for converting the pre-processed image data input from the pre-processing section 202 , and outputs converted image data. The selection section 216 outputs the image data input from the selection section 201 . In this case, the selection section 216 selects at least one of the first data transmission section 203 and the resizing section 217 . Then, the selection section 216 outputs the input image data to the selected output destination.
- the first data transmission section 203 stores (writes) the image data input from the selection section 216 in the DRAM 601 connected to the DRAM controller 600 by the DMA, for example.
- the resizing section 217 performs a resizing process on the image data input from the selection section 216 , and outputs image data of a reduced image (hereinafter, referred to as “reduced image data”), which is obtained as a result of resizing (here, reducing) the pre-processed image in a horizontal direction (lateral direction, first direction) and a vertical direction (longitudinal direction, second direction), to the second data transmission section 204 .
- the resizing section 217 can also output the image data (pre-processed image data, input image data, or original image data), which is input from the selection section 216 , to the second data transmission section 204 as it is without performing a resizing process. That is, the resizing section 217 can turn off the function of the resizing process.
- the second data transmission section 204 stores (writes) the image data input from the resizing section 217 , that is, reduced image data or image data input to the resizing section 217 in the DRAM 601 connected to the DRAM controller 600 by the DMA, for example.
- the image-processing unit 300 When the reduced image data is stored in the DRAM 601 , the image-processing unit 300 performs image processing on both the pre-processed image data and the reduced image data. For example, when the reduced image data is a thumbnail, the thumbnail is associated when performing a JPEG compression process.
- the CPU 700 controls the components of the imaging apparatus 11 , that is, the entire imaging apparatus 11 .
- the CPU 700 controls the operation of each component in the imaging apparatus 11 according to an imaging operation or a reproduction operation in the imaging apparatus 11 .
- the CPU 700 controls the start of the output of the image data from the image sensor 100 or the start of the acquisition of the image data by the image-capturing processing unit 210 when the imaging apparatus 11 performs an imaging operation.
- the CPU 700 controls the operation of each component in the image-capturing processing unit 210 in each operation after the start of the acquisition of the image data by the image-capturing processing unit 210 .
- the CPU 700 controls the selection of the output destination of image data by the selection sections 201 and 216 of the image-capturing processing unit 210 , or performs a setting of the resizing process by the resizing section 217 , a setting of the transmission of image data by the first and second data transmission sections 203 and 204 , or a setting of the data acquisition section 205 .
- the imaging apparatus 11 of the second embodiment similar to the imaging apparatus 10 of the first embodiment, it is possible to perform various kinds of pre-processing including a direct spatial filtering process on image data input in real time and various kinds of pre-processing including a spatial filtering process on the image data stored in the DRAM 601 .
- the imaging apparatus 11 of the second embodiment it is possible to generate a reduced image by performing a resizing process on the pre-processed image data in addition to various kinds of pre-processing including a spatial filtering process.
- the pre-processing section 202 includes a line memory having a storage capacity of 6000 pixels in the horizontal direction.
- FIGS. 6 to 8 are diagrams schematically showing an example of the operation of pre-processing and reduced image generation by the imaging apparatus 11 of the second embodiment.
- FIG. 6 shows an example of the first operation when the imaging apparatus 11 performs pre-processing (spatial filtering process) on the input image data of one half (left side) that is input from the image sensor 100 .
- FIG. 7 shows an example of the second operation when the imaging apparatus 11 performs pre-processing (spatial filtering process) on the input image data of the other half (right side) that is input from the image sensor 100 while generating a reduced image corresponding to the pre-processed image data of the other half (right side).
- FIG. 8 shows an example of the third operation when the imaging apparatus 11 generates a reduced image corresponding to the pre-processed image data of one half (left side).
- FIG. 6 shows the path of each piece of data in the first operation (operation of the first pass) on the block diagram of the imaging apparatus 11 shown in FIG. 5 .
- pre-processed image data obtained by performing pre-processing on the input image data in real time is transmitted (written) to the DRAM 601 .
- the image sensor 100 captures an image P1 having a horizontal width of 10000 pixels, and outputs the input image data of the captured image P1 to the image-capturing processing unit 210 .
- the image-capturing processing unit 210 input image data in all regions of the image P1 input from the image sensor 100 is acquired, and the acquired input image data of the image P1 is sequentially input to the selection section 201 of the image-capturing processing unit 210 .
- the selection section 201 outputs the input image data in all regions of the input image P1 sequentially to both the pre-processing section 202 and the selection section 216 .
- the pre-processing section 202 performs pre-processing on input image data of the image P2, which is a region of the left half, of the input image data of the image P1 that is sequentially input from the selection section 201 , and sequentially outputs the pre-processed image data of the image P2 after pre-processing to the selection section 216 .
- the selection section 216 outputs the pre-processed image data of the image P2, which is input from the pre-processing section 202 , to the first data transmission section 203 .
- the first data transmission section 203 transmits (writes) the pre-processed image data of the image P2, which is sequentially input from the selection section 216 , to the DRAM 601 through the DRAM controller 600 .
- the pre-processed image data of the image P2 after pre-processing is stored in the DRAM 601 .
- the selection section 216 sequentially outputs the input image data of the image P1, which is sequentially input from the selection section 201 , to the resizing section 217 .
- the function of the resizing process in the resizing section 217 is turned off.
- the resizing section 217 outputs input image data of the image P3, which is a region of the right half, of the input image data of the image P1 that is sequentially input from the selection section 216 , to the second data transmission section 204 as it is.
- the second data transmission section 204 transmits (writes) the input image data of the image P3 input sequentially from the resizing section 217 , as it is as original image data of the image P3, to the DRAM 601 through the DRAM controller 600 .
- the input image data of the image P3 of the right half of the image P1 output from the image sensor 100 is stored in the DRAM 601 as it is.
- pre-processed image data of the image P2 obtained by performing pre-processing on the region of the left half of the image P1 output from the image sensor 100 and the original image data of the image P3, which is a region of the right half of the image P1 output from the image sensor 100 , are stored in the DRAM 601 .
- the imaging apparatus 11 transmits (writes) to the DRAM 601 pre-processed image data obtained by performing pre-processing on the original image data of the image P3 that is the input image data of the right half of the image P1 input from the image sensor 100 , which could not be pre-processed in the first operation (operation of the first pass), and transmits (writes) to the DRAM 601 reduced image data obtained by performing a resizing process on the pre-processed image data of the image P3.
- FIG. 7 shows the path of data in the second operation (operation of the second pass) on the block diagram of the imaging apparatus 11 shown in FIG. 5 .
- pre-processed image data obtained by performing pre-processing on the original image data of the image P3 stored in the DRAM 601 is transmitted (written) to the DRAM 601 .
- the data acquisition section 205 sequentially acquires (reads) the original image data of the image P3 stored in the DRAM 601 , and sequentially outputs the acquired original image data of the image P3 to the selection section 201 .
- the selection section 201 sequentially outputs the input original image data of the image P3 to the pre-processing section 202 .
- the pre-processing section 202 performs pre-processing on the original image data of the image P3 input sequentially from the selection section 201 , and sequentially outputs the pre-processed image data of the image P4 after pre-processing to the selection section 216 .
- the selection section 216 sequentially outputs the pre-processed image data of the image P4, which is input from the pre-processing section 202 , to both the first data transmission section 203 and the resizing section 217 .
- the first data transmission section 203 transmits (writes) the pre-processed image data of the image P4, which is sequentially input from the selection section 216 , to the DRAM 601 through the DRAM controller 600 .
- the pre-processed image data of the image P4 after pre-processing is stored in the DRAM 601 .
- a reduced image is generated by performing a resizing process on the original image data of the image P3 stored in the DRAM 601 , and transmits (writes) the reduced image data of the generated reduced image to the DRAM 601 . More specifically, in the operation of the second pass, the function of the resizing process in the resizing section 217 is turned on.
- the resizing section 217 generates a reduced image by performing a resizing process on the pre-processed image data of the image P4 input sequentially from the selection section 216 , and outputs the reduced image data of a generated reduced image P5 to the second data transmission section 204 .
- the second data transmission section 204 transmits (writes) the reduced image data of the reduced image P5, which is sequentially input from the resizing section 217 , to the DRAM 601 through the DRAM controller 600 .
- the reduced image data of the reduced image P5 of the right half of the image P1 output from the image sensor 100 is stored in the DRAM 601 .
- pre-processed image data of the image P4 obtained by performing pre-processing on the image P3 of the right half of the image P1 output from the image sensor 100 , which is stored in the DRAM 601 , and reduced image data of the reduced image P5 obtained by performing a resizing process on the image P4 after pre-processing of the right half of the image P1, which is stored in the DRAM 601 , are stored in the DRAM 601 .
- pre-processed image data (pre-processed image data of the images P2 and P4) obtained by performing pre-processing on the input image data in all regions of one image P1 having a horizontal width of 10000 pixels, which is captured by the image sensor 100 , is stored in the DRAM 601 .
- the imaging apparatus 11 transmits (writes) reduced image data obtained by performing a resizing process on the pre-processed image data of the image P2, which is pre-processed image data obtained by performing pre-processing on the input image data of the left half of the image P1, to the DRAM 601 .
- FIG. 8 shows the path of data in the third operation (operation of the third pass) on the block diagram of the imaging apparatus 11 shown in FIG. 5 .
- a reduced image is generated by performing a resizing process on the pre-processed image data of the image P2 stored in the DRAM 601 , and transmits (writes) the reduced image data of the generated reduced image to the DRAM 601 .
- the data acquisition section 205 sequentially acquires (reads) the pre-processed image data of the image P2 stored in the DRAM 601 , and sequentially outputs the acquired pre-processed image data of the image P2 to the selection section 201 . Then, the selection section 201 sequentially outputs the input pre-processed image data of the image P2 to the selection section 216 .
- the selection section 216 sequentially outputs the pre-processed image data of the image P2, which is sequentially input from the selection section 201 , to the resizing section 217 .
- the function of the resizing process in the resizing section 217 is turned on.
- the resizing section 217 generates a reduced image by performing a resizing process on the pre-processed image data of the image P2 input sequentially from the selection section 216 , and outputs the reduced image data of a generated reduced image P6 to the second data transmission section 204 .
- the second data transmission section 204 transmits (writes) the reduced image data of the reduced image P6, which is sequentially input from the resizing section 217 , to the DRAM 601 through the DRAM controller 600 .
- the reduced image data of the reduced image P6 of the left half of the image P1 output from the image sensor 100 is stored in the DRAM 601 .
- reduced image data of the reduced image P6 obtained by performing a resizing process on the image P2 after pre-processing of the left half of the image P1 output from the image sensor 100 , which is stored in the DRAM 601 is stored in the DRAM 601 .
- reduced image data (reduced image data of the images P5 and P6) obtained by performing a resizing process after performing pre-processing on the input image data in all regions of one image P1 having a horizontal width of 10000 pixels, which is captured by the image sensor 100 is stored in the DRAM 601 .
- the imaging apparatus 11 of the second embodiment similar to the imaging apparatus 10 of the first embodiment, even when the amount of input image data in the horizontal direction that the image sensor 100 outputs for each row is larger than the amount of image data of one row that can be stored in a line memory provided in the pre-processing section 202 of the image-capturing processing unit 210 , pre-processing on all pieces of the input image data of one row can be performed by multiple operations. Therefore, also in the imaging apparatus 11 of the second embodiment, it is possible to achieve the same effect as in the imaging apparatus 10 of the first embodiment.
- a reduced image corresponding to an image after pre-processing can be generated by multiple operations, in addition to the pre-processing on the image input from the image sensor 100 . More specifically, in the second operation (operation of the second pass), the imaging apparatus 11 performs pre-processing on the image P3 of the right half of the image P1 input from the image sensor 100 , and generates the reduced image P5 by performing a resizing process on the pre-processed image data of the image P3 after pre-processing.
- the imaging apparatus 11 generates the reduced image P6 by performing a resizing process on the pre-processed image data of the image P2 obtained by performing pre-processing on the image of the left half of the image P1 input from the image sensor 100 in the first operation (operation of the first pass).
- the next processing can be performed from the region of the image on which pre-processing has been completed. Therefore, in the imaging apparatus 11 of the second embodiment, since the waiting time of each process on the image is shortened, an improvement in the processing speed of the entire imaging apparatus can be expected.
- the resizing process may be a process of enlarging an image.
- the case has been described in which not only the pre-processing but also the resizing process is performed on the image.
- the present invention is not limited to only the resizing process.
- the concept of the present invention can also be similarly applied to a process of changing the angle of view, such as a cutting process, or other various processes.
- the concept of the process of providing the overlapping region, the number of times to execute pre-processing, the execution order of pre-processing when the region of an image is divided, a direction in which the image sensor 100 outputs the input image data, and a direction in which pre-processing according to the image division direction is performed is the same as that in the imaging apparatus 10 of the first embodiment.
- the selection section, the second data transmission section, and the data acquisition section are provided in the image-capturing processing unit provided in the imaging apparatus, as a configuration to perform various kinds of pre-processing including a spatial filtering process on image data output from the image sensor multiple times in a range of the storage capacity of the line memory provided in the image-capturing processing unit. Then, in the embodiment of the present invention, pre-processing is directly performed on image data that is input in real time from the image sensor, and image data on which pre-processing cannot be directly performed is transmitted (written) to a storage unit (DRAM in the present embodiment) as it is by the second data transmission section.
- DRAM storage unit
- image data to be pre-processed is switched to image data stored in the storage unit, which has been acquired (read) by the data acquisition section, and pre-processing is performed.
- pre-processing on all pieces of image data of one image input from the image sensor is performed.
- the embodiment of the present invention it is possible to perform pre-processing corresponding to the image sensor in which the number of pixels has increased, without increasing the circuit size of the imaging apparatus, especially, the storage capacity of the line memory provided in the image-capturing processing unit. Therefore, in the embodiment of the present invention, it is possible to perform pre-processing corresponding to a multi-pixel image sensor with a smaller circuit size than that corresponding to an image sensor, in which the number of pixels has increased, in a known imaging apparatus. In the embodiment of the present invention, since pre-processing can be performed directly on the image data input from the image sensor, time required for the spatial filtering process can be made shorter than that in the known imaging apparatus.
- pre-processing is a process including the spatial filtering process.
- the content of the pre-processing is not limited to the embodiment of the present invention.
- the concept of the present invention can be similarly applied even if the spatial filtering process is not included in the pre-processing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
Abstract
Description
- 1. Field of the Invention
- The present invention relates to an imaging apparatus.
- Priority is claimed on Japanese Patent Application No. 2013-186471, filed Sep. 9, 2013, the content of which is incorporated herein by reference.
- 2. Description of Related Art
- In the imaging of an imaging apparatus, such as a digital camera, a process such as a spatial filtering process is performed on a signal of each pixel (hereinafter, referred to as a “pixel signal”) output from a solid-state imaging device provided in the imaging apparatus. In general, the solid-state imaging device performs a raster scan of pixels, and sequentially outputs pixel signals in a horizontal direction of the image. When performing a spatial filtering process directly on the pixel signals that are output as described above, an image-processing device that is provided in the imaging apparatus and performs the spatial filtering process needs to include a line memory that temporarily stores pixel signals in a horizontal direction (lateral direction), which are output from the solid-state imaging device, by the number of rows required for the spatial filtering process (refer to
FIGS. 9A and 9B ). - For example, as shown in
FIG. 9A , when performing the spatial filtering process on “3×3” pixels of an image, the image-processing device processes image data of “3 rows×3 columns” simultaneously. Therefore, as shown inFIG. 9B , the image-processing device needs to include two line memories in which pixel signals in the horizontal direction of the image are stored by one row.FIGS. 9A and 9B show an example of the configuration including two line memories (each for 6000 pixels) corresponding to an image having a horizontal width (image size in the horizontal direction) of 6000 pixels. - In the imaging apparatus in recent years, as the number of pixels of the solid-state imaging device has increased with an increase in the resolution of an image to be captured, the number of pixels in the horizontal direction of the solid-state imaging device has also increased every year. For this reason, in the image-processing device, the storage capacity of each line memory increases according to the corresponding number of pixels in the horizontal direction in the spatial filtering process. The increase in the storage capacity of each line memory in the image-processing device causes an increase in the circuit size of the imaging apparatus.
- For example, in the configuration shown in
FIG. 9B , a case of an image having a horizontal width of 10000 pixels is considered. In this case, it is possible to correspond to an image which has a horizontal image size of 10000 pixels by setting the storage capacity of each line memory to 10000 pixels. However, if the storage capacity of the line memory is set to 10000 pixels, the circuit size is increased by the increased amount. - Therefore, for example, Japanese Patent (Granted) Publication No. 4179701 discloses a technique of an image-processing device that performs image processing on one image a multiple of times. In the technique disclosed in Japanese Patent (Granted) Publication No. 4179701, pixel signals output from a solid-state imaging device are temporarily stored in a frame memory provided outside the image-processing device, and then image processing is performed multiple times while reading pixel signals required for processing from the frame memory, thereby generating one image. That is, in the technique disclosed in Japanese Patent (Granted) Publication No. 4179701, one image is divided into a plurality of blocks, and image processing is performed by the same number of times as the number of divided blocks, thereby generating one image.
- According to a first aspect of the present invention, an imaging apparatus is provided, including: a solid-state imaging device which has a plurality of pixels and outputs a pixel signal corresponding to a formed image of a subject as an image data; an image-capturing processing unit which outputs a pre-processed image data that is obtained by performing pre-processing on a part of the image data input from the solid-state imaging device, and an un-processed image data except the part of the image data from the image data input from the solid-state imaging device; and a storage unit which stores the pre-processed image data and the un-processed image data.
- According to a second aspect of the present invention, in the first aspect, the image-capturing processing unit may include: a pre-processing section that has a line memory, which stores the image data whose amount is smaller than a data amount of the image data in a first direction of the image, and outputs the pre-processed image data obtained by performing pre-processing on the image data of the image divided within a range of the data amount of the image data that is capable of being stored in the line memory; a first data transmission section that transmits the pre-processed image data to the storage unit; and a second data transmission section that transmits the image data in the first direction of the image that has not been pre-processed by the pre-processing section, to the storage unit as the un-processed image data.
- According to a third aspect of the present invention, in the second aspect, the image-capturing processing unit may further include: a data acquisition section that acquires image data stored in the storage unit; and a selection section that outputs the image data input from the solid-state imaging device or the image data acquired by the data acquisition section to at least one of the pre-processing section and the second data transmission section.
- According to a fourth aspect of the present invention, in the third aspect, in a first operation, the first data transmission section may transmit first pre-processed image data, which is obtained after the pre-processing section performs pre-processing on the image data of a first image obtained by dividing the image within a range of the data amount of the image data that is capable of being stored in the line memory, to the storage unit, and the second data transmission section may transmit the image data of a second image obtained by dividing the image that has not been pre-processed by the pre-processing section, as first un-processing image data, to the storage unit. In addition, in a second operation after the first operation is completed, the data acquisition section may acquire the first un-processing image data stored in the storage unit, the selection section may output the first un-processing image data to the pre-processing section, and the first data transmission section may transmit second pre-processed image data, which is obtained after the pre-processing section performs pre-processing on the first un-processing image data, to the storage unit.
- According to a fifth aspect of the present invention, in the fourth aspect, the image-capturing processing unit may further include a conversion section that performs conversion processing for converting the pre-processed image data and outputs converted image data after the conversion processing.
- According to a sixth aspect of the present invention, in the third aspect, the conversion section may include a resizing section that resizes an image corresponding to the pre-processed image data to an image, which has a size set in advance in the first direction and a second direction perpendicular to the first direction, and outputs the converted image data of the resized image.
- According to a seventh aspect of the present invention, in the sixth aspect, in the second operation, the first data transmission section may transmit second pre-processed image data, which is obtained after the pre-processing section performs pre-processing on the first un-processing image data, to the storage unit, and the second data transmission section may transmit first converted image data, which is obtained after the resizing section performs a resizing process on the second pre-processed image data, to the storage unit. In addition, in a third operation after the second operation is completed, the data acquisition section may acquire the first pre-processed image data stored in the storage unit, the selection section may output the first pre-processed image data to the resizing section, and the second data transmission section may transmit second converted image data, which is obtained after the resizing section performs a resizing process on the first pre-processed image data, to the storage unit.
-
FIG. 1 is a block diagram showing the schematic configuration of an imaging apparatus according to a first embodiment of the present invention. -
FIG. 2 is a diagram schematically showing an example of the first operation in a spatial filtering process by the imaging apparatus of the first embodiment. -
FIG. 3 is a diagram schematically showing an example of the second operation in the spatial filtering process by the imaging apparatus of the first embodiment. -
FIG. 4 is a diagram illustrating an example of an image on which a spatial filtering process is performed by the imaging apparatus of the first embodiment. -
FIG. 5 is a block diagram showing the schematic configuration of an imaging apparatus according to a second embodiment of the present invention. -
FIG. 6 is a diagram schematically showing an example of the first operation in a spatial filtering process by the imaging apparatus of the second embodiment. -
FIG. 7 is a diagram schematically showing an example of the second operation in the spatial filtering process by the imaging apparatus of the second embodiment. -
FIG. 8 is a diagram schematically showing an example of the third operation in the spatial filtering process by the imaging apparatus of the second embodiment. -
FIG. 9A is a diagram showing a spatial filtering process in a known imaging apparatus and the schematic configuration to execute the spatial filtering process. -
FIG. 9B is a diagram showing a spatial filtering process in a known imaging apparatus and the schematic configuration to execute the spatial filtering process. - Hereinafter, embodiments of the present invention will be described with reference to the accompanying diagrams.
FIG. 1 is a block diagram showing the schematic configuration of an imaging apparatus according to a first embodiment. Animaging apparatus 10 shown inFIG. 1 includes animage sensor 100, an image-capturing processing unit 200, an image-processing unit 300, a display-processing unit 400, adisplay device 401, a card interface (IF)unit 500, arecording medium 501, a dynamic random access memory (DRAM)controller 600, aDRAM 601, and aCPU 700. The image-capturing processing unit 200 of theimaging apparatus 10 includes aselection section 201, apre-processing section 202, a firstdata transmission section 203, a seconddata transmission section 204, and adata acquisition section 205. - The image-
capturing processing unit 200, the image-processing unit 300, the display-processing unit 400, thecard IF unit 500, theDRAM controller 600, and theCPU 700 in theimaging apparatus 10 are connected to each other through adata bus 800. For example, reading of data from theDRAM 601 connected to theDRAM controller 600 and writing of data into theDRAM 601 are performed by direct memory access (DMA). - The
image sensor 100 is a solid-state imaging device represented by a charge coupled device (CCD) image sensor or a complementary metal oxide semiconductor (CMOS) image sensor that photoelectrically converts an optical image of a subject that is formed by a lens (not shown). Theimage sensor 100 performs a raster scan of pixels provided in theimage sensor 100 itself, and sequentially outputs pixel signals corresponding to subject light to the image-capturing processing unit 200, as image data, in a horizontal direction of the image. - The image-
capturing processing unit 200 acquires the image data input from theimage sensor 100, performs pre-processing including a spatial filtering process on the acquired image data, and transmits (writes) image data obtained as a result of the pre-processing to theDRAM 601 through theDRAM controller 600. - The amount of image data that can be pre-processed by the image-
capturing processing unit 200, that is, the amount of data in a horizontal direction of the image is smaller than the amount of image data that theimage sensor 100 outputs as the amount of data of one row. Therefore, the image-capturing processing unit 200 transmits (writes) image data, which cannot be pre-processed, to theDRAM 601 through theDRAM controller 600 as it is, that is, without performing pre-processing. Then, the image-capturing processing unit 200 completes pre-processing on data, which can be pre-processed, and stores the pre-processed data in theDRAM 601 as it is. Then, the image-capturing processing unit 200 acquires image data that has not been pre-processed, performs pre-processing on the acquired image data, and transmits (writes) the pre-processed data to theDRAM 601 again. That is, when image data having an amount in the horizontal direction that cannot be pre-processed is input, the image-capturing processing unit 200 performs pre-processing on all pieces of the input image data by performing the pre-processing separately on the input image data divided in a range of the amount of data that can be pre-processed. - The
selection section 201 selects a component of the output destination of the input image data. Specifically, as an output destination of the image data input from theimage sensor 100 and the image data input from thedata acquisition section 205, theselection section 201 selects one or both of thepre-processing section 202 and the seconddata transmission section 204. Then, theselection section 201 outputs the input image data to the selected output destination. - The
pre-processing section 202 performs various kinds of pre-processing including a spatial filtering process, such as defect correction, shading correction, and pixel defect correction, on the image data input from theselection section 201, and outputs image data obtained as a result of the processing (hereinafter, referred to as “pre-processed image data”) to the firstdata transmission section 203. Thepre-processing section 202 is configured to include a plurality of line memories each of which stores image data, which corresponds to one row in the horizontal direction (lateral direction, first direction) of the image, of the image data input from theselection section 201, as shown inFIG. 9B . The configuration of thepre-processing section 202 is the same as the configuration of a pre-processing section provided in an imaging apparatus in the related art. The pre-processing operation of thepre-processing section 202 is also the same as the operation of the pre-processing section provided in the imaging apparatus in the related art. Therefore, detailed explanation regarding the configuration of thepre-processing section 202 or the pre-processing operation of thepre-processing section 202 will be omitted. - The first
data transmission section 203 stores (writes) the pre-processed image data input from thepre-processing section 202 in theDRAM 601 connected to theDRAM controller 600 by the DMA, for example. Since the configuration of the firstdata transmission section 203 or the operation when storing (writing) the pre-processed image data in theDRAM 601 by the DMA is the same as the normal configuration or operation to store data by the DMA, detailed explanation thereof will be omitted. The method used when the firstdata transmission section 203 stores (writes) the pre-processed image data in theDRAM 601 may be a method other than the DMA. - The second
data transmission section 204 stores (writes) the image data input from theselection section 201, that is, image data that has not been pre-processed by the pre-processing section 202 (hereinafter, referred to as “original image data”) in theDRAM 601 connected to theDRAM controller 600 by the DMA, for example. Similar to the firstdata transmission section 203, the configuration of the seconddata transmission section 204 or the operation when storing (writing) the original image data in theDRAM 601 by the DMA is the same as the normal configuration or operation to store data by the DMA. Accordingly, detailed explanation thereof will be omitted. Similar to the firstdata transmission section 203, the method used when the seconddata transmission section 204 stores (writes) the original image data in theDRAM 601 may be a method other than the DMA. - The
data acquisition section 205 acquires (reads) the image data stored in theDRAM 601 through theDRAM controller 600 by the DMA, for example, and outputs the acquired image data to theselection section 201. The image data acquired by thedata acquisition section 205 may not only be the pre-processed image data stored by the firstdata transmission section 203 or the original image data stored by the seconddata transmission section 204 but also image data stored in theDRAM 601 after image processing by the image-processing unit 300, for example. - The image-
processing unit 300 acquires (reads) the pre-processed image data stored in theDRAM 601, and generates image data for display or image data for recording by performing various kinds of image processing, such as noise removal processing, distortion aberration correction, YC conversion processing, resizing process, JPEG compression processing, and video compression processing (for example, MPEG compression processing or H.264 compression processing), on the acquired pre-processed image data. Then, the image-processing unit 300 stores (writes) the generated image data for display or image data for recording in theDRAM 601 again. - In addition, the image-
processing unit 300 acquires (reads) the image data for recording stored in theDRAM 601, and generates image data for display by performing various kinds of image processing, such as JPEG decompression processing and video decompression processing (for example, MPEG decompression processing or H.264 decompression processing). Then, the image-processing unit 300 stores (writes) the generated image data for display in theDRAM 601 again. - The display-
processing unit 400 acquires (reads) the image data for display stored in theDRAM 601, and performs display processing, such as a process of superimposing data for on-screen display (OSD) on the acquired image data for display. Then, the image data after display processing is output to thedisplay device 401. - The
display device 401 is a display device, such as a thin film transistor (TFT) liquid crystal display (LCD) or an organic electroluminescence (EL) display, and displays an image corresponding to the image data after display processing that is output from the display-processing unit 400. - The
display device 401 may be a display device, such as an electronic view finder (EVF), or an external display, such as a television. Although thedisplay device 401 is also a component of theimaging apparatus 10 inFIG. 1 , thedisplay device 401 may be configured to be detachable from theimaging apparatus 10. - The card IF
unit 500 acquires (reads) the image data for recording that is recorded in theDRAM 601, and records the acquired image data for recording on therecording medium 501. In addition, the card IFunit 500 reads the image data recorded on therecording medium 501, and transmits (writes) the read image data to theDRAM 601. - The
recording medium 501 is a recording medium, such as an SD memory card or a compact flash (CF: registered trademark), and records the image data for recording that is output from the card IFunit 500. In addition, the image data recorded by the card IFunit 500 is read. Although therecording medium 501 is also a component of theimaging apparatus 10 inFIG. 1 , therecording medium 501 may be configured to be detachable from theimaging apparatus 10. - The
DRAM controller 600 controls the storage (writing) of data into theconnected DRAM 601 and the acquisition (reading) of data from theDRAM 601 in response to an access request to theDRAM 601 from a plurality of components in theimaging apparatus 10 connected to thedata bus 800, for example, in response to the access request of DMA. - The
DRAM 601 is a memory (storage unit) whose access control is performed by theDRAM controller 600. TheDRAM 601 temporarily stores (memorizes) various kinds of data in the process of each component of theimaging apparatus 10. - The
CPU 700 controls the components of theimaging apparatus 10, that is, theentire imaging apparatus 10. - For example, the
CPU 700 controls the operation of each component in theimaging apparatus 10 according to an imaging operation or a reproduction operation in theimaging apparatus 10. For example, theCPU 700 controls the start of the output of the image data from theimage sensor 100 or the start of the acquisition of the image data by the image-capturingprocessing unit 200 when theimaging apparatus 10 performs an imaging operation. - In addition, the
CPU 700 controls the selection of the output destination of image data by theselection section 201 of the image-capturingprocessing unit 200, or performs a setting of transmission of image data by the first and seconddata transmission sections data acquisition section 205. - With such a configuration, in the
imaging apparatus 10 of the first embodiment, it is possible to acquire the image data output from theimage sensor 100 in real time while performing various kinds of pre-processing including a spatial filtering process directly on the acquired image data. In addition, in theimaging apparatus 10 of the first embodiment, various kinds of pre-processing including a spatial filtering process can be performed on the image data stored in theDRAM 601 by acquiring the image data stored in theDRAM 601 by thedata acquisition section 205. That is, in theimaging apparatus 10 of the first embodiment, various kinds of pre-processing including a spatial filtering process can also be performed on image data other than the image data that is input in real time. - Next, an example of the operation when pre-processing (spatial filtering process) on the image data output from the
image sensor 100 is performed in theimaging apparatus 10 will be described. In the following explanation, a case will be described in which the number of pixels in the horizontal direction of theimage sensor 100 is 10000, that is, theimage sensor 100 captures an image having a horizontal width of 10000 pixels and thepre-processing section 202 includes a line memory having a storage capacity of 6000 pixels in the horizontal direction. That is, the following explanation will be given for an operation relevant to how to perform pre-processing on the input image data of one row when the amount of image data in the horizontal direction that theimage sensor 100 outputs for each row (hereinafter, referred to as “input image data”) is larger than the amount of image data of one row that can be stored in a line memory provided in thepre-processing section 202, and accordingly, all pieces of the input image data of one row cannot be stored in the line memory. - When the amount of input image data is larger than the amount of data that can be stored in a line memory provided in the
pre-processing section 202, that is, when the amount of input image data that can be pre-processed by the image-capturingprocessing unit 200 is smaller than the amount of input image data that theimage sensor 100 outputs for each row, theimaging apparatus 10 performs pre-processing on all pieces of the input image data by performing the pre-processing separately on the input image data divided in a range of the amount of data that can be pre-processed (here, 6000 pixels) as described above. In order to facilitate the explanation, the following explanation will be given for a case where input image data, which is input from theimage sensor 100, is halved and pre-processing is performed separately on each of the divided halves (that is, twice). -
FIGS. 2 and 3 are diagrams schematically showing an example of the pre-processing operation of theimaging apparatus 10 of the first embodiment.FIG. 2 shows an example of the first operation when pre-processingsection 202 performs pre-processing (spatial filtering process) on the input image data of one half (left side) that is input from theimage sensor 100.FIG. 3 shows an example of the second operation when pre-processingsection 202 performs pre-processing (spatial filtering process) on the input image data of the other half (right side) that is input from theimage sensor 100. - In the first operation (operation of the first pass), pre-processed image data obtained by performing pre-processing in real time on input image data, which is output from the
image sensor 100, is transmitted (written) to theDRAM 601, and input image data that cannot be pre-processed is transmitted (written) to theDRAM 601 as it is.FIG. 2 shows the path of each piece of data in the first operation (operation of the first pass) on the block diagram of theimaging apparatus 10 shown inFIG. 1 . - First, in the operation of the first pass, in the path of image data of a path C11 shown in
FIG. 2 , pre-processed image data obtained by performing pre-processing on the input image data in real time is transmitted (written) to theDRAM 601. More specifically, in the operation of the first pass, theimage sensor 100 captures an image P1 having a horizontal width of 10000 pixels, and outputs the input image data of the captured image P1 to the image-capturingprocessing unit 200. In the image-capturingprocessing unit 200, input image data in all regions of the image P1 input from theimage sensor 100 is acquired, and the acquired input image data of the image P1 is sequentially input to theselection section 201 of the image-capturingprocessing unit 200. Then, theselection section 201 outputs the input image data in all regions of the input image P1 sequentially to both thepre-processing section 202 and the seconddata transmission section 204. - The
pre-processing section 202 performs pre-processing on input image data of an image P2, which is a region of the left half, of the input image data of the image P1 that is sequentially input from theselection section 201, and sequentially outputs pre-processed image data of the image P2 after pre-processing to the firstdata transmission section 203. Then, the firstdata transmission section 203 transmits (writes) the pre-processed image data of the image P2, which is sequentially input from thepre-processing section 202, to theDRAM 601 through theDRAM controller 600. As a result, the pre-processed image data of the image P2 after pre-processing is stored in theDRAM 601. - In addition, in the operation of the first pass, in the path of image data of a path C12 shown in
FIG. 2 , input image data is transmitted (written) to theDRAM 601 as it is. More specifically, in the operation of the first pass, the seconddata transmission section 204 transmits (writes) input image data of an image P3 that is a region of the right half, of the input image data of the image P1 that is sequentially input from theselection section 201, to theDRAM 601 through theDRAM controller 600 as it is as original image data of the image P3. As a result, the input image data of the image P3 of the right half of the image P1 output from theimage sensor 100 is stored in theDRAM 601 as it is. - Thus, in the operation of the first pass, pre-processed image data of the image P2 obtained by performing pre-processing on the region of the left half of the image P1 output from the
image sensor 100 and the original image data of the image P3, which is a region of the right half of the image P1 output from theimage sensor 100, are stored in theDRAM 601. - In the operation of the first pass, when storing the original image data in the
DRAM 601, input image data in all regions of the image P1 that is input from theimage sensor 100 can also be stored in theDRAM 601 as original image data of the image P1. However, as described above, in the operation of the first pass, only the input image data of the image P3 of the right half of the image P1 is stored in theDRAM 601 as original image data. This is because pre-processing in a subsequent second operation (operation of the second pass) is performed on only the input image data of the image P3, which is a region of the right half of the image P1 that could not be pre-processed in the operation of the first pass. Therefore, it is possible to reduce the storage capacity of theDRAM 601 and to avoid having to compress the bus band of thedata bus 800 than necessary. - Then, the
imaging apparatus 10 transmits (writes) to theDRAM 601 pre-processed image data obtained by performing pre-processing on the original image data of the image P3 that is the input image data of the right half of the image P1 input from theimage sensor 100, which could not be pre-processed in the first operation (operation of the first pass).FIG. 3 shows the path of data in the second operation (operation of the second pass) on the block diagram of theimaging apparatus 10 shown inFIG. 1 . - In the operation of the second pass, in the path of image data of a path C13 shown in
FIG. 3 , pre-processed image data obtained by performing pre-processing on the original image data of the image P3 stored in theDRAM 601 is transmitted (written) to theDRAM 601. More specifically, in the operation of the second pass, thedata acquisition section 205 sequentially acquires (reads) the original image data of the image P3 stored in theDRAM 601, and sequentially outputs the acquired original image data of the image P3 to theselection section 201. Then, theselection section 201 sequentially outputs the input original image data of the image P3 to thepre-processing section 202. - The
pre-processing section 202 performs pre-processing on the original image data of the image P3 that is sequentially input from theselection section 201, and sequentially outputs pre-processed image data of an image P4 after pre-processing to the firstdata transmission section 203. Then, the firstdata transmission section 203 transmits (writes) the pre-processed image data of the image P4, which is sequentially input from thepre-processing section 202, to theDRAM 601 through theDRAM controller 600. As a result, the pre-processed image data of the image P4 after pre-processing is stored in theDRAM 601. - Thus, in the operation of the second pass, pre-processed image data of the image P4 obtained by performing pre-processing on the image P3 of the right half of the image P1 output from the
image sensor 100, which is stored in theDRAM 601, is stored in theDRAM 601. In this manner, pre-processed image data (pre-processed image data of the images P2 and P4) obtained by performing pre-processing on the input image data in all regions of one image P1 having a horizontal width of 10000 pixels, which is captured by theimage sensor 100, is stored in theDRAM 601. - As described above, in the
imaging apparatus 10 of the first embodiment, even when the amount of input image data in the horizontal direction that theimage sensor 100 outputs for each row is larger than the amount of image data of one row that can be stored in a line memory provided in thepre-processing section 202 of the image-capturingprocessing unit 200, pre-processing on all pieces of the input image data of one row can be performed by two operations of the first and second operations. More specifically, in the first operation (operation of the first pass), theimaging apparatus 10 transmits (writes) pre-processed image data, which is obtained by performing pre-processing in real time on the input image data of the left half of the image P1 input from theimage sensor 100, and also transmits (writes) the input image data of the right half, which could not be pre-processed, to theDRAM 601 as it is. Then, in the second operation (operation of the second pass), theimaging apparatus 10 performs pre-processing on all pieces of the input image data of one image P1 by transmitting (writing) to theDRAM 601 pre-processed image data obtained by performing pre-processing on the input image data of the right half of the image P1 input from theimage sensor 100, which could not be pre-processed in the first operation (operation of the first pass). - Thus, in the
imaging apparatus 10 of the first embodiment, even when the amount of image data of one row that can be stored in a line memory provided in thepre-processing section 202 of the image-capturingprocessing unit 200 is smaller than the amount of image data that theimage sensor 100 outputs for each row, it is possible to perform pre-processing on all pieces of image data output from theimage sensor 100. That is, it is possible to perform pre-processing corresponding to theimage sensor 100 in which the number of pixels has increased, without increasing the circuit size of theimaging apparatus 10, especially, the storage capacity of the line memory. In other words, it is possible to meet an increase in the number of pixels of the image sensor with a small circuit size. - In the example of the pre-processing operation of the
imaging apparatus 10 of the first embodiment shown inFIGS. 2 and 3 , a case has been described in which the input image data that is input from theimage sensor 100 is halved and pre-processing is performed on each of the divided halves (that is, twice). In the actual spatial filtering process, however, it is thought that an incongruous image is obtained in a boundary where the process is divided if an image is divided in one place to perform the process. For this reason, when performing a spatial filtering process multiple times, a region where image data used in each process overlap, that is, an overlapping region is generally provided. In the pre-processing of theimaging apparatus 10 of the first embodiment, it is possible to perform the processing similarly by providing the overlapping region. -
FIG. 4 is a diagram illustrating an example of an image on which a spatial filtering process is performed by theimaging apparatus 10 of the first embodiment.FIG. 4 shows an example of the case where an overlapping region is provided in a region of the left half, which is processed in the first operation (operation of the first pass), and a region of the right half, which is processed in the second operation (operation of the second pass), at a position where the entire region of the image P1 is divided, when dividing the input image data of the entire region of the image P1 input from theimage sensor 100 in the pre-processing operation of theimaging apparatus 10 shown inFIGS. 2 and 3 . Thus, by providing an overlapping region when performing the spatial filtering process multiple times, it is possible to eliminate the inconvenience of having the image in a boundary where the process is divided. - In the example of the operation of the pre-processing by the
imaging apparatus 10 of the first embodiment shown inFIGS. 2 and 3 , a case has been described in which pre-processing on the input image data of the region of the left half of the image P1 input from theimage sensor 100 is performed in the first operation (operation of the first pass), and pre-processing on the input image data of the region (image P3) of the right half of the image P1 input from theimage sensor 100 is performed in the second operation (operation of the second pass). However, the order of regions of the image to be pre-processed in the first operation (operation of the first pass) and the second operation (operation of the second pass) is not limited to that shown inFIGS. 2 and 3 . For example, pre-processing on the input image data of the region of the right half of the image P1 may be performed in the first operation (operation of the first pass), and pre-processing on the input image data of the region of the left half of the image P1 may be performed in the second operation (operation of the second pass). - In the example of the operation of the pre-processing by the
imaging apparatus 10 of the first embodiment shown inFIGS. 2 and 3 , the case has been described in which the input image data that is input from theimage sensor 100 is halved and pre-processing is performed on each of the divided halves (that is, twice). However, the number of division times of pre-processing (spatial filtering process) performed on the input image data of one image is not limited to twice as described above, and may be 3 times or more. An operation when the number of division times of pre-processing (spatial filtering process) is different can be considered in the same manner as the operation when the pre-processing (spatial filtering process) is performed twice (operation of the pre-processing by theimaging apparatus 10 of the first embodiment shown inFIGS. 2 and 3 ). Accordingly, detailed explanation thereof will be omitted. - The operation of the pre-processing in the
imaging apparatus 10 of the first embodiment does not necessarily need to be performed twice or more. For example, when the number of pixels in the horizontal direction of the image sensor mounted in the imaging apparatus is smaller than the storage capacity (6000 pixels in the first embodiment) of the line memory provided in thepre-processing section 202 of the image-capturingprocessing unit 200, that is, when the amount of input image data is smaller than the amount of data that can be pre-processed by thepre-processing section 202, thepre-processing section 202 can perform pre-processing only once. Accordingly, the pre-processing does not necessarily need to be performed twice or more. Thus, the number of times by which thepre-processing section 202 performs pre-processing can be changed according to the number of pixels in the horizontal direction of the image sensor mounted in the imaging apparatus. - In general, an image sensor can be driven in a plurality of driving modes, such as a driving mode in which all pixel signals are output (hereinafter, referred to as an “all pixel output mode”) and a driving mode in which pixel signals are added and output (hereinafter, referred to as a “pixel addition mode”). The all pixel output mode is set when capturing a high-resolution image, such as a still image, and the pixel addition mode is set when performing video imaging or imaging of a so-called live view function of displaying a video for checking a subject to be imaged on a display device. Therefore, for example, when the
image sensor 100 is driven in the pixel addition mode and an image having 5000 pixels, which is ½ of the original horizontal width, is output due to the addition of pixel signals in the horizontal direction, the amount of image data that theimage sensor 100 outputs for one row is smaller than the amount of image data (6000 pixels) of one row that can be stored in the line memory provided in thepre-processing section 202 of the image-capturingprocessing unit 200. Also in this case, thepre-processing section 202 can perform pre-processing only once. Therefore, the pre-processing does not necessarily need to be performed twice or more. Thus, based on the driving mode of the image sensor mounted in the imaging apparatus, pre-processing is performed twice when the image sensor is driven in the all pixel output mode, that is, in the mode in which a still image is captured, and pre-processing is performed once in real time when the image sensor is driven in the pixel addition mode, that is, in the mode in which video imaging or live view imaging is performed, for example. Therefore, the number of times by which thepre-processing section 202 performs pre-processing can be changed according to an image to be captured. - In the first embodiment, the case has been described in which the
image sensor 100 outputs a pixel signal, which is obtained by performing a raster scan in the horizontal direction of the image, as image data. However, for example, when the image sensor outputs a pixel signal, which is obtained by performing a raster scan in a vertical direction of the image, as image data, the concept of the present invention can be similarly applied by regarding the horizontal direction in the first embodiment as a vertical direction, that is, by regarding the horizontal direction and the vertical direction of the image reversely. In the first embodiment, the case has been described in which pre-processing is performed multiple times by horizontally dividing the image input from theimage sensor 100. However, even if pre-processing is performed multiple times by vertically dividing the image input from theimage sensor 100, the concept of the present invention can be similarly applied by regarding the horizontal direction and the vertical direction of the image reversely. - In the imaging apparatus, the image-capturing processing unit may perform various processes other than the pre-processing. For example, a conversion section that generates a converted image by converting a normal image may be provided in addition to the configuration to record a normal image, which has the number of pixels that is almost the same as the number of effective pixels of the image sensor, as a still image. For example, it is possible to provide a resizing section that generates a reduced image having the same angle of view as a converted image with the number of pixels smaller than the normal number of pixels. For example, the reduced image generated by the resizing section (hereinafter, referred to as a “reduced image”) may be displayed on a display device so that the captured image can be immediately checked, or may be used as an image recorded together with a normal still image (so-called thumbnail) or an image for list display of captured images. That is, various applications can be considered.
- Next, an imaging apparatus of a second embodiment will be described. In the imaging apparatus of the second embodiment, only the image-capturing
processing unit 200 provided in theimaging apparatus 10 of the first embodiment shown inFIG. 1 is different. Therefore, in the explanation of the imaging apparatus of the second embodiment, the same reference numerals are given to the same components as in theimaging apparatus 10 of the first embodiment. In addition, only components and operations different from those in theimaging apparatus 10 of the first embodiment will be described, and detailed explanation regarding the same components and operations as in theimaging apparatus 10 of the first embodiment will be omitted. -
FIG. 5 is a block diagram showing the schematic configuration of the imaging apparatus according to the second embodiment. Animaging apparatus 11 shown inFIG. 5 includes animage sensor 100, an image-capturingprocessing unit 210, an image-processing unit 300, a display-processing unit 400, adisplay device 401, a card IFunit 500, arecording medium 501, aDRAM controller 600, aDRAM 601, and aCPU 700. The image-capturingprocessing unit 210 of theimaging apparatus 11 includes aselection section 201, apre-processing section 202, aselection section 216, a firstdata transmission section 203, aresizing section 217, a seconddata transmission section 204, and adata acquisition section 205. - The image-capturing
processing unit 210, the image-processing unit 300, the display-processing unit 400, the card IFunit 500, theDRAM controller 600, and theCPU 700 in theimaging apparatus 11 are connected to each other through adata bus 800, in the same manner as in theimaging apparatus 10 of the first embodiment shown in FIG. 1. For example, reading of data from theDRAM 601 connected to theDRAM controller 600 and writing of data into theDRAM 601 are performed by the DMA. - The
image sensor 100 performs a raster scan of pixels provided in theimage sensor 100 itself, and sequentially outputs pixel signals corresponding to subject light to the image-capturingprocessing unit 210, as image data, in a horizontal direction of the image. - Similar to the image-capturing
processing unit 200 provided in theimaging apparatus 10 of the first embodiment, the image-capturingprocessing unit 210 acquires input image data that is input from theimage sensor 100, and transmits (writes) pre-processed image data, which is obtained by performing pre-processing including a spatial filtering process on the acquired input image data, and input image data, which cannot be pre-processed, to theDRAM 601 through theDRAM controller 600. - Also in the image-capturing
processing unit 210, similar to the image-capturingprocessing unit 200 in the first embodiment, the amount of image data that can be pre-processed by the image-capturingprocessing unit 210, that is, the amount of data in a horizontal direction of the image is smaller than the amount of input image data that theimage sensor 100 outputs as the amount of data of one row. Therefore, when input image data having an amount in the horizontal direction that cannot be pre-processed is input, the image-capturingprocessing unit 210 also performs pre-processing on all pieces of the input image data by performing the pre-processing separately on the input image data divided in a range of the amount of data that can be pre-processed, in the same manner as in the image-capturingprocessing unit 200 in the first embodiment. - In addition, the image-capturing
processing unit 210 generates a reduced image by performing a reduction process, such as a resizing process or a cutting process, based on the pre-processed image data, and transmits (writes) the generated reduced image to theDRAM 601 through theDRAM controller 600. In the following explanation, the image-capturingprocessing unit 210 functions as a processing unit that generates a reduced image by performing a resizing process. - The
selection section 201 selects one or both of thepre-processing section 202 and theselection section 216 as an output destination of the input image data, which is input from theimage sensor 100, and the image data input from thedata acquisition section 205, and outputs the input image data to the selected output destination. - The
pre-processing section 202 outputs pre-processed image data, which is obtained by performing various kinds of pre-processing including a spatial filtering process on the image data input from theselection section 201, to theselection section 216. - The
selection section 216 selects a component of the output destination of the input image data. More specifically, theselection section 216 performs conversion processing for converting the pre-processed image data input from thepre-processing section 202, and outputs converted image data. Theselection section 216 outputs the image data input from theselection section 201. In this case, theselection section 216 selects at least one of the firstdata transmission section 203 and theresizing section 217. Then, theselection section 216 outputs the input image data to the selected output destination. - The first
data transmission section 203 stores (writes) the image data input from theselection section 216 in theDRAM 601 connected to theDRAM controller 600 by the DMA, for example. - The resizing
section 217 performs a resizing process on the image data input from theselection section 216, and outputs image data of a reduced image (hereinafter, referred to as “reduced image data”), which is obtained as a result of resizing (here, reducing) the pre-processed image in a horizontal direction (lateral direction, first direction) and a vertical direction (longitudinal direction, second direction), to the seconddata transmission section 204. The resizingsection 217 can also output the image data (pre-processed image data, input image data, or original image data), which is input from theselection section 216, to the seconddata transmission section 204 as it is without performing a resizing process. That is, the resizingsection 217 can turn off the function of the resizing process. - The second
data transmission section 204 stores (writes) the image data input from the resizingsection 217, that is, reduced image data or image data input to theresizing section 217 in theDRAM 601 connected to theDRAM controller 600 by the DMA, for example. - When the reduced image data is stored in the
DRAM 601, the image-processing unit 300 performs image processing on both the pre-processed image data and the reduced image data. For example, when the reduced image data is a thumbnail, the thumbnail is associated when performing a JPEG compression process. - The
CPU 700 controls the components of theimaging apparatus 11, that is, theentire imaging apparatus 11. For example, theCPU 700 controls the operation of each component in theimaging apparatus 11 according to an imaging operation or a reproduction operation in theimaging apparatus 11. For example, theCPU 700 controls the start of the output of the image data from theimage sensor 100 or the start of the acquisition of the image data by the image-capturingprocessing unit 210 when theimaging apparatus 11 performs an imaging operation. - In addition, the
CPU 700 controls the operation of each component in the image-capturingprocessing unit 210 in each operation after the start of the acquisition of the image data by the image-capturingprocessing unit 210. For example, theCPU 700 controls the selection of the output destination of image data by theselection sections processing unit 210, or performs a setting of the resizing process by the resizingsection 217, a setting of the transmission of image data by the first and seconddata transmission sections data acquisition section 205. - With such a configuration, also in the
imaging apparatus 11 of the second embodiment, similar to theimaging apparatus 10 of the first embodiment, it is possible to perform various kinds of pre-processing including a direct spatial filtering process on image data input in real time and various kinds of pre-processing including a spatial filtering process on the image data stored in theDRAM 601. In theimaging apparatus 11 of the second embodiment, it is possible to generate a reduced image by performing a resizing process on the pre-processed image data in addition to various kinds of pre-processing including a spatial filtering process. - Next, an example of the operation when pre-processing (spatial filtering process) on the image data output from the
image sensor 100 is performed and the operation of generating a reduced image in theimaging apparatus 11 will be described. In the following explanation, similar to theimaging apparatus 10 of the first embodiment, the number of pixels in the horizontal direction of theimage sensor 100 is 10000, and thepre-processing section 202 includes a line memory having a storage capacity of 6000 pixels in the horizontal direction. That is, the following explanation will be given for an operation relevant to how to perform pre-processing on the input image data of one row or how to generate a reduced image when the amount of input image data in the horizontal direction that theimage sensor 100 outputs for each row is larger than the amount of image data of one row that can be stored in a line memory provided in thepre-processing section 202, and accordingly, all pieces of the input image data of one row cannot be stored in the line memory. - As in the
imaging apparatus 10 of the first embodiment, in order to facilitate the explanation, the following explanation will be given for a case where input image data, which is input from theimage sensor 100, is halved and pre-processing is performed separately on each of the divided halves (that is, twice). -
FIGS. 6 to 8 are diagrams schematically showing an example of the operation of pre-processing and reduced image generation by theimaging apparatus 11 of the second embodiment.FIG. 6 shows an example of the first operation when theimaging apparatus 11 performs pre-processing (spatial filtering process) on the input image data of one half (left side) that is input from theimage sensor 100.FIG. 7 shows an example of the second operation when theimaging apparatus 11 performs pre-processing (spatial filtering process) on the input image data of the other half (right side) that is input from theimage sensor 100 while generating a reduced image corresponding to the pre-processed image data of the other half (right side).FIG. 8 shows an example of the third operation when theimaging apparatus 11 generates a reduced image corresponding to the pre-processed image data of one half (left side). - In the first operation (operation of the first pass), similar to the
imaging apparatus 10 of the first embodiment, pre-processed image data obtained by performing pre-processing in real time on input image data, which is output from theimage sensor 100, is transmitted (written) to theDRAM 601, and input image data that cannot be pre-processed is transmitted (written) to theDRAM 601 as it is.FIG. 6 shows the path of each piece of data in the first operation (operation of the first pass) on the block diagram of theimaging apparatus 11 shown inFIG. 5 . - First, in the operation of the first pass, in the path of image data of a path C21 shown in
FIG. 6 , pre-processed image data obtained by performing pre-processing on the input image data in real time is transmitted (written) to theDRAM 601. More specifically, in the operation of the first pass, theimage sensor 100 captures an image P1 having a horizontal width of 10000 pixels, and outputs the input image data of the captured image P1 to the image-capturingprocessing unit 210. In the image-capturingprocessing unit 210, input image data in all regions of the image P1 input from theimage sensor 100 is acquired, and the acquired input image data of the image P1 is sequentially input to theselection section 201 of the image-capturingprocessing unit 210. Then, theselection section 201 outputs the input image data in all regions of the input image P1 sequentially to both thepre-processing section 202 and theselection section 216. - The
pre-processing section 202 performs pre-processing on input image data of the image P2, which is a region of the left half, of the input image data of the image P1 that is sequentially input from theselection section 201, and sequentially outputs the pre-processed image data of the image P2 after pre-processing to theselection section 216. Theselection section 216 outputs the pre-processed image data of the image P2, which is input from thepre-processing section 202, to the firstdata transmission section 203. Then, the firstdata transmission section 203 transmits (writes) the pre-processed image data of the image P2, which is sequentially input from theselection section 216, to theDRAM 601 through theDRAM controller 600. As a result, the pre-processed image data of the image P2 after pre-processing is stored in theDRAM 601. - In addition, in the operation of the first pass, in the path of image data of a path C22 shown in
FIG. 6 , input image data is transmitted (written) to theDRAM 601 as it is. More specifically, in the operation of the first pass, theselection section 216 sequentially outputs the input image data of the image P1, which is sequentially input from theselection section 201, to theresizing section 217. In the operation of the first pass, the function of the resizing process in theresizing section 217 is turned off. Then, the resizingsection 217 outputs input image data of the image P3, which is a region of the right half, of the input image data of the image P1 that is sequentially input from theselection section 216, to the seconddata transmission section 204 as it is. Then, the seconddata transmission section 204 transmits (writes) the input image data of the image P3 input sequentially from the resizingsection 217, as it is as original image data of the image P3, to theDRAM 601 through theDRAM controller 600. As a result, the input image data of the image P3 of the right half of the image P1 output from theimage sensor 100 is stored in theDRAM 601 as it is. - Thus, in the operation of the first pass, pre-processed image data of the image P2 obtained by performing pre-processing on the region of the left half of the image P1 output from the
image sensor 100 and the original image data of the image P3, which is a region of the right half of the image P1 output from theimage sensor 100, are stored in theDRAM 601. - Then, the
imaging apparatus 11 transmits (writes) to theDRAM 601 pre-processed image data obtained by performing pre-processing on the original image data of the image P3 that is the input image data of the right half of the image P1 input from theimage sensor 100, which could not be pre-processed in the first operation (operation of the first pass), and transmits (writes) to theDRAM 601 reduced image data obtained by performing a resizing process on the pre-processed image data of the image P3.FIG. 7 shows the path of data in the second operation (operation of the second pass) on the block diagram of theimaging apparatus 11 shown inFIG. 5 . - First, in the operation of the second pass, in the path of image data of a path C23 shown in
FIG. 7 , pre-processed image data obtained by performing pre-processing on the original image data of the image P3 stored in theDRAM 601 is transmitted (written) to theDRAM 601. More specifically, in the operation of the second pass, thedata acquisition section 205 sequentially acquires (reads) the original image data of the image P3 stored in theDRAM 601, and sequentially outputs the acquired original image data of the image P3 to theselection section 201. Then, theselection section 201 sequentially outputs the input original image data of the image P3 to thepre-processing section 202. - The
pre-processing section 202 performs pre-processing on the original image data of the image P3 input sequentially from theselection section 201, and sequentially outputs the pre-processed image data of the image P4 after pre-processing to theselection section 216. Theselection section 216 sequentially outputs the pre-processed image data of the image P4, which is input from thepre-processing section 202, to both the firstdata transmission section 203 and theresizing section 217. Then, the firstdata transmission section 203 transmits (writes) the pre-processed image data of the image P4, which is sequentially input from theselection section 216, to theDRAM 601 through theDRAM controller 600. As a result, the pre-processed image data of the image P4 after pre-processing is stored in theDRAM 601. - In the operation of the second pass, in the path of image data of a path C24 shown in
FIG. 7 , a reduced image is generated by performing a resizing process on the original image data of the image P3 stored in theDRAM 601, and transmits (writes) the reduced image data of the generated reduced image to theDRAM 601. More specifically, in the operation of the second pass, the function of the resizing process in theresizing section 217 is turned on. Then, the resizingsection 217 generates a reduced image by performing a resizing process on the pre-processed image data of the image P4 input sequentially from theselection section 216, and outputs the reduced image data of a generated reduced image P5 to the seconddata transmission section 204. Then, the seconddata transmission section 204 transmits (writes) the reduced image data of the reduced image P5, which is sequentially input from the resizingsection 217, to theDRAM 601 through theDRAM controller 600. As a result, the reduced image data of the reduced image P5 of the right half of the image P1 output from theimage sensor 100 is stored in theDRAM 601. - Thus, in the operation of the second pass, pre-processed image data of the image P4 obtained by performing pre-processing on the image P3 of the right half of the image P1 output from the
image sensor 100, which is stored in theDRAM 601, and reduced image data of the reduced image P5 obtained by performing a resizing process on the image P4 after pre-processing of the right half of the image P1, which is stored in theDRAM 601, are stored in theDRAM 601. In this manner, pre-processed image data (pre-processed image data of the images P2 and P4) obtained by performing pre-processing on the input image data in all regions of one image P1 having a horizontal width of 10000 pixels, which is captured by theimage sensor 100, is stored in theDRAM 601. - Then, in order to generate a reduced image corresponding to the input image data of the left half of the image P1 input from the
image sensor 100 on which a reduced image has not been generated, theimaging apparatus 11 transmits (writes) reduced image data obtained by performing a resizing process on the pre-processed image data of the image P2, which is pre-processed image data obtained by performing pre-processing on the input image data of the left half of the image P1, to theDRAM 601.FIG. 8 shows the path of data in the third operation (operation of the third pass) on the block diagram of theimaging apparatus 11 shown inFIG. 5 . - In the operation of the third pass, in the path of image data of a path C25 shown in
FIG. 8 , a reduced image is generated by performing a resizing process on the pre-processed image data of the image P2 stored in theDRAM 601, and transmits (writes) the reduced image data of the generated reduced image to theDRAM 601. - More specifically, in the operation of the third pass, the
data acquisition section 205 sequentially acquires (reads) the pre-processed image data of the image P2 stored in theDRAM 601, and sequentially outputs the acquired pre-processed image data of the image P2 to theselection section 201. Then, theselection section 201 sequentially outputs the input pre-processed image data of the image P2 to theselection section 216. - The
selection section 216 sequentially outputs the pre-processed image data of the image P2, which is sequentially input from theselection section 201, to theresizing section 217. In the operation of the third pass, the function of the resizing process in theresizing section 217 is turned on. Then, the resizingsection 217 generates a reduced image by performing a resizing process on the pre-processed image data of the image P2 input sequentially from theselection section 216, and outputs the reduced image data of a generated reduced image P6 to the seconddata transmission section 204. Then, the seconddata transmission section 204 transmits (writes) the reduced image data of the reduced image P6, which is sequentially input from the resizingsection 217, to theDRAM 601 through theDRAM controller 600. As a result, the reduced image data of the reduced image P6 of the left half of the image P1 output from theimage sensor 100 is stored in theDRAM 601. - Thus, in the operation of the third pass, reduced image data of the reduced image P6 obtained by performing a resizing process on the image P2 after pre-processing of the left half of the image P1 output from the
image sensor 100, which is stored in theDRAM 601, is stored in theDRAM 601. In this manner, reduced image data (reduced image data of the images P5 and P6) obtained by performing a resizing process after performing pre-processing on the input image data in all regions of one image P1 having a horizontal width of 10000 pixels, which is captured by theimage sensor 100, is stored in theDRAM 601. - As described above, also in the
imaging apparatus 11 of the second embodiment, similar to theimaging apparatus 10 of the first embodiment, even when the amount of input image data in the horizontal direction that theimage sensor 100 outputs for each row is larger than the amount of image data of one row that can be stored in a line memory provided in thepre-processing section 202 of the image-capturingprocessing unit 210, pre-processing on all pieces of the input image data of one row can be performed by multiple operations. Therefore, also in theimaging apparatus 11 of the second embodiment, it is possible to achieve the same effect as in theimaging apparatus 10 of the first embodiment. - In the
imaging apparatus 11 of the second embodiment, a reduced image corresponding to an image after pre-processing can be generated by multiple operations, in addition to the pre-processing on the image input from theimage sensor 100. More specifically, in the second operation (operation of the second pass), theimaging apparatus 11 performs pre-processing on the image P3 of the right half of the image P1 input from theimage sensor 100, and generates the reduced image P5 by performing a resizing process on the pre-processed image data of the image P3 after pre-processing. Then, in the third operation (operation of the third pass), theimaging apparatus 11 generates the reduced image P6 by performing a resizing process on the pre-processed image data of the image P2 obtained by performing pre-processing on the image of the left half of the image P1 input from theimage sensor 100 in the first operation (operation of the first pass). - Thus, in the
imaging apparatus 11 of the second embodiment, even if pre-processing on the image data in all regions of one image has not been completed, the next processing (resizing process for reduction in the second embodiment) can be performed from the region of the image on which pre-processing has been completed. Therefore, in theimaging apparatus 11 of the second embodiment, since the waiting time of each process on the image is shortened, an improvement in the processing speed of the entire imaging apparatus can be expected. - In the
imaging apparatus 11 of the second embodiment, the case has been described in which an image is reduced by the resizing process. However, the resizing process may be a process of enlarging an image. In addition, in the example of the operation of theimaging apparatus 11 of the second embodiment shown inFIGS. 6 to 8 , the case has been described in which not only the pre-processing but also the resizing process is performed on the image. However, the present invention is not limited to only the resizing process. For example, the concept of the present invention can also be similarly applied to a process of changing the angle of view, such as a cutting process, or other various processes. - Also in the
imaging apparatus 11 of the second embodiment, the concept of the process of providing the overlapping region, the number of times to execute pre-processing, the execution order of pre-processing when the region of an image is divided, a direction in which theimage sensor 100 outputs the input image data, and a direction in which pre-processing according to the image division direction is performed is the same as that in theimaging apparatus 10 of the first embodiment. - As described above, in the embodiment of the present invention, the selection section, the second data transmission section, and the data acquisition section are provided in the image-capturing processing unit provided in the imaging apparatus, as a configuration to perform various kinds of pre-processing including a spatial filtering process on image data output from the image sensor multiple times in a range of the storage capacity of the line memory provided in the image-capturing processing unit. Then, in the embodiment of the present invention, pre-processing is directly performed on image data that is input in real time from the image sensor, and image data on which pre-processing cannot be directly performed is transmitted (written) to a storage unit (DRAM in the present embodiment) as it is by the second data transmission section. Then, in the embodiment of the present invention, image data to be pre-processed is switched to image data stored in the storage unit, which has been acquired (read) by the data acquisition section, and pre-processing is performed. As a result, pre-processing on all pieces of image data of one image input from the image sensor is performed.
- Thus, in the embodiment of the present invention, it is possible to perform pre-processing corresponding to the image sensor in which the number of pixels has increased, without increasing the circuit size of the imaging apparatus, especially, the storage capacity of the line memory provided in the image-capturing processing unit. Therefore, in the embodiment of the present invention, it is possible to perform pre-processing corresponding to a multi-pixel image sensor with a smaller circuit size than that corresponding to an image sensor, in which the number of pixels has increased, in a known imaging apparatus. In the embodiment of the present invention, since pre-processing can be performed directly on the image data input from the image sensor, time required for the spatial filtering process can be made shorter than that in the known imaging apparatus.
- In the embodiment of the present invention, in addition to various kinds of pre-processing including the spatial filtering process, other various processes (resizing process in the present embodiment) based on the image data after pre-processing are also performed multiple times similar to the pre-processing. Therefore, in the embodiment of the present invention, even if pre-processing on the image data in all regions of one image has not been completed, the next processing can be performed from the region of the image on which pre-processing has been completed. As a result, in the embodiment of the present invention, since the waiting time of each process on the image is shortened, it is possible to increase the processing speed of the entire imaging apparatus.
- In the present embodiment, the case has been described in which pre-processing is a process including the spatial filtering process. However, the content of the pre-processing is not limited to the embodiment of the present invention. As long as the relationship between the amount of image data that can be pre-processed and the amount of image data output from the
image sensor 100 is the relationship described above, the concept of the present invention can be similarly applied even if the spatial filtering process is not included in the pre-processing. - While preferred embodiments of the present invention have been described, the present invention is not limited to the embodiments. Additions, omissions, substitutions, and other variations may be made to the present invention without departing from the spirit and scope of the present invention. The present invention is not limited by the above description, but by the appended claims.
Claims (7)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-186471 | 2013-09-09 | ||
JP2013186471A JP2015053644A (en) | 2013-09-09 | 2013-09-09 | Imaging device |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150070530A1 true US20150070530A1 (en) | 2015-03-12 |
US9332212B2 US9332212B2 (en) | 2016-05-03 |
Family
ID=52625241
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/477,188 Expired - Fee Related US9332212B2 (en) | 2013-09-09 | 2014-09-04 | Imaging apparatus with improved pre-processing |
Country Status (2)
Country | Link |
---|---|
US (1) | US9332212B2 (en) |
JP (1) | JP2015053644A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108432237A (en) * | 2015-12-24 | 2018-08-21 | 三星电子株式会社 | Electronic equipment and control method for electronic equipment |
KR20200034504A (en) * | 2018-09-21 | 2020-03-31 | 삼성전자주식회사 | Image signal processor, method of operating the image signal processor, and application processor including the image signal processor |
CN113709494A (en) * | 2021-06-23 | 2021-11-26 | 珠海全志科技股份有限公司 | Image decompression method and device for super-resolution reconstruction |
US11315209B2 (en) * | 2020-05-08 | 2022-04-26 | Black Sesame Technolgies Inc. | In-line and offline staggered bandwidth efficient image signal processing |
US11368648B2 (en) * | 2018-04-25 | 2022-06-21 | Hangzhou Hikvision Digital Technology Co., Ltd. | Image storage method and apparatus, and electronic device and storage medium |
US11412148B2 (en) * | 2020-01-10 | 2022-08-09 | Samsung Electronics Co., Ltd. | Method of improving image quality in zoom scenario with single camera, and electronic device including the same |
US11675531B2 (en) | 2020-06-17 | 2023-06-13 | Samsung Electronics Co., Ltd. | Storage device for high speed link startup and storage system including the same |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5073959A (en) * | 1986-10-24 | 1991-12-17 | Canon Kabushiki Kaisha | Image processing apparatus with contour extraction |
US20030210335A1 (en) * | 2002-05-07 | 2003-11-13 | Carau Frank Paul | System and method for editing images on a digital still camera |
US20030210830A1 (en) * | 2002-04-22 | 2003-11-13 | Matsushita Elec. Ind. Co. Ltd. | Image signal distortion removal apparatus, and image signal distortion removal method |
US8466978B2 (en) * | 1998-05-29 | 2013-06-18 | Avago Technologies General Ip (Singapore) Pte. Ltd. | Digital image processor for a digital camera |
US20140146194A1 (en) * | 2012-11-23 | 2014-05-29 | Mediatek Inc. | Data processing system for transmitting compressed multimedia data over camera interface |
US8849090B2 (en) * | 2006-10-30 | 2014-09-30 | Sony Corporation | High speed image capturing apparatus and method |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4179701B2 (en) | 1999-04-28 | 2008-11-12 | オリンパス株式会社 | Image processing device |
-
2013
- 2013-09-09 JP JP2013186471A patent/JP2015053644A/en active Pending
-
2014
- 2014-09-04 US US14/477,188 patent/US9332212B2/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5073959A (en) * | 1986-10-24 | 1991-12-17 | Canon Kabushiki Kaisha | Image processing apparatus with contour extraction |
US8466978B2 (en) * | 1998-05-29 | 2013-06-18 | Avago Technologies General Ip (Singapore) Pte. Ltd. | Digital image processor for a digital camera |
US20030210830A1 (en) * | 2002-04-22 | 2003-11-13 | Matsushita Elec. Ind. Co. Ltd. | Image signal distortion removal apparatus, and image signal distortion removal method |
US20030210335A1 (en) * | 2002-05-07 | 2003-11-13 | Carau Frank Paul | System and method for editing images on a digital still camera |
US8849090B2 (en) * | 2006-10-30 | 2014-09-30 | Sony Corporation | High speed image capturing apparatus and method |
US20140146194A1 (en) * | 2012-11-23 | 2014-05-29 | Mediatek Inc. | Data processing system for transmitting compressed multimedia data over camera interface |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10701283B2 (en) | 2015-12-24 | 2020-06-30 | Samsung Electronics Co., Ltd. | Digital photographing apparatus and method of controlling the same |
EP3386185A4 (en) * | 2015-12-24 | 2018-11-14 | Samsung Electronics Co., Ltd. | Electronic device and control method for electronic device |
CN108432237A (en) * | 2015-12-24 | 2018-08-21 | 三星电子株式会社 | Electronic equipment and control method for electronic equipment |
US11368648B2 (en) * | 2018-04-25 | 2022-06-21 | Hangzhou Hikvision Digital Technology Co., Ltd. | Image storage method and apparatus, and electronic device and storage medium |
US20220084159A1 (en) * | 2018-09-21 | 2022-03-17 | Samsung Electronics Co., Ltd. | Image signal processor, method of operating the image signal processor, and application processor including the image signal processor |
US11189001B2 (en) * | 2018-09-21 | 2021-11-30 | Samsung Electronics Co., Ltd. | Image signal processor for generating a converted image, method of operating the image signal processor, and application processor including the image signal processor |
TWI757637B (en) * | 2018-09-21 | 2022-03-11 | 南韓商三星電子股份有限公司 | Image signal processor, method of operating thereof, and application processor |
CN110944110A (en) * | 2018-09-21 | 2020-03-31 | 三星电子株式会社 | Image signal processor, operating method thereof and application processor |
KR20200034504A (en) * | 2018-09-21 | 2020-03-31 | 삼성전자주식회사 | Image signal processor, method of operating the image signal processor, and application processor including the image signal processor |
KR102637732B1 (en) * | 2018-09-21 | 2024-02-19 | 삼성전자주식회사 | Image signal processor, method of operating the image signal processor, and application processor including the image signal processor |
US12020345B2 (en) * | 2018-09-21 | 2024-06-25 | Samsung Electronics Co., Ltd. | Image signal processor, method of operating the image signal processor, and application processor including the image signal processor |
US11412148B2 (en) * | 2020-01-10 | 2022-08-09 | Samsung Electronics Co., Ltd. | Method of improving image quality in zoom scenario with single camera, and electronic device including the same |
US11315209B2 (en) * | 2020-05-08 | 2022-04-26 | Black Sesame Technolgies Inc. | In-line and offline staggered bandwidth efficient image signal processing |
US11675531B2 (en) | 2020-06-17 | 2023-06-13 | Samsung Electronics Co., Ltd. | Storage device for high speed link startup and storage system including the same |
US11934691B2 (en) | 2020-06-17 | 2024-03-19 | Samsung Electronics Co., Ltd. | Storage device for high speed link startup and storage system including the same |
CN113709494A (en) * | 2021-06-23 | 2021-11-26 | 珠海全志科技股份有限公司 | Image decompression method and device for super-resolution reconstruction |
Also Published As
Publication number | Publication date |
---|---|
JP2015053644A (en) | 2015-03-19 |
US9332212B2 (en) | 2016-05-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9332212B2 (en) | Imaging apparatus with improved pre-processing | |
US20130021504A1 (en) | Multiple image processing | |
US9569160B2 (en) | Display processing device and imaging apparatus | |
JP2013146080A (en) | Techniques to facilitate use of small line buffers for processing small or large images | |
US9658815B2 (en) | Display processing device and imaging apparatus | |
US9554070B2 (en) | Imaging device for reducing pressure on data bus bandwidth | |
US9253401B2 (en) | Imaging device and imaging method where shake adjusted image is divided into regions for parallel processing | |
US20130162863A1 (en) | Imaging apparatus and evaluation value generation apparatus | |
US9288397B2 (en) | Imaging device, method for processing image, and program product for processing image | |
US8908060B2 (en) | Imaging apparatus generating evaluation values at a high frame rate and having a live view function of displaying a video smoothly at a low frame rate | |
US8970766B2 (en) | Imaging device | |
US9565378B2 (en) | Imaging device | |
US11202019B2 (en) | Display control apparatus with image resizing and method for controlling the same | |
US11165956B2 (en) | Imaging apparatus | |
US9225908B2 (en) | Imaging apparatus | |
US9979919B2 (en) | Solid-state imaging device and imaging system | |
US7855736B2 (en) | Method, circuit arrangement and camera for providing electronic scan reversal | |
US11310442B2 (en) | Display control apparatus for displaying image with/without a frame and control method thereof | |
US9277145B2 (en) | Imaging device dividing imaging region into first divided image data and second divided image data | |
JP2013102324A (en) | Image pickup device | |
JP2007189334A (en) | Imaging apparatus | |
JP2009165026A (en) | Imaging apparatus and control method of imaging apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: OLYMPUS CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TANAKA, YOSHINOBU;UENO, AKIRA;YANADA, TAKASHI;AND OTHERS;SIGNING DATES FROM 20140717 TO 20140804;REEL/FRAME:033670/0781 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: OLYMPUS CORPORATION, JAPAN Free format text: CHANGE OF ADDRESS;ASSIGNOR:OLYMPUS CORPORATION;REEL/FRAME:040223/0561 Effective date: 20160425 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20200503 |