US20100189308A1 - Image Measuring Apparatus and Computer Program - Google Patents
Image Measuring Apparatus and Computer Program Download PDFInfo
- Publication number
- US20100189308A1 US20100189308A1 US12/628,430 US62843009A US2010189308A1 US 20100189308 A1 US20100189308 A1 US 20100189308A1 US 62843009 A US62843009 A US 62843009A US 2010189308 A1 US2010189308 A1 US 2010189308A1
- Authority
- US
- United States
- Prior art keywords
- image
- measurement
- measurement result
- display
- information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000004590 computer program Methods 0.000 title claims description 39
- 238000005259 measurement Methods 0.000 claims abstract description 622
- 238000003384 imaging method Methods 0.000 claims description 39
- 230000008859 change Effects 0.000 claims description 21
- 238000000034 method Methods 0.000 description 24
- 230000015572 biosynthetic process Effects 0.000 description 20
- 230000006870 function Effects 0.000 description 15
- 230000008569 process Effects 0.000 description 13
- 238000010586 diagram Methods 0.000 description 8
- 238000004891 communication Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000003708 edge detection Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 2
- 238000012951 Remeasurement Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000000611 regression analysis Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
- G06T7/001—Industrial image inspection using an image reference approach
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
- G06T2207/30164—Workpiece; Machine component
Definitions
- the present invention relates to an image measuring apparatus and a computer program that allow measurement results to be displayed such that the measurement results are easily visible without depending on a state where a measurement object is mounted.
- Such an image apparatus applies light to a measurement object, acquires an image by image formation of transmitted light or reflected light of the applied light on an imaging device, such as a CCD (charge coupled device) and CMOS (complementary metal-oxide semiconductor), through a light receiving lens, and performs computing of the acquired image, thereby measuring the shape of the measurement object.
- an imaging device such as a CCD (charge coupled device) and CMOS (complementary metal-oxide semiconductor)
- the image of a measurement object becomes extremely accurately similar to the measurement object with the aid of a light receiving lens.
- the dimensions of an actual measurement object can be measured based on the dimensions of the image. Accordingly, accurately identifying the shape of a measurement object appearing on an image enables the shape of an actual measurement object to be accurately measured.
- the shape of a measurement object on an image is identified by detecting the boundary portion (hereinafter referred to as an “edge portion”) between the measurement object and the background image on the image.
- FIG. 17 shows an example of a conventional edge detection area specified on an image.
- FIG. 18 shows an example of a shape identified by using a method of least squares based on conventional edge points.
- FIG. 19 is a schematic view for illustrating a circle obtained by fitting conventional edge points to a geometrical figure.
- a periphery of an edge portion is surrounded by clicking or dragging it with a pointing device, such as a mouse, to specify an edge detection area as shown in FIG. 17 .
- the edge portion is a part with a sharp change in luminance value between the pixel of a measurement object and the pixel of a background image.
- a computer acquires, for example, a part (between pixels) with a luminance difference between adjacent pixels larger than a predetermined value in image data in the specified area, as a plurality of edge points shown in FIG. 17 .
- the plurality of acquired edge points are fitted to a geometrical figure, such as a straight line, by using a regression analysis method, such as a method of least squares, as shown in FIG. 18 .
- the straight line to which the edge points are fitted is detected as an edge. Even if points constitute a straight line, the points can be detected likewise.
- a circular shape like a straight line, a part with a sharp change in luminance value positioned between adjacent pixels is acquired as a plurality of edge points, and the plurality of acquired edge points are fitted to a geometrical figure by using a method of least squares or the like.
- a circle to which the edge points are fitted can be detected as an edge as shown in FIG. 19 .
- a size of a measurement object itself and a size of a measurement object part are various. Accordingly, if a measurement object having a hand-held size (size from several millimeters to about 100 mm square) is measured with a measuring device having the field of view of only a few millimeters square, the entire image of a measurement object cannot be collectively grasped. In this case, for example, an XY stage and the like are moved to relatively move a measurement object with respect to the field of view, so that a necessary portion is within the field of view, and thereafter measurement is performed, which is disclosed in Japanese Unexamined Patent Publication No. 11-132740.
- the distance between two straight lines that are not within the same field of view is measured, first, one straight line edge is specified, and then an XY stage is moved and the other straight line edge is specified.
- the distance between the specified straight line edges is measured by calculating from the position in the displayed image and the movement amount of the XY stage.
- a detected edge portion is displayed on a screen other than that for a measurement object, so that dimension lines and a measurement result are displayed on individual screens.
- the entire image of a measurement object cannot be displayed, which is the same as other conventional image measuring apparatuses.
- This apparatus is not different from other conventional image measuring apparatuses in that it is difficult to grasp what part of a measurement object is represented by a measured value and how the part is measured for obtaining the measured value if the shape of the measurement object is complicated.
- an object of the present invention is to provide an image measuring apparatus and a computer program to allow a user to accurately grasp what part of a measurement object is represented by a measured value of a measurement result and to allow a user to visually confirm what part of the measurement object is designated by the measurement result without depending on the attitude, position and other conditions of an image of the measurement object.
- an image measuring apparatus including: an imaging unit disposed facing a stage for mounting the measurement object and configured to acquire an image of a measurement object; a displaying unit configured to display the image of the measurement object acquired in the imaging unit; a feature image specification receiving unit configured to receive specification of any partial image on the displayed image as a feature image; a measurement position specification receiving unit configured to receive specification of a measurement position of the measurement object, on the displayed image; a relative position storing unit configured to store the feature image the specification of which is received in the feature image specification receiving unit, and information on a relative position for the feature image, the information representing the measurement position the specification of which is received in the measurement position specification receiving unit and a display position for displaying a dimension line indicating the measurement position; an attitude and position identifying unit configured to compare an image of the measurement object, the image being newly acquired by the imaging unit, with the stored feature image to identify information on an attitude and a position of the newly acquired image of the measurement object
- the relative position storing unit is configured to further store a display position for displaying the measurement result as the information on the relative position for the feature image
- the measurement result displaying unit is configured to display the measurement result at the predetermined position based on the information on the relative position of the display position for displaying the measurement result for the feature image, the information being stored in the relative position storing unit.
- the image measuring apparatus further includes: a display position changing unit configured to receive a change of the position for displaying the dimension line and/or the measurement result.
- the image measuring apparatus further includes: a measurement result specification receiving unit configured to receive specification of the measurement result to be an object of determination of whether the measurement result is within a range of a predetermined tolerance; and a display manner changing unit configured to change a display manner of the measurement result the specification of which is received in the measurement result specification receiving unit.
- the image measuring apparatus further includes a measurement result determining unit configured to determine whether the measurement result the specification of which is received in the measurement result specification receiving unit is within the range of the predetermined tolerance; and wherein the display manner changing unit is configured to change the display manner of the measurement result if it is determined in the measurement result determining unit that the measurement result is within the range of the predetermined tolerance.
- the image measuring apparatus further includes a display state determining unit configured to determine whether a portion displayed outside a display range exists in the measurement result; and wherein the measurement result displaying unit is configured to move the display position of the measurement result into the display range so that the measurement result is displayed within the display range, if it is determined in the display state determining unit that the portion displayed outside the display range exists.
- the image measuring apparatus further includes: a display angle detecting unit configured to detect an inclined angle of the display position of the measurement result; and an inclined angle determining unit configured to determine whether the detected inclined angle is equal to or greater than a predetermined angle; and wherein the measurement result displaying unit is configured to display the measurement result in an upside-down manner if it is determined in the inclined angle determining unit that the inclined angle is equal to or greater than the predetermined angle.
- the image measuring apparatus further includes a measurement result display position storing unit configured to store information on a relative position for the image of the measurement object, the information representing a display position of the displayed dimension line indicating the measurement position and the measurement result, in association with the attitude and the position of the image of the measurement object; and wherein the measurement result displaying unit is configured to change and display the information on the attitude and the position of the image of the measurement object identified in the attitude and position identifying unit based on the information on the relative position for the image of the measurement object stored in the measurement result display position storing unit, the information representing the display position of the dimension line and the measurement result.
- a computer program executable with an image measuring apparatus the computer program causing a computer to realize a function of the image measuring apparatus, the image measuring apparatus including: an imaging unit disposed facing a stage for mounting the measurement object and configured to acquire an image of a measurement object; a displaying unit configured to display the image of the measurement object acquired in the imaging unit; a feature image specification receiving unit configured to receive specification of any partial image on the displayed image as a feature image; a measurement position specification receiving unit configured to receive specification of a measurement position of the measurement object, on the displayed image; a relative position storing unit configured to store the feature image the specification of which is received in the feature image specification receiving unit, and information on a relative position for the feature image, the information representing the measurement position the specification of which is received in the measurement position specification receiving unit and a display position for displaying a dimension line indicating the measurement position; an attitude and position identifying unit configured to compare an image of the measurement object, the
- the computer program causing the computer to realize the function of the image measuring apparatus, the relative position storing unit is configured to further store a display position for displaying the measurement result as the information on the relative position for the feature image, and the measurement result displaying unit is configured to display the measurement result at the predetermined position based on the information on the relative position of the display position for displaying the measurement result for the feature image, the information being stored in the relative position storing unit.
- the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a display position changing unit configured to receive a change of the position for displaying the dimension line and/or the measurement result.
- the computer program causing the computer to realize the function of the image measuring apparatus the image measuring apparatus further includes: a measurement result specification receiving unit configured to receive specification of the measurement result to be an object of determination of whether the measurement result is within a range of a predetermined tolerance; and a display manner changing unit configured to change a display manner of the measurement result the specification of which is received in the measurement result specification receiving unit.
- the image measuring apparatus further includes: a measurement result determining unit configured to determine whether the measurement result the specification of which is received in the measurement result specification receiving unit is within the range of the predetermined tolerance; and wherein the display manner changing unit is configured to change the display manner of the measurement result if it is determined in the measurement result determining unit that the measurement result is within the range of the predetermined tolerance.
- the image measuring apparatus further includes: a display state determining unit configured to determine whether a portion displayed outside a display range exists in the measurement result; and wherein the measurement result displaying unit is configured to move the display position of the measurement result into the display range so that the measurement result is displayed within the display range, if it is determined in the display state determining unit that the portion displayed outside the display range exists.
- the computer program causing the computer to realize the function of the image measuring apparatus the image measuring apparatus further includes: a display angle detecting unit configured to detect an inclined angle of the display position of the measurement result; and an inclined angle determining unit configured to determine whether the detected inclined angle is equal to or greater than a predetermined angle; and wherein the measurement result displaying unit is configured to display the measurement result in an upside-down manner if it is determined in the inclined angle determining unit that the inclined angle is equal to or greater than the predetermined angle.
- the image measuring apparatus further includes: a measurement result display position storing unit configured to store information on a relative position for the image of the measurement object, the information representing a display position of the displayed dimension line indicating the measurement position and the measurement result, in association with the attitude and the position of the image of the measurement object; and wherein the measurement result displaying unit is configured to change and display the information on the attitude and the position of the image of the measurement object identified in the attitude and position identifying unit based on the information on the relative position for the image of the measurement object stored in the measurement result display position storing unit, the information representing the display position of the dimension line and the measurement result.
- an image of a measurement object that is obtained in an imaging unit is displayed, specification of any partial image on the displayed image as a feature image is received, and specification of a measurement position of the measurement object is received.
- the feature image the specification of which is received and information on a relative position for the feature image are stored. This information represents a measurement position and a display position for displaying a dimension line indicating the measurement position.
- An image of the measurement object newly acquired by the imaging unit is compared with the stored feature image to specify information on an attitude and a position of the newly acquired image of the measurement object.
- a measurement position is set for the image of the measurement object for which the attitude and the position are specified, and an edge portion at the measurement position is detected, and then a predetermined physical quantity is measured.
- the information representing the display position of the dimension line, a dimension line indicating the measurement position and a measurement result are displayed at a predetermined position. Storing the feature image, the specification of which is received, and the information on the relative position for the feature image, which represents the measurement position and the display position for displaying the dimension line indicating the measurement position, allows the image of the measurement object and the feature image to be compared.
- a display position for displaying the measurement result may be further stored as the information on the relative position for the feature image, and the measurement result may be displayed at the predetermined position based on the information on the relative position of the display position for displaying the measurement result for the feature image.
- its position can thus be changed according to an image of the measurement object without depending on the attitude, the position and the like of the image of the measurement object, and therefore a user can visually confirm what portion of the measurement object corresponds to the measurement result.
- a change of the position for displaying the dimension line and/or the measurement result may be received. Even when the displayed position of the dimension line and/or the measurement result is outside the field of view, the displayed position can thus be changed so as to be within the field of view. Therefore, a user can visually confirm what portion of the measurement object corresponds to the measurement result.
- specification of the measurement result to be an object of determination of whether the measurement result is within the range of a predetermined tolerance may be received, and the display manner of the measurement result the specification of which is received may be changed.
- the measurement result to be an object of determination of whether the measurement result is within the range of the predetermined tolerance can be easily confirmed with a difference of the display manner. Accordingly, it is possible to visually confirm the measurement result for which accuracy is required so as to be within the range of tolerance.
- the display manner of the measurement result may be changed if it is determined that the measurement result is within the range of the predetermined tolerance. Whether the measurement result is within a certain range considered to be appropriate can thus be easily confirmed with the display manner of the measurement result. Accordingly, an appropriate measurement result and an inappropriate measurement result can be visually confirmed. This makes it possible to easily confirm what portion of the measurement object has an inappropriate measurement result, and therefore remeasurement is made easier.
- the display position of the measurement result is moved into the display range so that the measurement result is displayed within the display range. This enables the display of the measurement result to be visually confirmed with reliability.
- an inclined angle of the display position of the measurement result may be detected, and the measurement result may be displayed in an upside-down manner if the detected inclined angle is equal to or greater than a predetermined angle. This allows the measurement result to be visually confirmed even if the measurement object is mounted at any angle on the stage.
- information on a relative position for the image of the measurement object may be stored in association with the attitude and the position of the image of the measurement object, and the identified information on the attitude and the position of the image of the measurement object may be changed and displayed based on the stored information on the relative position for the image of the measurement object, the information representing the dimension line and the display position of the measurement result.
- the measurement object can therefore be rotated and displayed in accordance with the display position based on the stored information on the dimension line and the position of the measurement result, even if the measurement object is mounted at any angle on the stage.
- the feature image, the specification of which is received, and the information on the relative position for the feature image, which represents the measurement position and the display position for displaying the dimension line indicating the measurement position are stored.
- the image of the measurement object and the feature image can therefore be compared. This makes it possible to determine the attitude and the position of the image of the measurement object, and the measurement position and the display position for displaying the dimension line indicating the measurement position as the relative position for the image of the measurement object. Accordingly, a user can visually confirm what portion of the measurement object corresponds to the measurement result without depending on the attitude, the position and the like of the image of the measurement object.
- FIG. 1 is a perspective view showing a main body of an image measuring apparatus according to a first embodiment of the present invention
- FIG. 2 is a schematic view showing a configuration of the image measuring apparatus according to the first embodiment of the present invention
- FIG. 3 is a block diagram showing a configuration of a control unit of the image measuring apparatus according to the first embodiment of the present invention
- FIG. 4 is a block diagram showing functions realized by the control unit of the image measuring apparatus according to the first embodiment of the present invention.
- FIG. 5 is a flow chart showing a procedure of a feature image storing process of a CPU of the control unit of the image measuring apparatus according to the first embodiment of the present invention
- FIGS. 6A to 6C are schematic views showing a relationship between an image of a measurement object and a range of a field of view and a relationship between the measurement object and a display position of the measurement result;
- FIGS. 7A to 7C are schematic views showing a way of receiving specification of a feature image and measurement positions
- FIG. 8 is a flow chart showing a procedure of an image measuring process of the CPU of the control unit of the image measuring apparatus according to the first embodiment of the present invention.
- FIGS. 9A and 9B show an example of a matching process between the feature image and an image of measurement object
- FIGS. 10A to 10C are schematic views showing a way of displaying measurement positions and dimension lines
- FIGS. 11A and 11B are schematic views for illustrating a way of changing display positions of a dimension line and/or a measurement result when the measurement result protrudes outside a display range;
- FIGS. 12A to 12D are schematic views for illustrating the ways of changing a display position when measurement results are displayed in an upside-down manner
- FIG. 13 is a block diagram showing a configuration of a control unit of an image measuring apparatus according to a second embodiment of the present invention.
- FIG. 14 is a function block diagram showing functions realized by a control unit of the image measuring apparatus according to the second embodiment of the present invention.
- FIG. 15 is a flow chart showing a procedure of an image measuring process of a CPU of the control unit of the image measuring apparatus according to the second embodiment of the present invention.
- FIGS. 16A to 16D are schematic views showing the way of displaying dimension lines and measurement results
- FIG. 17 shows an example of a conventional edge detection area specified on an image
- FIG. 18 shows an example of a shape identified by using a method of least squares based on conventional edge points.
- FIG. 19 is a schematic view for illustrating a circle obtained by fitting conventional edge points to a geometrical figure.
- FIG. 1 is a perspective view showing a main body 2 of an image measuring apparatus according to a first embodiment of the present invention
- FIG. 2 is a schematic view showing a configuration of an image measuring apparatus 1 according to the first embodiment of the present invention.
- the image measuring apparatus 1 according to the first embodiment includes the main body 2 and a control unit 3 .
- Image data is obtained by image formation in the main body 2 , and the obtained image data is computed in the control unit 3 , thereby measuring dimensions and the like of a desired shape of the image.
- the main body 2 includes a stage 21 having a mounting section 211 .
- the stage 21 is connected to the main body 2 with a supporting portion 212 interposed therebetween such that the stage 21 can be rotated around the mounting section 211 .
- the measurement object 20 is imaged by a imaging part 30 disposed above an arm 28 , and the measurement result is displayed on a display 27 .
- two sets of illuminating systems 22 and 24 are disposed on either side of a stage 21 for moving the mounted measurement object 20 .
- the ring-shaped epi-illuminating system 22 which illuminates the measurement object 20 of the stage 21 from the above, is provided in a light receiving lens unit 23 .
- Light applied by the epi-illuminating system 22 is reflected from the surface of the measurement object 20 , and is returned to the light receiving lens unit 23 . This permits irregularities, a pattern and the like of the surface of the measurement object 20 to be imaged.
- a transmission illuminating system 24 which illuminates the measurement object 20 from the below, is disposed under the stage 21 .
- the transmission illuminating system 24 includes at least a light source 241 , a reflecting mechanism 242 and a lens 243 .
- Light applied from the light source 241 is reflected from the reflecting mechanism 242 toward the stage 21 .
- the lens 243 the light is converted into parallel light rays in a direction approximately perpendicular to the stage 21 . In this way, it is possible to perform imaging in which light is transmitted only through a position without the measurement object 20 .
- the light receiving lens unit 23 includes at least a light receiving lens 231 , a beam splitter 232 , a high-magnification-side image formation lens part 233 and a low-magnification-side image formation lens part 236 .
- the high-magnification-side image formation lens part 233 includes a slit 234 for image formation and a high-magnification-side image formation lens 235
- the low-magnification-side image formation lens part 236 includes a slit 237 for image formation and a low-magnification-side image formation lens 238 .
- the beam splitter 232 is a prism to cause light from the light receiving lens 231 to branch in two directions. For example, cubic-type and plate-type beam splitters may be used.
- a cubic-type beam splitter is preferable compared to a plate-type beam splitter.
- FIG. 2 shows an example in which light applied from the epi-illuminating system 22 guides light reflected from the measurement object 20 and light applied from the transmission illuminating system 24 and transmitted through the measurement object 20 to the high-magnification-side image formation lens part 233 and the low-magnification-side image formation lens part 236 .
- Light rays in two directions obtained by branching by the beam splitter 232 are guided to both the low-magnification-side image formation lens part 236 and the high-magnification-side image formation lens part 233 .
- the high-magnification-side imaging apparatus 25 performs image formation of light guided to the high-magnification-side image formation lens part 233 using the imaging device 251 , such as a CCD or CMOS, and transmits the resultant image as high magnification image data to the control unit 3 .
- a low-magnification-side imaging apparatus 26 performs image formation of light guided to the low-magnification-side image formation lens part 236 using an imaging device 261 , such as a CCD or CMOS, and transmits the resultant image as low magnification image data to the control unit 3 .
- high magnification image data and low magnification image data can be simultaneously acquired without mechanically switching the optical system.
- Both high and low image data can be electronically switched and displayed on one screen, and can be individually displayed simultaneously on two screens.
- FIG. 3 is a block diagram showing the configuration of the control unit 3 of the image measuring apparatus 1 according to the first embodiment of the present invention.
- the control unit 3 of the image measuring apparatus 1 according to the first embodiment includes at least a CPU (central processing unit) 33 , a storing device 35 such as a RAM (random access memory) 34 or a disk, a communication device 36 , and an internal bus 37 that connects the hardware mentioned above.
- the control unit 3 is connected to a mouse 32 and a keyboard 31 , which are input devices, and a display device 27 , which is an output device.
- the CPU 33 is connected through the internal bus 37 to components as described above of hardware of the control unit 3 , and controls the operation of the components of hardware and executes various software functions in accordance with computer programs stored in the storing device 35 .
- the RAM 34 is a volatile memory, such as an SRAM (static random access memory) or an SDRAM (synchronous dynamic random access memory), and a load module is expanded during execution of a computer program to store temporary data and the like generated during execution of the computer program.
- Both of image data obtained by imaging using the high-magnification-side imaging apparatus 25 and the low-magnification-side imaging apparatus 26 are stored in the storing device 35 such that high-magnification image data and low-magnification image data are aligned each other.
- Each of the low-magnification image and the high-magnification image as used herein is obtained by causing light from the common light receiving lens 231 to branch in two directions with the beam splitter 232 , passing each branching light through the high-magnification-side image formation lens 235 or the low-magnification-side image formation lens 238 , and forming an image on the imaging device 251 or 261 . Both the images are taken on the same axis. That is, they are taken so that the center of the field of view of the low-magnification image agrees with the center of the field of view of the high-magnification image.
- the storing device 35 includes a feature image storing unit 351 and a relative position storing unit 352 .
- the feature image storing unit 351 stores feature image data representing a feature for each measurement object 20 . Matching the stored feature image data with the acquired image data makes it possible to grasp what attitude and position are taken when the measurement object 20 is mounted on the stage 21 .
- the relative position storing unit 352 stores information on relative positions for a feature image, the information representing a measurement position the specification of which has been received and a display position of a dimension line for indicating the measurement position. Based on the stored information on relative positions, the relative positions for displaying the measurement position and the dimension line are determined upon identifying the attitude and the position of an image of the measurement object 20 .
- the feature image storing unit 351 and the relative position storing unit 352 may be integrated for storing.
- the communication device 36 is connected to the internal bus 37 , and is connected through communication lines to high-magnification-side imaging apparatus 25 and the low-magnification-side imaging apparatus 26 to receive image data taken by the imaging apparatuses 25 and 26 .
- external networks such as the Internet, LAN (local area network) and WAN (wide area network)
- Computer programs stored in the storing device 35 are downloaded from external computers through the communication device 36 .
- FIG. 4 is a block diagram showing functions realized by the control unit 3 of the image measuring apparatus 1 according to the first embodiment of the present invention.
- an image data acquiring unit 331 acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 . Both of the acquired image data are stored in the storing device 35 .
- a displaying unit 332 displays the acquired low-magnification image data or high-magnification image data on the display 27 ;
- a feature image specification receiving unit 333 receives area specification to specify a partial image considered as an image that represents a portion characteristic of the image of the measurement object 20 displayed on the display 27 .
- the way of receiving area specification is not particularly limited. For example, specification of a rectangular area including a partial image that is considered as an image representing a characteristic portion is received by a drag operation and the like of the mouse 32 and the like.
- the partial image included in the rectangular area, the specification of which has been received, is stored as a feature image of the measurement object 20 in the feature image storing unit 351 of the storing device 35 .
- a measurement position specification receiving unit 334 receives specification of a measurement position for the image of the measurement object 20 displayed on the display 27 .
- the way of receiving specification of a measurement position is not particularly limited. For example, area specification to specify the vicinity of an edge portion is received by a drag operation and the like of the mouse 32 and the like.
- a tolerance specification receiving unit 346 to receive specification of an appropriate range of a tolerance may be included.
- the way of receiving specification of an appropriate range of a tolerance is not particularly limited. For example, specification may be received by pressing keys of the keyboard 31 from a setting window.
- An attitude and position identifying unit 335 performs a matching process (compare) between the acquired image data of the measurement object 20 and the feature image stored in the feature image storing unit 351 to identify information on the attitude and the position of the image of the measurement object 20 .
- An edge portion detecting unit 336 sets a measurement position for the image of the measurement object 20 , for which the attitude and the position have been identified, based on information on the relative position of a measurement position for the feature image stored in the relative position storing unit 352 , and detects an edge portion at the set measurement position.
- the measuring unit 337 measures predetermined physical quantities based on the detected edge portion and the identified information on the attitude and the position of the image of the measurement object 20 .
- a measurement result displaying unit 338 displays a measurement result and a dimension line indicating a measurement position at predetermined positions, based on information on the relative positions for the feature image stored in the relative position storing unit 352 , the information representing the measurement position and a display position of the dimension line.
- information on a relative position for displaying a measurement result for an image of the measurement object 20 is also stored in the relative position storing unit 352 , and the measurement result may be displayed at a predetermined position based on the attitude and the position of an image of the measurement object 20 identified in the attitude and position identifying unit 335 and the information on the relative position for displaying the measurement result for the image of the measurement object 20 . This allows the display position of the measurement result to be changed in accordance with the attitude and the position of the image of the measurement object 20 .
- a display position changing unit 339 receives a change in position at which a dimension line and/or a measurement result are displayed.
- the way of receiving a change in position for displaying is not particularly limited. For example, it is received by a drag operation and the like of the mouse 32 and the like.
- a measurement result specification receiving unit 340 receives specification of a measurement result to be an object of determination of whether the measurement result is within a predetermined tolerance.
- the way of receiving specification of a measurement result is not particularly limited. For example, specification of a display position of a measurement result may be received, and specification of a measurement position for measuring a measurement result may also be received, by a click operation and the like of the mouse 32 and the like.
- a measurement result determining unit 341 determines whether the measurement result, the specification of which has been received, is within a predetermined tolerance.
- a display manner changing unit 342 changes the display manner of the measurement result, the specification of which has been received in the measurement result specification receiving unit 340 , and the measurement result that is determined in the measurement result determining unit 341 to be within the tolerance. For example, changing the display manner, such as changing the display color and highlighting, allows a user to visually confirm the measurement result that is determined to be within the tolerance.
- a display state determining unit 343 determines whether a portion displayed outside a display range exists in a measurement result. If it is determined in the display state determining unit 343 that the portion displayed outside the display range exists in the measurement result, the measurement result displaying unit 338 moves the display position of the measurement result into the display range so that the measurement result is displayed within the display range. Accordingly, a user can visually confirm the measurement result.
- a display angle detecting unit 344 detects an inclined angle of a display position of a measurement result; an inclined angle determining unit 345 determines whether the inclined angle detected in the display angle detecting unit 344 is equal to or greater than a predetermined angle. If it is determined in the inclined angle determining unit 345 that the inclined angle is equal to or greater than the predetermined angle, the measurement result displaying unit 338 displays the measurement result in an upside-down manner.
- FIG. 5 is a flow chart showing the procedure of a feature image storing process of the CPU 33 of the control unit 3 of the image measuring apparatus 1 according to the first embodiment of the present invention.
- the CPU 33 of the control unit 3 acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 (step S 501 ). Both of the acquired image data are stored in the storing device 35 .
- the CPU 33 determines whether displaying the high-magnification image data within the range of the field of view is possible (step S 502 ). If it is determined by the CPU 33 that displaying the high-magnification image data within the range of the field of view is possible (step S 502 : YES), then the CPU 33 displays the high-magnification image data on the display 27 (step S 503 ); if it is determined by the CPU 33 that displaying the high-magnification image data within the range of the field of view is impossible (step S 502 : NO), the CPU 33 displays the low-magnification image data on the display 27 (step S 504 ).
- FIGS. 6A to 6C are schematic views showing a relationship between an image of the measurement object 20 and the range of the field of view and a relationship between the measurement object 20 and the display position of the measurement result.
- an image (image data) of the measurement object 20 is acquired as shown in FIG. 6A , it is not clear what portion of the measurement object 20 is displayed in the case where a range 61 of the field of view is too small for image data of the measurement object 20 as shown in FIG. 6B , such as the case of displaying high-magnification image data.
- the entire image of the measurement object 20 is within the range 61 of the field of view as shown in FIG. 6C in the case where the range 61 of the field of view is sufficiently large for the image data of the measurement object 20 , such as the case of displaying low-magnification image data.
- the CPU 33 of the control unit 3 receives specification of a rectangular area including a feature image considered as an image representing a characteristic portion on a low-magnification image or a high-magnification image (step S 505 ).
- the way of receiving specification of a rectangular area including a feature image is not particularly limited.
- specification of a display position of a measurement result may be received by a drag operation and the like of the mouse 32 and the like, and a partial image included in the rectangular area, the specification of which has been received, may be stored as a feature image of the measurement object 20 in the feature image storing unit 351 of the storing device 35 .
- the CPU 33 receives specification of a measurement position on the displayed low-magnification image or high-magnification image (step S 506 ).
- the way of receiving specification of the measurement position is not particularly limited. For example, area specification of the vicinity of the edge portion is received by a drag operation and the like of the mouse 32 and the like. Note that specification of an appropriate range of a tolerance may be received in step S 506 .
- the way of receiving specification of an appropriate range of a tolerance is not particularly limited. For example, specification may be received by pressing keys of the keyboard 31 from a setting window. The following may be mentioned as another example. With the measurement result displayed, specification of a measurement position is received by a click operation and the like of the mouse 32 and the like. Then, specification of an appropriate range of a tolerance is received for each measurement position the specification of which has been received.
- the CPU 33 stores the feature image, the specification of which has been received, in the feature image storing unit 351 ; the CPU 33 stores information on relative positions for the feature image, which represents the measurement position the specification of which has been received and the display position of the dimension line indicating the measurement position, in the relative position storing unit 352 (step S 507 ).
- FIGS. 7A to 7C are schematic views showing the way of receiving specification of a feature image and measurement positions.
- FIG. 7A shows an example of the way of receiving specification of a rectangular area 71 including a feature image and measurement positions 72 .
- FIG. 7B shows an example of a feature image the specification of which is received and stored in the feature image storing unit 351 .
- FIG. 7C shows an example of information on relative positions of the measurement positions 72 , the specification of which is received, and display positions of dimension lines 73 and measurement results 74 , the specification of which is received and stored in the relative position storing unit 352 .
- receiving specification of the rectangular area 71 as shown in FIG. 7A allows a feature image as shown in FIG. 7B to be cut out and stored in the feature image storing unit 351 .
- image data of the feature image but also coordinates, a rotation angle and the like at a relative position of the rectangular area 71 for an image of the measurement object 20 are stored in the feature image storing unit 351 .
- receiving specification of the measurement positions 72 as shown in FIG. 7A allows information on relative positions for the feature image, which represents the measurement positions and the display positions of the dimension lines 73 indicating the measurement positions, to be stored in the relative position storing unit 352 as shown in FIG. 7C .
- information on relative positions but also coordinates, rotation angles and the like at the relative positions for the image of the measurement object 20 which represent the measurement positions 72 and the display positions of the dimension lines 73 , are stored in the relative position storing unit 352 .
- measurement result display positions 74 at which displaying measurement results are displayed may be stored in the relative position storing unit 352 in the same way as those of other elements, as shown in FIG. 7C .
- FIG. 8 is a flow chart showing the procedure of an image measuring process of the CPU 33 of the control unit 3 of the image measuring apparatus 1 according to the first embodiment of the present invention.
- the CPU 33 of the control unit 3 newly acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 (step S 801 ). Both of the acquired image data are stored in the storing device 35 .
- the CPU 33 determines whether displaying the high-magnification image data within the range of the field of view is possible (step S 802 ). If it is determined by the CPU 33 that displaying the high-magnification image data within the range of the field of view is possible (step S 802 : YES), then the CPU 33 displays the high-magnification image data on the display 27 (step S 803 ); if it is determined by the CPU 33 that displaying the high-magnification image data within the range of the field of view is impossible (step S 802 : NO), the CPU 33 displays the low-magnification image data on the display 27 (step S 804 ).
- the CPU 33 performs a matching process (compare) between the newly acquired image data and the feature image stored in the feature image storing unit 351 (step S 805 ) to identify information on the attitude and the position of an image of the measurement object 20 (step S 806 ).
- the CPU 33 sets a measurement position for the image of the measurement object 20 , for which the attitude and the position have been identified, based on the information on the relative position of a measurement position for the feature image stored in the relative position storing unit 352 , and detects an edge portion of the set measurement position (step S 807 ).
- FIGS. 9A and 9B show an example of a matching process between a feature image and an image of the measurement object 20 .
- FIG. 9A shows an example of a newly acquired image data
- FIG. 9B shows an example of a feature image stored in the feature image storing unit 351 .
- the acquired image data is searched for a partial image that agrees with a feature image 91 shown in FIG. 9B . If the partial image that agrees with the feature image 91 is found, the rectangular area 71 for this partial image is identified. Coordinates, a relative rotation angle and the like of a relative position of the rectangular area 71 for an image of the measurement object 20 are stored in the feature image storing unit 351 and coordinates, relative rotation angles and the like of relative positions for the image of the measurement object 20 , which represent the measurement positions 72 and display positions of the dimension lines 73 , are stored in the relative position storing unit 352 . Therefore, specifying the rectangular area 71 allows accurate specification of what portion of the measurement object 20 is a measurement position. A desired edge portion can thus be detected.
- the CPU 33 of the control unit 3 measures predetermined physical quantities based on the detected edge portion (step S 808 ).
- the predetermined physical quantities to be measured include a length, a diameter, an angle and the like.
- the CPU 33 displays a dimension line indicating a measurement position and a measurement result at predetermined positions based on the information on the relative position for the feature image, which represents a display position of the dimension line, stored in the relative position storing unit 352 (step S 809 ).
- FIGS. 10A to 10C are schematic views showing the way of displaying measurement positions and dimension lines.
- predetermined physical quantities such as a length, a diameter and an angle
- measured for the measurement positions 62 are displayed as measurement results 64 at predetermined positions, based on the identified attitude and position of an image of the measurement object 20 and information on relative positions for displaying dimension lines 63 for a feature image stored in the relative position storing unit 352 .
- the measurement positions 62 and display positions of the dimension lines 63 are stored as relative positions for a feature image. Therefore, for example, even when the measurement object 20 is mounted at a given position on the stage 21 as shown in FIG. 10B , the measurement positions 62 and display positions of the dimension lines 63 are moved and rotated correspondingly to the identified attitude and the position of the measurement object 20 , allowing the measurement results 64 to be displayed at positions at which the measurement results are easily visible, as shown in FIG. 10C . Accordingly, even if a user arbitrarily mounts measurement object 20 on the stage 21 , the measurement positions 62 and the dimension lines 63 can be moved and rotated correspondingly to the attitude and the position of the measurement object 20 , and thus be displayed. The measurement results 64 measured at the stored measurement positions 62 can thus be displayed at positions at which they are easily visible.
- Information on relative positions for displaying the measurement results 64 for the feature image may be stored in the relative position storing unit 352 , and the CPU 33 of the control unit 3 may display the measurement results 64 at predetermined positions, based on the identified attitude and the position of the image of the measurement object 20 and information on relative positions for displaying the measurement results 64 for the feature image. This allows display positions of the measurement results 64 to be changed in accordance with the attitude and the position of the image of the measurement object 20 .
- the measurement result 64 might protrude outside the display range depending on how the measurement object 20 is mounted. If the measurement result 64 protrudes outside the display range, a user cannot be aware of the measurement result 64 and cannot visually confirm it.
- the dimension line 63 and/or the measurement result 64 are moved and displayed so that they are within the display range by receiving changes of positions for displaying the dimension line 63 and/or the measurement result 64 .
- the CPU 33 determines whether a portion displayed outside the display range exists in the measurement result 64 . If it is determined that a portion displayed outside the display range exists, then the CPU 33 may move the display position of the measurement result 64 to be within the display range.
- FIGS. 11A and 11B are schematic views for illustrating the way of changing display positions of the dimension line 63 and/or the measurement result 64 if the measurement result 64 protrudes outside the display range.
- the dimension line 63 indicating the length of an upper side of the measurement object 20 and the measurement result 64 are displayed in an area 81 outside the range 61 of the field of view. A user cannot visually confirm them.
- the CPU 33 determines whether the area 81 for displaying dimension line 63 and the measurement result 64 outside the range 61 of the field of view exist. If the existence is determined, then the display position is changed by decreasing the height of the dimension line 63 so that the measurement result 64 is within the range 61 of the field of view as shown in FIG. 11B .
- a user may explicitly move the display position by a drag operation and the like of the mouse 32 and the like.
- FIGS. 12A to 12D are schematic views for illustrating ways of changing a display position when the measurement result 64 is displayed in an upside-down manner.
- the numerical values of the measurement results 64 are displayed in an upside-down manner.
- This is a display manner that makes it difficult for a user to confirm the measurement results.
- the CPU 33 determines that the image of the measurement object 20 rotates clockwise or counterclockwise by 90 degrees or more, the measurement results 64 are displayed in an upside-down manner as shown in FIG. 12B . Accordingly, a user can easily confirm the measurement results 64 visually.
- the way to overcome the drawback is, of course, not limited to displaying the measurement results 64 in an upside-down manner.
- Any display manner may be employed as long as it allows the measurement results 64 to be displayed such that a user can easily confirm the measurement results.
- the measurement results 64 may be displayed in alignment in the transverse direction of a screen, not displayed in an upside-down manner.
- the image itself of the measurement object 20 may be rotated to display the measurement results 64 . That is, as shown in FIG. 12D , the relative position relationship is stored such that the image of the measurement object 20 is integrated with the dimension lines 63 and the measurement results 64 . If the CPU 33 determines that the image of the measurement object 20 rotates clockwise or counterclockwise by 90 degrees or more with respect to the stored feature image of the measurement object 20 , the image, the dimension lines 63 and the measurement results 64 of the measurement object 20 are integrally rotated until the image is in the same direction as that of the stored feature image. In this manner, the measurement results 64 can be displayed in such a display manner that the measurement results 64 are more easily confirmed.
- the display manner of the measurement results 64 may be changed. For example, changing the display manner, such as changing the display color and highlighting, permits a user to visually confirm the measurement results 64 that has not been within the range of the tolerance. Note that for the measurement results 64 to be objects of determination of whether the measurement results 64 are within the range of the tolerance, it is to be understood that the display manner of the measurement results 64 may be changed.
- a feature image the specification of which has been received and information on relative positions for the feature image, the information representing a measurement position and a display position of a dimension line indicating the measurement position are stored.
- the configuration of the image measuring apparatus 1 according to a second embodiment of the present invention is the same as that of the first embodiment, and therefore the components are denoted by the same reference numerals and the detailed description thereof will not be given.
- the second embodiment differs from the first embodiment in that information on display positions of dimension lines indicating the displayed measurement positions and measurement results is stored in association with the attitude and the position of an image of the measurement object 20 , and the image of the measurement object 20 is displayed in accordance with the displayed dimension lines, the measurement results and measurement positions functioning as the basis of measurement.
- FIG. 13 is a block diagram showing the configuration of the control unit 3 of the image measuring apparatus 1 according to the second embodiment of the present invention.
- the hardware configuration of the control unit 3 of the image measuring apparatus 1 according to the second embodiment is the same as that of the first embodiment, as shown in FIG. 13 , and therefore the same components are denoted by the same reference numerals and the detailed description thereof will not be given.
- the storing device 35 includes a measurement result display position storing unit 353 in addition to the feature image storing unit 351 and the relative position storing unit 352 .
- the measurement result display position storing unit 353 stores information on relative positions for an image of the measurement object 20 , the information representing the attitude, the position, the measurement positions and display positions of dimension lines of the image of the measurement object 20 . Based on the stored information on relative positions, the relative positions for displaying the corresponding measurement positions and the dimension lines are determined upon identifying information on the attitude and the position of the image of the measurement object 20 . Needless to say, information on the relative position may be stored for the positions for displaying measurement results.
- FIG. 14 is a function block diagram showing functions realized by the control unit 3 of the image measuring apparatus 1 according to the second embodiment of the present invention.
- the image data acquiring unit 331 acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 . Both of the acquired image data are stored in the feature image storing unit 351 of the storing device 35 .
- the displaying unit 332 displays the acquired low-magnification image data or high-magnification image data on the display 27 ;
- the feature image specification receiving unit 333 receives area specification to specify a partial image considered as an image that represents a portion characteristic of an image of the measurement object 20 displayed on the display 27 .
- the way of receiving area specification is not particularly limited. For example, specification of a rectangular area including a partial image that is considered as an image representing a characteristic portion is received by a drag operation and the like of the mouse 32 and the like.
- the partial image included in the rectangular area, the specification of which has been received, is stored as a feature image of the measurement object 20 in the feature image storing unit 351 of the storing device 35 .
- the measurement position specification receiving unit 334 receives specification of a measurement position for the image of the measurement object 20 displayed on the display 27 .
- the way of receiving specification of a measurement position is not particularly limited. For example, area specification to specify the vicinity of an edge portion is received by a drag operation and the like of the mouse 32 and the like.
- the tolerance specification receiving unit 346 to receive specification of an appropriate range of a tolerance may be included.
- the way of receiving specification of an appropriate range of a tolerance is not particularly limited. For example, specification may be received by pressing keys of the keyboard 31 from a setting window.
- the attitude and position identifying unit 335 performs a matching process (compare) between the acquired image data of the measurement object 20 and the feature image stored in the feature image storing unit 351 to identify information on the attitude and the position of the image of the measurement object 20 .
- the edge portion detecting unit 336 sets a measurement position for the image of the measurement object 20 , for which the attitude and the position have been identified, based on information on the relative position of a measurement position for the feature image stored in the relative position storing unit 352 , and detects an edge portion of the set measurement position; the measuring unit 337 measures predetermined physical quantities based on the detected edge portion and the identified information on the attitude and the position of the image of the measurement object 20 .
- the measurement result displaying unit 338 displays a measurement result and a dimension line indicating a measurement position at predetermined positions, based on the information on the relative positions for the feature image stored in the relative position storing unit 352 , the information representing the measurement position and a display position of the dimension line.
- Information on relative positions for the measurement object 20 that represents display positions of a dimension line and a measurement result are stored in the measurement result display position storing unit 353 in association with the feature image of the measurement object 20 being displayed.
- FIG. 15 is a flow chart showing the procedure of an image measuring process of the CPU 33 of the control unit 3 of the image measuring apparatus 1 according to the second embodiment of the present invention.
- the CPU 33 of the control unit 3 displays a dimension line indicating a measurement position and a measurement result at predetermined positions based on the information on the relative position for the feature image, which represents a display position of the dimension line, stored in the relative position storing unit 352 (step S 809 in FIG. 8 ), and stores in the measurement result display position storing unit 353 the displayed dimension line, measurement result, and display position of the image of the measurement object 20 as information on relative positions for the image of the measurement object 20 , the information representing the measurement position and the display position of the dimension line (step S 1501 ).
- the CPU 33 newly acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 (step S 1502 ). Both of the acquired image data are stored in the feature image storing unit 351 of the storing device 35 .
- the CPU 33 determines whether displaying the high-magnification image data within the range of the field of view is possible (step S 1503 ). If it is determined by the CPU 33 that displaying the high-magnification image data within the range of the field of view is possible (step S 1503 : YES), then the CPU 33 displays the high-magnification image data on the display 27 (step S 1504 ); if it is determined by the CPU 33 that displaying the high-magnification image data within the range of the field of view is impossible (step S 1503 : NO), the CPU 33 displays the low-magnification image data on the display 27 (step S 1505 ).
- the CPU 33 performs a matching process (compare) between the newly acquired image data and the feature image stored in the feature image storing unit 351 (step S 1506 ) to identify information on the attitude and the position of an image of the measurement object 20 (step S 1507 ).
- the CPU 33 sets a measurement position for the image of the measurement object 20 , for which the attitude and the position have been identified, based on the information on the relative position of a measurement position for the feature image stored in the relative position storing unit 352 , and detects an edge portion of the set measurement position (step S 1508 ).
- the CPU 33 measures predetermined physical quantities based on the detected edge portion (step S 1509 ).
- the predetermined physical quantities to be measured include a length, a diameter, an angle and the like.
- the CPU 33 displays a dimension line indicating a measurement position, a measurement result and the image of the measurement object 20 at predetermined positions based on the information on relative positions for the image of the measurement object 20 , the information representing measurement position and a display position of the dimension line, stored in the measurement result display position storing unit 353 (step S 1510 ).
- FIGS. 16A to 16D are schematic views showing the way of displaying dimension lines and measurement results.
- predetermined physical quantities measured such as a length, a diameter and an angle
- measurement results 64 are displayed as measurement results 64 at predetermined positions, based on the attitude and position of the image of the measurement object 20 and information on relative positions for displaying the measurement positions (measurement areas) 62 and dimension lines 63 for the image of the measurement object 20 stored in the relative position storing unit 352 .
- the displayed dimension lines 63 , measurement positions 62 , and preferably display positions 65 of the measurement results 64 are stored as relative positions with the image of the measurement object 20 .
- the measurement object 20 is removed from FIG. 16A , and the resultant is stored as a screen image excluding the display of the measurement results 64 in the measurement result display position storing unit 353 .
- the attitude and the position of the measurement object 20 are identified, and the image itself of the measurement object 20 is moved and rotated in accordance with the corresponding feature image, and the dimension lines 63 , the measurement positions 62 , and preferably display positions 65 of the measurement results 64 that correspond to the feature image. That is, as shown in FIG. 16D , the image of measurement object 20 is moved and rotated so that the image of the measurement object 20 overlaps the feature image. An edge portion is automatically detected at the stored measurement positions 62 to cause predetermined physical quantities to be automatically measured, and they are displayed as the measurement results 64 at the display positions 65 .
- the measurement object 20 even if the measurement object 20 is mounted at any angle on the stage 21 , an image of the measurement object 20 is rotated based on the information on relative positions for the image of the measurement object 20 , the information representing measurement positions and display positions of dimension line, stored in the measurement result display position storing unit 353 , and the measured measurement results can be displayed at the predetermined positions.
- necessary physical quantities can be automatically measured without newly receiving specification of a measurement position.
- the image measuring apparatus 1 includes the main body 2 and an external computer, and image data acquired by imaging in the main body 2 may be computed by the external computer to measure dimensions and the like of a desired shape.
- the control unit 3 may be built in the main body 2 .
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- Quality & Reliability (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
Description
- The present application claims foreign priority based on Japanese Patent Application No. 2009-013346, filed Jan. 23, 2009, the contents of which is incorporated herein by reference.
- 1. Field of the Invention
- The present invention relates to an image measuring apparatus and a computer program that allow measurement results to be displayed such that the measurement results are easily visible without depending on a state where a measurement object is mounted.
- 2. Description of the Related Art
- As an apparatus for measuring a shape of a measurement object, a number of image measuring apparatuses have been developed. Such an image apparatus applies light to a measurement object, acquires an image by image formation of transmitted light or reflected light of the applied light on an imaging device, such as a CCD (charge coupled device) and CMOS (complementary metal-oxide semiconductor), through a light receiving lens, and performs computing of the acquired image, thereby measuring the shape of the measurement object.
- The image of a measurement object becomes extremely accurately similar to the measurement object with the aid of a light receiving lens. By calibrating the dimensions of the image by using a magnification of the light receiving lens, the dimensions of an actual measurement object can be measured based on the dimensions of the image. Accordingly, accurately identifying the shape of a measurement object appearing on an image enables the shape of an actual measurement object to be accurately measured. The shape of a measurement object on an image is identified by detecting the boundary portion (hereinafter referred to as an “edge portion”) between the measurement object and the background image on the image.
FIG. 17 shows an example of a conventional edge detection area specified on an image.FIG. 18 shows an example of a shape identified by using a method of least squares based on conventional edge points.FIG. 19 is a schematic view for illustrating a circle obtained by fitting conventional edge points to a geometrical figure. - When detecting edges on an image, a periphery of an edge portion is surrounded by clicking or dragging it with a pointing device, such as a mouse, to specify an edge detection area as shown in
FIG. 17 . The edge portion is a part with a sharp change in luminance value between the pixel of a measurement object and the pixel of a background image. A computer acquires, for example, a part (between pixels) with a luminance difference between adjacent pixels larger than a predetermined value in image data in the specified area, as a plurality of edge points shown inFIG. 17 . The plurality of acquired edge points are fitted to a geometrical figure, such as a straight line, by using a regression analysis method, such as a method of least squares, as shown inFIG. 18 . For example, the straight line to which the edge points are fitted is detected as an edge. Even if points constitute a straight line, the points can be detected likewise. For a circular shape, like a straight line, a part with a sharp change in luminance value positioned between adjacent pixels is acquired as a plurality of edge points, and the plurality of acquired edge points are fitted to a geometrical figure by using a method of least squares or the like. For example, a circle to which the edge points are fitted can be detected as an edge as shown inFIG. 19 . - A size of a measurement object itself and a size of a measurement object part are various. Accordingly, if a measurement object having a hand-held size (size from several millimeters to about 100 mm square) is measured with a measuring device having the field of view of only a few millimeters square, the entire image of a measurement object cannot be collectively grasped. In this case, for example, an XY stage and the like are moved to relatively move a measurement object with respect to the field of view, so that a necessary portion is within the field of view, and thereafter measurement is performed, which is disclosed in Japanese Unexamined Patent Publication No. 11-132740.
- For example, when the distance between two straight lines that are not within the same field of view is measured, first, one straight line edge is specified, and then an XY stage is moved and the other straight line edge is specified. The distance between the specified straight line edges is measured by calculating from the position in the displayed image and the movement amount of the XY stage.
- It is difficult in the image measuring apparatus to grasp what measured value of what part of the measurement object is represented by the measurement result. Conventionally, for example as disclosed in Japanese Unexamined Patent Publication No. 11-132740, an image of a measurement object is displayed on a display screen, and dimension lines in accordance with measurement places specified for measurement of geometrical dimensions are also displayed. In this manner, a user can easily confirm the measurement result with the displayed geometrical dimensions.
- In conventional image measuring apparatuses including Japanese Unexamined Patent Publication No. 11-132740, if a measurement object is within the field of view, a user can easily grasp what measured value of what part of the measurement object is represented by the measurement result. However, if the measurement object is not within the field of view, the entire image of the measurement object cannot be displayed. This makes it difficult to grasp what part of the measurement object is represented by a measured value and how the part is measured for obtaining the measured value. Therefore, there has been a problem that it is difficult for a user to intuitively grasp a measured portion.
- For example, there is an apparatus in which a detected edge portion is displayed on a screen other than that for a measurement object, so that dimension lines and a measurement result are displayed on individual screens. However, the entire image of a measurement object cannot be displayed, which is the same as other conventional image measuring apparatuses. This apparatus is not different from other conventional image measuring apparatuses in that it is difficult to grasp what part of a measurement object is represented by a measured value and how the part is measured for obtaining the measured value if the shape of the measurement object is complicated.
- In view of such circumstances, an object of the present invention is to provide an image measuring apparatus and a computer program to allow a user to accurately grasp what part of a measurement object is represented by a measured value of a measurement result and to allow a user to visually confirm what part of the measurement object is designated by the measurement result without depending on the attitude, position and other conditions of an image of the measurement object.
- In order to achieve the above-described object, according to a first aspect of the present invention, there is provided an image measuring apparatus including: an imaging unit disposed facing a stage for mounting the measurement object and configured to acquire an image of a measurement object; a displaying unit configured to display the image of the measurement object acquired in the imaging unit; a feature image specification receiving unit configured to receive specification of any partial image on the displayed image as a feature image; a measurement position specification receiving unit configured to receive specification of a measurement position of the measurement object, on the displayed image; a relative position storing unit configured to store the feature image the specification of which is received in the feature image specification receiving unit, and information on a relative position for the feature image, the information representing the measurement position the specification of which is received in the measurement position specification receiving unit and a display position for displaying a dimension line indicating the measurement position; an attitude and position identifying unit configured to compare an image of the measurement object, the image being newly acquired by the imaging unit, with the stored feature image to identify information on an attitude and a position of the newly acquired image of the measurement object; an edge portion detecting unit configured to set a measurement position for the image of the measurement object, the image having the attitude and the position identified, based on the information on the relative position of the measurement position for the feature image, the information being stored in the relative position storing unit, and to detect an edge portion at the measurement position; a measuring unit configured to measure a predetermined physical quantity based on the detected edge portion; and a measurement result displaying unit configured to display the dimension line indicating the measurement position and a measurement result at a predetermined position based on the information on the relative position for the feature image stored in the relative position storing unit, the information representing the display position of the dimension line.
- According to a second aspect of the present invention, in the image measuring apparatus according to the first aspect, the relative position storing unit is configured to further store a display position for displaying the measurement result as the information on the relative position for the feature image, and the measurement result displaying unit is configured to display the measurement result at the predetermined position based on the information on the relative position of the display position for displaying the measurement result for the feature image, the information being stored in the relative position storing unit.
- According to a third aspect of the present invention, the image measuring apparatus according to the first or second aspect, further includes: a display position changing unit configured to receive a change of the position for displaying the dimension line and/or the measurement result.
- According to a fourth aspect of the present invention, the image measuring apparatus according to any one of the first to third aspects, further includes: a measurement result specification receiving unit configured to receive specification of the measurement result to be an object of determination of whether the measurement result is within a range of a predetermined tolerance; and a display manner changing unit configured to change a display manner of the measurement result the specification of which is received in the measurement result specification receiving unit.
- According to a fifth aspect of the present invention, the image measuring apparatus according to the fourth aspect, further includes a measurement result determining unit configured to determine whether the measurement result the specification of which is received in the measurement result specification receiving unit is within the range of the predetermined tolerance; and wherein the display manner changing unit is configured to change the display manner of the measurement result if it is determined in the measurement result determining unit that the measurement result is within the range of the predetermined tolerance.
- According to a sixth aspect of the present invention, the image measuring apparatus according to any one of the first to fifth aspects, further includes a display state determining unit configured to determine whether a portion displayed outside a display range exists in the measurement result; and wherein the measurement result displaying unit is configured to move the display position of the measurement result into the display range so that the measurement result is displayed within the display range, if it is determined in the display state determining unit that the portion displayed outside the display range exists.
- According to a seventh aspect of the present invention, the image measuring apparatus according to any one of the first to sixth aspects, further includes: a display angle detecting unit configured to detect an inclined angle of the display position of the measurement result; and an inclined angle determining unit configured to determine whether the detected inclined angle is equal to or greater than a predetermined angle; and wherein the measurement result displaying unit is configured to display the measurement result in an upside-down manner if it is determined in the inclined angle determining unit that the inclined angle is equal to or greater than the predetermined angle.
- According to an eighth aspect of the present invention, the image measuring apparatus according to the first aspect, further includes a measurement result display position storing unit configured to store information on a relative position for the image of the measurement object, the information representing a display position of the displayed dimension line indicating the measurement position and the measurement result, in association with the attitude and the position of the image of the measurement object; and wherein the measurement result displaying unit is configured to change and display the information on the attitude and the position of the image of the measurement object identified in the attitude and position identifying unit based on the information on the relative position for the image of the measurement object stored in the measurement result display position storing unit, the information representing the display position of the dimension line and the measurement result.
- In order to achieve the above-mentioned object, according to a ninth aspect of the present invention, there is provided a computer program executable with an image measuring apparatus, the computer program causing a computer to realize a function of the image measuring apparatus, the image measuring apparatus including: an imaging unit disposed facing a stage for mounting the measurement object and configured to acquire an image of a measurement object; a displaying unit configured to display the image of the measurement object acquired in the imaging unit; a feature image specification receiving unit configured to receive specification of any partial image on the displayed image as a feature image; a measurement position specification receiving unit configured to receive specification of a measurement position of the measurement object, on the displayed image; a relative position storing unit configured to store the feature image the specification of which is received in the feature image specification receiving unit, and information on a relative position for the feature image, the information representing the measurement position the specification of which is received in the measurement position specification receiving unit and a display position for displaying a dimension line indicating the measurement position; an attitude and position identifying unit configured to compare an image of the measurement object, the image being newly acquired by the imaging unit, with the stored feature image to identify information on an attitude and a position of the newly acquired image of the measurement object; an edge portion detecting unit configured to set a measurement position for the image of the measurement object, the image having the attitude and the position identified, based on the information on the relative position of the measurement position for the feature image, the information being stored in the relative position storing unit, and to detect an edge portion at the measurement position; a measuring unit configured to measure a predetermined physical quantity based on the detected edge portion; and a measurement result displaying unit configured to display the dimension line indicating the measurement position and a measurement result at a predetermined position based on the information on the relative position for the feature image stored in the relative position storing unit, the information representing the display position of the dimension line.
- According to a tenth aspect of the present invention, in the computer program according to the ninth aspect, the computer program causing the computer to realize the function of the image measuring apparatus, the relative position storing unit is configured to further store a display position for displaying the measurement result as the information on the relative position for the feature image, and the measurement result displaying unit is configured to display the measurement result at the predetermined position based on the information on the relative position of the display position for displaying the measurement result for the feature image, the information being stored in the relative position storing unit.
- According to an eleventh aspect of the present invention, in the computer program according to the ninth or tenth aspect, the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a display position changing unit configured to receive a change of the position for displaying the dimension line and/or the measurement result.
- According to a twelfth aspect of the present invention, in the computer program according to any one of the ninth to eleventh aspects, the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a measurement result specification receiving unit configured to receive specification of the measurement result to be an object of determination of whether the measurement result is within a range of a predetermined tolerance; and a display manner changing unit configured to change a display manner of the measurement result the specification of which is received in the measurement result specification receiving unit.
- According to a thirteenth aspect of the present invention, in the computer program according to the twelfth aspect, the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a measurement result determining unit configured to determine whether the measurement result the specification of which is received in the measurement result specification receiving unit is within the range of the predetermined tolerance; and wherein the display manner changing unit is configured to change the display manner of the measurement result if it is determined in the measurement result determining unit that the measurement result is within the range of the predetermined tolerance.
- According to a fourteenth aspect of the present invention, in the computer program according to any one of the ninth to thirteenth aspect, the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a display state determining unit configured to determine whether a portion displayed outside a display range exists in the measurement result; and wherein the measurement result displaying unit is configured to move the display position of the measurement result into the display range so that the measurement result is displayed within the display range, if it is determined in the display state determining unit that the portion displayed outside the display range exists.
- According to a fifteenth aspect of the present invention, in the computer program according to any one of the ninth to fourteenth aspects, the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a display angle detecting unit configured to detect an inclined angle of the display position of the measurement result; and an inclined angle determining unit configured to determine whether the detected inclined angle is equal to or greater than a predetermined angle; and wherein the measurement result displaying unit is configured to display the measurement result in an upside-down manner if it is determined in the inclined angle determining unit that the inclined angle is equal to or greater than the predetermined angle.
- According to a sixteenth aspect of the present invention, in the computer program according to the ninth aspect, the computer program causing the computer to realize the function of the image measuring apparatus, the image measuring apparatus further includes: a measurement result display position storing unit configured to store information on a relative position for the image of the measurement object, the information representing a display position of the displayed dimension line indicating the measurement position and the measurement result, in association with the attitude and the position of the image of the measurement object; and wherein the measurement result displaying unit is configured to change and display the information on the attitude and the position of the image of the measurement object identified in the attitude and position identifying unit based on the information on the relative position for the image of the measurement object stored in the measurement result display position storing unit, the information representing the display position of the dimension line and the measurement result.
- According to the first and ninth aspects of the present invention, an image of a measurement object that is obtained in an imaging unit is displayed, specification of any partial image on the displayed image as a feature image is received, and specification of a measurement position of the measurement object is received. The feature image the specification of which is received and information on a relative position for the feature image are stored. This information represents a measurement position and a display position for displaying a dimension line indicating the measurement position. An image of the measurement object newly acquired by the imaging unit is compared with the stored feature image to specify information on an attitude and a position of the newly acquired image of the measurement object. Based on the stored information on the relative position of the measurement position for the feature image, a measurement position is set for the image of the measurement object for which the attitude and the position are specified, and an edge portion at the measurement position is detected, and then a predetermined physical quantity is measured. Based on the stored information on the relative position for the feature image, the information representing the display position of the dimension line, a dimension line indicating the measurement position and a measurement result are displayed at a predetermined position. Storing the feature image, the specification of which is received, and the information on the relative position for the feature image, which represents the measurement position and the display position for displaying the dimension line indicating the measurement position, allows the image of the measurement object and the feature image to be compared. This makes it possible to determine the attitude and the position of the image of the measurement object, and the measurement position and the display position for displaying the dimension line indicating the measurement position as the relative position for the image of the measurement object. Accordingly, a user can visually confirm what portion of the measurement object corresponds to the measurement result without depending on the attitude, the position and the like of the image of the measurement object.
- According to the second and tenth aspects of the present invention, a display position for displaying the measurement result may be further stored as the information on the relative position for the feature image, and the measurement result may be displayed at the predetermined position based on the information on the relative position of the display position for displaying the measurement result for the feature image. Regarding display of the measurement result, its position can thus be changed according to an image of the measurement object without depending on the attitude, the position and the like of the image of the measurement object, and therefore a user can visually confirm what portion of the measurement object corresponds to the measurement result.
- According to the third and eleventh aspects of the present invention, a change of the position for displaying the dimension line and/or the measurement result may be received. Even when the displayed position of the dimension line and/or the measurement result is outside the field of view, the displayed position can thus be changed so as to be within the field of view. Therefore, a user can visually confirm what portion of the measurement object corresponds to the measurement result.
- According to the fourth and twelfth aspects of the present invention, specification of the measurement result to be an object of determination of whether the measurement result is within the range of a predetermined tolerance may be received, and the display manner of the measurement result the specification of which is received may be changed. The measurement result to be an object of determination of whether the measurement result is within the range of the predetermined tolerance can be easily confirmed with a difference of the display manner. Accordingly, it is possible to visually confirm the measurement result for which accuracy is required so as to be within the range of tolerance.
- According to the fifth and thirteenth aspects of the present invention, it may be determined whether the measurement result the specification of which is received is within the range of the predetermined tolerance, and the display manner of the measurement result may be changed if it is determined that the measurement result is within the range of the predetermined tolerance. Whether the measurement result is within a certain range considered to be appropriate can thus be easily confirmed with the display manner of the measurement result. Accordingly, an appropriate measurement result and an inappropriate measurement result can be visually confirmed. This makes it possible to easily confirm what portion of the measurement object has an inappropriate measurement result, and therefore remeasurement is made easier.
- According to the sixth and fourteenth aspects of the present invention, if a portion displayed outside a display range exists in the measurement result, the display position of the measurement result is moved into the display range so that the measurement result is displayed within the display range. This enables the display of the measurement result to be visually confirmed with reliability.
- According to the seventh and fifteenth aspects of the present invention, an inclined angle of the display position of the measurement result may be detected, and the measurement result may be displayed in an upside-down manner if the detected inclined angle is equal to or greater than a predetermined angle. This allows the measurement result to be visually confirmed even if the measurement object is mounted at any angle on the stage.
- According to the eighth and sixteenth aspects of the present invention, information on a relative position for the image of the measurement object, the information representing the displayed dimension line indicating the measurement position and the display position of the measurement result, may be stored in association with the attitude and the position of the image of the measurement object, and the identified information on the attitude and the position of the image of the measurement object may be changed and displayed based on the stored information on the relative position for the image of the measurement object, the information representing the dimension line and the display position of the measurement result. The measurement object can therefore be rotated and displayed in accordance with the display position based on the stored information on the dimension line and the position of the measurement result, even if the measurement object is mounted at any angle on the stage.
- With the above-described configuration, the feature image, the specification of which is received, and the information on the relative position for the feature image, which represents the measurement position and the display position for displaying the dimension line indicating the measurement position, are stored. The image of the measurement object and the feature image can therefore be compared. This makes it possible to determine the attitude and the position of the image of the measurement object, and the measurement position and the display position for displaying the dimension line indicating the measurement position as the relative position for the image of the measurement object. Accordingly, a user can visually confirm what portion of the measurement object corresponds to the measurement result without depending on the attitude, the position and the like of the image of the measurement object.
-
FIG. 1 is a perspective view showing a main body of an image measuring apparatus according to a first embodiment of the present invention; -
FIG. 2 is a schematic view showing a configuration of the image measuring apparatus according to the first embodiment of the present invention; -
FIG. 3 is a block diagram showing a configuration of a control unit of the image measuring apparatus according to the first embodiment of the present invention; -
FIG. 4 is a block diagram showing functions realized by the control unit of the image measuring apparatus according to the first embodiment of the present invention; -
FIG. 5 is a flow chart showing a procedure of a feature image storing process of a CPU of the control unit of the image measuring apparatus according to the first embodiment of the present invention; -
FIGS. 6A to 6C are schematic views showing a relationship between an image of a measurement object and a range of a field of view and a relationship between the measurement object and a display position of the measurement result; -
FIGS. 7A to 7C are schematic views showing a way of receiving specification of a feature image and measurement positions; -
FIG. 8 is a flow chart showing a procedure of an image measuring process of the CPU of the control unit of the image measuring apparatus according to the first embodiment of the present invention; -
FIGS. 9A and 9B show an example of a matching process between the feature image and an image of measurement object; -
FIGS. 10A to 10C are schematic views showing a way of displaying measurement positions and dimension lines; -
FIGS. 11A and 11B are schematic views for illustrating a way of changing display positions of a dimension line and/or a measurement result when the measurement result protrudes outside a display range; -
FIGS. 12A to 12D are schematic views for illustrating the ways of changing a display position when measurement results are displayed in an upside-down manner; -
FIG. 13 is a block diagram showing a configuration of a control unit of an image measuring apparatus according to a second embodiment of the present invention; -
FIG. 14 is a function block diagram showing functions realized by a control unit of the image measuring apparatus according to the second embodiment of the present invention; -
FIG. 15 is a flow chart showing a procedure of an image measuring process of a CPU of the control unit of the image measuring apparatus according to the second embodiment of the present invention; -
FIGS. 16A to 16D are schematic views showing the way of displaying dimension lines and measurement results; -
FIG. 17 shows an example of a conventional edge detection area specified on an image; -
FIG. 18 shows an example of a shape identified by using a method of least squares based on conventional edge points; and -
FIG. 19 is a schematic view for illustrating a circle obtained by fitting conventional edge points to a geometrical figure. - An image measuring apparatus according to embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
-
FIG. 1 is a perspective view showing amain body 2 of an image measuring apparatus according to a first embodiment of the present invention, andFIG. 2 is a schematic view showing a configuration of animage measuring apparatus 1 according to the first embodiment of the present invention. As shown inFIG. 2 , theimage measuring apparatus 1 according to the first embodiment includes themain body 2 and acontrol unit 3. Image data is obtained by image formation in themain body 2, and the obtained image data is computed in thecontrol unit 3, thereby measuring dimensions and the like of a desired shape of the image. - As shown in
FIG. 1 , themain body 2 includes astage 21 having a mountingsection 211. In the mountingsection 211, light is transmitted through the vicinity of the center portion. Thestage 21 is connected to themain body 2 with a supportingportion 212 interposed therebetween such that thestage 21 can be rotated around the mountingsection 211. By mounting ameasurement object 20 on the mountingsection 211 and receiving specification of a desired operation with anoperation panel 29, themeasurement object 20 is imaged by aimaging part 30 disposed above anarm 28, and the measurement result is displayed on adisplay 27. - As shown in
FIG. 2 , two sets of illuminatingsystems stage 21 for moving the mountedmeasurement object 20. The ring-shaped epi-illuminatingsystem 22, which illuminates themeasurement object 20 of thestage 21 from the above, is provided in a light receivinglens unit 23. Light applied by the epi-illuminatingsystem 22 is reflected from the surface of themeasurement object 20, and is returned to the light receivinglens unit 23. This permits irregularities, a pattern and the like of the surface of themeasurement object 20 to be imaged. - A
transmission illuminating system 24, which illuminates themeasurement object 20 from the below, is disposed under thestage 21. Thetransmission illuminating system 24 includes at least alight source 241, a reflectingmechanism 242 and alens 243. Light applied from thelight source 241 is reflected from the reflectingmechanism 242 toward thestage 21. Through thelens 243, the light is converted into parallel light rays in a direction approximately perpendicular to thestage 21. In this way, it is possible to perform imaging in which light is transmitted only through a position without themeasurement object 20. - The light receiving
lens unit 23 includes at least alight receiving lens 231, abeam splitter 232, a high-magnification-side imageformation lens part 233 and a low-magnification-side imageformation lens part 236. The high-magnification-side imageformation lens part 233 includes aslit 234 for image formation and a high-magnification-sideimage formation lens 235, and the low-magnification-side imageformation lens part 236 includes aslit 237 for image formation and a low-magnification-sideimage formation lens 238. Thebeam splitter 232 is a prism to cause light from thelight receiving lens 231 to branch in two directions. For example, cubic-type and plate-type beam splitters may be used. Light passing through a cubic-type beam splitter is never refracted, and therefore the optical axis does not deviate and alignment adjustment of a branch angle is easy. Thus, a cubic-type beam splitter is preferable compared to a plate-type beam splitter. -
FIG. 2 shows an example in which light applied from the epi-illuminatingsystem 22 guides light reflected from themeasurement object 20 and light applied from thetransmission illuminating system 24 and transmitted through themeasurement object 20 to the high-magnification-side imageformation lens part 233 and the low-magnification-side imageformation lens part 236. Light rays in two directions obtained by branching by thebeam splitter 232 are guided to both the low-magnification-side imageformation lens part 236 and the high-magnification-side imageformation lens part 233. - The high-magnification-
side imaging apparatus 25 performs image formation of light guided to the high-magnification-side imageformation lens part 233 using theimaging device 251, such as a CCD or CMOS, and transmits the resultant image as high magnification image data to thecontrol unit 3. Likewise, a low-magnification-side imaging apparatus 26 performs image formation of light guided to the low-magnification-side imageformation lens part 236 using animaging device 261, such as a CCD or CMOS, and transmits the resultant image as low magnification image data to thecontrol unit 3. With the above configuration of two-branch optical system using thelight receiving lens 231 and thebeam splitter 232, high magnification image data and low magnification image data can be simultaneously acquired without mechanically switching the optical system. Both high and low image data can be electronically switched and displayed on one screen, and can be individually displayed simultaneously on two screens. -
FIG. 3 is a block diagram showing the configuration of thecontrol unit 3 of theimage measuring apparatus 1 according to the first embodiment of the present invention. As shown inFIG. 3 , thecontrol unit 3 of theimage measuring apparatus 1 according to the first embodiment includes at least a CPU (central processing unit) 33, a storingdevice 35 such as a RAM (random access memory) 34 or a disk, acommunication device 36, and aninternal bus 37 that connects the hardware mentioned above. Through theinternal bus 37, thecontrol unit 3 is connected to amouse 32 and akeyboard 31, which are input devices, and adisplay device 27, which is an output device. - The
CPU 33 is connected through theinternal bus 37 to components as described above of hardware of thecontrol unit 3, and controls the operation of the components of hardware and executes various software functions in accordance with computer programs stored in thestoring device 35. TheRAM 34 is a volatile memory, such as an SRAM (static random access memory) or an SDRAM (synchronous dynamic random access memory), and a load module is expanded during execution of a computer program to store temporary data and the like generated during execution of the computer program. - Both of image data obtained by imaging using the high-magnification-
side imaging apparatus 25 and the low-magnification-side imaging apparatus 26 are stored in thestoring device 35 such that high-magnification image data and low-magnification image data are aligned each other. Each of the low-magnification image and the high-magnification image as used herein is obtained by causing light from the commonlight receiving lens 231 to branch in two directions with thebeam splitter 232, passing each branching light through the high-magnification-sideimage formation lens 235 or the low-magnification-sideimage formation lens 238, and forming an image on theimaging device - The storing
device 35 includes a featureimage storing unit 351 and a relativeposition storing unit 352. The featureimage storing unit 351 stores feature image data representing a feature for eachmeasurement object 20. Matching the stored feature image data with the acquired image data makes it possible to grasp what attitude and position are taken when themeasurement object 20 is mounted on thestage 21. The relativeposition storing unit 352 stores information on relative positions for a feature image, the information representing a measurement position the specification of which has been received and a display position of a dimension line for indicating the measurement position. Based on the stored information on relative positions, the relative positions for displaying the measurement position and the dimension line are determined upon identifying the attitude and the position of an image of themeasurement object 20. Needless to say, the featureimage storing unit 351 and the relativeposition storing unit 352 may be integrated for storing. - The
communication device 36 is connected to theinternal bus 37, and is connected through communication lines to high-magnification-side imaging apparatus 25 and the low-magnification-side imaging apparatus 26 to receive image data taken by theimaging apparatuses storing device 35 are downloaded from external computers through thecommunication device 36. -
FIG. 4 is a block diagram showing functions realized by thecontrol unit 3 of theimage measuring apparatus 1 according to the first embodiment of the present invention. InFIG. 4 , an imagedata acquiring unit 331 acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25. Both of the acquired image data are stored in thestoring device 35. - A displaying
unit 332 displays the acquired low-magnification image data or high-magnification image data on thedisplay 27; a feature imagespecification receiving unit 333 receives area specification to specify a partial image considered as an image that represents a portion characteristic of the image of themeasurement object 20 displayed on thedisplay 27. The way of receiving area specification is not particularly limited. For example, specification of a rectangular area including a partial image that is considered as an image representing a characteristic portion is received by a drag operation and the like of themouse 32 and the like. The partial image included in the rectangular area, the specification of which has been received, is stored as a feature image of themeasurement object 20 in the featureimage storing unit 351 of the storingdevice 35. - A measurement position
specification receiving unit 334 receives specification of a measurement position for the image of themeasurement object 20 displayed on thedisplay 27. The way of receiving specification of a measurement position is not particularly limited. For example, area specification to specify the vicinity of an edge portion is received by a drag operation and the like of themouse 32 and the like. Note that a tolerancespecification receiving unit 346 to receive specification of an appropriate range of a tolerance may be included. The way of receiving specification of an appropriate range of a tolerance is not particularly limited. For example, specification may be received by pressing keys of thekeyboard 31 from a setting window. - An attitude and
position identifying unit 335 performs a matching process (compare) between the acquired image data of themeasurement object 20 and the feature image stored in the featureimage storing unit 351 to identify information on the attitude and the position of the image of themeasurement object 20. An edgeportion detecting unit 336 sets a measurement position for the image of themeasurement object 20, for which the attitude and the position have been identified, based on information on the relative position of a measurement position for the feature image stored in the relativeposition storing unit 352, and detects an edge portion at the set measurement position. The measuringunit 337 measures predetermined physical quantities based on the detected edge portion and the identified information on the attitude and the position of the image of themeasurement object 20. - A measurement
result displaying unit 338 displays a measurement result and a dimension line indicating a measurement position at predetermined positions, based on information on the relative positions for the feature image stored in the relativeposition storing unit 352, the information representing the measurement position and a display position of the dimension line. Note that information on a relative position for displaying a measurement result for an image of themeasurement object 20 is also stored in the relativeposition storing unit 352, and the measurement result may be displayed at a predetermined position based on the attitude and the position of an image of themeasurement object 20 identified in the attitude andposition identifying unit 335 and the information on the relative position for displaying the measurement result for the image of themeasurement object 20. This allows the display position of the measurement result to be changed in accordance with the attitude and the position of the image of themeasurement object 20. - A display
position changing unit 339 receives a change in position at which a dimension line and/or a measurement result are displayed. The way of receiving a change in position for displaying is not particularly limited. For example, it is received by a drag operation and the like of themouse 32 and the like. - A measurement result
specification receiving unit 340 receives specification of a measurement result to be an object of determination of whether the measurement result is within a predetermined tolerance. The way of receiving specification of a measurement result is not particularly limited. For example, specification of a display position of a measurement result may be received, and specification of a measurement position for measuring a measurement result may also be received, by a click operation and the like of themouse 32 and the like. - A measurement
result determining unit 341 determines whether the measurement result, the specification of which has been received, is within a predetermined tolerance. A displaymanner changing unit 342 changes the display manner of the measurement result, the specification of which has been received in the measurement resultspecification receiving unit 340, and the measurement result that is determined in the measurementresult determining unit 341 to be within the tolerance. For example, changing the display manner, such as changing the display color and highlighting, allows a user to visually confirm the measurement result that is determined to be within the tolerance. - A display
state determining unit 343 determines whether a portion displayed outside a display range exists in a measurement result. If it is determined in the displaystate determining unit 343 that the portion displayed outside the display range exists in the measurement result, the measurementresult displaying unit 338 moves the display position of the measurement result into the display range so that the measurement result is displayed within the display range. Accordingly, a user can visually confirm the measurement result. - A display
angle detecting unit 344 detects an inclined angle of a display position of a measurement result; an inclinedangle determining unit 345 determines whether the inclined angle detected in the displayangle detecting unit 344 is equal to or greater than a predetermined angle. If it is determined in the inclinedangle determining unit 345 that the inclined angle is equal to or greater than the predetermined angle, the measurementresult displaying unit 338 displays the measurement result in an upside-down manner. -
FIG. 5 is a flow chart showing the procedure of a feature image storing process of theCPU 33 of thecontrol unit 3 of theimage measuring apparatus 1 according to the first embodiment of the present invention. TheCPU 33 of thecontrol unit 3 acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 (step S501). Both of the acquired image data are stored in thestoring device 35. - The
CPU 33 determines whether displaying the high-magnification image data within the range of the field of view is possible (step S502). If it is determined by theCPU 33 that displaying the high-magnification image data within the range of the field of view is possible (step S502: YES), then theCPU 33 displays the high-magnification image data on the display 27 (step S503); if it is determined by theCPU 33 that displaying the high-magnification image data within the range of the field of view is impossible (step S502: NO), theCPU 33 displays the low-magnification image data on the display 27 (step S504). -
FIGS. 6A to 6C are schematic views showing a relationship between an image of themeasurement object 20 and the range of the field of view and a relationship between themeasurement object 20 and the display position of the measurement result. When an image (image data) of themeasurement object 20 is acquired as shown inFIG. 6A , it is not clear what portion of themeasurement object 20 is displayed in the case where arange 61 of the field of view is too small for image data of themeasurement object 20 as shown inFIG. 6B , such as the case of displaying high-magnification image data. In contrast, the entire image of themeasurement object 20 is within therange 61 of the field of view as shown inFIG. 6C in the case where therange 61 of the field of view is sufficiently large for the image data of themeasurement object 20, such as the case of displaying low-magnification image data. - As shown in
FIG. 6C , specification of ameasurement area 62 of themeasurement object 20 whose entire image is displayed within therange 61 of the field of view is received by a drag operation and the like of themouse 32 and the like. This enables a user to specify themeasurement area 62 while viewing the entire image of themeasurement object 20. Accordingly, a user can specify the measurement area while visually confirming what portion is being measured. It is therefore possible to prevent a wrong portion from being measured. - Referring back to
FIG. 5 , theCPU 33 of thecontrol unit 3 receives specification of a rectangular area including a feature image considered as an image representing a characteristic portion on a low-magnification image or a high-magnification image (step S505). The way of receiving specification of a rectangular area including a feature image is not particularly limited. For example, specification of a display position of a measurement result may be received by a drag operation and the like of themouse 32 and the like, and a partial image included in the rectangular area, the specification of which has been received, may be stored as a feature image of themeasurement object 20 in the featureimage storing unit 351 of the storingdevice 35. - The
CPU 33 receives specification of a measurement position on the displayed low-magnification image or high-magnification image (step S506). The way of receiving specification of the measurement position is not particularly limited. For example, area specification of the vicinity of the edge portion is received by a drag operation and the like of themouse 32 and the like. Note that specification of an appropriate range of a tolerance may be received in step S506. The way of receiving specification of an appropriate range of a tolerance is not particularly limited. For example, specification may be received by pressing keys of thekeyboard 31 from a setting window. The following may be mentioned as another example. With the measurement result displayed, specification of a measurement position is received by a click operation and the like of themouse 32 and the like. Then, specification of an appropriate range of a tolerance is received for each measurement position the specification of which has been received. - The
CPU 33 stores the feature image, the specification of which has been received, in the featureimage storing unit 351; theCPU 33 stores information on relative positions for the feature image, which represents the measurement position the specification of which has been received and the display position of the dimension line indicating the measurement position, in the relative position storing unit 352 (step S507). -
FIGS. 7A to 7C are schematic views showing the way of receiving specification of a feature image and measurement positions.FIG. 7A shows an example of the way of receiving specification of arectangular area 71 including a feature image and measurement positions 72.FIG. 7B shows an example of a feature image the specification of which is received and stored in the featureimage storing unit 351.FIG. 7C shows an example of information on relative positions of the measurement positions 72, the specification of which is received, and display positions ofdimension lines 73 andmeasurement results 74, the specification of which is received and stored in the relativeposition storing unit 352. - Initially, receiving specification of the
rectangular area 71 as shown inFIG. 7A allows a feature image as shown inFIG. 7B to be cut out and stored in the featureimage storing unit 351. Not only image data of the feature image but also coordinates, a rotation angle and the like at a relative position of therectangular area 71 for an image of themeasurement object 20 are stored in the featureimage storing unit 351. - Next, receiving specification of the measurement positions 72 as shown in
FIG. 7A allows information on relative positions for the feature image, which represents the measurement positions and the display positions of the dimension lines 73 indicating the measurement positions, to be stored in the relativeposition storing unit 352 as shown inFIG. 7C . Not only the information on relative positions but also coordinates, rotation angles and the like at the relative positions for the image of themeasurement object 20, which represent the measurement positions 72 and the display positions of the dimension lines 73, are stored in the relativeposition storing unit 352. Note that measurement result display positions 74 at which displaying measurement results are displayed may be stored in the relativeposition storing unit 352 in the same way as those of other elements, as shown inFIG. 7C . -
FIG. 8 is a flow chart showing the procedure of an image measuring process of theCPU 33 of thecontrol unit 3 of theimage measuring apparatus 1 according to the first embodiment of the present invention. TheCPU 33 of thecontrol unit 3 newly acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 (step S801). Both of the acquired image data are stored in thestoring device 35. - The
CPU 33 determines whether displaying the high-magnification image data within the range of the field of view is possible (step S802). If it is determined by theCPU 33 that displaying the high-magnification image data within the range of the field of view is possible (step S802: YES), then theCPU 33 displays the high-magnification image data on the display 27 (step S803); if it is determined by theCPU 33 that displaying the high-magnification image data within the range of the field of view is impossible (step S802: NO), theCPU 33 displays the low-magnification image data on the display 27 (step S804). - The
CPU 33 performs a matching process (compare) between the newly acquired image data and the feature image stored in the feature image storing unit 351 (step S805) to identify information on the attitude and the position of an image of the measurement object 20 (step S806). TheCPU 33 sets a measurement position for the image of themeasurement object 20, for which the attitude and the position have been identified, based on the information on the relative position of a measurement position for the feature image stored in the relativeposition storing unit 352, and detects an edge portion of the set measurement position (step S807). -
FIGS. 9A and 9B show an example of a matching process between a feature image and an image of themeasurement object 20.FIG. 9A shows an example of a newly acquired image data, andFIG. 9B shows an example of a feature image stored in the featureimage storing unit 351. - When an image data of the
measurement object 20 is newly acquired as shown inFIG. 9A , the acquired image data is searched for a partial image that agrees with afeature image 91 shown inFIG. 9B . If the partial image that agrees with thefeature image 91 is found, therectangular area 71 for this partial image is identified. Coordinates, a relative rotation angle and the like of a relative position of therectangular area 71 for an image of themeasurement object 20 are stored in the featureimage storing unit 351 and coordinates, relative rotation angles and the like of relative positions for the image of themeasurement object 20, which represent the measurement positions 72 and display positions of the dimension lines 73, are stored in the relativeposition storing unit 352. Therefore, specifying therectangular area 71 allows accurate specification of what portion of themeasurement object 20 is a measurement position. A desired edge portion can thus be detected. - Referring back to
FIG. 8 , theCPU 33 of thecontrol unit 3 measures predetermined physical quantities based on the detected edge portion (step S808). The predetermined physical quantities to be measured include a length, a diameter, an angle and the like. TheCPU 33 displays a dimension line indicating a measurement position and a measurement result at predetermined positions based on the information on the relative position for the feature image, which represents a display position of the dimension line, stored in the relative position storing unit 352 (step S809). -
FIGS. 10A to 10C are schematic views showing the way of displaying measurement positions and dimension lines. As shown inFIG. 10A , predetermined physical quantities, such as a length, a diameter and an angle, measured for the measurement positions 62 are displayed as measurement results 64 at predetermined positions, based on the identified attitude and position of an image of themeasurement object 20 and information on relative positions for displayingdimension lines 63 for a feature image stored in the relativeposition storing unit 352. - Note that the measurement positions 62 and display positions of the dimension lines 63 are stored as relative positions for a feature image. Therefore, for example, even when the
measurement object 20 is mounted at a given position on thestage 21 as shown inFIG. 10B , the measurement positions 62 and display positions of the dimension lines 63 are moved and rotated correspondingly to the identified attitude and the position of themeasurement object 20, allowing the measurement results 64 to be displayed at positions at which the measurement results are easily visible, as shown inFIG. 10C . Accordingly, even if a user arbitrarily mountsmeasurement object 20 on thestage 21, the measurement positions 62 and the dimension lines 63 can be moved and rotated correspondingly to the attitude and the position of themeasurement object 20, and thus be displayed. The measurement results 64 measured at the storedmeasurement positions 62 can thus be displayed at positions at which they are easily visible. - Information on relative positions for displaying the measurement results 64 for the feature image may be stored in the relative
position storing unit 352, and theCPU 33 of thecontrol unit 3 may display the measurement results 64 at predetermined positions, based on the identified attitude and the position of the image of themeasurement object 20 and information on relative positions for displaying the measurement results 64 for the feature image. This allows display positions of the measurement results 64 to be changed in accordance with the attitude and the position of the image of themeasurement object 20. - When the display position of the
dimension line 63 and themeasurement result 64 are automatically identified in this way, themeasurement result 64 might protrude outside the display range depending on how themeasurement object 20 is mounted. If themeasurement result 64 protrudes outside the display range, a user cannot be aware of themeasurement result 64 and cannot visually confirm it. - To overcome this drawback, in the first embodiment, the
dimension line 63 and/or themeasurement result 64 are moved and displayed so that they are within the display range by receiving changes of positions for displaying thedimension line 63 and/or themeasurement result 64. This enables a user to visually confirm them. Of course, the following way may be used. TheCPU 33 determines whether a portion displayed outside the display range exists in themeasurement result 64. If it is determined that a portion displayed outside the display range exists, then theCPU 33 may move the display position of themeasurement result 64 to be within the display range.FIGS. 11A and 11B are schematic views for illustrating the way of changing display positions of thedimension line 63 and/or themeasurement result 64 if themeasurement result 64 protrudes outside the display range. - If the
measurement object 20 is mounted to be positioned at an upper position of the display screen as shown inFIG. 11A , thedimension line 63 indicating the length of an upper side of themeasurement object 20 and themeasurement result 64 are displayed in anarea 81 outside therange 61 of the field of view. A user cannot visually confirm them. To overcome this drawback, theCPU 33 determines whether thearea 81 for displayingdimension line 63 and themeasurement result 64 outside therange 61 of the field of view exist. If the existence is determined, then the display position is changed by decreasing the height of thedimension line 63 so that themeasurement result 64 is within therange 61 of the field of view as shown inFIG. 11B . Of course, a user may explicitly move the display position by a drag operation and the like of themouse 32 and the like. - Depending on how the
measurement object 20 is mounted on thestage 21, a case may occur where themeasurement result 64 is displayed in an upside-down manner to cause a measured value displayed as themeasurement result 64 to be illegible. To overcome this drawback, in the first embodiment, the inclined angle of a display position of the measurement results 64 is detected, and themeasurement result 64 is displayed in an upside-down manner if the detected inclined angle is equal to or greater than a predetermined angle. This enables themeasurement result 64 to be displayed as a measured value that can always be easily confirmed by a user.FIGS. 12A to 12D are schematic views for illustrating ways of changing a display position when themeasurement result 64 is displayed in an upside-down manner. - As shown in
FIG. 12A , for example, when an image of themeasurement object 20 rotates clockwise or counterclockwise by 90 degrees or more with respect to the stored feature image of themeasurement object 20, the numerical values of the measurement results 64 are displayed in an upside-down manner. This is a display manner that makes it difficult for a user to confirm the measurement results. To overcome this drawback, if theCPU 33 determines that the image of themeasurement object 20 rotates clockwise or counterclockwise by 90 degrees or more, the measurement results 64 are displayed in an upside-down manner as shown inFIG. 12B . Accordingly, a user can easily confirm the measurement results 64 visually. - The way to overcome the drawback is, of course, not limited to displaying the measurement results 64 in an upside-down manner. Any display manner may be employed as long as it allows the measurement results 64 to be displayed such that a user can easily confirm the measurement results. For example, as shown in
FIG. 12C , the measurement results 64 may be displayed in alignment in the transverse direction of a screen, not displayed in an upside-down manner. - Further, the image itself of the
measurement object 20 may be rotated to display the measurement results 64. That is, as shown inFIG. 12D , the relative position relationship is stored such that the image of themeasurement object 20 is integrated with the dimension lines 63 and the measurement results 64. If theCPU 33 determines that the image of themeasurement object 20 rotates clockwise or counterclockwise by 90 degrees or more with respect to the stored feature image of themeasurement object 20, the image, the dimension lines 63 and the measurement results 64 of themeasurement object 20 are integrally rotated until the image is in the same direction as that of the stored feature image. In this manner, the measurement results 64 can be displayed in such a display manner that the measurement results 64 are more easily confirmed. - In addition, regarding the measurement results 64 for which the
CPU 33 determines whether the measurement results 64 are within the range of a tolerance and theCPU 33 determines that they are within the range of the tolerance, the display manner of the measurement results 64 may be changed. For example, changing the display manner, such as changing the display color and highlighting, permits a user to visually confirm the measurement results 64 that has not been within the range of the tolerance. Note that for the measurement results 64 to be objects of determination of whether the measurement results 64 are within the range of the tolerance, it is to be understood that the display manner of the measurement results 64 may be changed. - As described above, according to the first embodiment, a feature image the specification of which has been received and information on relative positions for the feature image, the information representing a measurement position and a display position of a dimension line indicating the measurement position, are stored. This allows a matching process between an image of the
measurement object 20 and the feature image to be performed to determine the attitude and the position of the image of themeasurement object 20 and relative positions for the image of themeasurement object 20 that represent the measurement positions 62 and display positions of the dimension lines 63 indicating the measurement positions 62. Accordingly, a user can visually confirm what portions of themeasurement object 20 correspond to the measurement results 64 without depending on the attitude, the position and the like of the image of themeasurement object 20. - The configuration of the
image measuring apparatus 1 according to a second embodiment of the present invention is the same as that of the first embodiment, and therefore the components are denoted by the same reference numerals and the detailed description thereof will not be given. The second embodiment differs from the first embodiment in that information on display positions of dimension lines indicating the displayed measurement positions and measurement results is stored in association with the attitude and the position of an image of themeasurement object 20, and the image of themeasurement object 20 is displayed in accordance with the displayed dimension lines, the measurement results and measurement positions functioning as the basis of measurement. -
FIG. 13 is a block diagram showing the configuration of thecontrol unit 3 of theimage measuring apparatus 1 according to the second embodiment of the present invention. The hardware configuration of thecontrol unit 3 of theimage measuring apparatus 1 according to the second embodiment is the same as that of the first embodiment, as shown inFIG. 13 , and therefore the same components are denoted by the same reference numerals and the detailed description thereof will not be given. - The storing
device 35 includes a measurement result displayposition storing unit 353 in addition to the featureimage storing unit 351 and the relativeposition storing unit 352. The measurement result displayposition storing unit 353 stores information on relative positions for an image of themeasurement object 20, the information representing the attitude, the position, the measurement positions and display positions of dimension lines of the image of themeasurement object 20. Based on the stored information on relative positions, the relative positions for displaying the corresponding measurement positions and the dimension lines are determined upon identifying information on the attitude and the position of the image of themeasurement object 20. Needless to say, information on the relative position may be stored for the positions for displaying measurement results. -
FIG. 14 is a function block diagram showing functions realized by thecontrol unit 3 of theimage measuring apparatus 1 according to the second embodiment of the present invention. InFIG. 14 , the imagedata acquiring unit 331 acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25. Both of the acquired image data are stored in the featureimage storing unit 351 of the storingdevice 35. - The displaying
unit 332 displays the acquired low-magnification image data or high-magnification image data on thedisplay 27; the feature imagespecification receiving unit 333 receives area specification to specify a partial image considered as an image that represents a portion characteristic of an image of themeasurement object 20 displayed on thedisplay 27. The way of receiving area specification is not particularly limited. For example, specification of a rectangular area including a partial image that is considered as an image representing a characteristic portion is received by a drag operation and the like of themouse 32 and the like. The partial image included in the rectangular area, the specification of which has been received, is stored as a feature image of themeasurement object 20 in the featureimage storing unit 351 of the storingdevice 35. - The measurement position
specification receiving unit 334 receives specification of a measurement position for the image of themeasurement object 20 displayed on thedisplay 27. The way of receiving specification of a measurement position is not particularly limited. For example, area specification to specify the vicinity of an edge portion is received by a drag operation and the like of themouse 32 and the like. Note that the tolerancespecification receiving unit 346 to receive specification of an appropriate range of a tolerance may be included. The way of receiving specification of an appropriate range of a tolerance is not particularly limited. For example, specification may be received by pressing keys of thekeyboard 31 from a setting window. - The attitude and
position identifying unit 335 performs a matching process (compare) between the acquired image data of themeasurement object 20 and the feature image stored in the featureimage storing unit 351 to identify information on the attitude and the position of the image of themeasurement object 20. The edgeportion detecting unit 336 sets a measurement position for the image of themeasurement object 20, for which the attitude and the position have been identified, based on information on the relative position of a measurement position for the feature image stored in the relativeposition storing unit 352, and detects an edge portion of the set measurement position; the measuringunit 337 measures predetermined physical quantities based on the detected edge portion and the identified information on the attitude and the position of the image of themeasurement object 20. - The measurement
result displaying unit 338 displays a measurement result and a dimension line indicating a measurement position at predetermined positions, based on the information on the relative positions for the feature image stored in the relativeposition storing unit 352, the information representing the measurement position and a display position of the dimension line. Information on relative positions for themeasurement object 20 that represents display positions of a dimension line and a measurement result are stored in the measurement result displayposition storing unit 353 in association with the feature image of themeasurement object 20 being displayed. -
FIG. 15 is a flow chart showing the procedure of an image measuring process of theCPU 33 of thecontrol unit 3 of theimage measuring apparatus 1 according to the second embodiment of the present invention. TheCPU 33 of thecontrol unit 3 displays a dimension line indicating a measurement position and a measurement result at predetermined positions based on the information on the relative position for the feature image, which represents a display position of the dimension line, stored in the relative position storing unit 352 (step S809 inFIG. 8 ), and stores in the measurement result displayposition storing unit 353 the displayed dimension line, measurement result, and display position of the image of themeasurement object 20 as information on relative positions for the image of themeasurement object 20, the information representing the measurement position and the display position of the dimension line (step S1501). - The
CPU 33 newly acquires low-magnification image data of a low-magnification image taken by the low-magnification-side imaging apparatus 26 and high-magnification image data of a high-magnification image taken, on the same axis as that of the low-magnification image, by the high-magnification-side imaging apparatus 25 (step S1502). Both of the acquired image data are stored in the featureimage storing unit 351 of the storingdevice 35. - The
CPU 33 determines whether displaying the high-magnification image data within the range of the field of view is possible (step S1503). If it is determined by theCPU 33 that displaying the high-magnification image data within the range of the field of view is possible (step S1503: YES), then theCPU 33 displays the high-magnification image data on the display 27 (step S1504); if it is determined by theCPU 33 that displaying the high-magnification image data within the range of the field of view is impossible (step S1503: NO), theCPU 33 displays the low-magnification image data on the display 27 (step S1505). - The
CPU 33 performs a matching process (compare) between the newly acquired image data and the feature image stored in the feature image storing unit 351 (step S1506) to identify information on the attitude and the position of an image of the measurement object 20 (step S1507). TheCPU 33 sets a measurement position for the image of themeasurement object 20, for which the attitude and the position have been identified, based on the information on the relative position of a measurement position for the feature image stored in the relativeposition storing unit 352, and detects an edge portion of the set measurement position (step S1508). - The
CPU 33 measures predetermined physical quantities based on the detected edge portion (step S1509). The predetermined physical quantities to be measured include a length, a diameter, an angle and the like. TheCPU 33 displays a dimension line indicating a measurement position, a measurement result and the image of themeasurement object 20 at predetermined positions based on the information on relative positions for the image of themeasurement object 20, the information representing measurement position and a display position of the dimension line, stored in the measurement result display position storing unit 353 (step S1510). -
FIGS. 16A to 16D are schematic views showing the way of displaying dimension lines and measurement results. As shown inFIG. 16A , predetermined physical quantities measured, such as a length, a diameter and an angle, are displayed as measurement results 64 at predetermined positions, based on the attitude and position of the image of themeasurement object 20 and information on relative positions for displaying the measurement positions (measurement areas) 62 anddimension lines 63 for the image of themeasurement object 20 stored in the relativeposition storing unit 352. - Next, the displayed
dimension lines 63, measurement positions 62, and preferably displaypositions 65 of the measurement results 64 are stored as relative positions with the image of themeasurement object 20. For example, as shown inFIG. 16B , themeasurement object 20 is removed fromFIG. 16A , and the resultant is stored as a screen image excluding the display of the measurement results 64 in the measurement result displayposition storing unit 353. - Then, as shown in
FIG. 16C , when themeasurement object 20 is mounted at a given position on thestage 21, the attitude and the position of themeasurement object 20 are identified, and the image itself of themeasurement object 20 is moved and rotated in accordance with the corresponding feature image, and the dimension lines 63, the measurement positions 62, and preferably displaypositions 65 of the measurement results 64 that correspond to the feature image. That is, as shown inFIG. 16D , the image ofmeasurement object 20 is moved and rotated so that the image of themeasurement object 20 overlaps the feature image. An edge portion is automatically detected at the storedmeasurement positions 62 to cause predetermined physical quantities to be automatically measured, and they are displayed as the measurement results 64 at the display positions 65. - As described above, according to the second embodiment, even if the
measurement object 20 is mounted at any angle on thestage 21, an image of themeasurement object 20 is rotated based on the information on relative positions for the image of themeasurement object 20, the information representing measurement positions and display positions of dimension line, stored in the measurement result displayposition storing unit 353, and the measured measurement results can be displayed at the predetermined positions. For thesame measurement object 20, necessary physical quantities can be automatically measured without newly receiving specification of a measurement position. - It should be understood that the present invention is not limited to the first and second embodiments described above, and various modifications, replacements and the like may be made within the scope of the spirit of the present invention. For example, the
image measuring apparatus 1 includes themain body 2 and an external computer, and image data acquired by imaging in themain body 2 may be computed by the external computer to measure dimensions and the like of a desired shape. Further, it is to be understood thecontrol unit 3 may be built in themain body 2.
Claims (16)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009013346A JP5095644B2 (en) | 2009-01-23 | 2009-01-23 | Image measuring apparatus and computer program |
JP2009-013346 | 2009-01-23 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20100189308A1 true US20100189308A1 (en) | 2010-07-29 |
US8233665B2 US8233665B2 (en) | 2012-07-31 |
Family
ID=42354194
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/628,430 Active 2031-02-25 US8233665B2 (en) | 2009-01-23 | 2009-12-01 | Image measuring apparatus and computer program |
Country Status (3)
Country | Link |
---|---|
US (1) | US8233665B2 (en) |
JP (1) | JP5095644B2 (en) |
CN (1) | CN101832763B (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120194673A1 (en) * | 2011-02-01 | 2012-08-02 | Keyence Corporation | Dimension Measuring Apparatus, Dimension Measuring Method, And Program For Dimension Measuring Apparatus |
US20130141458A1 (en) * | 2011-12-02 | 2013-06-06 | Hon Hai Precision Industry Co., Ltd. | Image processing device and method |
US9122048B2 (en) | 2011-10-19 | 2015-09-01 | Keyence Corporation | Image processing apparatus and image processing program |
US20170316568A1 (en) * | 2016-04-28 | 2017-11-02 | Carl Zeiss Industrielle Messtechnik Gmbh | Method and apparatus for edge determination of a measurement object in optical metrology |
US9924912B2 (en) * | 2015-03-27 | 2018-03-27 | Vatech Co., Ltd. | X-ray image processing system and method |
EP3460754A4 (en) * | 2016-04-22 | 2020-02-26 | Nikon Corporation | Image measurement method, image measurement program, image measurement device, and object manufacture method |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5525953B2 (en) | 2010-07-29 | 2014-06-18 | 株式会社キーエンス | Dimension measuring apparatus, dimension measuring method and program for dimension measuring apparatus |
JP5597056B2 (en) | 2010-08-02 | 2014-10-01 | 株式会社キーエンス | Image measuring apparatus, image measuring method, and program for image measuring apparatus |
JP5679560B2 (en) | 2011-02-01 | 2015-03-04 | 株式会社キーエンス | Dimension measuring apparatus, dimension measuring method and program for dimension measuring apparatus |
US8570537B1 (en) | 2012-10-26 | 2013-10-29 | Nissan North America, Inc. | Method for bore chamfer measurement |
JP6325877B2 (en) * | 2014-04-18 | 2018-05-16 | 株式会社キーエンス | Optical coordinate measuring device |
JP2014197004A (en) * | 2014-04-21 | 2014-10-16 | 株式会社キーエンス | Image measurement device |
JP6348832B2 (en) * | 2014-12-12 | 2018-06-27 | ヤマハ発動機株式会社 | Component mounting apparatus, surface mounter, and component thickness detection method |
JP6474334B2 (en) | 2015-07-30 | 2019-02-27 | 株式会社キーエンス | Image inspection apparatus, image inspection method, and image inspection program |
WO2018117022A1 (en) * | 2016-12-20 | 2018-06-28 | 株式会社資生堂 | Application control device, application device, application control method, and recording medium |
JP7143057B2 (en) | 2016-12-28 | 2022-09-28 | 株式会社キーエンス | Three-dimensional measuring device |
JP7188870B2 (en) * | 2017-05-31 | 2022-12-13 | 株式会社キーエンス | Image inspection equipment |
US10885622B2 (en) | 2018-06-29 | 2021-01-05 | Photogauge, Inc. | System and method for using images from a commodity camera for object scanning, reverse engineering, metrology, assembly, and analysis |
JP7152223B2 (en) | 2018-08-31 | 2022-10-12 | 株式会社キーエンス | Image measuring device |
JP7154069B2 (en) | 2018-08-31 | 2022-10-17 | 株式会社キーエンス | Image inspection device |
JP7222764B2 (en) | 2019-03-18 | 2023-02-15 | 株式会社キーエンス | Image measuring device |
JP7252018B2 (en) | 2019-03-18 | 2023-04-04 | 株式会社キーエンス | Image measuring device |
JP7280810B2 (en) * | 2019-11-28 | 2023-05-24 | 株式会社キーエンス | Image measuring device |
JP7419164B2 (en) | 2020-05-29 | 2024-01-22 | 株式会社キーエンス | Image dimension measuring device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1132740A (en) * | 1997-07-22 | 1999-02-09 | Binshiyoo:Kk | Production of sesame processed product |
US20010040995A1 (en) * | 2000-05-10 | 2001-11-15 | Mitutoyo Corporation | Method and apparatus for generating part programs for use in image-measuring instruments, and image-measuring instrument and method of displaying measured results therefrom |
US20050048780A1 (en) * | 2003-06-13 | 2005-03-03 | Hitachi High-Technologies Corporation | Apparatus for inspecting three dimensional shape of a specimen and method of watching an etching process using the same |
US7411191B2 (en) * | 2000-06-27 | 2008-08-12 | Ebara Corporation | Inspection system by charged particle beam and method of manufacturing devices using the system |
US7499584B2 (en) * | 2004-10-21 | 2009-03-03 | Mitutoyo Corporation | Smear-limit based system and method for controlling vision systems for consistently accurate and high-speed inspection |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH03240879A (en) * | 1990-02-19 | 1991-10-28 | Nec Corp | Automatic dimension line generation system |
JPH07296023A (en) * | 1994-04-28 | 1995-11-10 | Hitachi Ltd | Dimension plotting method in cad system |
JPH10221054A (en) * | 1997-02-05 | 1998-08-21 | Shinko Electric Ind Co Ltd | Automatic generation apparatus for information on measuring place |
JP3887807B2 (en) | 1997-10-24 | 2007-02-28 | 株式会社ニコン | Image measuring device |
JP2000235594A (en) * | 1999-02-17 | 2000-08-29 | Toshiba Corp | Cad system and method for inspecting measured dimension value |
JP2002288633A (en) * | 2001-03-27 | 2002-10-04 | Matsushita Electric Works Ltd | Image processing device and its positional correction method |
JP2005017805A (en) * | 2003-06-27 | 2005-01-20 | Mitsutoyo Corp | Focus detecting method of image measuring device, focus detection mechanism, and image measuring device equipped with the focus detection mechanism |
JP4401126B2 (en) * | 2003-08-22 | 2010-01-20 | 株式会社日立国際電気 | Method for registering predetermined part of dimension measuring device |
JP2005190094A (en) * | 2003-12-25 | 2005-07-14 | Hitachi Ltd | Drawing editing method and device |
KR100752758B1 (en) * | 2005-10-19 | 2007-08-29 | (주) 인텍플러스 | Apparatus and method for measuring image |
CN101173853B (en) * | 2006-11-01 | 2011-02-02 | 鸿富锦精密工业(深圳)有限公司 | Positioning measurement method and device thereof |
JP5217193B2 (en) * | 2007-03-14 | 2013-06-19 | カシオ計算機株式会社 | Imaging apparatus, dimension measuring method, and program. |
-
2009
- 2009-01-23 JP JP2009013346A patent/JP5095644B2/en active Active
- 2009-12-01 US US12/628,430 patent/US8233665B2/en active Active
-
2010
- 2010-01-18 CN CN2010100028259A patent/CN101832763B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1132740A (en) * | 1997-07-22 | 1999-02-09 | Binshiyoo:Kk | Production of sesame processed product |
US20010040995A1 (en) * | 2000-05-10 | 2001-11-15 | Mitutoyo Corporation | Method and apparatus for generating part programs for use in image-measuring instruments, and image-measuring instrument and method of displaying measured results therefrom |
US7411191B2 (en) * | 2000-06-27 | 2008-08-12 | Ebara Corporation | Inspection system by charged particle beam and method of manufacturing devices using the system |
US20050048780A1 (en) * | 2003-06-13 | 2005-03-03 | Hitachi High-Technologies Corporation | Apparatus for inspecting three dimensional shape of a specimen and method of watching an etching process using the same |
US7230239B2 (en) * | 2003-06-13 | 2007-06-12 | Hitachi High-Technologies Corporation | Apparatus for inspecting three dimensional shape of a specimen and method of watching an etching process using the same |
US7499584B2 (en) * | 2004-10-21 | 2009-03-03 | Mitutoyo Corporation | Smear-limit based system and method for controlling vision systems for consistently accurate and high-speed inspection |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120194673A1 (en) * | 2011-02-01 | 2012-08-02 | Keyence Corporation | Dimension Measuring Apparatus, Dimension Measuring Method, And Program For Dimension Measuring Apparatus |
US9329024B2 (en) * | 2011-02-01 | 2016-05-03 | Keyence Corporation | Dimension measuring apparatus, dimension measuring method, and program for dimension measuring apparatus |
US9122048B2 (en) | 2011-10-19 | 2015-09-01 | Keyence Corporation | Image processing apparatus and image processing program |
US20130141458A1 (en) * | 2011-12-02 | 2013-06-06 | Hon Hai Precision Industry Co., Ltd. | Image processing device and method |
US9924912B2 (en) * | 2015-03-27 | 2018-03-27 | Vatech Co., Ltd. | X-ray image processing system and method |
EP3460754A4 (en) * | 2016-04-22 | 2020-02-26 | Nikon Corporation | Image measurement method, image measurement program, image measurement device, and object manufacture method |
US11195294B2 (en) | 2016-04-22 | 2021-12-07 | Nikon Corporation | Image measurement method, image measurement program, image measurement device, and object manufacture method |
US20170316568A1 (en) * | 2016-04-28 | 2017-11-02 | Carl Zeiss Industrielle Messtechnik Gmbh | Method and apparatus for edge determination of a measurement object in optical metrology |
US10360684B2 (en) * | 2016-04-28 | 2019-07-23 | Carl Zeiss Industrielle Messtechnik Gmbh | Method and apparatus for edge determination of a measurement object in optical metrology |
Also Published As
Publication number | Publication date |
---|---|
JP5095644B2 (en) | 2012-12-12 |
JP2010169584A (en) | 2010-08-05 |
CN101832763A (en) | 2010-09-15 |
US8233665B2 (en) | 2012-07-31 |
CN101832763B (en) | 2013-12-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8233665B2 (en) | Image measuring apparatus and computer program | |
US9959451B2 (en) | Image inspection device, image inspection method and image inspection program | |
US8994810B2 (en) | Magnification observation device | |
US8503758B2 (en) | Image measurement device, method for image measurement, and computer readable medium storing a program for image measurement | |
JP2004260785A (en) | Projector with distortion correction function | |
JP5997989B2 (en) | Image measuring apparatus, control method thereof, and program for image measuring apparatus | |
JP6726967B2 (en) | Brightness unevenness measuring device | |
JP5270971B2 (en) | Image measuring apparatus, image measuring method, and computer program | |
JP5074319B2 (en) | Image measuring apparatus and computer program | |
KR100499764B1 (en) | Method and system of measuring an object in a digital | |
US6304680B1 (en) | High resolution, high accuracy process monitoring system | |
KR20010113515A (en) | Device for detecting tilt angle of optical axis and image measuring apparatus equipped therewith | |
JP2010032330A (en) | Image measuring device and computer program | |
US11009691B2 (en) | Microscope apparatus and method for correcting position of light source | |
JP2010032329A (en) | Image measuring device and computer program | |
JP5531071B2 (en) | Image measuring apparatus and computer program | |
JP4163654B2 (en) | Scanning electron microscope | |
KR20100034039A (en) | Micro-dimension measuring method and measuring device | |
CN114518217B (en) | Method for determining center distance between lenses, microscope control device, and storage medium | |
Wolf et al. | 33‐1: Invited Paper: Metrological Challenges of Curved Displays | |
JP2014197004A (en) | Image measurement device | |
US6115450A (en) | X-ray fluorescence analyzer capable of determining the center of a sample | |
US7274471B2 (en) | Systems and methods for measuring distance of semiconductor patterns | |
CN111707446A (en) | Method and system for adjusting alignment of light spot center and detector receiving surface center | |
CN113891068B (en) | Optical axis precision detection method based on camera uniformity test auxiliary device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KEYENCE CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NAKATSUKASA, TAKASHI;REEL/FRAME:023585/0244 Effective date: 20090722 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |