US20170351920A1 - System and method - Google Patents
System and method Download PDFInfo
- Publication number
- US20170351920A1 US20170351920A1 US15/685,774 US201715685774A US2017351920A1 US 20170351920 A1 US20170351920 A1 US 20170351920A1 US 201715685774 A US201715685774 A US 201715685774A US 2017351920 A1 US2017351920 A1 US 2017351920A1
- Authority
- US
- United States
- Prior art keywords
- user
- wearable device
- display
- image
- light
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title description 43
- 230000008859 change Effects 0.000 claims description 14
- 230000003287 optical effect Effects 0.000 claims description 10
- 238000001514 detection method Methods 0.000 abstract description 43
- 238000004519 manufacturing process Methods 0.000 description 86
- 238000004891 communication Methods 0.000 description 66
- 238000013523 data management Methods 0.000 description 28
- 230000006870 function Effects 0.000 description 28
- 238000012423 maintenance Methods 0.000 description 23
- 239000010410 layer Substances 0.000 description 15
- 238000007726 management method Methods 0.000 description 12
- 230000007423 decrease Effects 0.000 description 11
- 230000007613 environmental effect Effects 0.000 description 8
- 239000011521 glass Substances 0.000 description 8
- 210000003128 head Anatomy 0.000 description 8
- 230000001133 acceleration Effects 0.000 description 7
- 239000012790 adhesive layer Substances 0.000 description 7
- 238000003745 diagnosis Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 210000001508 eye Anatomy 0.000 description 5
- 238000004020 luminiscence type Methods 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 230000008439 repair process Effects 0.000 description 5
- 239000000853 adhesive Substances 0.000 description 3
- 210000000554 iris Anatomy 0.000 description 3
- 239000000725 suspension Substances 0.000 description 3
- 239000002390 adhesive tape Substances 0.000 description 2
- 208000003464 asthenopia Diseases 0.000 description 2
- 210000005252 bulbus oculi Anatomy 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 239000011368 organic material Substances 0.000 description 2
- 238000010248 power generation Methods 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 208000019695 Migraine disease Diseases 0.000 description 1
- 230000001070 adhesive effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000013013 elastic material Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 210000004209 hair Anatomy 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 206010027599 migraine Diseases 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000284 resting effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000012780 transparent material Substances 0.000 description 1
- 238000012384 transportation and delivery Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G06K9/00671—
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/169—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated pointing device, e.g. trackball in the palm rest area, mini-joystick integrated between keyboard keys, touch pads or touch stripes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1694—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a single or a set of motion sensors for pointer control or gesture input obtained by sensing movements of the portable computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61K—PREPARATIONS FOR MEDICAL, DENTAL OR TOILETRY PURPOSES
- A61K39/00—Medicinal preparations containing antigens or antibodies
- A61K2039/51—Medicinal preparations containing antigens or antibodies comprising whole cells, viruses or DNA/RNA
- A61K2039/525—Virus
- A61K2039/5254—Virus avirulent or attenuated
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61K—PREPARATIONS FOR MEDICAL, DENTAL OR TOILETRY PURPOSES
- A61K39/00—Medicinal preparations containing antigens or antibodies
- A61K2039/70—Multivalent vaccine
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61K—PREPARATIONS FOR MEDICAL, DENTAL OR TOILETRY PURPOSES
- A61K39/00—Medicinal preparations containing antigens or antibodies
- A61K39/12—Viral antigens
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61K—PREPARATIONS FOR MEDICAL, DENTAL OR TOILETRY PURPOSES
- A61K9/00—Medicinal preparations characterised by special physical form
- A61K9/14—Particulate form, e.g. powders, Processes for size reducing of pure drugs or the resulting products, Pure drug nanoparticles
- A61K9/19—Particulate form, e.g. powders, Processes for size reducing of pure drugs or the resulting products, Pure drug nanoparticles lyophilised, i.e. freeze-dried, solutions or dispersions
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B65—CONVEYING; PACKING; STORING; HANDLING THIN OR FILAMENTARY MATERIAL
- B65B—MACHINES, APPARATUS OR DEVICES FOR, OR METHODS OF, PACKAGING ARTICLES OR MATERIALS; UNPACKING
- B65B3/00—Packaging plastic material, semiliquids, liquids or mixed solids and liquids, in individual containers or receptacles, e.g. bags, sacks, boxes, cartons, cans, or jars
- B65B3/003—Filling medical containers such as ampoules, vials, syringes or the like
-
- C—CHEMISTRY; METALLURGY
- C12—BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
- C12N—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA
- C12N2770/00—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA ssRNA viruses positive-sense
- C12N2770/00011—Details
- C12N2770/24011—Flaviviridae
- C12N2770/24111—Flavivirus, e.g. yellow fever virus, dengue, JEV
- C12N2770/24134—Use of virus or viral component as vaccine, e.g. live-attenuated or inactivated virus, VLP, viral protein
-
- C—CHEMISTRY; METALLURGY
- C12—BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
- C12N—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA
- C12N2770/00—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA ssRNA viruses positive-sense
- C12N2770/00011—Details
- C12N2770/24011—Flaviviridae
- C12N2770/24111—Flavivirus, e.g. yellow fever virus, dengue, JEV
- C12N2770/24151—Methods of production or purification of viral material
-
- C—CHEMISTRY; METALLURGY
- C12—BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
- C12N—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA
- C12N2770/00—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA ssRNA viruses positive-sense
- C12N2770/00011—Details
- C12N2770/24011—Flaviviridae
- C12N2770/24111—Flavivirus, e.g. yellow fever virus, dengue, JEV
- C12N2770/24161—Methods of inactivation or attenuation
- C12N2770/24164—Methods of inactivation or attenuation by serial passage
-
- C—CHEMISTRY; METALLURGY
- C12—BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
- C12N—MICROORGANISMS OR ENZYMES; COMPOSITIONS THEREOF; PROPAGATING, PRESERVING, OR MAINTAINING MICROORGANISMS; MUTATION OR GENETIC ENGINEERING; CULTURE MEDIA
- C12N7/00—Viruses; Bacteriophages; Compositions thereof; Preparation or purification thereof
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A50/00—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE in human health protection, e.g. against extreme weather
- Y02A50/30—Against vector-borne diseases, e.g. mosquito-borne, fly-borne, tick-borne or waterborne diseases whose impact is exacerbated by climate change
Definitions
- Embodiments described herein relate generally to a system and a method using an eyeglasses-type wearable device.
- a checklist showing a workflow of each work step.
- a system for supporting an operator by using a head-mounted display with a built-in camera there is a medical-equipment management system which supports an operator of a medical device such as a used and contaminated endoscope or a piece of medical equipment such as a scalpel or forceps.
- the system includes a head-mounted camera for capturing an image of the sight of an operator of a medical device or a piece of medical equipment; storage means for storing an image of the sight of an operator captured when the operator demonstrates the medical device or the medical equipment as a reference image; first determination means for comparing the image captured by the camera and the reference image read from the storage means in order to determine whether a predetermined operation is performed normally based on similarity between these two images; data output means for outputting data indicating an alarm or an instruction based on a determination result of the determination means; and output means for outputting the alarm or the instruction to the operator based on the data indicating the alarm or the instruction.
- the system automatically recognizes the operation of the medical equipment operator by comparing the operation of the operator captured by the camera with the reference image prepared in advance.
- this method since the recognition accuracy of simple pattern match between images is low, complicated image processing such as feature extraction from the images is further required. Consequently, as the automatic recognition processing becomes highly complicated, a considerable amount of time will be required for the image processing.
- the present embodiment aims to provide a system and a method for recognizing movements of a user of a wearable device quickly, simply and accurately and displaying content to support the user based on a recognition result.
- FIG. 1 is a perspective diagram showing an example of a wearable device of an embodiment.
- FIG. 2A shows a front view of an example of the wearable device.
- FIG. 2B shows a cross-section structure of an example of the wearable device.
- FIG. 3 shows an example of position detection of the wearable device.
- FIG. 4 exemplarily shows the principle of the position detection of the wearable device.
- FIGS. 5A, 5B, and 5C show an example of operation periods of the wearable devices.
- FIGS. 5D, 5E, 5F, and 5G show an example of communication periods of the photo detectors.
- FIGS. 5H, 51, 5J, and 5K show an example of signal waveforms of the photo detectors.
- FIG. 6 shows an example of a system including the wearable device and a data management server.
- FIG. 7 exemplarily shows an electrical configuration of the wearable device.
- FIG. 8A shows an example of detection of the state of an apparatus.
- FIG. 8B shows an example of a sensor for detecting the state of an apparatus.
- FIG. 9 shows an example of a sensor for detecting movements of a user.
- FIGS. 10A and 10B show an example of a usage environment of the system.
- FIG. 11 is an exploded view showing an example of the structure of the sensor used in the system and for detecting movements of a user.
- FIG. 12 is an exploded view showing another example of the structure of the sensor used in the system and for detecting movements of a user.
- FIG. 13A shows an example of a workflow displayed by the system.
- FIG. 13B shows an example of a work record made by the system.
- a system in general, includes a wearable device on a head of a user and including a display in a line of vision of the user, a first detector configured to detect a movement of the user, a second detector configured to detect a state of an apparatus operated by the user, and a server connected to the wearable device, the first detector and the second detector.
- the server is configured to display information about work contents of the user on the display based on a detection result of the first detector and a detection result of the second detector.
- Wearable devices include head-mounted type wearable devices (such as eyeglasses, goggles and helmet types which may also be called an eyeglasses type collectively), wristband-type wearable devices, pendant-type wearable devices and the like.
- head-mounted type wearable devices such as eyeglasses, goggles and helmet types which may also be called an eyeglasses type collectively
- wristband-type wearable devices such as eyeglasses, goggles and helmet types which may also be called an eyeglasses type collectively
- pendant-type wearable devices and the like.
- the following description is based on the assumption that the wearable device of the present embodiment is an eyeglasses-type wearable device.
- Eyeglasses-type wearable devices include optical head-mounted displays, which allow the user to see through their transparent lenses, and non-optical head-mounted displays, which block the view of the user and do not allow the user to see through them. In the following, optical head-mounted displays, which allow the user to see through them, will be taken as an example.
- FIG. 1 is a perspective view of an eyeglasses-type wearable device (hereinafter referred to simply as a wearable device) 10
- FIG. 2A is a front view
- FIG. 2B is a diagram showing a cross-section structure viewed from above.
- the wearable device 10 has a shape substantially the same as that of an ordinary pair of glasses, but here a projector 12 is attached to the outside of the right-eye temple. Glasses 14 and 16 are set in the frame.
- the left-eye glass 14 is a normal transparent glass so that the user can see through the glass.
- the right-eye glass 16 is at least partly a screen 16 .
- the screen 16 is configured to show an image projected by the projector 12 to the user. When the projector 12 is not projecting an image, the screen 16 is transparent and allows the user to see through the right-eye glass (screen) 16 .
- the projector 12 includes a power supply 22 and a controller 24 as electronic components.
- the power supply 22 includes a button battery, a rechargeable battery, a non-contact power supply secondary battery and the like.
- the projector 12 may not include a built-in battery but may be supplied with power from an external power supply via a power-supply line or a wireless channel.
- the controller 24 is configured to perform a communication with a server or another electronic device via a network which will be described later and thereby transmit and receive data. This communication may be a wired communication or may be a wireless communication.
- Bluetooth registered trademark
- ZigBee registered trademark
- a short-range wireless communication such as UWB
- a medium-range wireless communication such as WiFi (registered trademark)
- a long-range wireless communication such as 3G/4G or WiMAX (registered trademark)
- 3G/4G or WiMAX registered trademark
- the projector 12 further includes a light source 28 , a display 30 , a prism 32 , a set of lenses 34 and the like as optical components.
- the light source 28 may be a dimming white LED light source having several, for example, three light emitting diodes having luminescent colors different from each other and amounts of output light respectively variable. According to the dimming white LED light source, even if the wearable device 10 is used in such an environment as a clean room using light having a luminescent color consisting principally of orange, a clear projection image can be obtained by changing the luminescent color of the LED light source based on the usage environment.
- the dimming white LED light source it is possible to output a display color easy for the user to see, and thus as compared to the case of outputting a display color difficult for the user to see, the causes of troubles to the user such as eye strain and migraine associated with eye strain can be prevented.
- the display 30 is, for example, a reflective liquid crystal display (LCD) module and configured to display a predetermined text, image and the like (hereinafter referred to also as a display image collectively) based on display control executed by the controller 24 .
- Non-parallel light (hereinafter referred to also as diverging light) output from the light source 28 is reflected on a half mirror surface 32 a of the prism 32 , and thereby illuminates a display image of the display 30 .
- the reflected light of the display 30 is, after passing through the half mirror surface 32 a as light indicative of the display image (hereinafter referred to also as image light), output from the outgoing surface 32 c and then projected on the screen 16 as a projection image in a predetermined size via the set of lenses 34 .
- the screen 16 includes a near-side transparent refractor 42 , a Fresnel-lens-type half mirror surface 44 and a back-side transparent refractor 46 .
- the image light reaching the half mirror surface 44 is partly reflected on the half mirror surface 44 and forms a visual image (projection image) indicative of the display image of the display 30 at a few meters away. Note that, since the screen 16 allows the user to partly see through the screen 16 , it is also possible to configure the screen 16 to show the projection image as well as the view in front of the user.
- a part of the image light (diverging light) output from the light source 28 and passing thorough the half mirror surface 32 a is totally reflected on the total-reflection surface 32 b and becomes leaking light 50 of the diverging light from the light source 28 refracted in the outgoing surface 32 c.
- the leaking light 50 is output in a direction different from the direction of the screen 16 through an opening or a gap (leading portion) 52 formed on the front side of the projector 12 .
- the wearable device 10 includes a speaker 54 A, an earphone jack 54 B, a microphone jack 56 , a sliding switch 57 , a rotating switch 58 and the like in a predetermined portion, for example, in a bottom portion of the projector 12 .
- the microphone jack 56 is connected to a hands free microphone (not shown in the drawing) and collects the user's voice.
- the sliding switch 57 is configured, for example, to adjust the brightness, color tone and the like of the projection image of the projector 12 .
- the rotating switch 58 is configured, for example, to adjust the projection angle and the like of the projection image.
- the user can adjust the projection image by performing touch operations while looking at the projection image.
- the sliding switch 57 it is possible to provide the projection image having the display brightness and color tone of the user's taste.
- the rotation switch 58 it is possible to adjust the projection angle so that the projection image is displayed in the most appropriate position based on the shape or size of the user's head.
- the objects to be adjusted by the sliding switch 57 and the rotating switch 58 may be opposite to each other, the positions of the sliding switch 57 and the rotating switch 58 may be opposite to each other, or their functions may be assigned to a single operation member configured to perform two kinds of operations.
- a menu item may not be selected by an operation on the switch 57 or 58 but may be selected by a touch operation. Therefore, a touchpad 55 is further provided on the outside of the projector 12 .
- the user can input an operation easily and efficiently by touching a position of the touchpad 55 corresponding to the display position of an item in the menu.
- a camera 59 is provided in the center front on the outside and configured to capture an image of the front view (still image and moving image). Note that, although not shown in the drawing, it is possible to provide another camera in the center front on the inner side to face the user and configure to capture the eyeballs of the user to detect the irises of the user.
- the irises can be used for user authentication.
- the state of the wearable device 10 that is, the state of the user can be detected.
- the state includes a position, a shift of the position and the like.
- FIG. 3 An example of the use of the wearable device is shown in FIG. 3 .
- a work area 60 of, for example, a component yard of a plant, a product warehouse of a mail-order firm or a delivery department of a retailer, a given number of work spaces or product shelves A 01 to Axy (x and y are both positive integers), B 01 to Bxy and C 01 to Cxy are arranged.
- the work spaces or the product shelves may be, for example, work tables in a plant, manufacturing apparatuses in a production line, desks at school, seating positions in a conference room, and the like.
- At least one photo detector 62 - 1 to 62 -n (n is a positive integer) is installed.
- the at least one photo detector 62 - 1 to 62 -n is configured to detect the positions (x, y, z), the numbers, the shifts of the positions, the changes of the directions and the like of the wearable devices 10 - 1 to 10 -m (m is a positive integer) respectively by a detection method shown in FIGS. 4 and 5A-5K .
- the states such as the positions and the shifts of the positions of a given number of the users of the wearable devices 10 - 1 to 10 -m can be recognized.
- the users can move around the work area 60 freely.
- the users perform predetermined work in predetermined positions, namely, work spaces 64 such as stations (carts), or containers or movable tables equivalent thereto.
- work spaces 64 such as stations (carts), or containers or movable tables equivalent thereto.
- the work space 64 is not necessarily movable but may be a fixed desk, a seating position or the like.
- a detection system includes at least one wearable device 10 and at least one photo detector 62 .
- the photo detector 62 has a function of detecting the leaking light 50 and a function of performing communication to transmit a detection result to a server or the like.
- the communication function may be a wired communication function or may be a wireless communication function as in the case of the communication function of the wearable device 10 .
- Bluetooth registered trademark
- ZigBee registered trademark
- a short-range wireless communication such as UWB
- a medium-range wireless communication such as WiFi (registered trademark)
- a long-range wireless communication such as 3G/4G or WiMAX (registered trademark)
- 3G/4G or WiMAX registered trademark
- Bluetooth registered trademark
- ZigBee registered trademark
- a short-range wireless communication such as UWB
- a medium-range wireless communication such as WiFi (registered trademark)
- a long-range wireless communication such as 3G/4G or WiMAX (registered trademark)
- 3G/4G or WiMAX registered trademark
- the wearable device 10 intermittently modulates the leaking light 50 by using data including identification data of the device (hereinafter referred to also as a device ID) so that the photo detector 62 can identify the wearable device 10 based on the received leaking light 50 .
- a typical example of the modulation method is a chopper modulation method of decreasing an amount of luminescence to zero
- the following description is based on the assumption that the wearable device 10 adopts a modulation method of ensuring a predetermined or more amount of luminescence even in the case of light having a small amount of luminescence. In this way, the strain on the user's eyes can be reduced.
- DSV digital sum value
- the human eyes can perceive a change up to about 0.02 second, it is possible to achieve the effect of reducing the strain of the user's eyes by setting the reference frequency of the above-described modulation to, for example, greater than or equal to 20 Hz, more preferable, greater than or equal to 60 Hz.
- the LED used for the light source 28 has an internal impedance and a connecting capacity, the frequency of less than 100 MHz, more preferably, less than or equal to 10 MHz is desirable for performing highly-accurate modulation. From the above, it is desirable that the modulation frequency of the light source 28 used in the detection system of the present embodiment be 10 Hz to 100 MHz, more preferable, 10 Hz to 10 MHz.
- the amount of light detected by the photo detector 62 varies depending on the distance between the wearable device 10 and the photo detector 62 .
- the distance between the wearable device 10 and the photo detector 62 or the direction of the wearable device 10 with respect to the photo detector 62 can be obtained. Since the position (including the level) of the photo detector 62 is fixed, as the distance between the photo detector 62 and the wearable device 10 is obtained, the position of the wearable device 10 (x, y, z) can be detected accordingly.
- the leaking light 50 of the diverging light from the light source 28 detection of the leaking light 50 can be performed in a relatively wide area.
- the positions of the wearable devices 10 - 1 to 10 -m in the work area 60 the distances between the wearable devices 10 and the photo detectors 62 , the directions of the wearable devices 10 , or the directions of the wearable devices 10 with respect to the photo detectors 62 can be detected. Consequently, the installation cost required for installing the detection system can be reduced.
- the data of amounts of the leaking light 50 detected by the photo detectors 62 is transmitted from the photo detectors 62 to a server which will be described later at a predetermined time.
- the server analyzes the data collected from the photo detectors 62 . In this way, the positions and the states of the desired wearable devices 10 - 1 to 10 -m, more specifically, the states of the users can be detected.
- FIG. 4 is a schematic diagram showing a specific example of the use of the system for recognizing the wearable device of the embodiment. The following description is based on the assumption that there are three users wearing wearable devices 10 - 1 to 10 - 3 around four photo detectors 62 - 1 to 62 - 4 .
- the leaking light 50 from the wearable devices 10 - 1 and 10 - 2 is detected by the photo detectors 62 - 1 to 60 - 4 .
- the photo detectors 62 - 1 to 62 - 4 perform analog-to-digital conversion of the amounts of the leaking light 50 detected respectively and transmit them to a server as light amount data indicative of the amounts of light at a predetermined time, for example, by a short-range wireless communication.
- the following description is based on the assumption that the position of the wearable device 10 - 1 is shifted toward the photo detector 62 - 1 as the user moves toward the photo detector 62 - 1 and meanwhile the direction of the wearable device 10 - 2 is temporarily changed as the user makes a given movement such as turning of the head (rotating of the head).
- the changes in the detection data occurring at this time is shown in FIGS. 5A-5K .
- FIGS. 5A-5K illustrate the case of using an intermittent time-shift method as a modulation method of the leaking light 50 of the wearable devices 10 - 1 to 10 - 3 . That is, ID modulation times are set respectively to the wearable devices 10 - 1 to 10 - 3 in a staggered manner.
- intermittent ID modulation times are set respectively to the first to third wearable devices 10 - 1 to 10 - 3 , and the other times are set as non-modulation times.
- a synchronization signal SYNC is paired with each of the device IDs of the wearable devices 10 - 1 to 10 - 3 (on a one-to-one basis), and the pairs are repeated for several times (multiples of four times in the case where there are four sensors as shown in FIGS. 5D-5G ).
- the ID modulation time of the second wearable device 10 - 2 starts.
- the ID modulation time of the third wearable device 10 - 3 starts.
- the synchronization signal SYNC and the device ID of the wearable device 10 - 2 or 10 - 3 are repeatedly modulated.
- the device ID of the wearable device 10 By superimposing the device ID of the wearable device 10 on a modulation signal in this way, the device ID can be detected.
- the modulation times of the respective wearable devices 10 - 1 to 10 - 3 are set on a time-division basis (on an intermittent basis). However, for example, it is also possible to perform modulation successively for all the wearable devices 10 - 1 to 10 - 3 and change the modulation reference frequencies of the wearable devices 10 - 1 to 10 - 3 respectively. Further, it is also possible to change the characteristics of the frequency spectrums in spread spectrum, respectively.
- each ID modulation time is divided into sections of the data communication times (COMs) with the photo detectors 62 - 1 to 62 - 4 .
- the photo detector 62 - 4 detects the leaking light from the wearable device 10 - 1 at the beginning.
- the modulation signal amplitude of the leaking light from the wearable device 10 - 1 detected by the photo detector 62 - 4 decreases.
- the modulation signal amplitude of the leaking light from the wearable device 10 - 1 detected by the photo detector 62 - 1 increases as time advances.
- the changes (shifts) of the positions of the detection targets By comparing the changes of the modulation signal amplitudes detected by the photo detectors 62 - 1 to 62 -n with time, the changes (shifts) of the positions of the detection targets, namely, the wearable devices 10 - 1 to 10 -m with time can be detected.
- the wearable device 10 - 2 is directed to the photo detector 62 - 3 at the beginning, and thus with regard to the modulation signal amplitude of the leaking light, the detection value of the photo detector 62 - 3 is greater than the detection value of the photo detector 62 - 2 .
- the detection output of the wearable device 10 - 2 output from the photo detector 62 - 2 temporarily increases and then decreases as shown in FIG. 51 .
- the detection output of the wearable device 10 - 2 output from the photo detector 62 - 3 temporarily decreases and then increases as shown in FIG. 5J .
- the position or state of the wearable device 10 is detected by executing comparative processing of device identification data output from a number of wearable devices 10 as modulated light and received by a number of photo detectors 62 .
- the wearable device 10 which receives a beacon to be in a position substantially the same as the position of the transmitter having transmitted that beacon. Further, it is also possible to detect the position of a wearable device by using the GPS.
- the position detection is not necessarily based on a single method but may be based on a plurality of methods to improve detection accuracy.
- FIG. 6 is a diagram showing an example of the whole system using the wearable device.
- a plurality of wearable devices 10 , a plurality of photo detectors 62 of FIG. 3 , at least one supervisor's device 104 , a plurality of manufacturing apparatuses 106 , at least one camera 114 , a data management server 116 are connected to a network 102 .
- the network 102 may be provided, for example, on a plant building, a department, a floor or a business office basis, or may be a network installed in each plant, each building or each company or the Internet.
- the network of a manufacturing site of FIG. 6 may constitute a LAN, and a plurality of LANs may be connected to the network of the whole plant.
- the network 102 may be a wireless network or may be a wired network.
- the wearable devices 10 may not be prepared for all the operators, but only a predetermined number of wearable devices 10 may be prepared and the operators wear available shared wearable devices when needed.
- the system needs to identify the user if the user puts on the wearable device. This is because the system displays, for the user working on a specific manufacturing apparatus, the workflow of the manufacturing apparatus or makes a report on the work based on the user's movements. There are various methods of identifying the user, but the user may input the user's ID and the device ID from a device not shown in the drawing when the user puts on the wearable device 10 .
- the input operation is not necessarily a key input but may be an audio input from a microphone or a scan input using a bar code. Further, since it is likely that the user has his or her own unique way of putting the device on, it is possible to detect the user by detecting the user's movements made at this time.
- the feature quantities indicating the user's movements can be obtained from acceleration or angular velocity of the wearable device 10 , movements of the face, hands or fingers of the user, or environmental sounds collected by a microphone. For example, it is possible to recognize the state of the wearable device based on friction sounds between the temple and the skin or the hairs produced when the user puts on or takes off the wearable device 10 .
- the supervisor uses the supervisor's device 104 . Since the supervisor does not need to move around, the supervisor's device 104 may have a structure the same as that of the wearable device 10 or may have a structure the same as that of an ordinary personal computer or an ordinary tablet computer, and description of the supervisor's device 104 will be omitted.
- an apparatus state sensor 108 and a user movement sensor 110 are attached to each manufacturing apparatus 106 . These sensors 108 and 110 have communication functions and are connected to the network 102 .
- the camera 114 constantly captures moving images of the users in the manufacturing site. By analyzing the images, the movements of the users can be detected. For example, the user of the wearable device 10 can be identified by storing reference images for the respective users in advance and comparing an image of the user putting the wearable device 10 on or an image of the user taking the wearable device 10 off with the reference images. When it is difficult to install an enough number of cameras to cover the whole manufacturing site at a time, a few number of cameras 114 each having a variable angle of view and configured to capture an image of the users in a wider area may be installed instead.
- the data management server 116 includes a controller 118 , a communication unit 120 , a position management unit 122 , a user movement management unit 124 , an apparatus state management unit 126 and the like.
- the communication functions of the sensors 108 and 110 , the communication function of the supervisor's device 104 , the communication function of the camera 114 and the communication function of the communication unit 120 may be wired communication functions or may be wireless communication functions as in the case of the communication function of the wearable device 10 .
- Bluetooth registered trademark
- ZigBee registered trademark
- a short-range wireless communication such as UWB
- a medium-range wireless communication such as WiFi (registered trademark)
- a long-range wireless communication such as 3G/4G or WiMAX (registered trademark)
- 3G/4G or WiMAX registered trademark
- the position management unit 122 is configured to collect data of the positions of the wearable device 10 and the supervisor's device 104 based on the outputs of the photo detector 62 and various sensors of the wearable device 10 and the supervisor's device 104 at regular intervals. Further, the position management unit 122 is configured to identify the user of the wearable device 10 or the supervisor's device 104 , and manage the device ID, the user ID and the position of the wearable device 10 or the management device 104 .
- the user movement management unit 124 is configured to collect data of the movements and state of the user of the wearable device 10 based on the outputs of the photo detectors 62 , various sensors of the wearable devices 10 , and the user movement sensor 110 of the manufacturing apparatus 106 and manage the device ID, the user ID, and the movements and state of the wearable device 10 .
- the apparatus state management unit 126 is configured to collect data of the state of the manufacturing apparatus based on the output of the apparatus state sensor 108 of the manufacturing apparatus 106 at regular intervals and manage the data. Note that it is possible to configure the apparatus state sensor 108 to notify, if there is a change in the state of the apparatus, the change to the apparatus state management unit 126 and collect data of the state of the manufacturing apparatus.
- the data management server 116 is configured to notify, if the apparatus state management unit 126 detects that a manufacturing apparatus has a problem, data of the position having the problem and state of the manufacturing apparatus to the management device 104 . At the same time, the states of the operators are determined, and candidate operators who can deal with the apparatus having a problem most efficiently are extracted and presented to the supervisor's device 104 .
- the present embodiment relates generally to a technique of automatically making a work checklist and presenting it to the user and of automatically checking off a corresponding item in the checklist as the user completes each work step. Therefore, the data management server 116 integrates the data obtained from the plurality of sensors 108 and 110 connected to the network 102 or various sensors of the devices 10 and 104 and performs a computation to automatically detect and recognize the movements of each operator.
- the data management server 116 makes a workflow (checklist) based on the result and supports an automatic input (automatic entry) to a corresponding portion in the checklist.
- the data management server 116 automatically makes a work report when completing an automatic input (automatic entry) to the last item in the work checklist.
- the data management server 116 collects data related to the manufacturing apparatus requiring maintenance from the manufacturing state sensor 108 , and automatically detects and recognizes the portion in the manufacturing apparatus having a problem. The data management server 116 then automatically identifies the wearable device 10 of an operator who is to perform maintenance and displays the contents of the maintenance on the device 10 in a work checklist form.
- FIG. 7 is a diagram showing an example of the electrical configuration of the wearable device 10 .
- the wearable device 10 includes a CPU 140 , a system controller 142 , a main memory 144 , a storage device 146 , a microphone 148 , the speaker 54 , a projection processor 150 (configured to control the light source 28 and the display 30 ), the camera 59 , a wireless communication device 152 , a motion sensor 154 , a sight line (line of vision) detector 156 , a gesture sensor 158 , the touchpad 55 , a vibrator 68 , a position data receiver 159 , a GPS unit 155 and the like.
- the CPU 140 is a processor configured to control various modules in the wearable device 10 and execute computer programs loaded from the storage device 146 including a nonvolatile semiconductor memory such as an SSD or a flash array to the main memory 144 .
- the programs include an operating system (OS) and various application programs.
- the CPU 140 executes, for example, the following processing by executing the application programs and performing communication with the data management server 116 via the network 102 using the wireless communication device 152 .
- the CPU 140 executes various kinds of control such as control to input a voice by the microphone 148 and transmit the audio data to the data management server 116 , control to capture an image by the camera 59 and transmit the image data to the data management server 116 , control to transmit input data from the motion sensor 154 , the sight line detector 156 , the gesture sensor 158 , the touchpad 55 or the position data receiver 159 to the data management server 116 , control to play a sound by the speaker 54 or stereo earphones (not shown) connected to the earphone jack 54 B, and control to vibrate the vibrator 68 .
- control to input a voice by the microphone 148 and transmit the audio data to the data management server 116 control to capture an image by the camera 59 and transmit the image data to the data management server 116 , control to transmit input data from the motion sensor 154 , the sight line detector 156 , the gesture sensor 158 , the touchpad 55 or the position data receiver 159 to the data management server 116 , control to play a sound
- the system controller 142 is a device configured to connect the local bus of the CPU 140 and various components.
- the microphone 148 is connected to the microphone jack 56 and configured to collect user's voices or environmental sounds. By recognizing user's voices or analyzing environmental sounds, it is possible to detect movements of the user and thereby identify the user. For example, by storing reference voices of respective users in advance and comparing the voice of the wearer with the reference voices, the wearer can be identified. Further, the work area the wearer is in can be identified by analyzing environmental sounds.
- the speaker 54 is configured to output an alarm or the like to attract the user's attention.
- the projection processor 150 is configured to output an image signal to the display 30 and project an image of the display 30 on the screen 16 by lighting the light source 28 . The image includes not only a still image but also a moving image.
- the wireless communication device 152 includes, for example, a wireless LAN function and wirelessly connects the wearable device 10 and an access point 112 .
- the motion sensor 154 is a sensor including a three-axis acceleration sensor, a three-axis gyroscope sensor and a three-axis geomagnetic sensor integrated with each other and is configured to detect movements of the head of the user of the wearable device 10 and determine the direction of the user's head base on the detection result.
- the state of the operator may also be detected by the microphone 148 , a barometer or the like.
- the state of the operator includes work content, work progress and the like in addition to a walking state and a resting state.
- the sight line detector 156 is provided in the center on the inner side of the frame of the eyeglasses and directed to the user's face, and is configured to capture an image of the eyeballs of the user and detect a line of vision. Further, it is also possible to configure the sight line detector 156 to detect the irises of the user.
- the gesture sensor 158 is a sensor configured to determine a gesture such as movements of the fingers. More specifically, the gesture sensor 158 is a sensor configured to determine the user's gesture by analyzing movements of the fingers made on the touchpad 55 provided in the projector 12 or movements of the hands or the fingers shown in an image captured by the camera 59 .
- the vibrator 68 is configured to vibrate the temple of the wearable device 10 by vibrating the projector 12 and communicate certain information to the user.
- the position data receiver 159 is configured to receive beacons including position data transmitted from a plurality of the position data transmitters 113 installed in the area of the LAN 102 using a short-range wireless communication such as RF-ID. In the case of a short-range wireless communication, the position of the transmitter and the position of the receiver (wearable device) can be estimated to be substantially the same as each other.
- the GPS unit 155 is configured to detect the position (x, y, z) of the wearable device 10 . By generalizing this result, the detection result of the position data receiver 159 and the detection result of the photo detector 62 of FIG. 3 , the position of the user and the shift of the position can be detected more accurately.
- the display 30 is configured to display an instruction or an incoming call from the supervisor's device 104 or the data management server 116 , the work state of an operator detected by the motion sensor 154 and the like.
- the display image is displayed on the screen 16 by the projection processor 150 .
- the supervisor's device 104 may have a structure the same as that of the wearable device 10 or may have a structure the same as that of an ordinary personal computer or tablet computer.
- the electrical configuration of an ordinary personal computer or tablet computer is equivalent to the electrical configuration of the wearable device 10 except that the projection processor 150 , the camera 59 , the motion sensor 154 , the sight line detector 156 , the gesture sensor 158 and the like are omitted.
- the position of the supervisor's device 104 is detected by the GPS.
- FIG. 8A shows attachment positions to the apparatus
- FIG. 8B shows the structure of the sensor 108 .
- the data management server 116 automatically detects or recognizes the portion in the manufacturing apparatus having a problem by collecting and integrating the apparatus state data obtained from the apparatus state sensor 108 connected to the network 102 .
- the portion in the manufacturing apparatus having a problem can be automatically diagnosed, it is possible to significantly decrease the maintenance time of the manufacturing apparatus (operation suspension time of the manufacturing apparatus) and thereby prevent a decrease in the productivity.
- the apparatus state sensor 108 includes an acceleration sensor 108 a and a wireless communication device 108 b and is configured to transmit an acceleration signal detected by the acceleration sensor 108 a to the data management server 116 via the wireless communication device 108 b and the network 102 .
- the apparatus state sensor 108 is provided with an attachment portion or a fixing portion so that the apparatus state sensor 108 can be easily attached to an existing manufacturing apparatus.
- An adhesive layer may be formed on the attachment portion in advance or an adhesive agent may be applied thereto at the time of attachment.
- the apparatus state sensor 108 may be attached to an existing manufacturing apparatus by screwing the fixing portion into the manufacturing apparatus.
- the apparatus state sensor 108 is fixed, for example, to a part of a moving belt 136 , a movable arm 124 configured to hold products or a part of a movable shaft 132 . Then, if a portion which moves in a normal operation stands still, it is determined that the movable portion has a problem.
- the controller 118 in the data management server 116 stores handbooks for repairing, maintaining and checking the respective portions of various manufacturing apparatuses, namely, maintenance procedure handbooks in advance, and makes an appropriate work checklist based on a result of the above-described automatic diagnosis.
- FIGS. 8A and 8B as an example of the apparatus state sensor 108 , an acceleration detection method has been described.
- the above-described method is in no way restrictive, and various physical quantities or chemical quantities such as a temperature or a conducting current value may also be used. Further, it is also possible to perform the automatic diagnosis of the portion in the manufacturing apparatus having a problem by comparing images captured by a camera or environmental sounds collected by a microphone.
- the data management server 116 automatically selects an operator who is to perform maintenance of the manufacturing apparatus and displays maintenance procedure or a work checklist derived from the maintenance procedure on the wearable device 10 of the operator.
- the data management server 116 selects an operator, for example, (i) who is near the manufacturing apparatus having a problem, (ii) who can stop the work the operator is currently engaging with and (iii) who can perform the maintenance work. In this way, it is possible to minimize a time loss in dispatching an operator to the manufacturing device having a problem.
- a photo detector 106 a similar to the photo detector 62 of FIG. 4 and a wireless communication device 106 b are attached to a part of the manufacturing apparatus 106 as shown in FIG. 9 .
- the leaking light 50 radiating from the wearable device 10 includes the device ID data of the device 10 . Therefore, if the data included in the leaking light 50 is detected by the photo detector 106 a and transmitted to the data management server 116 via the wireless communication device 106 b and the network 102 , the data management server 116 can recognize the wearable devices 10 , that is, the operators near the manufacturing apparatus having a problem.
- the data management server 116 selects an operator who is to perform maintenance of the target manufacturing apparatus based on the data and transmits a work checklist to the wearable device 10 of the operator. As shown in FIG. 13A , the work checklist is displayed on the screen 16 of the device 10 . Note that, although the work checklist is simplified in FIG. 13A for the sake of convenience, the actual work checklist is as follows.
- the leaking light 50 radiating from the wearable device 10 is detected, collected and summarized in real time, it is possible to identify an operator near the manufacturing apparatus 106 subjected to maintenance easily and accurately. Consequently, it becomes possible to save the time of dispatching an operator to the apparatus and reduce the maintenance time, and thereby prevent a decrease in the manufacturing efficiency.
- an image sensor 114 a in the camera 114 captures an image around the manufacturing apparatus 106 and transmits the image data to the data management server 116 via a wireless communication 114 b.
- the data management server 116 analyzes the received image data and automatically identifies operators therein.
- FIGS. 10A and 10B an example of a case where an operator performs work in accordance with a work checklist will be described.
- a work checklist of FIG. 13A is displayed on the screen 16 .
- an operator starts work.
- the whole checklist is displayed at a time, there are some cases where the letters in the checklist become too small to see. In that case, only first or first few steps may be displayed by larger letters, and then the checklist may be updated as each step ends by automatically recognizing the progress of work step by step.
- the 10A puts a thing 162 in a cart 164 , closes a valve 170 (or turns a handle 170 to a specified angle), flicks off an on/off switch 172 , flick off a first light switch 176 and a third light switch 180 according to the work checklist.
- the movements of the operator are automatically recognized and identified in real time by the user movement sensor 110 attached to the manufacturing apparatus 106 and the work completion times are automatically written in the work checklist (see FIG. 13B ). If the last work step ends, a work report is automatically made in the data management server 116 , and the contents are displayed on the supervisor's device 104 .
- the work checklist ( FIG. 13B ) corresponds to the work checklist ( FIG. 13A ) containing the completion times input thereto.
- various detection techniques and the combinations thereof may be used. For example, it is possible to perform automatic recognition and identification of user's movements by using the camera 114 or 59 and analyzing an image of the user's movements. Note that, in the method of analyzing an image of the user's movements captured by the camera 114 , depending on the situation, the user's movements may be hidden behind in the image. Alternatively, it is possible to use a sound recognition technique.
- the wearable device 10 If it is determined in advance that the operator produces a specific sound as the operator finishes with the work of each item of the work checklist (maintenance work procedure) displayed on the wearable device 10 , it is possible to perform automatic recognition and identification of the user's movements by detecting an input of the specific sound with the microphone 148 . Alternatively, it is also possible to perform automatic recognition and identification of the user's movements by detecting environmental sounds produced in specific work using the microphone 148 or a built-in microphone of the apparatus state sensor 108 . Further, there is a method of identifying a predetermined operator's gesture and thereby performing automatic recognition and identification of operator's movements.
- images of operator's movements captured by the cameras 59 and 114 may be analyzed or detection results of the leaking light 50 of the wearable devices 10 by a plurality of the photo detectors 62 or the photo detectors 106 a attached to the manufacturing apparatuses 106 may be compared with each other.
- a pair of a light emitting device 166 a and a photo detector 166 b is provided in an opening portion of the cart 164 , and it is automatically detected that the thing 162 is put in or taken out of the cart 164 by detecting the interception of light caused if the thing 162 passes through the opening portion of the cart 164 .
- FIG. 10B shows the signal characteristics detected by the photo detector 166 b if the thing 162 is put in or taken out of the cart 164 .
- the vertical axis shows the amount of light detected by the photo detector 166 b while the horizontal axis shows the time passed. If the thing 162 passes through the opening portion of the cart 164 , the amount of light the photo detector 166 b detects decreases.
- a method of detecting that the thing 162 is put in or taken out of the cart 164 not the above-described method using light but various other methods may be used.
- a contact sensor 168 is attached to the valve 170 , and transparent contact sensors are attached respectively to the on/off switch 172 and a light switch board 174 .
- the light switch board 174 includes the first, second and third light switches 176 , 178 and 180 .
- the contact sensor as an example of the user movement sensor 110 includes a wireless communication function (for example, a short-range wireless communication) and a detection function of detecting the contact state of the operator.
- a wireless communication function for example, a short-range wireless communication
- various elements configured to perform contact detection such as a piezoelectric element, a photo interrupter and an acceleration sensor (gyroscope sensor) can be used.
- the contact sensor of this type is attachable to existing facilities such as existing manufacturing apparatuses and is available at an extremely low cost. Therefore, it is possible to add a short-range wireless communication network environment inexpensively by simply attaching the contact sensor (user movement sensor 110 ) to an existing manufacturing apparatus while maintaining the existing infrastructure.
- FIGS. 11 and 12 An example of the user movement sensor 100 is shown in FIGS. 11 and 12 .
- FIG. 11 shows the user movement sensor 110 attached to an existing infrastructure, namely, the on/off switch 172 or the user movement sensor 110 (contact sensor 168 ) attached to the valve 170
- FIG. 12 shows the user movement sensor 110 attached to an existing infrastructure, namely, the light switch board 174 .
- the user movement sensor 110 includes an adhesive layer 202 at the bottom, and further includes a control and communication circuit 204 and a solar cell 206 formed in this order on the adhesive layer 202 .
- a transparent conductive layer 208 On the solar cell 206 , a transparent conductive layer 208 , a transparent intermediate layer 210 , a transparent conductive layer 212 and a transparent uneven layer 214 are stacked one after another.
- the control and communication circuit 204 includes a function of performing wireless communication (short-range wireless communication) and a function of detecting contact with the operator.
- the circuit 204 is driven by the solar cell 206 to perform these functions. In the case of using a battery as a power supply, there is the trouble of battery replacement.
- the cable blocks the operator from contacting.
- the solar cell 206 it is possible to drive the user movement sensor 110 for a long period of time without giving the trouble of battery replacement or obstructing the operator from contacting.
- control and communication circuit 204 configured to perform a short-range wireless communication and execute control below the solar cell 206 , it is possible to increase the power generation efficiency of the solar cell 206 and reduce the plane size of the user movement sensor 110 .
- the solar cell 206 needs to be irradiated with surrounding light. Meanwhile, it is preferable that the portion configured to detect the user's contact is provided on the surface of the user movement sensor 110 . As a method of satisfying both demands at the same time, a capacitance type detection method is adopted and the contact detection portion is made transparent. To detect the operator's contact or pressure by using a change in capacitance, the following structure is adopted: the transparent intermediate layer 210 formed of a transparent and elastic material (for example, a transparent organic material sheet) is sandwiched between the two transparent conductive layers 208 and 212 (for example, transparent organic material sheets).
- a transparent and elastic material for example, a transparent organic material sheet
- the transparent conductive layers 208 and 212 are resonated with each other. If the operator contacts the surface of the user movement sensor 110 , a change occurs in the capacitance, and consequently a change occurs in the above-described AC resonance state. By detecting a change in the AC resonance state, the operator's contact is detected. Note that this capacitance type detection method may not necessarily be used but any element may be used as long as the element allows at least a part of surrounding light to reach the solar cell 206 in the user movement sensor 110 and is configured to detect contact or pressure.
- the transparent layer on the surface of the user movement sensor 110 is, for example, provided with small asperities. This functions as a non-slip surface, but it is possible to record information in Braille using the asperities for the sake of people with impaired vision.
- the bonding or attaching method not only a method of directly bonding with an adhesive agent but also a method of using an adhesive sheet or an adhesive tape may be used.
- the adhesive layer 202 having characteristics of a double-faced adhesive tape may be used for the on/off switch 172 and the light switch board 174 , and on the other hand, the adhesive layer 202 formed of a transparent adhesive layer may be used for the valve 170 .
- FIG. 12 shows the user movement sensor 110 attached to the light switch board 174 .
- the layers provided above the light switches 176 , 178 and 180 preferably be transparent. Further, it is necessary to detect the contact states of the plurality of light switches 176 , 178 and 180 , respectively.
- the light switch board 174 there is a space 182 left in a portion not provided with the light switches 176 , 178 and 180 . To conform to such a situation, the user movement sensor 110 of FIG.
- the transparent sheet 12 includes the transparent sheet 208 , the transparent intermediate layer 210 , the transparent sheet 212 and the transparent uneven layer 214 stacked in series on the adhesive layer 202 .
- the transparent sheets 208 and 212 correspond to the transparent conductive layers 208 and 212 of FIG. 11 , and the transparent sheet 208 includes three transparent conductive regions 208 a, 208 b and 208 c and the transparent sheet 212 includes three transparent conductive regions 212 a, 212 b and 212 c.
- the transparent conductive regions 208 a and 212 a are provided in the positions corresponding to the first light switch 176
- the transparent conductive regions 208 b and 212 b are provided in the positions corresponding to the second light switch 178
- the transparent conductive regions 208 c and 212 c are provided in the positions corresponding to the third light switch 180 .
- the AC voltage 216 is applied between the transparent sheets 208 and 212 .
- a control circuit 204 a and a communication circuit 204 b are formed, and the solar cell 206 is formed thereon. Since the solar cell 206 is provided on the top, the power generation efficiency is high. Further, since the control circuit 204 a, the communication circuit 204 b and the solar cell 206 are located in the vertical direction, the plane size of the user movement sensor 110 is reduced.
- the embodiment by detecting the states of wearable devices and manufacturing apparatuses and displaying, based on the detection result, a workflow on the wearable device of an operator who is near a manufacturing apparatus requiring maintenance, checkups and repairs and who can perform the maintenance work, it is possible to provide the operator with useful information. Further, since completion of each step of the work is determined and a work report recording the progress of the work is made automatically based on the detection result of the states of the wearable device and the manufacturing apparatus, it is possible to save the operator the troubles thereof. Note that, since the detection of completion of work is realized simply by attaching a contact sensor to a manufacturing apparatus, it is possible to detect and recognize movements of the operator quite easily, inexpensively and accurately without making modifications to an existing manufacturing apparatus.
- the present embodiment describes the case of performing maintenance of a manufacturing apparatus.
- the present embodiment is not necessarily limited to this case but may be applied to a case of monitoring user's movements corresponding to other purposes and displaying work contents according to the purposes.
- the present embodiment describes providing a contact sensor for monitoring the user's movements in a portion which the user is likely to contact, but other sensors may be used instead.
- the present embodiment describes the case of an eyeglasses-type wearable device
- the present embodiment is also applicable to head-mounted type wearable devices of other types such as goggles and helmet types as well as to a wristband-type wearable device, a pendant-type wearable device and the like.
- head-mounted type wearable devices of other types such as goggles and helmet types
- wristband-type wearable device such as a bracelet or goggles-type wearable device
- a pendant-type wearable device such as to a wearable device.
- the projector 12 and the camera 59 can be attached to the helmet or the goggles, people with eye glasses can also use the wearable device.
- the speaker 54 can be attached to the inner side of the helmet, the user can hear a sound clearly, and since a microphone can be attached to the helmet and the position of the microphone can be adjusted, the sound collection performance of the microphone improves.
- sensors configured to detect the states of a wearable device and a manufacturing apparatus
- various other sensors may be used appropriately instead of the sensors described above.
- the present embodiment is applicable to wearable devices other than head-mounted type wearable devices.
- the present embodiment is also applicable to portable and light electronic devices carried with the users at all times as notebook computers, tablet computers and smart phones.
- the above description is in no way restrictive, but instead, some of the above-described functions of the wearable device may be realized as those of the data management server or some of the above-described functions of the data management server may be realized as those of the wearable device.
- the various modules of the systems described herein can be implemented as software applications, hardware and/or software modules, or components on one or more computers, such as servers. While the various modules are illustrated separately, they may share some or all of the same underlying logic or code.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- User Interface Of Digital Computer (AREA)
- Chemical & Material Sciences (AREA)
- Medicinal Chemistry (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Genetics & Genomics (AREA)
- Public Health (AREA)
- Immunology (AREA)
- Microbiology (AREA)
- Wood Science & Technology (AREA)
- Zoology (AREA)
- Veterinary Medicine (AREA)
- Virology (AREA)
- Organic Chemistry (AREA)
- Animal Behavior & Ethology (AREA)
- Pharmacology & Pharmacy (AREA)
- Epidemiology (AREA)
- Mycology (AREA)
- Biotechnology (AREA)
- Biochemistry (AREA)
- Mechanical Engineering (AREA)
- Biomedical Technology (AREA)
- General Factory Administration (AREA)
- Testing And Monitoring For Control Systems (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
According to one embodiment, a system includes a wearable device on a head of a user and including a display in a line of vision of the user, a first detector configured to detect a movement of the user, a second detector configured to detect a state of an apparatus operated by the user, and a server connected to the wearable device, the first detector and the second detector. The server is configured to display information about work contents of the user on the display based on a detection result of the first detector and a detection result of the second detector.
Description
- This application is a divisional of U.S. patent application Ser. No. 14/979,169, filed Dec. 22, 2015, which claims the benefit of priority from Japanese Patent Application No. 2015-171933, filed Sep. 1, 2015, the entire contents of which are incorporated herein by reference.
- Embodiments described herein relate generally to a system and a method using an eyeglasses-type wearable device.
- In manufacturing sites or manufacturing plants in which a large number of manufacturing apparatuses are operated, the operation rates of the manufacturing apparatuses have a great impact on production volumes. In a case where a usually-avoidable problem resulting from neglect of regular maintenance and checkups or an unexpected problem has occurred in a manufacturing apparatus but the problem has not been handled efficiently, the manufacturing apparatus cannot be operated for a long time, which leads to decreases in the operation rate and the production volume. Therefore, it is desired that the operation suspension time of a manufacturing apparatus is reduced as much as possible. In performing maintenance, checkups and repairs, since the maintenance, checkups and repairs vary from manufacturing apparatus to manufacturing apparatus, there are some cases where an operator refers to an instruction manual or a checklist (hereinafter referred to as a checklist) showing a workflow of each work step.
- Recently, wearable devices have been actively introduced into manufacturing sites. In such a manufacturing site, for example, operators wear eyeglasses-type wearable devices and refer to checklists displayed on the lens surfaces. In this way, the operators no longer need to refer to paper checklists while working on apparatuses, and consequently the operators can have their work done efficiently even in the case of unfamiliar and complicated work without interruption of the work.
- However, the operators still need to check against checklists to ensure completion of their work at each work step. Therefore, paper checklist are still prepared even although the operators electronically refer to checklists on the screens while working on the apparatuses, and the operators stop their work to fill in the paper checklists at the end of each work step. Since the operations of manufacturing apparatuses are kept stopped during that time, the production volumes decrease. Further, after returning their office, the operators write reports on their work based on the checklists. It is quite cumbersome for the operators to write such reports on their work.
- There is a system for supporting an operator by using a head-mounted display with a built-in camera. As an example of the system, there is a medical-equipment management system which supports an operator of a medical device such as a used and contaminated endoscope or a piece of medical equipment such as a scalpel or forceps.
- The system includes a head-mounted camera for capturing an image of the sight of an operator of a medical device or a piece of medical equipment; storage means for storing an image of the sight of an operator captured when the operator demonstrates the medical device or the medical equipment as a reference image; first determination means for comparing the image captured by the camera and the reference image read from the storage means in order to determine whether a predetermined operation is performed normally based on similarity between these two images; data output means for outputting data indicating an alarm or an instruction based on a determination result of the determination means; and output means for outputting the alarm or the instruction to the operator based on the data indicating the alarm or the instruction.
- The system automatically recognizes the operation of the medical equipment operator by comparing the operation of the operator captured by the camera with the reference image prepared in advance. However, in this method, since the recognition accuracy of simple pattern match between images is low, complicated image processing such as feature extraction from the images is further required. Consequently, as the automatic recognition processing becomes highly complicated, a considerable amount of time will be required for the image processing.
- The present embodiment aims to provide a system and a method for recognizing movements of a user of a wearable device quickly, simply and accurately and displaying content to support the user based on a recognition result.
- A general architecture that implements the various features of the embodiments will now be described with reference to the drawings. The drawings and the associated descriptions are provided to illustrate the embodiments and not to limit the scope of the invention.
-
FIG. 1 is a perspective diagram showing an example of a wearable device of an embodiment. -
FIG. 2A shows a front view of an example of the wearable device. -
FIG. 2B shows a cross-section structure of an example of the wearable device. -
FIG. 3 shows an example of position detection of the wearable device. -
FIG. 4 exemplarily shows the principle of the position detection of the wearable device. -
FIGS. 5A, 5B, and 5C show an example of operation periods of the wearable devices. -
FIGS. 5D, 5E, 5F, and 5G show an example of communication periods of the photo detectors. -
FIGS. 5H, 51, 5J, and 5K show an example of signal waveforms of the photo detectors. -
FIG. 6 shows an example of a system including the wearable device and a data management server. -
FIG. 7 exemplarily shows an electrical configuration of the wearable device. -
FIG. 8A shows an example of detection of the state of an apparatus. -
FIG. 8B shows an example of a sensor for detecting the state of an apparatus. -
FIG. 9 shows an example of a sensor for detecting movements of a user. -
FIGS. 10A and 10B show an example of a usage environment of the system. -
FIG. 11 is an exploded view showing an example of the structure of the sensor used in the system and for detecting movements of a user. -
FIG. 12 is an exploded view showing another example of the structure of the sensor used in the system and for detecting movements of a user. -
FIG. 13A shows an example of a workflow displayed by the system. -
FIG. 13B shows an example of a work record made by the system. - Various embodiments will be described hereinafter with reference to the accompanying drawings.
- In general, according to one embodiment, a system includes a wearable device on a head of a user and including a display in a line of vision of the user, a first detector configured to detect a movement of the user, a second detector configured to detect a state of an apparatus operated by the user, and a server connected to the wearable device, the first detector and the second detector. The server is configured to display information about work contents of the user on the display based on a detection result of the first detector and a detection result of the second detector.
- Wearable devices include head-mounted type wearable devices (such as eyeglasses, goggles and helmet types which may also be called an eyeglasses type collectively), wristband-type wearable devices, pendant-type wearable devices and the like. The following description is based on the assumption that the wearable device of the present embodiment is an eyeglasses-type wearable device. Eyeglasses-type wearable devices include optical head-mounted displays, which allow the user to see through their transparent lenses, and non-optical head-mounted displays, which block the view of the user and do not allow the user to see through them. In the following, optical head-mounted displays, which allow the user to see through them, will be taken as an example.
-
FIG. 1 is a perspective view of an eyeglasses-type wearable device (hereinafter referred to simply as a wearable device) 10, andFIG. 2A is a front view andFIG. 2B is a diagram showing a cross-section structure viewed from above. - The
wearable device 10 has a shape substantially the same as that of an ordinary pair of glasses, but here aprojector 12 is attached to the outside of the right-eye temple.Glasses eye glass 14 is a normal transparent glass so that the user can see through the glass. The right-eye glass 16 is at least partly ascreen 16. Thescreen 16 is configured to show an image projected by theprojector 12 to the user. When theprojector 12 is not projecting an image, thescreen 16 is transparent and allows the user to see through the right-eye glass (screen) 16. - The
projector 12 includes apower supply 22 and acontroller 24 as electronic components. Thepower supply 22 includes a button battery, a rechargeable battery, a non-contact power supply secondary battery and the like. Alternatively, theprojector 12 may not include a built-in battery but may be supplied with power from an external power supply via a power-supply line or a wireless channel. Thecontroller 24 is configured to perform a communication with a server or another electronic device via a network which will be described later and thereby transmit and receive data. This communication may be a wired communication or may be a wireless communication. In the case of performing a wireless communication, Bluetooth (registered trademark), ZigBee (registered trademark), a short-range wireless communication such as UWB, a medium-range wireless communication such as WiFi (registered trademark) or a long-range wireless communication such as 3G/4G or WiMAX (registered trademark) may be used according to the usage environment. - The
projector 12 further includes alight source 28, adisplay 30, aprism 32, a set oflenses 34 and the like as optical components. Thelight source 28 may be a dimming white LED light source having several, for example, three light emitting diodes having luminescent colors different from each other and amounts of output light respectively variable. According to the dimming white LED light source, even if thewearable device 10 is used in such an environment as a clean room using light having a luminescent color consisting principally of orange, a clear projection image can be obtained by changing the luminescent color of the LED light source based on the usage environment. Further, according to the dimming white LED light source, it is possible to output a display color easy for the user to see, and thus as compared to the case of outputting a display color difficult for the user to see, the causes of troubles to the user such as eye strain and migraine associated with eye strain can be prevented. - The
display 30 is, for example, a reflective liquid crystal display (LCD) module and configured to display a predetermined text, image and the like (hereinafter referred to also as a display image collectively) based on display control executed by thecontroller 24. Non-parallel light (hereinafter referred to also as diverging light) output from thelight source 28 is reflected on ahalf mirror surface 32 a of theprism 32, and thereby illuminates a display image of thedisplay 30. The reflected light of thedisplay 30 is, after passing through thehalf mirror surface 32 a as light indicative of the display image (hereinafter referred to also as image light), output from theoutgoing surface 32 c and then projected on thescreen 16 as a projection image in a predetermined size via the set oflenses 34. - The
screen 16 includes a near-sidetransparent refractor 42, a Fresnel-lens-typehalf mirror surface 44 and a back-sidetransparent refractor 46. The image light reaching thehalf mirror surface 44 is partly reflected on thehalf mirror surface 44 and forms a visual image (projection image) indicative of the display image of thedisplay 30 at a few meters away. Note that, since thescreen 16 allows the user to partly see through thescreen 16, it is also possible to configure thescreen 16 to show the projection image as well as the view in front of the user. - A part of the image light (diverging light) output from the
light source 28 and passing thorough thehalf mirror surface 32 a is totally reflected on the total-reflection surface 32 b and becomes leakinglight 50 of the diverging light from thelight source 28 refracted in theoutgoing surface 32 c. The leakinglight 50 is output in a direction different from the direction of thescreen 16 through an opening or a gap (leading portion) 52 formed on the front side of theprojector 12. - The
wearable device 10 includes a speaker 54A, anearphone jack 54B, amicrophone jack 56, a slidingswitch 57, arotating switch 58 and the like in a predetermined portion, for example, in a bottom portion of theprojector 12. Themicrophone jack 56 is connected to a hands free microphone (not shown in the drawing) and collects the user's voice. The slidingswitch 57 is configured, for example, to adjust the brightness, color tone and the like of the projection image of theprojector 12. Therotating switch 58 is configured, for example, to adjust the projection angle and the like of the projection image. With such a configuration as to set different adjustment values by different operations such as by operating the slidingswitch 57 and therotating switch 58, the user can adjust the projection image by performing touch operations while looking at the projection image. For example, by operating the slidingswitch 57, it is possible to provide the projection image having the display brightness and color tone of the user's taste. By operating therotation switch 58, it is possible to adjust the projection angle so that the projection image is displayed in the most appropriate position based on the shape or size of the user's head. Note that the objects to be adjusted by the slidingswitch 57 and therotating switch 58 may be opposite to each other, the positions of the slidingswitch 57 and therotating switch 58 may be opposite to each other, or their functions may be assigned to a single operation member configured to perform two kinds of operations. - Although it is possible to perform adjustment using these
switches display 30 displays a menu image, the menu image is projected on thescreen 16. - Further, a menu item may not be selected by an operation on the
switch touchpad 55 is further provided on the outside of theprojector 12. When a menu or the like is displayed by thedisplay 30, the user can input an operation easily and efficiently by touching a position of thetouchpad 55 corresponding to the display position of an item in the menu. - A
camera 59 is provided in the center front on the outside and configured to capture an image of the front view (still image and moving image). Note that, although not shown in the drawing, it is possible to provide another camera in the center front on the inner side to face the user and configure to capture the eyeballs of the user to detect the irises of the user. The irises can be used for user authentication. - By using the leaking light 50 from the
wearable device 10, the state of thewearable device 10, that is, the state of the user can be detected. With reference toFIGS. 3, 4 and 5A-5K , the principle of detecting the state of the wearable device will be described. Here, the state includes a position, a shift of the position and the like. - An example of the use of the wearable device is shown in
FIG. 3 . For example, in awork area 60 of, for example, a component yard of a plant, a product warehouse of a mail-order firm or a delivery department of a retailer, a given number of work spaces or product shelves A01 to Axy (x and y are both positive integers), B01 to Bxy and C01 to Cxy are arranged. The work spaces or the product shelves may be, for example, work tables in a plant, manufacturing apparatuses in a production line, desks at school, seating positions in a conference room, and the like. - In the
work area 60, at least one photo detector 62-1 to 62-n (n is a positive integer) is installed. The at least one photo detector 62-1 to 62-n is configured to detect the positions (x, y, z), the numbers, the shifts of the positions, the changes of the directions and the like of the wearable devices 10-1 to 10-m (m is a positive integer) respectively by a detection method shown inFIGS. 4 and 5A-5K . By detecting the positions, the numbers, the shifts of the positions, the changes of the directions and the like of the wearable devices 10-1 to 10-m, the states such as the positions and the shifts of the positions of a given number of the users of the wearable devices 10-1 to 10-m can be recognized. - The users can move around the
work area 60 freely. The users perform predetermined work in predetermined positions, namely,work spaces 64 such as stations (carts), or containers or movable tables equivalent thereto. Note that thework space 64 is not necessarily movable but may be a fixed desk, a seating position or the like. - As shown in
FIGS. 3 and 4 , a detection system includes at least onewearable device 10 and at least onephoto detector 62. Thephoto detector 62 has a function of detecting the leakinglight 50 and a function of performing communication to transmit a detection result to a server or the like. The communication function may be a wired communication function or may be a wireless communication function as in the case of the communication function of thewearable device 10. In the case of a wireless communication, Bluetooth (registered trademark), ZigBee (registered trademark), a short-range wireless communication such as UWB, a medium-range wireless communication such as WiFi (registered trademark) or a long-range wireless communication such as 3G/4G or WiMAX (registered trademark) may be used according to the usage environment. In the present embodiment, various units and modules having communication functions will be described below, and these communication functions may be wired communication functions or may be wireless communication functions similarly. In the case of a wireless communication, Bluetooth (registered trademark), ZigBee (registered trademark), a short-range wireless communication such as UWB, a medium-range wireless communication such as WiFi (registered trademark) or a long-range wireless communication such as 3G/4G or WiMAX (registered trademark) may be used according to the usage environment. - The
wearable device 10 intermittently modulates the leakinglight 50 by using data including identification data of the device (hereinafter referred to also as a device ID) so that thephoto detector 62 can identify thewearable device 10 based on the received leakinglight 50. Although a typical example of the modulation method is a chopper modulation method of decreasing an amount of luminescence to zero, the following description is based on the assumption that thewearable device 10 adopts a modulation method of ensuring a predetermined or more amount of luminescence even in the case of light having a small amount of luminescence. In this way, the strain on the user's eyes can be reduced. In the case of adopting a digital sum value (DSV) free modulation method (that is, a method of calculating the DSV of a modulation signal constantly, inserting an appropriate bit inversion code and setting a direct-current component to zero) as a modulation method, it is possible to prevent a change in the amount of luminescence over a relatively long range and thereby keep a change in the amount of luminescence macroscopically zero, and thus the strain on the user's eyes can be further reduced. Since the human eyes can perceive a change up to about 0.02 second, it is possible to achieve the effect of reducing the strain of the user's eyes by setting the reference frequency of the above-described modulation to, for example, greater than or equal to 20 Hz, more preferable, greater than or equal to 60 Hz. On the other hand, since the LED used for thelight source 28 has an internal impedance and a connecting capacity, the frequency of less than 100 MHz, more preferably, less than or equal to 10 MHz is desirable for performing highly-accurate modulation. From the above, it is desirable that the modulation frequency of thelight source 28 used in the detection system of the present embodiment be 10 Hz to 100 MHz, more preferable, 10 Hz to 10 MHz. - Since the leaking
light 50 of the diverging light from thelight source 28 is used, the amount of light detected by thephoto detector 62 varies depending on the distance between thewearable device 10 and thephoto detector 62. By using this phenomenon, the distance between thewearable device 10 and thephoto detector 62 or the direction of thewearable device 10 with respect to thephoto detector 62 can be obtained. Since the position (including the level) of thephoto detector 62 is fixed, as the distance between thephoto detector 62 and thewearable device 10 is obtained, the position of the wearable device 10 (x, y, z) can be detected accordingly. - Further, by using the leaking
light 50 of the diverging light from thelight source 28, detection of the leakinglight 50 can be performed in a relatively wide area. As a result, by installing only a relatively small number of the photo detectors 62-1 to 62-n, the positions of the wearable devices 10-1 to 10-m in thework area 60, the distances between thewearable devices 10 and thephoto detectors 62, the directions of thewearable devices 10, or the directions of thewearable devices 10 with respect to thephoto detectors 62 can be detected. Consequently, the installation cost required for installing the detection system can be reduced. - The data of amounts of the leaking
light 50 detected by thephoto detectors 62 is transmitted from thephoto detectors 62 to a server which will be described later at a predetermined time. The server analyzes the data collected from thephoto detectors 62. In this way, the positions and the states of the desired wearable devices 10-1 to 10-m, more specifically, the states of the users can be detected. -
FIG. 4 is a schematic diagram showing a specific example of the use of the system for recognizing the wearable device of the embodiment. The following description is based on the assumption that there are three users wearing wearable devices 10-1 to 10-3 around four photo detectors 62-1 to 62-4. The leaking light 50 from the wearable devices 10-1 and 10-2 is detected by the photo detectors 62-1 to 60-4. The photo detectors 62-1 to 62-4 perform analog-to-digital conversion of the amounts of the leakinglight 50 detected respectively and transmit them to a server as light amount data indicative of the amounts of light at a predetermined time, for example, by a short-range wireless communication. - The following description is based on the assumption that the position of the wearable device 10-1 is shifted toward the photo detector 62-1 as the user moves toward the photo detector 62-1 and meanwhile the direction of the wearable device 10-2 is temporarily changed as the user makes a given movement such as turning of the head (rotating of the head). The changes in the detection data occurring at this time is shown in
FIGS. 5A-5K . -
FIGS. 5A-5K illustrate the case of using an intermittent time-shift method as a modulation method of the leakinglight 50 of the wearable devices 10-1 to 10-3. That is, ID modulation times are set respectively to the wearable devices 10-1 to 10-3 in a staggered manner. - As shown in
FIGS. 5A, 5B, and 5C , intermittent ID modulation times are set respectively to the first to third wearable devices 10-1 to 10-3, and the other times are set as non-modulation times. In each ID modulation time, a synchronization signal SYNC is paired with each of the device IDs of the wearable devices 10-1 to 10-3 (on a one-to-one basis), and the pairs are repeated for several times (multiples of four times in the case where there are four sensors as shown inFIGS. 5D-5G ). - As the non-modulation time of the first wearable device 10-1 starts, the ID modulation time of the second wearable device 10-2 starts. Similarly, as the non-modulation time of the second wearable device 10-2 starts, the ID modulation time of the third wearable device 10-3 starts.
- In the ID modulation time of the second wearable device 10-2 and the ID modulation time of the third wearable device 10-3, the synchronization signal SYNC and the device ID of the wearable device 10-2 or 10-3 are repeatedly modulated. By superimposing the device ID of the
wearable device 10 on a modulation signal in this way, the device ID can be detected. - In the above-described case, the modulation times of the respective wearable devices 10-1 to 10-3 are set on a time-division basis (on an intermittent basis). However, for example, it is also possible to perform modulation successively for all the wearable devices 10-1 to 10-3 and change the modulation reference frequencies of the wearable devices 10-1 to 10-3 respectively. Further, it is also possible to change the characteristics of the frequency spectrums in spread spectrum, respectively.
- As shown in
FIGS. 5D-5G , each ID modulation time is divided into sections of the data communication times (COMs) with the photo detectors 62-1 to 62-4. - As shown in
FIG. 4 , a part of the leaking light from the wearable device 10-1 reaches the photo detector 62-4 at the beginning. Therefore, as shown inFIG. 5K , the photo detector 62-4 detects the leaking light from the wearable device 10-1 at the beginning. However, as the position of the wearable device 10-1 is shifted toward the photo detector 62-1, the modulation signal amplitude of the leaking light from the wearable device 10-1 detected by the photo detector 62-4 decreases. On the other hand, as shown inFIG. 5H , the modulation signal amplitude of the leaking light from the wearable device 10-1 detected by the photo detector 62-1 increases as time advances. By comparing the changes of the modulation signal amplitudes detected by the photo detectors 62-1 to 62-n with time, the changes (shifts) of the positions of the detection targets, namely, the wearable devices 10-1 to 10-m with time can be detected. - Meanwhile, the wearable device 10-2 is directed to the photo detector 62-3 at the beginning, and thus with regard to the modulation signal amplitude of the leaking light, the detection value of the photo detector 62-3 is greater than the detection value of the photo detector 62-2. Here, suppose that the second user then turns the head and the wearable device 10-2 is temporarily directed to the photo detector 62-2. In this case, the detection output of the wearable device 10-2 output from the photo detector 62-2 temporarily increases and then decreases as shown in
FIG. 51 . On the other hand, the detection output of the wearable device 10-2 output from the photo detector 62-3 temporarily decreases and then increases as shown inFIG. 5J . - In this way, by comparing changes in the modulation signal amplitudes detected by the
photo detectors 62 with time, changes in the directions of the detection targets, namely, the wearable devices 10-1 to 10-m can be detected. - In the above-described detection, such movements of the user as moving from one place to another or turning the head are used. However, the above-described case is in no way restrictive, and various other movements of the user may also be used for detection. For example, as the user makes such movements as moving his or her hands or twisting the upper part of the body, the leaking light may be blocked temporarily. In that case, all the modulation signal amplitudes of the photo detectors 62-1 to 62-4 temporarily decrease in the same period of time. In this way, by comparing the relationships among changes in the modulation signal amplitudes of all the photo detectors 62-1 to 62-4, various movement patterns of the user can be identified.
- According to the above-described method, not only the movements of the user but also the will of the user can be recognized.
- Note that, as a method of detecting the position (x, y, z) of the
wearable device 10, it is also possible to use a beacon. In the above-described case, the position or state of thewearable device 10 is detected by executing comparative processing of device identification data output from a number ofwearable devices 10 as modulated light and received by a number ofphoto detectors 62. However, by installing a number of position data transmitters in thework area 60 and transmitting beacons according to the installation positions from the transmitters by, for example, a short-range wireless communication with a communication range of a few meters such as RF-ID, it is also possible to detect thewearable device 10 which receives a beacon to be in a position substantially the same as the position of the transmitter having transmitted that beacon. Further, it is also possible to detect the position of a wearable device by using the GPS. The position detection is not necessarily based on a single method but may be based on a plurality of methods to improve detection accuracy. -
FIG. 6 is a diagram showing an example of the whole system using the wearable device. Here, a case where the system is applied to a manufacturing site of a manufacturing plant will be described. A plurality ofwearable devices 10, a plurality ofphoto detectors 62 ofFIG. 3 , at least one supervisor'sdevice 104, a plurality ofmanufacturing apparatuses 106, at least onecamera 114, adata management server 116 are connected to anetwork 102. Thenetwork 102 may be provided, for example, on a plant building, a department, a floor or a business office basis, or may be a network installed in each plant, each building or each company or the Internet. In a case where there is a plurality of manufacturing sites in a single plant, the network of a manufacturing site ofFIG. 6 may constitute a LAN, and a plurality of LANs may be connected to the network of the whole plant. Thenetwork 102 may be a wireless network or may be a wired network. - There are a number of operators in a manufacturing site, but not all the operators wear the
wearable devices 10. Therefore, thewearable devices 10 may not be prepared for all the operators, but only a predetermined number ofwearable devices 10 may be prepared and the operators wear available shared wearable devices when needed. The system needs to identify the user if the user puts on the wearable device. This is because the system displays, for the user working on a specific manufacturing apparatus, the workflow of the manufacturing apparatus or makes a report on the work based on the user's movements. There are various methods of identifying the user, but the user may input the user's ID and the device ID from a device not shown in the drawing when the user puts on thewearable device 10. The input operation is not necessarily a key input but may be an audio input from a microphone or a scan input using a bar code. Further, since it is likely that the user has his or her own unique way of putting the device on, it is possible to detect the user by detecting the user's movements made at this time. The feature quantities indicating the user's movements can be obtained from acceleration or angular velocity of thewearable device 10, movements of the face, hands or fingers of the user, or environmental sounds collected by a microphone. For example, it is possible to recognize the state of the wearable device based on friction sounds between the temple and the skin or the hairs produced when the user puts on or takes off thewearable device 10. - There is at least one supervisor for the operators in the manufacturing site, and the supervisor uses the supervisor's
device 104. Since the supervisor does not need to move around, the supervisor'sdevice 104 may have a structure the same as that of thewearable device 10 or may have a structure the same as that of an ordinary personal computer or an ordinary tablet computer, and description of the supervisor'sdevice 104 will be omitted. - To each
manufacturing apparatus 106, anapparatus state sensor 108 and auser movement sensor 110 are attached. Thesesensors network 102. - The
camera 114 constantly captures moving images of the users in the manufacturing site. By analyzing the images, the movements of the users can be detected. For example, the user of thewearable device 10 can be identified by storing reference images for the respective users in advance and comparing an image of the user putting thewearable device 10 on or an image of the user taking thewearable device 10 off with the reference images. When it is difficult to install an enough number of cameras to cover the whole manufacturing site at a time, a few number ofcameras 114 each having a variable angle of view and configured to capture an image of the users in a wider area may be installed instead. - The
data management server 116 includes acontroller 118, a communication unit 120, a position management unit 122, a user movement management unit 124, an apparatus state management unit 126 and the like. The communication functions of thesensors device 104, the communication function of thecamera 114 and the communication function of the communication unit 120 may be wired communication functions or may be wireless communication functions as in the case of the communication function of thewearable device 10. In the case of a wireless communication, Bluetooth (registered trademark), ZigBee (registered trademark), a short-range wireless communication such as UWB, a medium-range wireless communication such as WiFi (registered trademark) or a long-range wireless communication such as 3G/4G or WiMAX (registered trademark) may be used according to the usage environment. - The position management unit 122 is configured to collect data of the positions of the
wearable device 10 and the supervisor'sdevice 104 based on the outputs of thephoto detector 62 and various sensors of thewearable device 10 and the supervisor'sdevice 104 at regular intervals. Further, the position management unit 122 is configured to identify the user of thewearable device 10 or the supervisor'sdevice 104, and manage the device ID, the user ID and the position of thewearable device 10 or themanagement device 104. - The user movement management unit 124 is configured to collect data of the movements and state of the user of the
wearable device 10 based on the outputs of thephoto detectors 62, various sensors of thewearable devices 10, and theuser movement sensor 110 of themanufacturing apparatus 106 and manage the device ID, the user ID, and the movements and state of thewearable device 10. The apparatus state management unit 126 is configured to collect data of the state of the manufacturing apparatus based on the output of theapparatus state sensor 108 of themanufacturing apparatus 106 at regular intervals and manage the data. Note that it is possible to configure theapparatus state sensor 108 to notify, if there is a change in the state of the apparatus, the change to the apparatus state management unit 126 and collect data of the state of the manufacturing apparatus. - The
data management server 116 is configured to notify, if the apparatus state management unit 126 detects that a manufacturing apparatus has a problem, data of the position having the problem and state of the manufacturing apparatus to themanagement device 104. At the same time, the states of the operators are determined, and candidate operators who can deal with the apparatus having a problem most efficiently are extracted and presented to the supervisor'sdevice 104. - The present embodiment relates generally to a technique of automatically making a work checklist and presenting it to the user and of automatically checking off a corresponding item in the checklist as the user completes each work step. Therefore, the
data management server 116 integrates the data obtained from the plurality ofsensors network 102 or various sensors of thedevices data management server 116 makes a workflow (checklist) based on the result and supports an automatic input (automatic entry) to a corresponding portion in the checklist. Thedata management server 116 automatically makes a work report when completing an automatic input (automatic entry) to the last item in the work checklist. - The contents of the above-described work checklist vary depending on the manufacturing apparatus subjected to maintenance. Further, the contents of the above-described work checklist also vary depending on the portion in a manufacturing apparatus having a problem. Therefore, the
data management server 116 collects data related to the manufacturing apparatus requiring maintenance from themanufacturing state sensor 108, and automatically detects and recognizes the portion in the manufacturing apparatus having a problem. Thedata management server 116 then automatically identifies thewearable device 10 of an operator who is to perform maintenance and displays the contents of the maintenance on thedevice 10 in a work checklist form. -
FIG. 7 is a diagram showing an example of the electrical configuration of thewearable device 10. Thewearable device 10 includes aCPU 140, asystem controller 142, amain memory 144, astorage device 146, amicrophone 148, thespeaker 54, a projection processor 150 (configured to control thelight source 28 and the display 30), thecamera 59, awireless communication device 152, amotion sensor 154, a sight line (line of vision)detector 156, agesture sensor 158, thetouchpad 55, avibrator 68, aposition data receiver 159, aGPS unit 155 and the like. - The
CPU 140 is a processor configured to control various modules in thewearable device 10 and execute computer programs loaded from thestorage device 146 including a nonvolatile semiconductor memory such as an SSD or a flash array to themain memory 144. The programs include an operating system (OS) and various application programs. TheCPU 140 executes, for example, the following processing by executing the application programs and performing communication with thedata management server 116 via thenetwork 102 using thewireless communication device 152. For example, theCPU 140 executes various kinds of control such as control to input a voice by themicrophone 148 and transmit the audio data to thedata management server 116, control to capture an image by thecamera 59 and transmit the image data to thedata management server 116, control to transmit input data from themotion sensor 154, thesight line detector 156, thegesture sensor 158, thetouchpad 55 or theposition data receiver 159 to thedata management server 116, control to play a sound by thespeaker 54 or stereo earphones (not shown) connected to theearphone jack 54B, and control to vibrate thevibrator 68. Although the description is based on the assumption that thespeaker 54 is a monaural speaker, it is possible to further provide a speaker (not shown in theFIGS. 1 and 2 ) in the left-eye temple when a stereo speaker is required. - The
system controller 142 is a device configured to connect the local bus of theCPU 140 and various components. Themicrophone 148 is connected to themicrophone jack 56 and configured to collect user's voices or environmental sounds. By recognizing user's voices or analyzing environmental sounds, it is possible to detect movements of the user and thereby identify the user. For example, by storing reference voices of respective users in advance and comparing the voice of the wearer with the reference voices, the wearer can be identified. Further, the work area the wearer is in can be identified by analyzing environmental sounds. Thespeaker 54 is configured to output an alarm or the like to attract the user's attention. Theprojection processor 150 is configured to output an image signal to thedisplay 30 and project an image of thedisplay 30 on thescreen 16 by lighting thelight source 28. The image includes not only a still image but also a moving image. Thewireless communication device 152 includes, for example, a wireless LAN function and wirelessly connects thewearable device 10 and an access point 112. - The
motion sensor 154 is a sensor including a three-axis acceleration sensor, a three-axis gyroscope sensor and a three-axis geomagnetic sensor integrated with each other and is configured to detect movements of the head of the user of thewearable device 10 and determine the direction of the user's head base on the detection result. Note that the state of the operator may also be detected by themicrophone 148, a barometer or the like. The state of the operator includes work content, work progress and the like in addition to a walking state and a resting state. By using movements detected by themotion sensor 154, a barometric altitude or the like, it is determined whether the feature quantities obtained from the detection result correspond to the feature quantities of each work step obtained from an operator or the like beforehand, and it is thereby determined which step in a plurality of work steps the operator is performing or has finished with. Further, it is also possible to determine which step in a plurality of work steps the operator is performing or has finished with by determining whether the feature quantities of environmental sounds input from themicrophone 148 correspond to the feature quantities of environmental sounds unique to each work step obtained beforehand. - The
sight line detector 156 is provided in the center on the inner side of the frame of the eyeglasses and directed to the user's face, and is configured to capture an image of the eyeballs of the user and detect a line of vision. Further, it is also possible to configure thesight line detector 156 to detect the irises of the user. Thegesture sensor 158 is a sensor configured to determine a gesture such as movements of the fingers. More specifically, thegesture sensor 158 is a sensor configured to determine the user's gesture by analyzing movements of the fingers made on thetouchpad 55 provided in theprojector 12 or movements of the hands or the fingers shown in an image captured by thecamera 59. Thevibrator 68 is configured to vibrate the temple of thewearable device 10 by vibrating theprojector 12 and communicate certain information to the user. Theposition data receiver 159 is configured to receive beacons including position data transmitted from a plurality of theposition data transmitters 113 installed in the area of theLAN 102 using a short-range wireless communication such as RF-ID. In the case of a short-range wireless communication, the position of the transmitter and the position of the receiver (wearable device) can be estimated to be substantially the same as each other. TheGPS unit 155 is configured to detect the position (x, y, z) of thewearable device 10. By generalizing this result, the detection result of theposition data receiver 159 and the detection result of thephoto detector 62 ofFIG. 3 , the position of the user and the shift of the position can be detected more accurately. - The
display 30 is configured to display an instruction or an incoming call from the supervisor'sdevice 104 or thedata management server 116, the work state of an operator detected by themotion sensor 154 and the like. The display image is displayed on thescreen 16 by theprojection processor 150. - It is possible to take an incoming call by using the
microphone 148 and thespeaker 54. - The supervisor's
device 104 may have a structure the same as that of thewearable device 10 or may have a structure the same as that of an ordinary personal computer or tablet computer. The electrical configuration of an ordinary personal computer or tablet computer is equivalent to the electrical configuration of thewearable device 10 except that theprojection processor 150, thecamera 59, themotion sensor 154, thesight line detector 156, thegesture sensor 158 and the like are omitted. The position of the supervisor'sdevice 104 is detected by the GPS. - With reference to
FIGS. 8A and 8B , an example of theapparatus state sensor 108 attached to themanufacturing apparatus 106 will be described below.FIG. 8A shows attachment positions to the apparatus, whileFIG. 8B shows the structure of thesensor 108. Conventionally, each time a problem occurs in a manufacturing apparatus, an operator checks the portion in the manufacturing apparatus having a problem and repairs the apparatus, and investigates the cause of the problem. Therefore, the maintenance time of the manufacturing apparatus (operation suspension time of the manufacturing apparatus) increases and this leads to a decrease in the productivity. In the present embodiment, thedata management server 116 automatically detects or recognizes the portion in the manufacturing apparatus having a problem by collecting and integrating the apparatus state data obtained from theapparatus state sensor 108 connected to thenetwork 102. As a result, since the portion in the manufacturing apparatus having a problem can be automatically diagnosed, it is possible to significantly decrease the maintenance time of the manufacturing apparatus (operation suspension time of the manufacturing apparatus) and thereby prevent a decrease in the productivity. - The
apparatus state sensor 108 includes anacceleration sensor 108 a and awireless communication device 108 b and is configured to transmit an acceleration signal detected by theacceleration sensor 108 a to thedata management server 116 via thewireless communication device 108 b and thenetwork 102. Theapparatus state sensor 108 is provided with an attachment portion or a fixing portion so that theapparatus state sensor 108 can be easily attached to an existing manufacturing apparatus. An adhesive layer may be formed on the attachment portion in advance or an adhesive agent may be applied thereto at the time of attachment. Alternatively, theapparatus state sensor 108 may be attached to an existing manufacturing apparatus by screwing the fixing portion into the manufacturing apparatus. - To realize the automatic diagnosis of a portion in a manufacturing apparatus having a problem, it is necessary to automatically collect the operation state data of each unit of a manufacturing apparatus. In the case of achieving the automatic diagnosis by buying or replacing with a new manufacturing apparatus, the cost increases significantly. However, in the present embodiment, it is possible to realize the automatic diagnosis by simply attaching a sensor available at a significantly low cost to each unit of an existing manufacturing apparatus. Therefore, it is possible to add the environment of the automatic problem diagnosis inexpensively while maintaining the environment of an existing apparatus.
- As shown in
FIG. 8A , theapparatus state sensor 108 is fixed, for example, to a part of a movingbelt 136, a movable arm 124 configured to hold products or a part of amovable shaft 132. Then, if a portion which moves in a normal operation stands still, it is determined that the movable portion has a problem. - The
controller 118 in thedata management server 116 stores handbooks for repairing, maintaining and checking the respective portions of various manufacturing apparatuses, namely, maintenance procedure handbooks in advance, and makes an appropriate work checklist based on a result of the above-described automatic diagnosis. - In
FIGS. 8A and 8B , as an example of theapparatus state sensor 108, an acceleration detection method has been described. However, the above-described method is in no way restrictive, and various physical quantities or chemical quantities such as a temperature or a conducting current value may also be used. Further, it is also possible to perform the automatic diagnosis of the portion in the manufacturing apparatus having a problem by comparing images captured by a camera or environmental sounds collected by a microphone. - If a manufacturing apparatus having a problem is detected by the method described above with reference to
FIGS. 8A and 8B , thedata management server 116 automatically selects an operator who is to perform maintenance of the manufacturing apparatus and displays maintenance procedure or a work checklist derived from the maintenance procedure on thewearable device 10 of the operator. Thedata management server 116 selects an operator, for example, (i) who is near the manufacturing apparatus having a problem, (ii) who can stop the work the operator is currently engaging with and (iii) who can perform the maintenance work. In this way, it is possible to minimize a time loss in dispatching an operator to the manufacturing device having a problem. - As a method of most efficiently searching an operator near the manufacturing apparatus having a problem, in the present embodiment, a
photo detector 106 a similar to thephoto detector 62 ofFIG. 4 and awireless communication device 106 b are attached to a part of themanufacturing apparatus 106 as shown inFIG. 9 . As described above with reference toFIGS. 5A-5K , the leakinglight 50 radiating from thewearable device 10 includes the device ID data of thedevice 10. Therefore, if the data included in the leakinglight 50 is detected by thephoto detector 106 a and transmitted to thedata management server 116 via thewireless communication device 106 b and thenetwork 102, thedata management server 116 can recognize thewearable devices 10, that is, the operators near the manufacturing apparatus having a problem. Thedata management server 116 selects an operator who is to perform maintenance of the target manufacturing apparatus based on the data and transmits a work checklist to thewearable device 10 of the operator. As shown inFIG. 13A , the work checklist is displayed on thescreen 16 of thedevice 10. Note that, although the work checklist is simplified inFIG. 13A for the sake of convenience, the actual work checklist is as follows. - Put the thing in the cart.
- Close the valve.
- Flick off the on/off switch.
- Flick off the first light switch.
- Flick off the second light switch.
- As described above, since the leaking
light 50 radiating from thewearable device 10 is detected, collected and summarized in real time, it is possible to identify an operator near themanufacturing apparatus 106 subjected to maintenance easily and accurately. Consequently, it becomes possible to save the time of dispatching an operator to the apparatus and reduce the maintenance time, and thereby prevent a decrease in the manufacturing efficiency. - Note that, as another method of recognizing operators near the target manufacturing apparatus, there is a method of using the
camera 114 provided near themanufacturing apparatus 106. Animage sensor 114 a in thecamera 114 captures an image around themanufacturing apparatus 106 and transmits the image data to thedata management server 116 via awireless communication 114 b. Thedata management server 116 analyzes the received image data and automatically identifies operators therein. - With reference to
FIGS. 10A and 10B , an example of a case where an operator performs work in accordance with a work checklist will be described. If a work checklist ofFIG. 13A is displayed on thescreen 16, an operator starts work. Here, if the whole checklist is displayed at a time, there are some cases where the letters in the checklist become too small to see. In that case, only first or first few steps may be displayed by larger letters, and then the checklist may be updated as each step ends by automatically recognizing the progress of work step by step. The operator in the operation site ofFIG. 10A puts athing 162 in acart 164, closes a valve 170 (or turns ahandle 170 to a specified angle), flicks off an on/offswitch 172, flick off a first light switch 176 and a third light switch 180 according to the work checklist. In the present embodiment, the movements of the operator are automatically recognized and identified in real time by theuser movement sensor 110 attached to themanufacturing apparatus 106 and the work completion times are automatically written in the work checklist (seeFIG. 13B ). If the last work step ends, a work report is automatically made in thedata management server 116, and the contents are displayed on the supervisor'sdevice 104. The work checklist (FIG. 13B ) corresponds to the work checklist (FIG. 13A ) containing the completion times input thereto. - As a method of automatic recognition and identification of user's movements by the
user movement sensor 110, various detection techniques and the combinations thereof may be used. For example, it is possible to perform automatic recognition and identification of user's movements by using thecamera camera 114, depending on the situation, the user's movements may be hidden behind in the image. Alternatively, it is possible to use a sound recognition technique. If it is determined in advance that the operator produces a specific sound as the operator finishes with the work of each item of the work checklist (maintenance work procedure) displayed on thewearable device 10, it is possible to perform automatic recognition and identification of the user's movements by detecting an input of the specific sound with themicrophone 148. Alternatively, it is also possible to perform automatic recognition and identification of the user's movements by detecting environmental sounds produced in specific work using themicrophone 148 or a built-in microphone of theapparatus state sensor 108. Further, there is a method of identifying a predetermined operator's gesture and thereby performing automatic recognition and identification of operator's movements. As a method of identifying an operator's gesture, images of operator's movements captured by thecameras light 50 of thewearable devices 10 by a plurality of thephoto detectors 62 or thephoto detectors 106 a attached to themanufacturing apparatuses 106 may be compared with each other. - A pair of a
light emitting device 166 a and aphoto detector 166 b is provided in an opening portion of thecart 164, and it is automatically detected that thething 162 is put in or taken out of thecart 164 by detecting the interception of light caused if thething 162 passes through the opening portion of thecart 164.FIG. 10B shows the signal characteristics detected by thephoto detector 166 b if thething 162 is put in or taken out of thecart 164. The vertical axis shows the amount of light detected by thephoto detector 166 b while the horizontal axis shows the time passed. If thething 162 passes through the opening portion of thecart 164, the amount of light thephoto detector 166 b detects decreases. As a method of detecting that thething 162 is put in or taken out of thecart 164, not the above-described method using light but various other methods may be used. - An example of the detection method which realizes the real-time automatic recognition and identification of movements other than the putting in or taking out of the thing such as the closing the valve, the flicking off the on/off switch and the flicking off the light switch will be described below. In general, to perform maintenance (maintenance, checkups and repairs) of the manufacturing apparatus, the operator needs to directly contact a predetermined portion in the manufacturing apparatus. By using this feature, in the present embodiment, if it is detected that the operator contacts a predetermined portion in the manufacturing apparatus, the detection result is reflected in the automatic recognition and identification of the operator's movements. According to this method, it is possible to perform detection easily and perform automatic recognition and identification with high accuracy. In the case of
FIG. 10A , acontact sensor 168 is attached to thevalve 170, and transparent contact sensors are attached respectively to the on/offswitch 172 and alight switch board 174. Thelight switch board 174 includes the first, second and thirdlight switches 176, 178 and 180. - The contact sensor as an example of the
user movement sensor 110 includes a wireless communication function (for example, a short-range wireless communication) and a detection function of detecting the contact state of the operator. In the detection of the contact state, various elements configured to perform contact detection such as a piezoelectric element, a photo interrupter and an acceleration sensor (gyroscope sensor) can be used. The contact sensor of this type is attachable to existing facilities such as existing manufacturing apparatuses and is available at an extremely low cost. Therefore, it is possible to add a short-range wireless communication network environment inexpensively by simply attaching the contact sensor (user movement sensor 110) to an existing manufacturing apparatus while maintaining the existing infrastructure. - An example of the user movement sensor 100 is shown in
FIGS. 11 and 12 .FIG. 11 shows theuser movement sensor 110 attached to an existing infrastructure, namely, the on/offswitch 172 or the user movement sensor 110 (contact sensor 168) attached to thevalve 170, whileFIG. 12 shows theuser movement sensor 110 attached to an existing infrastructure, namely, thelight switch board 174. - As shown in
FIG. 11 , theuser movement sensor 110 includes anadhesive layer 202 at the bottom, and further includes a control andcommunication circuit 204 and asolar cell 206 formed in this order on theadhesive layer 202. On thesolar cell 206, a transparentconductive layer 208, a transparentintermediate layer 210, a transparentconductive layer 212 and a transparentuneven layer 214 are stacked one after another. The control andcommunication circuit 204 includes a function of performing wireless communication (short-range wireless communication) and a function of detecting contact with the operator. Thecircuit 204 is driven by thesolar cell 206 to perform these functions. In the case of using a battery as a power supply, there is the trouble of battery replacement. Further, in the case of using an external power supply connected to a cable as a power supply, the cable blocks the operator from contacting. However, in the case of thesolar cell 206, it is possible to drive theuser movement sensor 110 for a long period of time without giving the trouble of battery replacement or obstructing the operator from contacting. - By stacking the control and
communication circuit 204 configured to perform a short-range wireless communication and execute control below thesolar cell 206, it is possible to increase the power generation efficiency of thesolar cell 206 and reduce the plane size of theuser movement sensor 110. - To use the
solar cell 206, thesolar cell 206 needs to be irradiated with surrounding light. Meanwhile, it is preferable that the portion configured to detect the user's contact is provided on the surface of theuser movement sensor 110. As a method of satisfying both demands at the same time, a capacitance type detection method is adopted and the contact detection portion is made transparent. To detect the operator's contact or pressure by using a change in capacitance, the following structure is adopted: the transparentintermediate layer 210 formed of a transparent and elastic material (for example, a transparent organic material sheet) is sandwiched between the two transparentconductive layers 208 and 212 (for example, transparent organic material sheets). By applying an alternating-current voltage 216 between the two transparentconductive layers conductive layers user movement sensor 110, a change occurs in the capacitance, and consequently a change occurs in the above-described AC resonance state. By detecting a change in the AC resonance state, the operator's contact is detected. Note that this capacitance type detection method may not necessarily be used but any element may be used as long as the element allows at least a part of surrounding light to reach thesolar cell 206 in theuser movement sensor 110 and is configured to detect contact or pressure. - The transparent layer on the surface of the
user movement sensor 110 is, for example, provided with small asperities. This functions as a non-slip surface, but it is possible to record information in Braille using the asperities for the sake of people with impaired vision. - As a method of fixing the
user movement sensor 110 to a part of an existing manufacturing apparatus, although various fixing methods such as screwing may be adopted, it is possible to save space by directly bonding or attaching thesensor 110 thereto. As the bonding or attaching method, not only a method of directly bonding with an adhesive agent but also a method of using an adhesive sheet or an adhesive tape may be used. Theadhesive layer 202 having characteristics of a double-faced adhesive tape may be used for the on/offswitch 172 and thelight switch board 174, and on the other hand, theadhesive layer 202 formed of a transparent adhesive layer may be used for thevalve 170. -
FIG. 12 shows theuser movement sensor 110 attached to thelight switch board 174. In thelight switch board 174, since letters such aslight 1,light 2 and light 3 are written on the surfaces of the first, second and thirdlight switches 176, 178 and 180, it is preferable that these letters be seen directly even if theuser movement sensor 110 is attached. Therefore, the layers provided above thelight switches 176, 178 and 180 preferably be transparent. Further, it is necessary to detect the contact states of the plurality oflight switches 176, 178 and 180, respectively. Meanwhile, in thelight switch board 174, there is aspace 182 left in a portion not provided with thelight switches 176, 178 and 180. To conform to such a situation, theuser movement sensor 110 ofFIG. 12 includes thetransparent sheet 208, the transparentintermediate layer 210, thetransparent sheet 212 and the transparentuneven layer 214 stacked in series on theadhesive layer 202. Thetransparent sheets conductive layers FIG. 11 , and thetransparent sheet 208 includes three transparentconductive regions transparent sheet 212 includes three transparentconductive regions 212 a, 212 b and 212 c. The transparentconductive regions 208 a and 212 a are provided in the positions corresponding to the first light switch 176, the transparentconductive regions 208 b and 212 b are provided in the positions corresponding to the secondlight switch 178, and the transparentconductive regions AC voltage 216 is applied between thetransparent sheets uneven layer 214. - In a portion on the transparent
uneven layer 214 corresponding to theleft space 182 not provided with the light switches, acontrol circuit 204 a and acommunication circuit 204 b are formed, and thesolar cell 206 is formed thereon. Since thesolar cell 206 is provided on the top, the power generation efficiency is high. Further, since thecontrol circuit 204 a, thecommunication circuit 204 b and thesolar cell 206 are located in the vertical direction, the plane size of theuser movement sensor 110 is reduced. - According to the embodiment, by detecting the states of wearable devices and manufacturing apparatuses and displaying, based on the detection result, a workflow on the wearable device of an operator who is near a manufacturing apparatus requiring maintenance, checkups and repairs and who can perform the maintenance work, it is possible to provide the operator with useful information. Further, since completion of each step of the work is determined and a work report recording the progress of the work is made automatically based on the detection result of the states of the wearable device and the manufacturing apparatus, it is possible to save the operator the troubles thereof. Note that, since the detection of completion of work is realized simply by attaching a contact sensor to a manufacturing apparatus, it is possible to detect and recognize movements of the operator quite easily, inexpensively and accurately without making modifications to an existing manufacturing apparatus.
- The present embodiment describes the case of performing maintenance of a manufacturing apparatus. However, the present embodiment is not necessarily limited to this case but may be applied to a case of monitoring user's movements corresponding to other purposes and displaying work contents according to the purposes. Further, the present embodiment describes providing a contact sensor for monitoring the user's movements in a portion which the user is likely to contact, but other sensors may be used instead.
- Although the present embodiment describes the case of an eyeglasses-type wearable device, the present embodiment is also applicable to head-mounted type wearable devices of other types such as goggles and helmet types as well as to a wristband-type wearable device, a pendant-type wearable device and the like. For example, in the case of a helmet or goggles-type wearable device, since the
projector 12 and thecamera 59 can be attached to the helmet or the goggles, people with eye glasses can also use the wearable device. Further, in the case of a helmet-type wearable device, since thespeaker 54 can be attached to the inner side of the helmet, the user can hear a sound clearly, and since a microphone can be attached to the helmet and the position of the microphone can be adjusted, the sound collection performance of the microphone improves. - As the sensors configured to detect the states of a wearable device and a manufacturing apparatus, various other sensors may be used appropriately instead of the sensors described above.
- The present embodiment is applicable to wearable devices other than head-mounted type wearable devices. The present embodiment is also applicable to portable and light electronic devices carried with the users at all times as notebook computers, tablet computers and smart phones.
- As to the share of functions between the wearable device and the data management server, the above description is in no way restrictive, but instead, some of the above-described functions of the wearable device may be realized as those of the data management server or some of the above-described functions of the data management server may be realized as those of the wearable device.
- The various modules of the systems described herein can be implemented as software applications, hardware and/or software modules, or components on one or more computers, such as servers. While the various modules are illustrated separately, they may share some or all of the same underlying logic or code.
- While certain embodiments have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel embodiments described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of the embodiments described herein may be made without departing from the spirit of the inventions. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the inventions.
Claims (6)
1. A wearable device comprising:
a display configured to display an image; and
an optical system configured to project the image displayed by the display at a distance away from a user of the wearable device, wherein the image displayed by the display changes in response to a movement of the user.
2. The wearable device of claim 1 , wherein
the display and the optical system are formed at a pair of eyeglasses,
the display comprises a flat panel display formed at a temple of the eyeglasses, and
the optical system comprises a lens configured to project the image displayed by the flat panel display and a half mirror configured to form a projected image at the distance away from the user.
3. The wearable device of claim 1 , wherein the display is connected to an external server and is configured to display the image based on an image signal transmitted from the external server, and the external server configured to change the display signal in response to a change in the movement of the user.
4. The wearable device of claim 1 , wherein the external server is further connected to a sensor configured to detect a movement of the user.
5. A system comprising:
a wearable device; and
a server connected to the wearable device, wherein the wearable device comprises:
a display configured to display an image; and
an optical system configured to project the image displayed by the display at a distance away from a user of the wearable device, and
the server is configured to detect a movement of the user and causes the display change the image in response to the movement of the user.
6. The system of claim 5 , wherein
the display and the optical system are formed at a pair of eyeglasses,
the display comprises a flat panel display formed at a temple of the eyeglasses, and
the optical system comprises a lens configured to project the image displayed by the flat panel display and a half mirror configured to project the image projected to by the lens at the distance away from the user.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/685,774 US20170351920A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
US15/917,332 US20180197011A1 (en) | 2015-09-01 | 2018-03-09 | System and method |
US15/917,328 US10679059B2 (en) | 2015-09-01 | 2018-03-09 | System and method for visual image adjustment |
US16/592,664 US10682405B2 (en) | 2015-09-01 | 2019-10-03 | System and method and device for adjusting image positioning |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2015-171933 | 2015-09-01 | ||
JP2015171933A JP2017049762A (en) | 2015-09-01 | 2015-09-01 | System and method |
US14/979,169 US10685232B2 (en) | 2015-09-01 | 2015-12-22 | Wearable device for displaying checklist of a work |
US15/685,774 US20170351920A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/979,169 Division US10685232B2 (en) | 2015-09-01 | 2015-12-22 | Wearable device for displaying checklist of a work |
Related Child Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/917,332 Division US20180197011A1 (en) | 2015-09-01 | 2018-03-09 | System and method |
US15/917,328 Division US10679059B2 (en) | 2015-09-01 | 2018-03-09 | System and method for visual image adjustment |
US16/592,664 Division US10682405B2 (en) | 2015-09-01 | 2019-10-03 | System and method and device for adjusting image positioning |
Publications (1)
Publication Number | Publication Date |
---|---|
US20170351920A1 true US20170351920A1 (en) | 2017-12-07 |
Family
ID=58095747
Family Applications (12)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/979,169 Active US10685232B2 (en) | 2015-09-01 | 2015-12-22 | Wearable device for displaying checklist of a work |
US15/685,782 Abandoned US20170351921A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
US15/685,766 Abandoned US20170351919A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
US15/685,774 Abandoned US20170351920A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
US15/917,328 Active US10679059B2 (en) | 2015-09-01 | 2018-03-09 | System and method for visual image adjustment |
US15/917,332 Abandoned US20180197011A1 (en) | 2015-09-01 | 2018-03-09 | System and method |
US16/592,611 Active US10671849B2 (en) | 2015-09-01 | 2019-10-03 | System and method for sensor based visual adjustments |
US16/592,664 Active US10682405B2 (en) | 2015-09-01 | 2019-10-03 | System and method and device for adjusting image positioning |
US16/851,088 Active US11002975B2 (en) | 2015-09-01 | 2020-04-16 | System and method for image generation based on a display-attachable wearable device |
US17/210,420 Active US11428944B2 (en) | 2015-09-01 | 2021-03-23 | Wearable device and method for visual image adjustment |
US17/897,968 Active US11789279B2 (en) | 2015-09-01 | 2022-08-29 | System and method for virtual image adjustment |
US18/451,535 Active US12135432B2 (en) | 2015-09-01 | 2023-08-17 | System and method directed to an eyeglasses-type wearable device |
Family Applications Before (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/979,169 Active US10685232B2 (en) | 2015-09-01 | 2015-12-22 | Wearable device for displaying checklist of a work |
US15/685,782 Abandoned US20170351921A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
US15/685,766 Abandoned US20170351919A1 (en) | 2015-09-01 | 2017-08-24 | System and method |
Family Applications After (8)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/917,328 Active US10679059B2 (en) | 2015-09-01 | 2018-03-09 | System and method for visual image adjustment |
US15/917,332 Abandoned US20180197011A1 (en) | 2015-09-01 | 2018-03-09 | System and method |
US16/592,611 Active US10671849B2 (en) | 2015-09-01 | 2019-10-03 | System and method for sensor based visual adjustments |
US16/592,664 Active US10682405B2 (en) | 2015-09-01 | 2019-10-03 | System and method and device for adjusting image positioning |
US16/851,088 Active US11002975B2 (en) | 2015-09-01 | 2020-04-16 | System and method for image generation based on a display-attachable wearable device |
US17/210,420 Active US11428944B2 (en) | 2015-09-01 | 2021-03-23 | Wearable device and method for visual image adjustment |
US17/897,968 Active US11789279B2 (en) | 2015-09-01 | 2022-08-29 | System and method for virtual image adjustment |
US18/451,535 Active US12135432B2 (en) | 2015-09-01 | 2023-08-17 | System and method directed to an eyeglasses-type wearable device |
Country Status (2)
Country | Link |
---|---|
US (12) | US10685232B2 (en) |
JP (1) | JP2017049762A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10671849B2 (en) | 2015-09-01 | 2020-06-02 | Kabushiki Kaisha Toshiba | System and method for sensor based visual adjustments |
US11983959B2 (en) | 2019-04-18 | 2024-05-14 | Beckman Coulter, Inc. | Securing data of objects in a laboratory environment |
US12001600B2 (en) | 2018-11-09 | 2024-06-04 | Beckman Coulter, Inc. | Service glasses with selective data provision |
Families Citing this family (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USD800158S1 (en) * | 2015-05-15 | 2017-10-17 | Metabeauty, Inc. | Display screen or portion thereof with a graphical user interface |
JP1559731S (en) * | 2016-02-25 | 2016-10-03 | ||
JP1560139S (en) * | 2016-02-25 | 2016-10-03 | ||
US11004355B2 (en) * | 2016-03-25 | 2021-05-11 | Shenzhen Augmented Reality Technologies Co., Ltd. | Intelligent wearable device, and working assistance method and system based thereon |
US20170309152A1 (en) * | 2016-04-20 | 2017-10-26 | Ulysses C. Dinkins | Smart safety apparatus, system and method |
CN106059615A (en) * | 2016-08-11 | 2016-10-26 | 京东方科技集团股份有限公司 | Positioning method and system, and wearable device |
JP6512452B2 (en) * | 2016-09-23 | 2019-05-15 | 京セラドキュメントソリューションズ株式会社 | Adjustment work support system |
US10782775B2 (en) * | 2017-01-13 | 2020-09-22 | Atheer, Inc. | Methods and apparatus for providing procedure guidance |
JP6845072B2 (en) | 2017-04-21 | 2021-03-17 | ファナック株式会社 | Factory equipment maintenance support equipment and maintenance support system |
DE102017108622A1 (en) * | 2017-04-23 | 2018-10-25 | Goodly Innovations GmbH | SYSTEM FOR SUPPORTING TEAMWORK BY AUGMENTED REALITY |
WO2018217910A1 (en) * | 2017-05-25 | 2018-11-29 | Spot You More, Inc. | Task monitoring |
JP6817153B2 (en) * | 2017-06-16 | 2021-01-20 | 株式会社日立製作所 | Alarm notification system and alarm notification method |
WO2019104260A1 (en) | 2017-11-27 | 2019-05-31 | Spot You More, Inc. | Smart shelf sensor |
JP7114885B2 (en) * | 2017-11-29 | 2022-08-09 | 沖電気工業株式会社 | Worksite monitoring devices and programs |
JP2019106158A (en) * | 2018-01-18 | 2019-06-27 | 株式会社デンソー九州 | production management system |
CN108322674A (en) * | 2018-01-25 | 2018-07-24 | 上海康斐信息技术有限公司 | A kind of method and system that video switching is launched |
GB2583678A (en) * | 2018-01-26 | 2020-11-04 | Elements Tech Platforms Limited | Process management with location identification |
US10863812B2 (en) * | 2018-07-18 | 2020-12-15 | L'oreal | Makeup compact with eye tracking for guidance of makeup application |
CN110133852A (en) * | 2019-05-13 | 2019-08-16 | 苏州经贸职业技术学院 | Intelligence beautification glasses |
JP7248516B2 (en) * | 2019-06-20 | 2023-03-29 | 株式会社日立製作所 | Work support device, work support method, and work support program |
US11816526B2 (en) | 2019-06-24 | 2023-11-14 | Suncorporation | Function execution system |
EP4028826A4 (en) * | 2019-09-09 | 2023-10-11 | Securetest LLC | Secure testing device with optical element |
JP2021157284A (en) * | 2020-03-25 | 2021-10-07 | 富士フイルムビジネスイノベーション株式会社 | Information processing device, information processing system, and program |
US11151898B1 (en) * | 2020-04-15 | 2021-10-19 | Klatt Works, Inc. | Techniques for enhancing workflows relating to equipment maintenance |
JP7222387B2 (en) * | 2020-10-13 | 2023-02-15 | 横河電機株式会社 | Apparatus, method and program |
US11520980B2 (en) | 2020-11-25 | 2022-12-06 | Klatt Works, Inc. | Techniques for enhancing an electronic document with an interactive workflow |
US11771162B2 (en) * | 2020-11-25 | 2023-10-03 | Brett Krupps | Test environment monitoring device for online proctoring |
CN112764657B (en) * | 2020-12-31 | 2022-05-20 | 维沃移动通信有限公司 | Electronic device and control method of electronic device |
JP7575341B2 (en) | 2021-05-17 | 2024-10-29 | 株式会社日立製作所 | Work support device and work support method |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6452544B1 (en) * | 2001-05-24 | 2002-09-17 | Nokia Corporation | Portable map display system for presenting a 3D map image and method thereof |
US20150260993A1 (en) * | 2012-10-11 | 2015-09-17 | Sony Computer Entertainment Europe Limited | Head mountable device |
US20170059869A1 (en) * | 2014-05-15 | 2017-03-02 | Jun Hee Lee | Optical system for head mount display |
US20170235152A1 (en) * | 2014-09-26 | 2017-08-17 | Osterhout Group, Inc. | See-through computer display systems |
Family Cites Families (118)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2862019B2 (en) | 1990-08-01 | 1999-02-24 | 富士通株式会社 | Video superimposition device |
JP3032369B2 (en) | 1992-01-30 | 2000-04-17 | 株式会社フジタ | Two-way information transmission system |
JPH0793020A (en) | 1993-09-22 | 1995-04-07 | Nissan Motor Co Ltd | Maintenance staff positioning managing device |
JPH07121119A (en) | 1993-10-26 | 1995-05-12 | Olympus Optical Co Ltd | Head mounting type video display device |
US6600461B1 (en) * | 1994-10-12 | 2003-07-29 | Canon Kabushiki Kaisha | Display apparatus and control method therefor |
JP3324945B2 (en) | 1996-11-15 | 2002-09-17 | 日本輸送機株式会社 | Picking work method |
JPH10198289A (en) | 1997-01-08 | 1998-07-31 | Fuji Xerox Co Ltd | Picture display device |
JP2000354943A (en) | 1999-06-09 | 2000-12-26 | Nippon Telegr & Teleph Corp <Ntt> | Work management and support method and device, and recording medium having program therefor recorded |
JP2002065099A (en) * | 2000-09-04 | 2002-03-05 | Kyowa Kikai Kk | Tending system for automatic egg treatment apparatus, automatic egg treatment apparatus and tending computer used for the system |
JP2002162597A (en) | 2000-11-22 | 2002-06-07 | Toshiba Corp | Wearable display device |
JP2002288294A (en) | 2001-03-26 | 2002-10-04 | Mitsubishi Heavy Ind Ltd | Site-supporting system |
US8719661B1 (en) * | 2010-11-15 | 2014-05-06 | Vmware, Inc. | Transparent and lightweight recovery from hardware memory errors |
JP2003196681A (en) | 2001-12-26 | 2003-07-11 | Sharp Corp | Work instruction system, work instruction method, and recording medium storing program for realizing the same |
JP2003216687A (en) | 2002-01-28 | 2003-07-31 | Yamaha Livingtec Corp | Execution support system |
JP2004101197A (en) | 2002-09-04 | 2004-04-02 | Sony Corp | Mobile type positional information transmitting device and navigation method |
JP2004102727A (en) | 2002-09-10 | 2004-04-02 | Mitsubishi Heavy Ind Ltd | Work support system |
WO2004061519A1 (en) * | 2002-12-24 | 2004-07-22 | Nikon Corporation | Head mount display |
US7135971B2 (en) * | 2003-09-22 | 2006-11-14 | Walter Taehwan Kim | Anti-theft and security system for computers |
JP2005250990A (en) | 2004-03-05 | 2005-09-15 | Mitsubishi Electric Corp | Operation support apparatus |
JP2006003157A (en) | 2004-06-16 | 2006-01-05 | Nippon Telegr & Teleph Corp <Ntt> | Position detection system |
AU2005304281A1 (en) * | 2004-11-12 | 2006-05-18 | Garrett Thermal Systems Limited | Particle detector, system and method |
JP2006146803A (en) | 2004-11-24 | 2006-06-08 | Olympus Corp | Operation device, and remote operation system |
WO2006098097A1 (en) | 2005-03-14 | 2006-09-21 | Nikon Corporation | Image display optical system and image display |
US7664116B2 (en) * | 2005-04-12 | 2010-02-16 | Fujitsu Limited | Network based routing scheme |
JP2007121600A (en) | 2005-10-27 | 2007-05-17 | Yokogawa Electric Corp | Image information display device |
JP4411547B2 (en) | 2006-03-20 | 2010-02-10 | ソニー株式会社 | Image display device |
JP4622937B2 (en) | 2006-06-14 | 2011-02-02 | 株式会社島津製作所 | Head-mounted display device |
JP4927631B2 (en) | 2006-06-27 | 2012-05-09 | パナソニック株式会社 | Display device, control method therefor, program, recording medium, and integrated circuit |
US7928926B2 (en) | 2006-06-27 | 2011-04-19 | Panasonic Corporation | Display apparatus and method for hands free operation that selects a function when window is within field of view |
JP4884256B2 (en) | 2007-02-22 | 2012-02-29 | 株式会社日立製作所 | Work management system, work management method, and management computer |
JP2009128565A (en) | 2007-11-22 | 2009-06-11 | Toshiba Corp | Display device, display method and head-up display |
JP5493287B2 (en) | 2008-04-15 | 2014-05-14 | 株式会社リコー | Detection device and position information management system |
JP2009279193A (en) | 2008-05-22 | 2009-12-03 | Fujifilm Corp | Medical apparatus management system |
JP5215098B2 (en) | 2008-09-17 | 2013-06-19 | オリンパス株式会社 | Information processing system, program, and information storage medium |
US8350871B2 (en) * | 2009-02-04 | 2013-01-08 | Motorola Mobility Llc | Method and apparatus for creating virtual graffiti in a mobile virtual and augmented reality system |
JP2010271928A (en) | 2009-05-21 | 2010-12-02 | Kanto Auto Works Ltd | Work assist system, work assist method, and recording medium recording the same |
JP2010272041A (en) * | 2009-05-25 | 2010-12-02 | Toa Engineering Co Ltd | Centralized management system for timber precut operation |
JP2011081737A (en) | 2009-10-09 | 2011-04-21 | Toshiba Tec Corp | Cooking assistance terminal and program |
JP4755733B2 (en) | 2009-10-14 | 2011-08-24 | パイオニア株式会社 | Sheet-type remote control device, setting method, setting program, and storage medium |
JP5193988B2 (en) | 2009-12-03 | 2013-05-08 | 東芝テック株式会社 | Cooking assistance terminal and program |
US8984665B2 (en) * | 2010-02-02 | 2015-03-24 | Wilcox Industries Corp. | Helmet mounting system and mounting shoe interface |
US8964298B2 (en) | 2010-02-28 | 2015-02-24 | Microsoft Corporation | Video display modification based on sensor input for a see-through near-to-eye display |
CN102906623A (en) * | 2010-02-28 | 2013-01-30 | 奥斯特豪特集团有限公司 | Local advertising content on an interactive head-mounted eyepiece |
JP4913913B2 (en) | 2010-04-28 | 2012-04-11 | 新日鉄ソリューションズ株式会社 | Information processing system, information processing method, and program |
JP5651386B2 (en) | 2010-06-23 | 2015-01-14 | ソフトバンクモバイル株式会社 | Eyeglass type display device |
JP5496030B2 (en) | 2010-09-16 | 2014-05-21 | オリンパス株式会社 | Head-mounted image display device |
EP2619749A4 (en) * | 2010-09-21 | 2017-11-15 | 4IIII Innovations Inc. | Head-mounted peripheral vision display systems and methods |
JP4945691B2 (en) | 2011-03-10 | 2012-06-06 | 株式会社東芝 | Display device, display method, and head-up display |
US9217867B2 (en) * | 2011-03-24 | 2015-12-22 | Seiko Epson Corporation | Head-mounted display device and control method for the head-mounted display device |
JP5423716B2 (en) | 2011-03-30 | 2014-02-19 | ブラザー工業株式会社 | Head mounted display |
JP5807803B2 (en) | 2011-04-28 | 2015-11-10 | 大日本印刷株式会社 | Projection device and projection control device |
JP5638464B2 (en) * | 2011-05-18 | 2014-12-10 | 三菱重工業株式会社 | Work process management system and work process management method |
JP5348192B2 (en) | 2011-07-11 | 2013-11-20 | 日本電気株式会社 | Work support system, terminal, method and program |
WO2013049248A2 (en) | 2011-09-26 | 2013-04-04 | Osterhout Group, Inc. | Video display modification based on sensor input for a see-through near-to-eye display |
WO2013111267A1 (en) * | 2012-01-24 | 2013-08-01 | パイオニア株式会社 | Head-mounted display and display method |
EP3270194B1 (en) | 2012-01-24 | 2020-10-14 | The Arizona Board of Regents on behalf of The University of Arizona | Compact eye-tracked head-mounted display |
JP6097703B2 (en) * | 2012-02-08 | 2017-03-15 | 学校法人 埼玉医科大学 | Axisymmetric polarization conversion element |
US20130217294A1 (en) | 2012-02-17 | 2013-08-22 | Arjuna Ragunath Karunaratne | Toy brick with sensing, actuation and control |
US20160361662A1 (en) | 2012-02-17 | 2016-12-15 | Technologyone, Inc. | Interactive lcd display back light and triangulating toy brick baseplate |
CN103930182B (en) | 2012-02-17 | 2017-02-15 | 技术第一公司 | Baseplate assembly for use with toy pieces |
JP6028357B2 (en) | 2012-03-22 | 2016-11-16 | ソニー株式会社 | Head mounted display and surgical system |
JP6186689B2 (en) | 2012-09-26 | 2017-08-30 | セイコーエプソン株式会社 | Video display system |
JP6364735B2 (en) | 2013-10-04 | 2018-08-01 | セイコーエプソン株式会社 | Display device, head-mounted display device, display device control method, and head-mounted display device control method |
US9542958B2 (en) | 2012-12-18 | 2017-01-10 | Seiko Epson Corporation | Display device, head-mount type display device, method of controlling display device, and method of controlling head-mount type display device |
JP6076099B2 (en) | 2013-01-18 | 2017-02-08 | オリンパス株式会社 | Glasses-type operating device and glasses-type operating system |
TWI614813B (en) | 2013-01-21 | 2018-02-11 | 半導體能源研究所股份有限公司 | Method for manufacturing semiconductor device |
JP6318462B2 (en) | 2013-01-30 | 2018-05-09 | 株式会社ニコン | Display device |
USD748203S1 (en) | 2013-02-06 | 2016-01-26 | Technologyone, Inc. | Baseplate for use with toy pieces |
EP2957816B1 (en) * | 2013-02-14 | 2018-03-14 | Factory Inc. | Illuminating instrument |
US20140240349A1 (en) * | 2013-02-22 | 2014-08-28 | Nokia Corporation | Method and apparatus for presenting task-related objects in an augmented reality display |
JP6082272B2 (en) | 2013-02-25 | 2017-02-15 | 東京エレクトロン株式会社 | Support information display method, substrate processing apparatus maintenance support method, support information display control apparatus, substrate processing system, and program |
US10866952B2 (en) | 2013-03-04 | 2020-12-15 | Fisher-Rosemount Systems, Inc. | Source-independent queries in distributed industrial system |
US9397836B2 (en) | 2014-08-11 | 2016-07-19 | Fisher-Rosemount Systems, Inc. | Securing devices to process control systems |
US10223327B2 (en) | 2013-03-14 | 2019-03-05 | Fisher-Rosemount Systems, Inc. | Collecting and delivering data to a big data machine in a process control system |
US10649424B2 (en) | 2013-03-04 | 2020-05-12 | Fisher-Rosemount Systems, Inc. | Distributed industrial performance monitoring and analytics |
US10678225B2 (en) | 2013-03-04 | 2020-06-09 | Fisher-Rosemount Systems, Inc. | Data analytic services for distributed industrial performance monitoring |
US9665088B2 (en) | 2014-01-31 | 2017-05-30 | Fisher-Rosemount Systems, Inc. | Managing big data in process control systems |
US10282676B2 (en) | 2014-10-06 | 2019-05-07 | Fisher-Rosemount Systems, Inc. | Automatic signal processing-based learning in a process plant |
US9804588B2 (en) | 2014-03-14 | 2017-10-31 | Fisher-Rosemount Systems, Inc. | Determining associations and alignments of process elements and measurements in a process |
US9558220B2 (en) | 2013-03-04 | 2017-01-31 | Fisher-Rosemount Systems, Inc. | Big data in process control systems |
US9823626B2 (en) | 2014-10-06 | 2017-11-21 | Fisher-Rosemount Systems, Inc. | Regional big data in process control systems |
US10386827B2 (en) | 2013-03-04 | 2019-08-20 | Fisher-Rosemount Systems, Inc. | Distributed industrial performance monitoring and analytics platform |
US10909137B2 (en) | 2014-10-06 | 2021-02-02 | Fisher-Rosemount Systems, Inc. | Streaming data for analytics in process control systems |
US10649449B2 (en) | 2013-03-04 | 2020-05-12 | Fisher-Rosemount Systems, Inc. | Distributed industrial performance monitoring and analytics |
US20160132046A1 (en) | 2013-03-15 | 2016-05-12 | Fisher-Rosemount Systems, Inc. | Method and apparatus for controlling a process plant with wearable mobile control devices |
DE112014001381T5 (en) | 2013-03-15 | 2016-03-03 | Fisher-Rosemount Systems, Inc. Emerson Process Management | Data Modeling Studio |
US9678484B2 (en) | 2013-03-15 | 2017-06-13 | Fisher-Rosemount Systems, Inc. | Method and apparatus for seamless state transfer between user interface devices in a mobile control room |
JP5939191B2 (en) * | 2013-03-29 | 2016-06-22 | ブラザー工業株式会社 | Head mounted display |
JP6197366B2 (en) | 2013-05-23 | 2017-09-20 | ソニー株式会社 | Information processing apparatus and storage medium |
JP2014228817A (en) | 2013-05-27 | 2014-12-08 | セイコーエプソン株式会社 | Image display device and head-mounted display |
JP2015001468A (en) * | 2013-06-17 | 2015-01-05 | 株式会社松井製作所 | Molding inspection device |
US9199128B2 (en) * | 2013-06-28 | 2015-12-01 | Acushnet Company | Non-ionomeric polymer compositions for golf ball constructions and methods relating thereto |
US9457228B2 (en) * | 2013-07-09 | 2016-10-04 | Aditi Sinha | Sport training equipment |
US8961328B1 (en) * | 2013-08-18 | 2015-02-24 | Robert Dean Schmok | Multiple light beam method and system for golf swing alignment and calibration |
WO2015030099A1 (en) | 2013-08-30 | 2015-03-05 | ブラザー工業株式会社 | Image display device, and head-mounted display |
WO2015030100A1 (en) * | 2013-08-30 | 2015-03-05 | ブラザー工業株式会社 | Image display device, and head-mounted display |
US20150062164A1 (en) | 2013-09-05 | 2015-03-05 | Seiko Epson Corporation | Head mounted display, method of controlling head mounted display, computer program, image display system, and information processing apparatus |
JP6229381B2 (en) | 2013-09-05 | 2017-11-15 | セイコーエプソン株式会社 | Head-mounted display device, method for controlling head-mounted display device, image display system, and information processing device |
JP6222440B2 (en) | 2013-10-07 | 2017-11-01 | コニカミノルタ株式会社 | AR display system, AR display device, information processing device, and program |
CN105765504A (en) * | 2013-11-21 | 2016-07-13 | 3M创新有限公司 | Touch systems and methods employing force direction determination |
US9529465B2 (en) | 2013-12-02 | 2016-12-27 | At&T Intellectual Property I, L.P. | Secure interaction with input devices |
US9787934B2 (en) | 2014-01-15 | 2017-10-10 | Apple Inc. | Wireless devices with touch sensors and solar cells |
AU2015210991B2 (en) * | 2014-01-29 | 2017-08-10 | Becton, Dickinson And Company | System and method for assuring patient medication and fluid delivery at the clinical point of use |
EP3141954B1 (en) * | 2014-05-06 | 2024-04-24 | Ningbo Sunny Opotech Co., Ltd. | Light-deflection three-dimensional imaging device and projection device, and application thereof |
TWI519817B (en) * | 2014-05-06 | 2016-02-01 | 中強光電股份有限公司 | Smart helmet |
JP5696262B1 (en) * | 2014-06-13 | 2015-04-08 | 株式会社日立システムズ | Work management device, work defect prevention program, and work defect prevention method |
US9697656B2 (en) | 2014-08-19 | 2017-07-04 | Sensormatic Electronics, LLC | Method and system for access control proximity location |
JP5646108B1 (en) | 2014-08-26 | 2014-12-24 | 宮川工機株式会社 | Business support system for wood precut factory |
JP6455520B2 (en) * | 2014-09-30 | 2019-01-23 | コニカミノルタ株式会社 | Head mounted display and wearable computer |
US10168691B2 (en) | 2014-10-06 | 2019-01-01 | Fisher-Rosemount Systems, Inc. | Data pipeline for process control system analytics |
US20160140868A1 (en) * | 2014-11-13 | 2016-05-19 | Netapp, Inc. | Techniques for using augmented reality for computer systems maintenance |
JP6635048B2 (en) * | 2014-12-25 | 2020-01-22 | ソニー株式会社 | Information processing device |
EP3261338A4 (en) | 2015-02-17 | 2018-03-28 | Konica Minolta, Inc. | Head mounted display |
CN107409189B (en) | 2015-04-07 | 2020-06-16 | 奥林巴斯株式会社 | Glasses type display device |
US20160343168A1 (en) | 2015-05-20 | 2016-11-24 | Daqri, Llc | Virtual personification for augmented reality system |
US10019625B2 (en) * | 2015-07-23 | 2018-07-10 | Orcam Technologies Ltd. | Wearable camera for reporting the time based on wrist-related trigger |
JP2017049762A (en) | 2015-09-01 | 2017-03-09 | 株式会社東芝 | System and method |
EP3301940A1 (en) | 2016-09-30 | 2018-04-04 | Advanced Digital Broadcast S.A. | A method and a system for registering shutter glasses in an image generating device |
-
2015
- 2015-09-01 JP JP2015171933A patent/JP2017049762A/en active Pending
- 2015-12-22 US US14/979,169 patent/US10685232B2/en active Active
-
2017
- 2017-08-24 US US15/685,782 patent/US20170351921A1/en not_active Abandoned
- 2017-08-24 US US15/685,766 patent/US20170351919A1/en not_active Abandoned
- 2017-08-24 US US15/685,774 patent/US20170351920A1/en not_active Abandoned
-
2018
- 2018-03-09 US US15/917,328 patent/US10679059B2/en active Active
- 2018-03-09 US US15/917,332 patent/US20180197011A1/en not_active Abandoned
-
2019
- 2019-10-03 US US16/592,611 patent/US10671849B2/en active Active
- 2019-10-03 US US16/592,664 patent/US10682405B2/en active Active
-
2020
- 2020-04-16 US US16/851,088 patent/US11002975B2/en active Active
-
2021
- 2021-03-23 US US17/210,420 patent/US11428944B2/en active Active
-
2022
- 2022-08-29 US US17/897,968 patent/US11789279B2/en active Active
-
2023
- 2023-08-17 US US18/451,535 patent/US12135432B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6452544B1 (en) * | 2001-05-24 | 2002-09-17 | Nokia Corporation | Portable map display system for presenting a 3D map image and method thereof |
US20150260993A1 (en) * | 2012-10-11 | 2015-09-17 | Sony Computer Entertainment Europe Limited | Head mountable device |
US20170059869A1 (en) * | 2014-05-15 | 2017-03-02 | Jun Hee Lee | Optical system for head mount display |
US20170235152A1 (en) * | 2014-09-26 | 2017-08-17 | Osterhout Group, Inc. | See-through computer display systems |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10671849B2 (en) | 2015-09-01 | 2020-06-02 | Kabushiki Kaisha Toshiba | System and method for sensor based visual adjustments |
US10679059B2 (en) | 2015-09-01 | 2020-06-09 | Kabushiki Kaisha Toshiba | System and method for visual image adjustment |
US10685232B2 (en) | 2015-09-01 | 2020-06-16 | Kabushiki Kaisha Toshiba | Wearable device for displaying checklist of a work |
US10682405B2 (en) | 2015-09-01 | 2020-06-16 | Kabushiki Kaisha Toshiba | System and method and device for adjusting image positioning |
US11002975B2 (en) | 2015-09-01 | 2021-05-11 | Kabushiki Kaisha Toshiba | System and method for image generation based on a display-attachable wearable device |
US11428944B2 (en) | 2015-09-01 | 2022-08-30 | Kabushiki Kaisha Toshiba | Wearable device and method for visual image adjustment |
US11789279B2 (en) | 2015-09-01 | 2023-10-17 | Kabushiki Kaisha Toshiba | System and method for virtual image adjustment |
US12135432B2 (en) | 2015-09-01 | 2024-11-05 | Kbushiki Kaisha Toshiba | System and method directed to an eyeglasses-type wearable device |
US12001600B2 (en) | 2018-11-09 | 2024-06-04 | Beckman Coulter, Inc. | Service glasses with selective data provision |
US11983959B2 (en) | 2019-04-18 | 2024-05-14 | Beckman Coulter, Inc. | Securing data of objects in a laboratory environment |
Also Published As
Publication number | Publication date |
---|---|
US10685232B2 (en) | 2020-06-16 |
US20180197010A1 (en) | 2018-07-12 |
US10671849B2 (en) | 2020-06-02 |
US20210208410A1 (en) | 2021-07-08 |
US11428944B2 (en) | 2022-08-30 |
US20220413307A1 (en) | 2022-12-29 |
US20200034625A1 (en) | 2020-01-30 |
US11002975B2 (en) | 2021-05-11 |
US20170351919A1 (en) | 2017-12-07 |
US20170061212A1 (en) | 2017-03-02 |
US12135432B2 (en) | 2024-11-05 |
US20200030434A1 (en) | 2020-01-30 |
US10682405B2 (en) | 2020-06-16 |
US20170351921A1 (en) | 2017-12-07 |
US20200293780A1 (en) | 2020-09-17 |
US20180197011A1 (en) | 2018-07-12 |
JP2017049762A (en) | 2017-03-09 |
US11789279B2 (en) | 2023-10-17 |
US10679059B2 (en) | 2020-06-09 |
US20230384603A1 (en) | 2023-11-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12135432B2 (en) | System and method directed to an eyeglasses-type wearable device | |
US10564919B2 (en) | Display system, display apparatus, method for controlling display apparatus, and program | |
KR20170137476A (en) | Mobile device and method for controlling thereof | |
CN108370488A (en) | Audio providing method and its equipment | |
JP7119164B2 (en) | WEARABLE TERMINAL, SYSTEM AND DISPLAY METHOD | |
JP6253619B2 (en) | Electronic device and support method | |
JP2017049449A (en) | Electronic apparatus and support method | |
JP2018081697A (en) | System and wearable terminal | |
JP7263603B2 (en) | WEARABLE TERMINAL, SYSTEM AND DISPLAY METHOD | |
JP6641055B2 (en) | Wearable terminal, system and display method | |
JP2018045707A (en) | System, method, and sensor device | |
JP6352874B2 (en) | Wearable terminal, method and system | |
JP2023076619A (en) | Wearable terminal, system, and display method | |
JP6445118B2 (en) | Wearable terminal, method and system | |
JP2018037107A (en) | System and method | |
JP2017049868A (en) | Wearable terminal, method, and system | |
JP2018101425A (en) | Electronic apparatus, display method and system | |
CN117784931A (en) | Intelligent glasses | |
KR20220028572A (en) | Bio signal notification device and notification system comprising the same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KABUSHIKI KAISHA TOSHIBA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TANAKA, AKIRA;KANISHIMA, YASUHIRO;DONIWA, KENICHI;AND OTHERS;SIGNING DATES FROM 20151214 TO 20151215;REEL/FRAME:043573/0711 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |