US20170206673A1 - Information processing apparatus, information processing method, and image display system - Google Patents
Information processing apparatus, information processing method, and image display system Download PDFInfo
- Publication number
- US20170206673A1 US20170206673A1 US15/328,089 US201515328089A US2017206673A1 US 20170206673 A1 US20170206673 A1 US 20170206673A1 US 201515328089 A US201515328089 A US 201515328089A US 2017206673 A1 US2017206673 A1 US 2017206673A1
- Authority
- US
- United States
- Prior art keywords
- posture
- information
- image display
- image
- display apparatus
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 84
- 238000003672 processing method Methods 0.000 title claims description 7
- 238000004891 communication Methods 0.000 claims description 34
- 230000000007 visual effect Effects 0.000 claims description 10
- 238000013461 design Methods 0.000 claims description 9
- 230000008859 change Effects 0.000 claims description 7
- 238000013507 mapping Methods 0.000 claims description 4
- 230000004807 localization Effects 0.000 claims description 3
- 230000036316 preload Effects 0.000 claims description 3
- 230000001131 transforming effect Effects 0.000 claims description 3
- 210000003128 head Anatomy 0.000 description 218
- 238000012545 processing Methods 0.000 description 39
- 238000005516 engineering process Methods 0.000 description 34
- 238000000034 method Methods 0.000 description 34
- 238000010586 diagram Methods 0.000 description 16
- 230000003287 optical effect Effects 0.000 description 10
- 230000005540 biological transmission Effects 0.000 description 8
- 239000003550 marker Substances 0.000 description 7
- MWUXSHHQAYIFBG-UHFFFAOYSA-N Nitric oxide Chemical compound O=[N] MWUXSHHQAYIFBG-UHFFFAOYSA-N 0.000 description 6
- 238000005401 electroluminescence Methods 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 239000004973 liquid crystal related substance Substances 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 210000001747 pupil Anatomy 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 239000003086 colorant Substances 0.000 description 3
- 238000012937 correction Methods 0.000 description 3
- 230000007613 environmental effect Effects 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 230000002207 retinal effect Effects 0.000 description 3
- CURLTUGMZLYLDI-UHFFFAOYSA-N Carbon dioxide Chemical compound O=C=O CURLTUGMZLYLDI-UHFFFAOYSA-N 0.000 description 2
- 241001025261 Neoraja caerulea Species 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000006073 displacement reaction Methods 0.000 description 2
- 239000000428 dust Substances 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 230000035807 sensation Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 239000012855 volatile organic compound Substances 0.000 description 2
- SNICXCGAKADSCV-JTQLQIEISA-N (-)-Nicotine Chemical compound CN1CCC[C@H]1C1=CC=CN=C1 SNICXCGAKADSCV-JTQLQIEISA-N 0.000 description 1
- 239000004215 Carbon black (E152) Substances 0.000 description 1
- UGFAIRIUMAVXCW-UHFFFAOYSA-N Carbon monoxide Chemical compound [O+]#[C-] UGFAIRIUMAVXCW-UHFFFAOYSA-N 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000002730 additional effect Effects 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 239000010425 asbestos Substances 0.000 description 1
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 1
- 230000036760 body temperature Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000001569 carbon dioxide Substances 0.000 description 1
- 229910002092 carbon dioxide Inorganic materials 0.000 description 1
- 229910002091 carbon monoxide Inorganic materials 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000005284 excitation Effects 0.000 description 1
- 210000000744 eyelid Anatomy 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 239000007789 gas Substances 0.000 description 1
- 229930195733 hydrocarbon Natural products 0.000 description 1
- 150000002430 hydrocarbons Chemical class 0.000 description 1
- 150000002500 ions Chemical class 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000007257 malfunction Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000006996 mental state Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 230000001114 myogenic effect Effects 0.000 description 1
- 229960002715 nicotine Drugs 0.000 description 1
- SNICXCGAKADSCV-UHFFFAOYSA-N nicotine Natural products CN1CCCC1C1=CC=CN=C1 SNICXCGAKADSCV-UHFFFAOYSA-N 0.000 description 1
- 229910017464 nitrogen compound Inorganic materials 0.000 description 1
- 150000002830 nitrogen compounds Chemical class 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 239000001301 oxygen Substances 0.000 description 1
- 229910052760 oxygen Inorganic materials 0.000 description 1
- 238000006552 photochemical reaction Methods 0.000 description 1
- 230000035790 physiological processes and functions Effects 0.000 description 1
- 230000002285 radioactive effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 229910052895 riebeckite Inorganic materials 0.000 description 1
- 238000010079 rubber tapping Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000000779 smoke Substances 0.000 description 1
- 230000000392 somatic effect Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1686—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
- G06F3/0325—Detection arrangements using opto-electronic means using a plurality of light emitters or reflectors or a plurality of detectors forming a reference frame from which to derive the orientation of the object, e.g. by triangulation or on the basis of reference deformation in the picked up image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/183—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
- H04N7/185—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source from a mobile camera, e.g. for remote control
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0132—Head-up displays characterised by optical features comprising binocular systems
- G02B2027/0134—Head-up displays characterised by optical features comprising binocular systems of stereoscopic type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1454—Digital output to display device ; Cooperation and interconnection of the display device with other functional units involving copying of the display data of a local workstation or window to a remote workstation or window so that an actual copy of the data is displayed simultaneously on two or more displays, e.g. teledisplay
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
- G09G2340/0464—Positioning
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/12—Overlay of images, i.e. displayed pixel being the result of switching between the corresponding input pixels
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2356/00—Detection of the display position w.r.t. other display screens
Definitions
- the technology disclosed in the present specification relates to an information processing apparatus, an information processing method, and an image display system that provide screen information to be displayed on an image display apparatus that is worn by a user on the head or face and used for viewing an image.
- the head mount display includes, for example, image display units which are arranged for the respective right and left eyes and forms an enlarged virtual image of a display image by a virtual image optical system so that a user can observe an image with a realistic feeling.
- the head mount display is equipped with a high-resolution display panel which includes, for example, a liquid crystal or an organic electro-luminescence (EL) element as each of the display units for right and left eyes. Further, head mount displays may be classified into a transmissive type and a light-shielding type.
- the transmissive head mount display enables a wearer to observe surrounding scenery also when the wearer wears the head mount display on the head and an image is displayed thereon (refer to Patent Document 1, for example).
- a user can avoid the risk of, for example, collision with an obstacle when using the head mount display in the outside or during walking.
- the light-shielding head mount display is configured to directly cover the eyes of a wearer when worn on the head, which increases an immersive feeling during viewing of an image. It is possible to duplicate a realistic feeling like viewing in a movie theater by projecting a display screen in an enlarged manner using the virtual optical system to allow a user to observe the display screen as an enlarged virtual image having an appropriate angle of view and duplicating multiple channels by a headphone (refer to Patent Document 2, for example).
- the light-shielding type head mount display there is known a video see-through method that has a built-in camera capable of capturing an image of the front side of a wearer and displays an external image obtained by the image capturing, and the wearer can observe surrounding scenery through the external image (refer to Patent Documents 3, 4, for example).
- the former transmissive head mount display is called optical see-through or merely see-through contrary to the video see-through.
- the head mount display restricts the sense of sight and the sense of hearing of a user wearing the head mount display.
- the view of a user wearing the light-shielding head mount display is completely obstructed.
- the user cannot perform any operation to the outside such as answering a phone or input to the screen of a computer.
- a view image of a user captured by a camera is displayed to allow the user to operate an object which is present in the view through the displayed image.
- a portable terminal system which is provided with a portable terminal which includes a display apparatus which performs display according to display information and a head mount display which displays a virtual screen in a display area in the view of a wearer on the basis of the display information acquired from the portable terminal (refer to Patent Document 5, for example).
- the portable terminal system as described above is premised on that the head mount display is equipped with a camera which captures an image of the outside world. It is difficult for a user wearing a head mount display equipped with no camera to perform an operation with respect to the outside unless the head mount display is detached from the head to completely suspend viewing.
- Patent Document 2 Japanese Patent Application Laid-Open No. 2012-141461
- Patent Document 3 Japanese Patent Application Laid-Open No. 2005-38321
- Patent Document 4 Japanese Patent Application Laid-Open No. 2011-242591
- Patent Document 5 Japanese Patent Application Laid-Open No. 2011-186856
- Patent Document 6 Japanese Patent Application Laid-Open No. 2012-186660
- An object of the technology disclosed in the present specification is to provide an excellent information processing apparatus, an excellent information processing method, and an excellent image display system capable of appropriately providing screen information to be displayed on an image display apparatus that is worn by a user on the head or face and used for viewing of an image.
- a technology disclosed in the present specification has been made in consideration of the problem, and a first aspect thereof is an information processing apparatus including:
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus;
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information relating to the information processing apparatus in a screen of the image display apparatus.
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus;
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
- the information processing apparatus further includes a communication unit that communicates with the image display apparatus. Further, screen information of the display unit and information of the embedding position are configured to be transmitted to the image display apparatus through the communication unit.
- the position/posture calculation unit of the information processing apparatus is configured to calculate the position and posture of the image display apparatus in the first reference coordinate system on the basis of a reference index mounted on the image display apparatus.
- the information processing apparatus further includes a reference index arrangement information acquisition unit that acquires information relating to a relationship between posture of the image display apparatus and posture of the reference index.
- the reference index arrangement information acquisition unit of the information processing apparatus is configured to preload design information relating to the relationship between the posture of the image display apparatus and the posture of the reference index
- the position/posture calculation unit is configured to calculate the position and posture of the image display apparatus in the first reference coordinate system on the basis of the design information from an image captured by the image capturing unit when the information processing apparatus is arranged in a predetermined direction with respect to the image display apparatus.
- the reference index arrangement information acquisition unit of the information processing apparatus is configured to receive information from the reference index.
- the reference index arrangement information acquisition unit of the information processing apparatus is configured to receive information encoded in a flashing pattern from the reference index.
- the reference index arrangement information acquisition unit of the information processing apparatus is configured to make an inquiry about information obtained by decoding the flashing pattern to a database to acquire coordinate information of the reference index, information of the reference index, or information of the image display apparatus.
- the position/posture calculation unit of the information processing apparatus is configured to calculate the position and posture of the image display apparatus in the first reference coordinate system using visual SLAM.
- a thirteenth aspect of the technology disclosed in the present specification is an image processing method including:
- an embedding position calculation step of calculating, on the basis of the position and posture of the information terminal in the second reference coordinate system, a position of embedding screen information of the information terminal with respect to a screen of the image display apparatus.
- an image display system including:
- an information processing apparatus that includes an image capturing unit and a display unit, calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus, transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus, and calculates a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
- system indicates a logical aggregate of a plurality of apparatuses (or functional modules which achieve a specific function), and it does not particularly matter whether the apparatuses or the functional modules are present in a single housing.
- FIG. 2 is a diagram illustrating a top view of the user wearing the head mount display 100 illustrated in FIG. 1 .
- FIG. 3 is a diagram illustrating an example of an internal configuration of the head mount display 100 .
- FIG. 4 is a diagram illustrating a state in which the mutual positional relationship and posture of an information terminal are grasped on the basis of an image captured by an outer camera 312 of the head mount display 100 and an image of the information terminal is displayed in a superimposed manner.
- FIG. 6 is a diagram illustrating a state in which the mutual positional relationship and posture between the head mount display 100 and the information terminal are grasped on the basis of an image captured by a camera of the information terminal and an image of the information terminal is displayed in a superimposed manner.
- FIG. 7 is a diagram illustrating a procedure for achieving screen display illustrated in FIG. 6 .
- FIG. 8 is a diagram illustrating a state in which an AR marker 801 is mounted on the head mount display 100 , and a camera of an information terminal which is held in a principal axis front direction captures an image of the head mount display 100 .
- FIG. 10 is a diagram illustrating a state in which four or more reference indexes 1001 A to D having different colors and a known shape are mounted on the head mount display 100 , and the camera of the information terminal which is held in the principal axis front direction captures an image of the head mount display 100 .
- FIG. 11 is a diagram illustrating a state in which reference indexes 1101 A to D each of which flashes are mounted on the head mount display 100 , and the camera of the information terminal which is held in the principal axis front direction captures an image of the head mount display 100 .
- FIG. 12 is a diagram illustrating a procedure for embedding screen information of the information terminal in the screen of the head mount display 100 on the basis of object coordinates received from the reference index.
- FIG. 13 is a diagram illustrating a procedure for embedding screen information of the information terminal in the screen of the head mount display 100 on the basis of a retrieval key received from the reference index.
- FIG. 14 is a diagram illustrating a procedure for initializing the position of the information terminal.
- FIG. 15 is a diagram illustrating a procedure for embedding screen information of the information terminal in the screen of the head mount display 100 using the position of the information terminal initialized by the procedure illustrated in FIG. 14 .
- FIG. 16 is a diagram illustrating an example of the configuration of the information terminal.
- the head mount display 100 When a user wears the head mount display 100 on the head or face, the head mount display 100 directly covers the eyes of the user, which enables an immersive feeling to be given to the user viewing an image. Further, since the display image is invisible from the outside (that is, other people), privacy is easily protected in information display. Differently from a see-through type, a user wearing the head mount display 100 cannot directly look at scenery in the real world.
- the head mount display 100 is equipped with an outer camera 312 which captures an image of scenery in a sight-line direction of a user, the user can indirectly look at scenery in the real world (that is, display the scenery by video see-through) by displaying the captured image.
- the technology disclosed in the present specification is not premised on the equipment of the outer camera 312 or the video see-through display.
- the head mount display 100 illustrated in FIG. 1 is a structure which is similar to a cap shape and configured to directly cover the right and left eyes of a user wearing the head mount display 100 .
- a display panel (not illustrated in FIG. 1 ) which is observed by a user is disposed on the inner side of a body of the head mount display 100 at a position facing the right and left eyes.
- the display panel includes, for example, a micro display such as an organic EL element or a liquid crystal display and a laser scanning display such as a retinal direct drawing display.
- Microphones 103 L, 103 R are respectively placed near left and right ends of the body of the head mount display 100 .
- the microphones 103 L, 103 R which are substantially symmetrically placed left and right enables only a voice (a voice of a user) which is localized on the center to be recognized and separated from surrounding noises and voices of other people. Accordingly, for example, it is possible to prevent a malfunction during an operation by audio input.
- a touch panel 315 through which a user can perform touch input using, for example, the fingertip is disposed near the back face of the display panel on the front face of the body of the head mount display 100 .
- a pair of right and left touch panels 315 are provided, one or three or more touch panels 315 may be provided.
- FIG. 2 illustrates a top view of the user wearing the head mount display 100 illustrated in FIG. 1 .
- the illustrated head mount display 100 includes a display panel 104 L for left eye and a display panel 104 R for right eye on a side face facing the face of the user.
- Each of the display panels 104 L, 104 R includes, for example, a micro display such as an organic EL element or a liquid crystal display and a laser scanning display such as a retinal direct drawing display. Display images on the display panels 104 L, 104 R are observed as enlarged virtual images by a user by passing through virtual image optical units 101 L, 101 R.
- a pupil distance adjustment mechanism 105 is disposed between the display panel for right eye and the display panel for left eye.
- FIG. 3 illustrates an example of an internal configuration of the head mount display 100 . Hereinbelow, each unit will be described.
- the program stored in the ROM 301 A includes a display control program for displaying a reproducing moving image on the screen, and displaying, in a super imposed manner, image information taken from an external apparatus (e.g., a smartphone) on the screen.
- the data stored in the ROM 301 A includes identification information (including a model number and a manufacturer's serial number) which is unique to the head mount display 100 , user attribute information such as authentication information (e.g., a personal identification number, a password, or biological information) for authenticating a user who uses the head mount display 100 , and design information of the head mount display 100 (the shape or the arrangement of a reference index mounted on the head mount display 100 ).
- An input interface (IF) unit 302 is provided with one or more operators (not illustrated) such as a key, a button, and a switch to which a user performs an input operation, and receives an instruction of the user through the operators and outputs the received instruction to the control unit 301 . Further, the input operation unit 302 receives an instruction of a user which includes a remote-control command received by a remote-control reception unit 303 and outputs the received instruction to the control unit 301 .
- operators such as a key, a button, and a switch to which a user performs an input operation
- the input interface (IF) unit 302 outputs input information such as coordinate data of a position touched by the fingertip to the control unit 301 .
- the touch panel 315 is arranged exactly on the back face of a display image (an enlarged virtual image observed through a virtual image optical unit 310 ) of a display unit 309 on the front face of the body of the head mount display 100 (refer to FIG. 1 )
- a user can perform a touch operation with a feeling like touching the display image with the fingertip.
- a state information acquisition unit 304 is a functional module which acquires state information of the body of the head mount display 100 or a user wearing the head mount display 100 .
- the state information acquisition unit 304 may be equipped with various sensors for detecting state information by itself or may acquire state information from an external apparatus (e.g., a smartphone or a wristwatch worn by the user, or another multifunction terminal) which is provided with some or all of these sensors through a communication unit 305 (described below).
- an external apparatus e.g., a smartphone or a wristwatch worn by the user, or another multifunction terminal
- the state information acquisition unit 304 acquires, for example, information of the position and posture or information of the posture of the head of a user to track a head motion of the user.
- the state information acquisition unit 304 is a sensor capable of detecting nine axes in total including a three-axis gyro sensor, a three-axis accelerometer, and a three-axis magnetic field sensor.
- the state information acquisition unit 304 may further use, in combination, any one or two or more sensors such as a global positioning system (GPS) sensor, a Doppler sensor, an infrared sensor, and a radio field intensity sensor.
- GPS global positioning system
- the state information acquisition unit 304 may further use, in combination, information provided from various infrastructures such as cellular phone base station information and Place Engine (registered trademark) information (e.g., electrically measured information from a wireless LAN access point) to acquire the position/posture information.
- the state acquisition unit 304 for head motion tracking is incorporated in the head mount display 100
- the state acquisition unit 304 may include an accessory component which is externally attached to the head mount display 100 .
- the externally-attached state acquisition unit 304 transmits posture information of the head to the body of the head mount display 100 through wireless communication such as Bluetooth (registered trademark) communication or a high-speed wired interface such as a universal serial bus (USB).
- wireless communication such as Bluetooth (registered trademark) communication or a high-speed wired interface such as a universal serial bus (USB).
- the state information acquisition unit 304 acquires, as state information of a user wearing the head mount display 100 , for example, a working state of a user (wearing/unwearing of the head mount display 100 ), an action state of a user (a moving state such as standing still, walking, or running, a gesture by the hand or fingertip, an opening/closing state of the eyelid, a sight-line direction, or the large/small of the pupil), a mental state (the degree of impression such as whether a user is immersed in or concentrated on the observation of a display image, the degree of excitation, the degree of awareness, feeling, or emotion), and a physiological state.
- a working state of a user wearing/unwearing of the head mount display 100
- an action state of a user a moving state such as standing still, walking, or running, a gesture by the hand or fingertip, an opening/closing state of the eyelid, a sight-line direction, or the large/small of the pupil
- a mental state the degree of impression such
- the state information acquisition unit 304 may use a wearing sensor (refer to Patent Document 6, for example) which detects, when a user has worn the head mount display on the head, the wearing of the head mount display by the user in conjunction with the motion of abutment against the forehead in the head to detect wearing/unwearing of the head mount display 100 .
- a wearing sensor (refer to Patent Document 6, for example) which detects, when a user has worn the head mount display on the head, the wearing of the head mount display by the user in conjunction with the motion of abutment against the forehead in the head to detect wearing/unwearing of the head mount display 100 .
- An environment information acquisition unit 316 is a functional module which acquires information relating to the environment surrounding the body of the head mount display 100 or a user wearing the head mount display 100 .
- the information relating to the environment described herein includes sound; the volume of air; air temperature; air pressure; atmosphere (smoke; dense fog; electromagnetic waves (ultraviolet rays, blue light, radio waves), heat waves (infrared rays) and radioactive rays applied to the head mount display 100 or a user; carbon monoxide, carbon dioxide, oxygen and nitrogen compound (nicotine) in the atmosphere; nitrogen oxide (NO x ) and hydrocarbon (volatile organic compounds (VOC)) floating in the atmosphere or photochemical smog generated by a photochemical reaction of these materials due to the influence of ultraviolet rays; particulate matters; pollen; dust such as house dust; and harmful chemical substances such as asbestos); and other environmental factors.
- the environment information acquisition unit 316 may be equipped with various environmental sensors such as a sound sensor and an air volume sensor to detect environment information.
- the environmental sensors may include the microphone and the outer camera 312 described above.
- the environment information acquisition unit 316 may acquire environment information from an external apparatus (e.g., a smartphone or a wristwatch worn by the user, or another multifunction terminal) which is provided with some or all of these sensors through the communication unit 305 (described below).
- the outer camera 312 is arranged, for example, on substantially the center in the front face of the body of the head mount display 100 (refer to FIG. 2 ) and capable of capturing surrounding images.
- the technology disclosed in the present specification is not premised on the equipment or the use of the outer camera 312 .
- the communication unit 305 performs communication processing with an external apparatus (not illustrated), and modulation/demodulation and encoding/decoding processing of a communication signal.
- the external apparatus includes a contents reproducing apparatus (a Blue-ray Disk player or a DVD player) which supplies viewing contents when a user uses the head mount display 100 , a streaming server, and information terminals such as a smartphone and a personal computer used by a user. Further, the control unit 301 transmits transmission data to the external apparatus through the communication unit 305 .
- An image processing unit 307 further performs signal processing such as image quality correction and performs transformation to a resolution corresponding to the screen of the display unit 309 with respect to an image signal output from the control unit 301 . Then, a display drive unit 308 sequentially selects pixels of the display unit 309 for each row and line-sequentially scans the pixels to supply a pixel signal based on the signal-processed image signal.
- the display unit 309 includes a display panel which includes, for example, a micro display such as an organic EL element or a liquid crystal display or a laser scanning display such as a retinal direct drawing display.
- the virtual image optical unit 310 projects, in an enlarged manner, a display image of the display unit 309 and allows a user to observe the display image as an enlarged virtual image.
- the display image output by the display unit 309 includes commercial contents (virtual world) which are supplied from a contents reproducing apparatus (a Blue-ray Disk player or a DVD player) or a streaming server and a captured image of the outer camera 312 (an image in the real world such as a view image of a user). Further, image information transmitted from an information terminal such as a smartphone or a personal computer may be displayed in a superimposed manner on the contents reproducing screen.
- a contents reproducing apparatus a Blue-ray Disk player or a DVD player
- a streaming server a captured image of the outer camera 312 (an image in the real world such as a view image of a user).
- image information transmitted from an information terminal such as a smartphone or a personal computer may be displayed in a superimposed manner on the contents reproducing screen.
- An audio processing unit 313 further performs audio quality correction and audio amplification with respect to an audio signal output from the control unit 301 and performs signal processing of an audio signal input thereto. Then, an audio input/output unit 314 outputs the processed audio to the outside and performs voice input from the microphone (described above).
- a user wearing the head mount display 100 may wish to operate another information terminal such as a smartphone when the view of the user is completely obstructed during viewing of contents. For example, the user may wish to check a record of incoming mails or calls or wish to know the current time.
- the head mount display 100 inputs screen information of an external information terminal, and changes an original contents reproducing screen to the screen of the external information terminal so as to be displayed on the display unit 309 or displays, in an superimposed manner, the screen of the external information terminal screen on the original screen, a user can check the contents of the screen of the external information terminal without detaching the head mount display 100 from the head.
- the superimposition display described herein includes separately displaying a small screen in a picture-in-picture form and blending two screens at a predetermined ratio.
- the operation lacks reality.
- the position of the screen of the information terminal displayed by the display unit 309 is not completely linked with the position of the screen of the actual information terminal, it is extremely difficult for a user to perform an operation with respect to the screen of the actual information terminal while looking at the screen of the information terminal displayed on the head mount display 100 .
- an image of an information terminal which is held by a user in his or her hand is captured using the outer camera 312 (reference sign 401 ). Then, the captured image of the information terminal is displayed on the contents reproducing screen of the head mount display 100 with the same position and posture as the actual information terminal on the view of the user (reference sign 402 ). Accordingly, the user can easily perform an operation with respect to the screen of the actual information terminal while looking at the screen of the information terminal which is displayed by video see-through on the head mount display 100 .
- FIG. 5 illustrates a procedure for achieving the screen display illustrated in FIG. 4 .
- viewing contents are displayed on the display unit 309 (F 501 ).
- an image is captured using the outer camera 312 (F 511 ). Then, object recognition processing is performed with respect to the captured image to detect the information terminal (F 512 ). Then, the position and posture of the information terminal are calculated (F 513 ).
- a reference coordinate system of the display unit 309 which displays the information terminal on the head mount display 100 differs from a reference coordinate system of the outer camera 312 which captures an image of the information terminal in a strict sense.
- F 513 it is necessary to transform the position and posture of the information terminal detected in the camera reference coordinate system to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 .
- a visual reference index may be applied to the information terminal.
- a known pattern displayed on the screen of the information terminal may be used as the reference index.
- the position and posture of the information terminal may be calculated with no marker (described below). Then, the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F 514 ).
- the head mount display 100 acquires screen information from the information terminal (F 521 ), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F 514 to generate an output screen (F 530 ).
- the screen of the information terminal is present on the contents reproducing screen of the head mount display 100 with the same position and posture as the actual information terminal on the view of the user.
- the user can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents.
- the outer camera 312 is required to correct the displayed position and posture of the information terminal using visual information.
- the above method cannot be applied to a product that is not equipped with the outer camera 312 .
- FIGS. 1 to 3 illustrate, as an example, the head mount display 100 equipped with the outer camera 312 .
- an outer camera is not necessarily provided as standard equipment, and a considerable number of products equipped with no outer camera have already become widespread.
- the head mount display 100 includes no means for grasping the mutual positional relationship and posture between the display body and an information terminal as a current operation target. Thus, it is not possible to display the information terminal on the contents reproducing screen of the head mount display 100 with the same position and posture as the actual information terminal on the view of a user.
- the present applicant focuses on this point and proposes a method that conversely captures an image of the head mount display 100 using a camera mounted on an information terminal (reference sign 601 ) to grasp the mutual positional relationship and posture between the head mount display 100 and the information terminal as the current operation target and displays screen information transmitted from the information terminal on the contents reproducing screen of the head mount display 100 with the same position and posture as the actual information terminal on the view of a user (reference sign 602 ) as illustrated in FIG. 6 .
- FIG. 7 illustrates a procedure for achieving the screen display illustrated in FIG. 6 .
- viewing contents are displayed on the display unit 309 (F 701 ).
- an image is captured using the camera which is provided as standard equipment (F 711 )
- the information terminal is held by a user wearing the head mount display 100 in his or her hand, and the camera faces the front face of the head mount display 100 .
- object recognition processing is performed with respect to the captured image to detect the head mount display 100 (F 712 ).
- the position and posture of the head mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated (F 713 ).
- the position and posture of the head mount display 100 in the reference coordinate system of the information terminal are transformed to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 (F 714 ), and the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F 715 ).
- the processes (F 712 to F 715 ) after the processing of the captured image may be executed by either the information terminal or the head mount display 100 , or may be executed by an external device such as a cloud computer.
- the head mount display 100 acquires screen information from the information terminal (F 721 ), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F 715 to generate an output screen (F 730 ).
- the screen of the information terminal with the same position and posture as the actual information terminal on the view of the user is displayed in the contents reproducing screen (an image in a virtual world) of the head mount display 100 .
- the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents.
- FIG. 16 illustrates an example of the configuration of an information terminal capable of achieving the processing as illustrated in FIG. 7 .
- the illustrated information terminal is configured by connecting a display unit 1620 , an audio processing unit 1630 , a communication unit 1640 , a storage unit 1650 , a camera unit 1660 , a sensor unit 1670 and the like to a control unit 1610 .
- the control unit 1610 includes, for example, a central processing unit (CPU) 211 , a ROM 1612 , and a RAM 1613 .
- CPU central processing unit
- ROM read-only memory
- RAM random access memory
- the CPU 1611 loads the program code to the RAM 1613 from the ROM 1612 or the storage unit 1640 and executes the program code.
- the program executed by the CPU 1611 includes operating systems such as Android and iOS, an application for transferring screen information to an external image display apparatus (e.g., the head mount display 100 ), and various application programs which operate under an execution environment provided by the operating system.
- the display unit 1620 is provided with a display panel 1621 which includes, for example, a liquid crystal element or an organic EL element and a transparent touch panel 1623 which is stuck to the upper face of the display panel 1622 .
- the display panel 1621 is connected to the control unit 1610 through a display interface 1622 , and displays and outputs image information generated by the control unit 210 .
- the touch panel 1623 is connected to the control unit 1610 through a touch interface 1624 and outputs coordinate information obtained by an operation of a user on the display panel 1621 with the fingertip to the control unit 1610 .
- the control unit 1610 detects a user operation such as a tap, long press, flick, or swipe on the basis of the input coordinate information and activates processing corresponding to the user operation.
- the audio processing unit 1630 is provided with an audio output unit 1631 such as a speaker, an audio input unit 1632 such as a microphone, and an audio codec 1633 which encodes and decodes an audio signal which is input and output. Further, the audio processing unit 1630 may further be provided with an output terminal 1634 for outputting an audio signal to a headphone (not illustrated).
- the communication unit 1640 performs information communication processing between an application executed by the control unit 1610 and an external apparatus.
- the external apparatus described herein includes a television set, an information terminal (not illustrated) handled by another user, and server which is present on the Internet.
- the communication unit 1640 is equipped with a physical layer module such as Wi-Fi, near field communication (NFC), or Bluetooth (registered trademark) according to a commutation medium to be used and performs modulation/demodulation processing or encoding/decoding processing of a communication signal which is transmitted and received through the physical layer module. Further, screen information displayed on the display unit 1620 may be transferred to an external image display apparatus (e.g., the head mount display 100 ) through the communication unit 1640 .
- an external image display apparatus e.g., the head mount display 100
- the storage unit 1650 includes a mass storage device such as a solid state drive (SSD) or a hard disc drive (HDD).
- SSD solid state drive
- HDD hard disc drive
- an application program and contents downloaded through the communication unit 1640 and image data such as a still image and a moving image captured by the camera unit 1660 are stored in the storage unit 1650 .
- the camera unit 1660 is provided with an image sensor 1661 which photoelectrically converts light taken through a lens (not illustrated) such as a charge coupled device (CCD) or a complementary metal oxide semiconductor (CMOS) and an analog front end (AFE) processing unit 1662 which performs noise reduction and digitization of a detection signal of the image sensor 1661 to generate image data, and outputs the generated image data to the control unit 1610 from the camera interface 1663 .
- a lens such as a charge coupled device (CCD) or a complementary metal oxide semiconductor (CMOS)
- AFE analog front end
- the sensor unit 1670 includes, for example, a global positioning system (GPS) sensor for acquiring positional information of the information terminal, a gyro sensor for detecting the posture or an acting force of the body of the information terminal, and an accelerometer.
- GPS global positioning system
- one or more reference indexes such as an augmented reality (AR) marker may be mounted on the head mount display 100 so as to more accurately grasp the mutual positional relationship and posture between the head mount display 100 and the information terminal on the basis of an image captured by the camera which is mounted on the information terminal.
- AR augmented reality
- detailed information e.g., product information such as the model number and the manufacturer's serial number
- product information such as the model number and the manufacturer's serial number
- Using the detailed information further improves the accuracy of recognizing the mutual position and posture between the head mount display 100 and the information terminal.
- the following (1) and (2) can be described as examples of a use method of the reference index.
- the direction and unit length of an optical principal axis are specified from the arrangement of a reference index based on a certain agreement to precisely perform coordinate axis alignment and position measurement.
- Identification information of a predetermined bit length (e.g., 32-bit) is encoded in the reference index, and a database is searched using the identification information to obtain information similar to (1) to precisely perform coordinate axis alignment and position measurement.
- a method for sending a reset signal includes tapping a touch panel screen of the information terminal and operating a switch on the head mount display 100 . A relative displacement thereafter is measured and reflected to rendering on the screen of the head mount display 100 .
- the position/posture calculation may be performed using visual simultaneous localization and mapping (SLAM).
- the visual SLAM is a technology capable of simultaneously performing own position estimation of the camera and map creation under an unknown environment.
- An example of visual SLAM includes integrated augmented reality technology “SmartAR” (registered trademark of Sony Corporation). According to the marker-less method, it is not necessary to apply the reference index to the head mount display 100 , and the method can be applied to any product.
- the reference coordinate system of the head mount display 100 differs from the reference coordinate system of the outer camera 312 in a strict sense.
- F 513 it is necessary to transform the position and posture of the information terminal detected in the camera reference coordinate system to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 .
- the position of the outer camera 312 with respect to the head mount display 100 is fixed, that is, known.
- the position of the camera with respect to the information terminal is fixed, that is, known. Thus, it is easy to transform the position and posture of the head mount display 100 whose image is captured by the camera to the position and posture of the head mount display 100 in the reference coordinate system of the information terminal.
- reference index arrangement information transmission means for transmitting information of the posture relationship between the head mount display 100 and the reference index to the information terminal is required to more precisely perform transform processing of the position and posture in F 714 .
- the following (1) to (3) will be described as examples of the reference index arrangement information transmission means.
- Design information of the head mount display 100 is used as known information of the information terminal.
- the reference index transmits necessary information to the information terminal.
- Design information of the head mount display 100 is used as known information of the information terminal.
- a dedicated application of the information terminal e.g., an application for calculating the position and posture of the head mount display 100 from an image captured by the camera
- information corresponding to the model number of the head mount display 100 is preloaded to the information terminal.
- This case is premised on that the shape and arrangement of the reference index which is mounted on the head mount display 100 are all previously designed. Further, defining, for example, a principal axis of the head mount display 100 and arranging the reference index so as to be aligned with the principal axis is a promise or specification. This transmission means enables the shape of the reference index to have flexibility.
- FIG. 8 illustrates a state in which an AR marker 801 is mounted on the head mount display 100 , and the camera of the information terminal which is held in a principal axis front direction captures an image of the head mount display 100 .
- Arrangement information for sticking the AR marker is known.
- the relative position and posture of the information terminal can be calculated by observing the position and posture (shape) of the AR marker on the captured image.
- FIG. 9 illustrates a state in which reference indexes 901 A to D having a known shape are mounted on the head mount display 100 , and the camera of the information terminal which is held in the principal axis front direction captures an image of the head mount display 100 .
- Arrangement information of each of the reference indexes 901 A to D is known.
- the relative position and posture of the information terminal can be calculated by observing the position of each of the reference indexes 901 A to D on the captured image. However, it is necessary to install three or more reference indexes.
- a method that captures an image of three or more points having known three-dimensional coordinates and obtains the position/posture of the camera from the position of each of the points on the image that is, a method that calculates mapping from a known three-dimensional coordinate system to two-dimensional coordinates defined on a camera image is an established technology which is called 3D-2D registration in the field of, for example, image processing. Observation of at least three or more points is required to obtain six variables corresponding to position/posture called an external parameter of the camera. Typically, estimation using the least squares method is performed using more observation points (note that, here, an internal parameter of the camera is known).
- the relative position and posture of the information terminal can be recognized by the camera on the information terminal, and it is not necessary to use the outer camera 312 on the head mount display 100 .
- the reference index transmits necessary information to the information terminal.
- FIG. 10 illustrates a state in which three or more reference indexes 1001 A to D having different colors and a known shape are mounted on the head mount display 100 , and the camera of the information terminal which is held in the principal axis front direction captures an image of the head mount display 100 .
- the reference indexes 1001 A to D have different colors, textures and shapes, and axial information of the reference coordinate system of the head mount display 100 is transmitted to the information terminal according to the combination of thereof. Arrangement information is known.
- discrimination between the reference indexes 1001 A to D can be made on the basis of the color, texture, and shape of each of the reference indexes 1001 A to D on the captured image.
- a line formed by connecting the reference indexes 1001 B and 1001 D is defined as a reference horizontal line.
- a combination of the reference indexes 1001 A and 100 B or a combination of the reference indexes 1001 C and 1001 D represents a reference vertical line.
- FIG. 11 illustrates a state in which reference indexes 1101 A to D each of which flashes are mounted on the head mount display 100 , and the camera of the information terminal which is held in the principal axis front direction captures an image of the head mount display 100 .
- information which is desired to be transmitted such as positional information of each of the indexes in the reference coordinate system of the head mount display 100 , model number information of the head mount display 100 in which the indexes are installed, or a retrieval key of a database which stores other necessary data is encoded in a flashing pattern, and the information terminal can obtain information by decoding the flashing pattern.
- the relative position and posture of the information terminal can be recognized by the camera on the information terminal, and it is not necessary to use the outer camera 312 on the head mount display 100 .
- FIG. 12 illustrates a procedure for embedding screen information of the information terminal in the screen of the head mount display 100 on the basis of object coordinates received from the reference index.
- viewing contents are displayed on the display unit 309 (F 1201 ).
- an image is captured using a camera which is provided as standard equipment (F 1211 ).
- the information terminal is held by a user wearing the head mount display 100 in his or her hand, and the camera faces the front face of the head mount display 100 .
- object recognition processing is performed with respect to the captured image to detect a reference index of the head mount display 100 (F 1212 ).
- the calculation processing is the 3D-2D registration (described above) or processing equivalent to the 3D-2D registration.
- the position/posture of the HMD (coordinate system) viewed from the camera coordinate system or the position/posture of the camera (coordinate system) viewed from the HMD coordinate system by associating the two-dimensional position of the reference index on the camera coordinates detected in F 1212 with the three-dimensional position of each of the reference indexes in the reference coordinate system of the head mount display 100 received in F 1241 .
- the position/posture of the HMD viewed from the camera coordinate system is calculated.
- the position and posture of the head mount display 100 in the reference coordinate system of the information terminal are transformed to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 (F 1214 ), and the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F 1215 ).
- the processes (F 1212 to F 1215 ) after the processing of the captured image may be executed by either the information terminal or the head mount display 100 , or may be executed by an external device such as a cloud computer.
- the head mount display 100 acquires screen information from the information terminal (F 1221 ), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F 1215 to generate an output screen (F 1230 ).
- the screen of the information terminal is displayed in the contents reproducing screen (an image in a virtual world) of the head mount display 100 with the same position and posture as the actual information terminal on the view of the user.
- the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents.
- FIG. 13 illustrates a procedure for embedding screen information of the information terminal in the screen of the head mount display 100 on the basis of a retrieval key received from the reference index.
- viewing contents are displayed on the display unit 309 (F 1301 ).
- an image is captured using a camera which is provided as standard equipment (F 1311 ).
- the information terminal is held by a user wearing the head mount display 100 in his or her hand, and the camera faces the front face of the head mount display 100 .
- object recognition processing is performed with respect to the captured image to detect a reference index of the head mount display 100 (F 1312 ).
- a retrieval key is received from the reference index (F 1341 ), and an inquiry is made using the retrieval key (F 1342 ) to acquire, for example, coordinate information of the reference index, information of the reference index, and information of the head mount display 100 . Then, the position and posture of the head mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated from the position and posture of the detected reference index (F 1313 ).
- the position and posture of the head mount display 100 in the reference coordinate system of the information terminal are transformed to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 (F 1314 ), and the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F 1315 ).
- the processes (F 1312 to F 1315 ) after the processing of the captured image may be executed either the information terminal or the head mount display 100 , or may be executed by an external device such as a cloud computer.
- the head mount display 100 acquires screen information from the information terminal (F 1321 ), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F 1315 to generate an output screen (F 1330 ).
- the screen of the information terminal is displayed in the contents reproducing screen (an image in a virtual world) of the head mount display 100 with the same position and posture as the actual information terminal on the view of the user.
- the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents.
- a user wearing the head mount display 100 holds the information terminal with a previously determined initial posture, for example, the front side of the eye-line and performs reset (initialization).
- the head mount display 100 whose image has been captured by the camera is recognized by applying, for example, the visual SLAM technology, and the position of the information terminal is initialized corresponding to the posture thereof.
- FIG. 14 illustrates a procedure for initializing the position of the information terminal.
- a user wearing the head mount display 100 holds the information terminal with the initial posture, for example, the front side of the eye-line.
- an image of the head mount display 100 is captured by the camera of the information terminal (F 1401 ), and the head mount display 100 is detected from the captured image (F 1402 ).
- the position and posture of the head mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated (F 1403 ).
- the calculated position and posture of the head mount display 100 are stored as the position and posture of the head mount display 100 in a reference state (F 1410 ).
- FIG. 15 illustrates a procedure for embedding screen information of the information terminal in the screen of the head mount display 100 using the position of the information terminal initialized by the procedure illustrated in FIG. 14 .
- viewing contents are displayed on the display unit 309 (F 1501 ).
- an image is captured using a camera which is provided as standard equipment (F 1511 ).
- the information terminal is held by a user wearing the head mount display 100 in his or her hand, and the camera faces the front face of the head mount display 100 .
- object recognition processing is performed with respect to the captured image to detect the head mount display 100 (F 1512 ).
- the position and posture of the head mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated (F 1513 ).
- the position and posture of the head mount display 100 in a reference state are read (F 1541 ), and a change in the posture of the head mount display 100 from the reference state is calculated (F 1514 ).
- How the position/posture of the information terminal has changed from the reference state can be calculated by performing reverse calculation of a change in the position and posture of the information terminal (camera) that produce the above change in the position/posture of the head mount display 100 assuming that the head mount display 100 remains stationary (F 1515 ).
- the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated change in the position and posture of the information terminal (F 1516 ).
- the processes (F 1512 to F 1516 ) after the processing of the captured image may be executed either the information terminal or the head mount display 100 , or may be executed by an external device such as a cloud computer.
- the head mount display 100 acquires screen information from the information terminal (F 1521 ), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F 1516 to generate an output screen (F 1530 ).
- the screen of the information terminal is displayed in the contents reproducing screen (an image in a virtual world) of the head mount display 100 with the same position and posture as the actual information terminal on the view of the user.
- the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents.
- the technology disclosed in the present specification can be suitably applied to an immersive head mount display equipped with no camera. It is needless to say that the technology can also be applied to a video see-through head mount display equipped with a camera, a transmissive, that is, optical see-through head mount display, and, in addition, various types of image display apparatuses other than a head mount display in a similar manner.
- An information processing apparatus including:
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus;
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information relating to the information processing apparatus in a screen of the image display apparatus.
- An information processing apparatus including:
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus;
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
- the information processing apparatus further including a communication unit that communicates with the image display apparatus,
- screen information of the display unit and information of the embedding position are transmitted to the image display apparatus through the communication unit.
- the reference index arrangement information acquisition unit preloads design information relating to the relationship between the posture of the image display apparatus and the posture of the reference index
- the position/posture calculation unit calculates the position and posture of the image display apparatus in the first reference coordinate system on the basis of the design information from an image captured by the image capturing unit when the information processing apparatus is arranged in a predetermined direction with respect to the image display apparatus.
- the reference index arrangement information acquisition unit makes discrimination between the reference indexes on the basis of at least one of color, texture, and shape and acquires axial information of the reference coordinate system of the image display apparatus.
- the information processing apparatus receives information encoded in a flashing pattern from the reference index.
- the reference index arrangement information acquisition unit makes an inquiry about information obtained by decoding the flashing pattern to a database to acquire coordinate information of the reference index, information of the reference index, or information of the image display apparatus.
- the position/posture calculation unit calculates the position and posture of the image display apparatus in the first reference coordinate system using visual simultaneous localization and mapping (SLAM).
- the position/posture transform unit calculates a change in the posture of the image display apparatus from the reference state and transforms the position and posture of the image display apparatus in the first reference coordinate system to the position and posture of the information processing apparatus in the second reference coordinate system.
- An image processing method including:
- an embedding position calculation step of calculating, on the basis of the position and posture of the information terminal in the second reference coordinate system, a position of embedding screen information of the information terminal with respect to a screen of the image display apparatus.
- An image display system including:
- a display apparatus fixed to a head or face of an observer
- an information processing apparatus that includes an image capturing unit and a display unit, calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus, transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus, and calculates a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Hardware Design (AREA)
- Optics & Photonics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- User Interface Of Digital Computer (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
- This application is a U.S. National Phase of International Patent Application No. PCT/JP2015/062934 filed on Apr. 30, 2015, which claims priority benefit of Japanese Patent Application No. JP 2014-159169 filed in the Japan Patent Office on Aug. 5, 2014. Each of the above-referenced applications is hereby incorporated herein by reference in its entirety.
- The technology disclosed in the present specification relates to an information processing apparatus, an information processing method, and an image display system that provide screen information to be displayed on an image display apparatus that is worn by a user on the head or face and used for viewing an image.
- There is known an image display apparatus that is worn on the head or face and used for viewing an image, that is, a head mount display. The head mount display includes, for example, image display units which are arranged for the respective right and left eyes and forms an enlarged virtual image of a display image by a virtual image optical system so that a user can observe an image with a realistic feeling.
- The head mount display is equipped with a high-resolution display panel which includes, for example, a liquid crystal or an organic electro-luminescence (EL) element as each of the display units for right and left eyes. Further, head mount displays may be classified into a transmissive type and a light-shielding type.
- The transmissive head mount display enables a wearer to observe surrounding scenery also when the wearer wears the head mount display on the head and an image is displayed thereon (refer to Patent Document 1, for example). Thus, a user can avoid the risk of, for example, collision with an obstacle when using the head mount display in the outside or during walking.
- On the other hand, the light-shielding head mount display is configured to directly cover the eyes of a wearer when worn on the head, which increases an immersive feeling during viewing of an image. It is possible to duplicate a realistic feeling like viewing in a movie theater by projecting a display screen in an enlarged manner using the virtual optical system to allow a user to observe the display screen as an enlarged virtual image having an appropriate angle of view and duplicating multiple channels by a headphone (refer to Patent Document 2, for example). Further, also in the light-shielding type head mount display, there is known a video see-through method that has a built-in camera capable of capturing an image of the front side of a wearer and displays an external image obtained by the image capturing, and the wearer can observe surrounding scenery through the external image (refer to Patent Documents 3, 4, for example). Note that the former transmissive head mount display is called optical see-through or merely see-through contrary to the video see-through.
- In either of the light-shielding and the transmissive types, the head mount display restricts the sense of sight and the sense of hearing of a user wearing the head mount display. In particular, the view of a user wearing the light-shielding head mount display is completely obstructed. Thus, the user cannot perform any operation to the outside such as answering a phone or input to the screen of a computer.
- In the above video see-through head mount display, a view image of a user captured by a camera is displayed to allow the user to operate an object which is present in the view through the displayed image.
- For example, there has been proposed a portable terminal system which is provided with a portable terminal which includes a display apparatus which performs display according to display information and a head mount display which displays a virtual screen in a display area in the view of a wearer on the basis of the display information acquired from the portable terminal (refer to Patent Document 5, for example).
- However, the portable terminal system as described above is premised on that the head mount display is equipped with a camera which captures an image of the outside world. It is difficult for a user wearing a head mount display equipped with no camera to perform an operation with respect to the outside unless the head mount display is detached from the head to completely suspend viewing.
- Patent Document 1: Japanese Patent Application Laid-Open No. 2012-42654
- Patent Document 2: Japanese Patent Application Laid-Open No. 2012-141461
- Patent Document 3: Japanese Patent Application Laid-Open No. 2005-38321
- Patent Document 4: Japanese Patent Application Laid-Open No. 2011-242591
- Patent Document 5: Japanese Patent Application Laid-Open No. 2011-186856
- Patent Document 6: Japanese Patent Application Laid-Open No. 2012-186660
- An object of the technology disclosed in the present specification is to provide an excellent information processing apparatus, an excellent information processing method, and an excellent image display system capable of appropriately providing screen information to be displayed on an image display apparatus that is worn by a user on the head or face and used for viewing of an image.
- A technology disclosed in the present specification has been made in consideration of the problem, and a first aspect thereof is an information processing apparatus including:
- an image capturing unit;
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus; and
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information relating to the information processing apparatus in a screen of the image display apparatus.
- Further, a second aspect of the technology disclosed in the present specification is an information processing apparatus including:
- an image capturing unit;
- a display unit;
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus; and
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
- According to a third aspect of the technology disclosed in the present specification, the information processing apparatus according to the second aspect further includes a communication unit that communicates with the image display apparatus. Further, screen information of the display unit and information of the embedding position are configured to be transmitted to the image display apparatus through the communication unit.
- According to a fourth aspect of the technology disclosed in the present specification, the position/posture calculation unit of the information processing apparatus according to the second or third aspect is configured to calculate the position and posture of the image display apparatus in the first reference coordinate system on the basis of a reference index mounted on the image display apparatus.
- According to a fifth aspect of the technology disclosed in the present specification, the information processing apparatus according to the fourth aspect further includes a reference index arrangement information acquisition unit that acquires information relating to a relationship between posture of the image display apparatus and posture of the reference index.
- According to a sixth aspect of the technology disclosed in the present specification, the reference index arrangement information acquisition unit of the information processing apparatus according to the fifth aspect is configured to preload design information relating to the relationship between the posture of the image display apparatus and the posture of the reference index, and the position/posture calculation unit is configured to calculate the position and posture of the image display apparatus in the first reference coordinate system on the basis of the design information from an image captured by the image capturing unit when the information processing apparatus is arranged in a predetermined direction with respect to the image display apparatus.
- According to a seventh aspect of the technology disclosed in the present specification, the reference index arrangement information acquisition unit of the information processing apparatus according to the fifth aspect is configured to receive information from the reference index.
- According to an eighth aspect of the technology disclosed in the present specification, three or more reference indexes whose arrangement information is known are arranged on the image display apparatus. Further, the reference index arrangement information acquisition unit of the information processing apparatus according to the seventh aspect is configured to make discrimination between the reference indexes on the basis of at least one of color, texture, and shape and acquire axial information of the reference coordinate system of the image display apparatus.
- According to a ninth aspect of the technology disclosed in the present specification, the reference index arrangement information acquisition unit of the information processing apparatus according to the seventh aspect is configured to receive information encoded in a flashing pattern from the reference index.
- According to a tenth aspect of the technology disclosed in the present specification, the reference index arrangement information acquisition unit of the information processing apparatus according to the ninth aspect is configured to make an inquiry about information obtained by decoding the flashing pattern to a database to acquire coordinate information of the reference index, information of the reference index, or information of the image display apparatus.
- According to a eleventh aspect of the technology disclosed in the present specification, the position/posture calculation unit of the information processing apparatus according to the second or third aspect is configured to calculate the position and posture of the image display apparatus in the first reference coordinate system using visual SLAM.
- According to a twelfth aspect of the technology disclosed in the present specification, the information processing apparatus according to the eleventh aspect is configured to store position and posture of the image display apparatus calculated by the position/posture calculation unit in a reference state in which the information processing apparatus is held in an initial posture, and the position/posture transform unit is configured to calculate a change in the posture of the image display apparatus from the reference state and transform the position and posture of the image display apparatus in the first reference coordinate system to the position and posture of the information processing apparatus in the second reference coordinate system.
- Further, a thirteenth aspect of the technology disclosed in the present specification is an image processing method including:
- a position/posture calculation step of calculating, on the basis of an image of another image display apparatus captured by an image capturing unit mounted on an information terminal, position and posture of the image display apparatus in a first reference coordinate system of the information terminal;
- a position/posture transform step of transforming the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information terminal in a second reference coordinate system of the image display apparatus; and
- an embedding position calculation step of calculating, on the basis of the position and posture of the information terminal in the second reference coordinate system, a position of embedding screen information of the information terminal with respect to a screen of the image display apparatus.
- Further, a fourteenth aspect of the technology disclosed in the present specification is an image display system including:
- a display apparatus fixed to a head or face of an observer; and
- an information processing apparatus that includes an image capturing unit and a display unit, calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus, transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus, and calculates a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
- Note that the “system” described herein indicates a logical aggregate of a plurality of apparatuses (or functional modules which achieve a specific function), and it does not particularly matter whether the apparatuses or the functional modules are present in a single housing.
- According to the technology disclosed in the present specification, it is possible to provide an excellent information processing apparatus, an excellent information processing method, and an excellent image display system capable of appropriately providing screen information to be displayed on an image display apparatus that is worn by a user on the head or face and used for viewing an image.
- Note that the effects described in the present specification are merely examples, and the effects of the present invention are not limited thereto. Further, the present invention may achieve an additional effect other than the above effects.
- Other objects, features, and advantages of the technology disclosed in the present specification will become apparent from the more detailed description based on the following embodiment and the accompanying drawings.
-
FIG. 1 is a diagram illustrating a front view of a user wearing ahead mount display 100 on the head. -
FIG. 2 is a diagram illustrating a top view of the user wearing thehead mount display 100 illustrated inFIG. 1 . -
FIG. 3 is a diagram illustrating an example of an internal configuration of thehead mount display 100. -
FIG. 4 is a diagram illustrating a state in which the mutual positional relationship and posture of an information terminal are grasped on the basis of an image captured by anouter camera 312 of thehead mount display 100 and an image of the information terminal is displayed in a superimposed manner. -
FIG. 5 is a diagram illustrating a procedure for achieving screen display illustrated inFIG. 4 . -
FIG. 6 is a diagram illustrating a state in which the mutual positional relationship and posture between thehead mount display 100 and the information terminal are grasped on the basis of an image captured by a camera of the information terminal and an image of the information terminal is displayed in a superimposed manner. -
FIG. 7 is a diagram illustrating a procedure for achieving screen display illustrated inFIG. 6 . -
FIG. 8 is a diagram illustrating a state in which anAR marker 801 is mounted on thehead mount display 100, and a camera of an information terminal which is held in a principal axis front direction captures an image of thehead mount display 100. -
FIG. 9 is a diagram illustrating a state in whichreference indexes 901A to D having a known shape are mounted on thehead mount display 100, and the camera of the information terminal which is held in the principal axis front direction captures an image of thehead mount display 100. -
FIG. 10 is a diagram illustrating a state in which four ormore reference indexes 1001A to D having different colors and a known shape are mounted on thehead mount display 100, and the camera of the information terminal which is held in the principal axis front direction captures an image of thehead mount display 100. -
FIG. 11 is a diagram illustrating a state in whichreference indexes 1101A to D each of which flashes are mounted on thehead mount display 100, and the camera of the information terminal which is held in the principal axis front direction captures an image of thehead mount display 100. -
FIG. 12 is a diagram illustrating a procedure for embedding screen information of the information terminal in the screen of thehead mount display 100 on the basis of object coordinates received from the reference index. -
FIG. 13 is a diagram illustrating a procedure for embedding screen information of the information terminal in the screen of thehead mount display 100 on the basis of a retrieval key received from the reference index. -
FIG. 14 is a diagram illustrating a procedure for initializing the position of the information terminal. -
FIG. 15 is a diagram illustrating a procedure for embedding screen information of the information terminal in the screen of thehead mount display 100 using the position of the information terminal initialized by the procedure illustrated inFIG. 14 . -
FIG. 16 is a diagram illustrating an example of the configuration of the information terminal. - Hereinbelow, an embodiment of the technology disclosed in the present specification will be specifically described with reference to the drawings.
-
FIG. 1 illustrates a front view of a user wearing ahead mount display 100 to which the technology disclosed in the present specification is applied on the head. - When a user wears the
head mount display 100 on the head or face, thehead mount display 100 directly covers the eyes of the user, which enables an immersive feeling to be given to the user viewing an image. Further, since the display image is invisible from the outside (that is, other people), privacy is easily protected in information display. Differently from a see-through type, a user wearing thehead mount display 100 cannot directly look at scenery in the real world. When thehead mount display 100 is equipped with anouter camera 312 which captures an image of scenery in a sight-line direction of a user, the user can indirectly look at scenery in the real world (that is, display the scenery by video see-through) by displaying the captured image. However, as described below, the technology disclosed in the present specification is not premised on the equipment of theouter camera 312 or the video see-through display. - The
head mount display 100 illustrated inFIG. 1 is a structure which is similar to a cap shape and configured to directly cover the right and left eyes of a user wearing thehead mount display 100. A display panel (not illustrated inFIG. 1 ) which is observed by a user is disposed on the inner side of a body of thehead mount display 100 at a position facing the right and left eyes. The display panel includes, for example, a micro display such as an organic EL element or a liquid crystal display and a laser scanning display such as a retinal direct drawing display. -
Microphones head mount display 100. Themicrophones - Further, a
touch panel 315 through which a user can perform touch input using, for example, the fingertip is disposed near the back face of the display panel on the front face of the body of thehead mount display 100. Although, in the illustrated example, a pair of right and lefttouch panels 315 are provided, one or three ormore touch panels 315 may be provided. -
FIG. 2 illustrates a top view of the user wearing thehead mount display 100 illustrated inFIG. 1 . The illustratedhead mount display 100 includes adisplay panel 104L for left eye and adisplay panel 104R for right eye on a side face facing the face of the user. Each of thedisplay panels display panels optical units head mount display 100 with the right and left display systems. In the example illustrated inFIG. 2 , a pupil distance adjustment mechanism 105 is disposed between the display panel for right eye and the display panel for left eye. -
FIG. 3 illustrates an example of an internal configuration of thehead mount display 100. Hereinbelow, each unit will be described. - A
control unit 301 is provided with a read only memory (ROM) 301A and a random access memory (RAM) 301B. A program code to be executed by thecontrol unit 301 and various data are stored in theROM 301A. Thecontrol unit 301 starts display control of an image by executing a program loaded to theRAM 301B and totally controls the operation of the entirehead mount display 100. - Note that the program stored in the
ROM 301A includes a display control program for displaying a reproducing moving image on the screen, and displaying, in a super imposed manner, image information taken from an external apparatus (e.g., a smartphone) on the screen. Further, the data stored in theROM 301A includes identification information (including a model number and a manufacturer's serial number) which is unique to thehead mount display 100, user attribute information such as authentication information (e.g., a personal identification number, a password, or biological information) for authenticating a user who uses thehead mount display 100, and design information of the head mount display 100 (the shape or the arrangement of a reference index mounted on the head mount display 100). - An input interface (IF)
unit 302 is provided with one or more operators (not illustrated) such as a key, a button, and a switch to which a user performs an input operation, and receives an instruction of the user through the operators and outputs the received instruction to thecontrol unit 301. Further, theinput operation unit 302 receives an instruction of a user which includes a remote-control command received by a remote-control reception unit 303 and outputs the received instruction to thecontrol unit 301. - Further, when a user performs a touch operation to the
touch panel 315 which is disposed on the outer side of the body of thehead mount display 100 with the fingertip, the input interface (IF)unit 302 outputs input information such as coordinate data of a position touched by the fingertip to thecontrol unit 301. For example, when thetouch panel 315 is arranged exactly on the back face of a display image (an enlarged virtual image observed through a virtual image optical unit 310) of adisplay unit 309 on the front face of the body of the head mount display 100 (refer toFIG. 1 ), a user can perform a touch operation with a feeling like touching the display image with the fingertip. - A state
information acquisition unit 304 is a functional module which acquires state information of the body of thehead mount display 100 or a user wearing thehead mount display 100. The stateinformation acquisition unit 304 may be equipped with various sensors for detecting state information by itself or may acquire state information from an external apparatus (e.g., a smartphone or a wristwatch worn by the user, or another multifunction terminal) which is provided with some or all of these sensors through a communication unit 305 (described below). - The state
information acquisition unit 304 acquires, for example, information of the position and posture or information of the posture of the head of a user to track a head motion of the user. In order to track a head motion of the user, the stateinformation acquisition unit 304 is a sensor capable of detecting nine axes in total including a three-axis gyro sensor, a three-axis accelerometer, and a three-axis magnetic field sensor. Further, the stateinformation acquisition unit 304 may further use, in combination, any one or two or more sensors such as a global positioning system (GPS) sensor, a Doppler sensor, an infrared sensor, and a radio field intensity sensor. Further, the stateinformation acquisition unit 304 may further use, in combination, information provided from various infrastructures such as cellular phone base station information and Place Engine (registered trademark) information (e.g., electrically measured information from a wireless LAN access point) to acquire the position/posture information. Although, in the example illustrated inFIG. 3 , thestate acquisition unit 304 for head motion tracking is incorporated in thehead mount display 100, thestate acquisition unit 304 may include an accessory component which is externally attached to thehead mount display 100. In the latter case, the externally-attachedstate acquisition unit 304 transmits posture information of the head to the body of thehead mount display 100 through wireless communication such as Bluetooth (registered trademark) communication or a high-speed wired interface such as a universal serial bus (USB). - Further, in addition to the tracking of the head motion of a user described above, the state
information acquisition unit 304 acquires, as state information of a user wearing thehead mount display 100, for example, a working state of a user (wearing/unwearing of the head mount display 100), an action state of a user (a moving state such as standing still, walking, or running, a gesture by the hand or fingertip, an opening/closing state of the eyelid, a sight-line direction, or the large/small of the pupil), a mental state (the degree of impression such as whether a user is immersed in or concentrated on the observation of a display image, the degree of excitation, the degree of awareness, feeling, or emotion), and a physiological state. Further, the stateinformation acquisition unit 304 may be provided with theouter camera 312, a wearing sensor which includes, for example, a mechanical switch, an inner camera which captures an image of the face of a user, various state sensors such as a gyro sensor, an accelerometer, a speed sensor, a pressure sensor, a temperature sensor which senses the body temperature or the ambient temperature, a perspiration sensor, a pulse sensor, a myogenic potential sensor, an ocular potential sensor, a brain wave sensor, an exhalation sensor, and a gas/ion concentration sensor, and a timer (all not illustrated) to acquire these state information items from a user. Further, the stateinformation acquisition unit 304 may use a wearing sensor (refer to Patent Document 6, for example) which detects, when a user has worn the head mount display on the head, the wearing of the head mount display by the user in conjunction with the motion of abutment against the forehead in the head to detect wearing/unwearing of thehead mount display 100. - An environment
information acquisition unit 316 is a functional module which acquires information relating to the environment surrounding the body of thehead mount display 100 or a user wearing thehead mount display 100. The information relating to the environment described herein includes sound; the volume of air; air temperature; air pressure; atmosphere (smoke; dense fog; electromagnetic waves (ultraviolet rays, blue light, radio waves), heat waves (infrared rays) and radioactive rays applied to thehead mount display 100 or a user; carbon monoxide, carbon dioxide, oxygen and nitrogen compound (nicotine) in the atmosphere; nitrogen oxide (NOx) and hydrocarbon (volatile organic compounds (VOC)) floating in the atmosphere or photochemical smog generated by a photochemical reaction of these materials due to the influence of ultraviolet rays; particulate matters; pollen; dust such as house dust; and harmful chemical substances such as asbestos); and other environmental factors. The environmentinformation acquisition unit 316 may be equipped with various environmental sensors such as a sound sensor and an air volume sensor to detect environment information. The environmental sensors may include the microphone and theouter camera 312 described above. Alternatively, the environmentinformation acquisition unit 316 may acquire environment information from an external apparatus (e.g., a smartphone or a wristwatch worn by the user, or another multifunction terminal) which is provided with some or all of these sensors through the communication unit 305 (described below). - The
outer camera 312 is arranged, for example, on substantially the center in the front face of the body of the head mount display 100 (refer toFIG. 2 ) and capable of capturing surrounding images. However, as described below, the technology disclosed in the present specification is not premised on the equipment or the use of theouter camera 312. - The
communication unit 305 performs communication processing with an external apparatus (not illustrated), and modulation/demodulation and encoding/decoding processing of a communication signal. The external apparatus includes a contents reproducing apparatus (a Blue-ray Disk player or a DVD player) which supplies viewing contents when a user uses thehead mount display 100, a streaming server, and information terminals such as a smartphone and a personal computer used by a user. Further, thecontrol unit 301 transmits transmission data to the external apparatus through thecommunication unit 305. - The
communication unit 305 has any configuration. For example, thecommunication unit 305 may be configured according to a communication mode which is used in a transmission/reception operation with the external apparatus as a communication partner. The communication mode may be a wired form or a wireless form. The communication standard described herein includes mobile high-definition link (MHL), universal serial bus (USB), high definition multimedia interface (HDMI, registered trademark), Wi-Fi (registered trademark), Bluetooth (registered trademark) communication, Bluetooth (registered trademark) low energy (BLE) communication, ultra-low power consumption wireless communication such as ANT, and mesh network standardized by, for example, IEEE802.11s. Alternatively, thecommunication unit 305 is, for example, a cellular wireless transmitter receiver which operates in accordance with a standard specification such as wideband code division multiple access (W-CDMA) or long term evolution (LTE). - A
storage unit 306 is a mass storage device which includes, for example, a solid state drive (SSD). Thestorage unit 306 stores therein an application program to be executed by thecontrol unit 301 and various data. For example, contents viewed by a user using thehead mount display 100 are stored in thestorage unit 306. - An
image processing unit 307 further performs signal processing such as image quality correction and performs transformation to a resolution corresponding to the screen of thedisplay unit 309 with respect to an image signal output from thecontrol unit 301. Then, adisplay drive unit 308 sequentially selects pixels of thedisplay unit 309 for each row and line-sequentially scans the pixels to supply a pixel signal based on the signal-processed image signal. - The
display unit 309 includes a display panel which includes, for example, a micro display such as an organic EL element or a liquid crystal display or a laser scanning display such as a retinal direct drawing display. The virtual imageoptical unit 310 projects, in an enlarged manner, a display image of thedisplay unit 309 and allows a user to observe the display image as an enlarged virtual image. - Note that the display image output by the
display unit 309 includes commercial contents (virtual world) which are supplied from a contents reproducing apparatus (a Blue-ray Disk player or a DVD player) or a streaming server and a captured image of the outer camera 312 (an image in the real world such as a view image of a user). Further, image information transmitted from an information terminal such as a smartphone or a personal computer may be displayed in a superimposed manner on the contents reproducing screen. - An
audio processing unit 313 further performs audio quality correction and audio amplification with respect to an audio signal output from thecontrol unit 301 and performs signal processing of an audio signal input thereto. Then, an audio input/output unit 314 outputs the processed audio to the outside and performs voice input from the microphone (described above). - A user wearing the
head mount display 100 may wish to operate another information terminal such as a smartphone when the view of the user is completely obstructed during viewing of contents. For example, the user may wish to check a record of incoming mails or calls or wish to know the current time. - When the
head mount display 100 inputs screen information of an external information terminal, and changes an original contents reproducing screen to the screen of the external information terminal so as to be displayed on thedisplay unit 309 or displays, in an superimposed manner, the screen of the external information terminal screen on the original screen, a user can check the contents of the screen of the external information terminal without detaching thehead mount display 100 from the head. The superimposition display described herein includes separately displaying a small screen in a picture-in-picture form and blending two screens at a predetermined ratio. - However, when a user displays the screen of an information terminal held in his/her hand on the
head mount display 100 and the position of the screen of the information terminal is not changed by moving the hand, the operation lacks reality. Further, when the position of the screen of the information terminal displayed by thedisplay unit 309 is not completely linked with the position of the screen of the actual information terminal, it is extremely difficult for a user to perform an operation with respect to the screen of the actual information terminal while looking at the screen of the information terminal displayed on thehead mount display 100. - The present applicant thinks that, in order for a user to grasp the positional relationship between the hand holding the information terminal and an operation finger, not only feedback of somatic sensation such as cutaneous sensation, but also correction by visual information is required.
- For example, as illustrated in
FIG. 4 , an image of an information terminal which is held by a user in his or her hand is captured using the outer camera 312 (reference sign 401). Then, the captured image of the information terminal is displayed on the contents reproducing screen of thehead mount display 100 with the same position and posture as the actual information terminal on the view of the user (reference sign 402). Accordingly, the user can easily perform an operation with respect to the screen of the actual information terminal while looking at the screen of the information terminal which is displayed by video see-through on thehead mount display 100. Specifically, it is possible to display, in a superimposed manner, an image of the information terminal on the contents reproducing screen of thehead mount display 100 with the same position and posture as the actual information terminal on the view of the user by grasping the mutual positional relationship and posture between thehead mount display 100 and the information terminal on the basis of the image captured by theouter camera 312. -
FIG. 5 illustrates a procedure for achieving the screen display illustrated inFIG. 4 . - In the
head mount display 100, viewing contents are displayed on the display unit 309 (F501). - Further, in the
head mount display 100, an image is captured using the outer camera 312 (F511). Then, object recognition processing is performed with respect to the captured image to detect the information terminal (F512). Then, the position and posture of the information terminal are calculated (F513). - Here, a reference coordinate system of the
display unit 309 which displays the information terminal on the head mount display 100 (hereinbelow, merely referred to as the reference coordinate system of the head mount display 100) differs from a reference coordinate system of theouter camera 312 which captures an image of the information terminal in a strict sense. Thus, in F513, it is necessary to transform the position and posture of the information terminal detected in the camera reference coordinate system to the position and posture of the information terminal in the reference coordinate system of thehead mount display 100. - Further, in order to improve the accuracy of object recognition and position/posture calculation, a visual reference index (marker) may be applied to the information terminal. For example, a known pattern displayed on the screen of the information terminal may be used as the reference index. Alternatively, the position and posture of the information terminal may be calculated with no marker (described below). Then, the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F514).
- Then, the
head mount display 100 acquires screen information from the information terminal (F521), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F514 to generate an output screen (F530). - As a result of such processing, the screen of the information terminal is present on the contents reproducing screen of the
head mount display 100 with the same position and posture as the actual information terminal on the view of the user. Thus, the user can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents. - However, in the above method as illustrated in
FIGS. 4 and 5 , theouter camera 312 is required to correct the displayed position and posture of the information terminal using visual information. In other words, disadvantageously, the above method cannot be applied to a product that is not equipped with theouter camera 312. -
FIGS. 1 to 3 illustrate, as an example, thehead mount display 100 equipped with theouter camera 312. However, an outer camera is not necessarily provided as standard equipment, and a considerable number of products equipped with no outer camera have already become widespread. In this kind of existing products, thehead mount display 100 includes no means for grasping the mutual positional relationship and posture between the display body and an information terminal as a current operation target. Thus, it is not possible to display the information terminal on the contents reproducing screen of thehead mount display 100 with the same position and posture as the actual information terminal on the view of a user. - On the other hand, many of information terminals to be the current operation target such as smartphones and personal computers are products equipped with cameras. Thus, the present applicant focuses on this point and proposes a method that conversely captures an image of the
head mount display 100 using a camera mounted on an information terminal (reference sign 601) to grasp the mutual positional relationship and posture between thehead mount display 100 and the information terminal as the current operation target and displays screen information transmitted from the information terminal on the contents reproducing screen of thehead mount display 100 with the same position and posture as the actual information terminal on the view of a user (reference sign 602) as illustrated inFIG. 6 . -
FIG. 7 illustrates a procedure for achieving the screen display illustrated inFIG. 6 . - In the
head mount display 100, viewing contents are displayed on the display unit 309 (F701). - On the other hand, in the information terminal, an image is captured using the camera which is provided as standard equipment (F711) The information terminal is held by a user wearing the
head mount display 100 in his or her hand, and the camera faces the front face of thehead mount display 100. Then, object recognition processing is performed with respect to the captured image to detect the head mount display 100 (F712). Then, the position and posture of thehead mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated (F713). - Then, the position and posture of the
head mount display 100 in the reference coordinate system of the information terminal are transformed to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 (F714), and the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F715). Note that the processes (F712 to F715) after the processing of the captured image may be executed by either the information terminal or thehead mount display 100, or may be executed by an external device such as a cloud computer. - Then, the
head mount display 100 acquires screen information from the information terminal (F721), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F715 to generate an output screen (F730). - As a result of such processing, the screen of the information terminal with the same position and posture as the actual information terminal on the view of the user is displayed in the contents reproducing screen (an image in a virtual world) of the
head mount display 100. Thus, the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents. -
FIG. 16 illustrates an example of the configuration of an information terminal capable of achieving the processing as illustrated inFIG. 7 . The illustrated information terminal is configured by connecting adisplay unit 1620, anaudio processing unit 1630, acommunication unit 1640, astorage unit 1650, acamera unit 1660, asensor unit 1670 and the like to acontrol unit 1610. - The
control unit 1610 includes, for example, a central processing unit (CPU) 211, aROM 1612, and aRAM 1613. A program code to be executed by theCPU 1611 and information which is essential for the information terminal are stored in theROM 1612. - The
CPU 1611 loads the program code to theRAM 1613 from theROM 1612 or thestorage unit 1640 and executes the program code. The program executed by theCPU 1611 includes operating systems such as Android and iOS, an application for transferring screen information to an external image display apparatus (e.g., the head mount display 100), and various application programs which operate under an execution environment provided by the operating system. - The
display unit 1620 is provided with adisplay panel 1621 which includes, for example, a liquid crystal element or an organic EL element and atransparent touch panel 1623 which is stuck to the upper face of thedisplay panel 1622. Thedisplay panel 1621 is connected to thecontrol unit 1610 through adisplay interface 1622, and displays and outputs image information generated by the control unit 210. Further, thetouch panel 1623 is connected to thecontrol unit 1610 through atouch interface 1624 and outputs coordinate information obtained by an operation of a user on thedisplay panel 1621 with the fingertip to thecontrol unit 1610. Thecontrol unit 1610 detects a user operation such as a tap, long press, flick, or swipe on the basis of the input coordinate information and activates processing corresponding to the user operation. - The
audio processing unit 1630 is provided with anaudio output unit 1631 such as a speaker, anaudio input unit 1632 such as a microphone, and anaudio codec 1633 which encodes and decodes an audio signal which is input and output. Further, theaudio processing unit 1630 may further be provided with anoutput terminal 1634 for outputting an audio signal to a headphone (not illustrated). - The
communication unit 1640 performs information communication processing between an application executed by thecontrol unit 1610 and an external apparatus. The external apparatus described herein includes a television set, an information terminal (not illustrated) handled by another user, and server which is present on the Internet. Thecommunication unit 1640 is equipped with a physical layer module such as Wi-Fi, near field communication (NFC), or Bluetooth (registered trademark) according to a commutation medium to be used and performs modulation/demodulation processing or encoding/decoding processing of a communication signal which is transmitted and received through the physical layer module. Further, screen information displayed on thedisplay unit 1620 may be transferred to an external image display apparatus (e.g., the head mount display 100) through thecommunication unit 1640. - The
storage unit 1650 includes a mass storage device such as a solid state drive (SSD) or a hard disc drive (HDD). For example, an application program and contents downloaded through thecommunication unit 1640 and image data such as a still image and a moving image captured by thecamera unit 1660 are stored in thestorage unit 1650. - The
camera unit 1660 is provided with animage sensor 1661 which photoelectrically converts light taken through a lens (not illustrated) such as a charge coupled device (CCD) or a complementary metal oxide semiconductor (CMOS) and an analog front end (AFE) processing unit 1662 which performs noise reduction and digitization of a detection signal of theimage sensor 1661 to generate image data, and outputs the generated image data to thecontrol unit 1610 from thecamera interface 1663. - The
sensor unit 1670 includes, for example, a global positioning system (GPS) sensor for acquiring positional information of the information terminal, a gyro sensor for detecting the posture or an acting force of the body of the information terminal, and an accelerometer. - In the method illustrated in
FIGS. 6 and 7 , one or more reference indexes such as an augmented reality (AR) marker may be mounted on thehead mount display 100 so as to more accurately grasp the mutual positional relationship and posture between thehead mount display 100 and the information terminal on the basis of an image captured by the camera which is mounted on the information terminal. - For example, detailed information (e.g., product information such as the model number and the manufacturer's serial number) of the
head mount display 100 can be acquired depending on the kind of the reference index. Using the detailed information further improves the accuracy of recognizing the mutual position and posture between thehead mount display 100 and the information terminal. The following (1) and (2) can be described as examples of a use method of the reference index. - (1) The direction and unit length of an optical principal axis are specified from the arrangement of a reference index based on a certain agreement to precisely perform coordinate axis alignment and position measurement.
- (2) Identification information of a predetermined bit length (e.g., 32-bit) is encoded in the reference index, and a database is searched using the identification information to obtain information similar to (1) to precisely perform coordinate axis alignment and position measurement.
- Further, in the case of marker-less, that is, when no reference index is mounted on the
head mount display 100, it is not possible to acquire absolute position/posture information. It is merely possible to measure a displacement in position/posture relative to an initial position in the start of processing. Thus, when the camera mounted on the information terminal captures an image of thehead mount display 100, a user may hold the information terminal at a previously determined initial position and send a reset signal which notifies the initial position to the information terminal. The previously-determined initial position described herein is, for example, the front side of the eye-line of a user wearing thehead mount display 100. Further, a method for sending a reset signal includes tapping a touch panel screen of the information terminal and operating a switch on thehead mount display 100. A relative displacement thereafter is measured and reflected to rendering on the screen of thehead mount display 100. - Note that, as a specific marker-less method, the position/posture calculation may be performed using visual simultaneous localization and mapping (SLAM). The visual SLAM is a technology capable of simultaneously performing own position estimation of the camera and map creation under an unknown environment. An example of visual SLAM includes integrated augmented reality technology “SmartAR” (registered trademark of Sony Corporation). According to the marker-less method, it is not necessary to apply the reference index to the
head mount display 100, and the method can be applied to any product. - Here, a method for calculating the position and posture of the information terminal to be embedded in the screen of the
display unit 309 of thehead mount display 100 will further be considered. - As illustrated in
FIGS. 4 and 5 , when the position and posture of the information terminal to be embedded in the screen are calculated from an image captured by theouter camera 312 which is mounted on thehead mount display 100, the reference coordinate system of thehead mount display 100 differs from the reference coordinate system of theouter camera 312 in a strict sense. Thus, in F513, it is necessary to transform the position and posture of the information terminal detected in the camera reference coordinate system to the position and posture of the information terminal in the reference coordinate system of thehead mount display 100. The position of theouter camera 312 with respect to thehead mount display 100 is fixed, that is, known. Thus, it is easy to transform the position and posture of the information terminal detected in the camera reference coordinate system to the position and posture of the information terminal in the reference coordinate system of thehead mount display 100. - Next, a case in which the position and posture of the information terminal to be embedded in the screen are calculated using the camera which is mounted on the information terminal as illustrated in
FIGS. 6 and 7 will be considered. - The position of the camera with respect to the information terminal is fixed, that is, known. Thus, it is easy to transform the position and posture of the
head mount display 100 whose image is captured by the camera to the position and posture of thehead mount display 100 in the reference coordinate system of the information terminal. - Further, it is possible to measure the position and posture of the reference index which is mounted on the
head mount display 100 in the reference coordinate system of the camera which is mounted on the information terminal. - However, the relationship between the posture of the
head mount display 100 and the posture of the reference index mounted on thehead mount display 100 cannot be acquired by measurement from the outside. Thus, reference index arrangement information transmission means for transmitting information of the posture relationship between thehead mount display 100 and the reference index to the information terminal is required to more precisely perform transform processing of the position and posture in F714. The following (1) to (3) will be described as examples of the reference index arrangement information transmission means. - (1) Design information of the
head mount display 100 is used as known information of the information terminal. - (2) The reference index transmits necessary information to the information terminal.
- (3) An initialization process by a user is gone through.
- Hereinbelow, each transmission means will be specifically described.
- (1) Design information of the
head mount display 100 is used as known information of the information terminal. - For example, a dedicated application of the information terminal (e.g., an application for calculating the position and posture of the
head mount display 100 from an image captured by the camera) or information corresponding to the model number of thehead mount display 100 is preloaded to the information terminal. This case is premised on that the shape and arrangement of the reference index which is mounted on thehead mount display 100 are all previously designed. Further, defining, for example, a principal axis of thehead mount display 100 and arranging the reference index so as to be aligned with the principal axis is a promise or specification. This transmission means enables the shape of the reference index to have flexibility. -
FIG. 8 illustrates a state in which anAR marker 801 is mounted on thehead mount display 100, and the camera of the information terminal which is held in a principal axis front direction captures an image of thehead mount display 100. Arrangement information for sticking the AR marker is known. In the information terminal, the relative position and posture of the information terminal can be calculated by observing the position and posture (shape) of the AR marker on the captured image. - Further,
FIG. 9 illustrates a state in whichreference indexes 901A to D having a known shape are mounted on thehead mount display 100, and the camera of the information terminal which is held in the principal axis front direction captures an image of thehead mount display 100. Arrangement information of each of thereference indexes 901A to D is known. In the information terminal, the relative position and posture of the information terminal can be calculated by observing the position of each of thereference indexes 901A to D on the captured image. However, it is necessary to install three or more reference indexes. A method that captures an image of three or more points having known three-dimensional coordinates and obtains the position/posture of the camera from the position of each of the points on the image, that is, a method that calculates mapping from a known three-dimensional coordinate system to two-dimensional coordinates defined on a camera image is an established technology which is called 3D-2D registration in the field of, for example, image processing. Observation of at least three or more points is required to obtain six variables corresponding to position/posture called an external parameter of the camera. Typically, estimation using the least squares method is performed using more observation points (note that, here, an internal parameter of the camera is known). - In both the transmission means illustrated in
FIGS. 8 and 9 , the relative position and posture of the information terminal can be recognized by the camera on the information terminal, and it is not necessary to use theouter camera 312 on thehead mount display 100. - (2) The reference index transmits necessary information to the information terminal.
-
FIG. 10 illustrates a state in which three ormore reference indexes 1001A to D having different colors and a known shape are mounted on thehead mount display 100, and the camera of the information terminal which is held in the principal axis front direction captures an image of thehead mount display 100. Thereference indexes 1001A to D have different colors, textures and shapes, and axial information of the reference coordinate system of thehead mount display 100 is transmitted to the information terminal according to the combination of thereof. Arrangement information is known. In the information terminal, discrimination between thereference indexes 1001A to D can be made on the basis of the color, texture, and shape of each of thereference indexes 1001A to D on the captured image. For example, a line formed by connecting thereference indexes reference indexes 1001A and 100B or a combination of thereference indexes - Further,
FIG. 11 illustrates a state in whichreference indexes 1101A to D each of which flashes are mounted on thehead mount display 100, and the camera of the information terminal which is held in the principal axis front direction captures an image of thehead mount display 100. In each of thereference indexes 1101A to D, information which is desired to be transmitted such as positional information of each of the indexes in the reference coordinate system of thehead mount display 100, model number information of thehead mount display 100 in which the indexes are installed, or a retrieval key of a database which stores other necessary data is encoded in a flashing pattern, and the information terminal can obtain information by decoding the flashing pattern. In both the transmission means, the relative position and posture of the information terminal can be recognized by the camera on the information terminal, and it is not necessary to use theouter camera 312 on thehead mount display 100. -
FIG. 12 illustrates a procedure for embedding screen information of the information terminal in the screen of thehead mount display 100 on the basis of object coordinates received from the reference index. - In the
head mount display 100, viewing contents are displayed on the display unit 309 (F1201). - On the other hand, in the information terminal, an image is captured using a camera which is provided as standard equipment (F1211). The information terminal is held by a user wearing the
head mount display 100 in his or her hand, and the camera faces the front face of thehead mount display 100. Then, object recognition processing is performed with respect to the captured image to detect a reference index of the head mount display 100 (F1212). - Then, information of object coordinates is received from the reference index (F1241), and the position and posture of the
head mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated from the position and posture of the detected reference index (F1213). The calculation processing is the 3D-2D registration (described above) or processing equivalent to the 3D-2D registration. That is, it is possible to calculate the position/posture of the HMD (coordinate system) viewed from the camera coordinate system or the position/posture of the camera (coordinate system) viewed from the HMD coordinate system by associating the two-dimensional position of the reference index on the camera coordinates detected in F1212 with the three-dimensional position of each of the reference indexes in the reference coordinate system of thehead mount display 100 received in F1241. Here, the position/posture of the HMD viewed from the camera coordinate system is calculated. - Then, the position and posture of the
head mount display 100 in the reference coordinate system of the information terminal are transformed to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 (F1214), and the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F1215). Note that the processes (F1212 to F1215) after the processing of the captured image may be executed by either the information terminal or thehead mount display 100, or may be executed by an external device such as a cloud computer. - Then, the
head mount display 100 acquires screen information from the information terminal (F1221), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F1215 to generate an output screen (F1230). - As a result of such processing, the screen of the information terminal is displayed in the contents reproducing screen (an image in a virtual world) of the
head mount display 100 with the same position and posture as the actual information terminal on the view of the user. Thus, the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents. -
FIG. 13 illustrates a procedure for embedding screen information of the information terminal in the screen of thehead mount display 100 on the basis of a retrieval key received from the reference index. - In the
head mount display 100, viewing contents are displayed on the display unit 309 (F1301). - On the other hand, in the information terminal, an image is captured using a camera which is provided as standard equipment (F1311). The information terminal is held by a user wearing the
head mount display 100 in his or her hand, and the camera faces the front face of thehead mount display 100. Then, object recognition processing is performed with respect to the captured image to detect a reference index of the head mount display 100 (F1312). - Then, a retrieval key is received from the reference index (F1341), and an inquiry is made using the retrieval key (F1342) to acquire, for example, coordinate information of the reference index, information of the reference index, and information of the
head mount display 100. Then, the position and posture of thehead mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated from the position and posture of the detected reference index (F1313). - Then, the position and posture of the
head mount display 100 in the reference coordinate system of the information terminal are transformed to the position and posture of the information terminal in the reference coordinate system of the head mount display 100 (F1314), and the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated position and posture (F1315). Note that the processes (F1312 to F1315) after the processing of the captured image may be executed either the information terminal or thehead mount display 100, or may be executed by an external device such as a cloud computer. - Then, the
head mount display 100 acquires screen information from the information terminal (F1321), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F1315 to generate an output screen (F1330). - As a result of such processing, the screen of the information terminal is displayed in the contents reproducing screen (an image in a virtual world) of the
head mount display 100 with the same position and posture as the actual information terminal on the view of the user. Thus, the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents. - (3) An initialization process by a user is gone through.
- A user wearing the
head mount display 100 holds the information terminal with a previously determined initial posture, for example, the front side of the eye-line and performs reset (initialization). In the information terminal, thehead mount display 100 whose image has been captured by the camera is recognized by applying, for example, the visual SLAM technology, and the position of the information terminal is initialized corresponding to the posture thereof. -
FIG. 14 illustrates a procedure for initializing the position of the information terminal. - A user wearing the
head mount display 100 holds the information terminal with the initial posture, for example, the front side of the eye-line. In this state, an image of thehead mount display 100 is captured by the camera of the information terminal (F1401), and thehead mount display 100 is detected from the captured image (F1402). Then, the position and posture of thehead mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated (F1403). - Here, for example, when the user performs an initial position trigger, the calculated position and posture of the
head mount display 100 are stored as the position and posture of thehead mount display 100 in a reference state (F1410). -
FIG. 15 illustrates a procedure for embedding screen information of the information terminal in the screen of thehead mount display 100 using the position of the information terminal initialized by the procedure illustrated inFIG. 14 . - In the
head mount display 100, viewing contents are displayed on the display unit 309 (F1501). - On the other hand, in the information terminal, an image is captured using a camera which is provided as standard equipment (F1511). The information terminal is held by a user wearing the
head mount display 100 in his or her hand, and the camera faces the front face of thehead mount display 100. Then, object recognition processing is performed with respect to the captured image to detect the head mount display 100 (F1512). Then, the position and posture of thehead mount display 100 in a coordinate system based on the information terminal (or a camera coordinate system) are calculated (F1513). - Then, the position and posture of the
head mount display 100 in a reference state (refer toFIG. 14 ) are read (F1541), and a change in the posture of thehead mount display 100 from the reference state is calculated (F1514). How the position/posture of the information terminal has changed from the reference state can be calculated by performing reverse calculation of a change in the position and posture of the information terminal (camera) that produce the above change in the position/posture of thehead mount display 100 assuming that thehead mount display 100 remains stationary (F1515). Then, the position of embedding screen information of the information terminal with respect to the screen of the viewing contents is calculated on the basis of information of the calculated change in the position and posture of the information terminal (F1516). Note that the processes (F1512 to F1516) after the processing of the captured image may be executed either the information terminal or thehead mount display 100, or may be executed by an external device such as a cloud computer. - Then, the
head mount display 100 acquires screen information from the information terminal (F1521), and embeds the screen information in the contents reproducing screen at the embedding position calculated in F1516 to generate an output screen (F1530). - As a result of such processing, the screen of the information terminal is displayed in the contents reproducing screen (an image in a virtual world) of the
head mount display 100 with the same position and posture as the actual information terminal on the view of the user. Thus, the user whose view is completely obstructed can perform a screen check and an input operation in the information terminal without losing reality while continuously viewing contents. - In the above, the technology disclosed in the present specification has been specifically described with reference to the specific embodiment. However, it is obvious that a person skilled in the art may perform amendments and substitutions for the embodiment without departing from the gist of the technology disclosed in the present specification.
- The technology disclosed in the present specification can be suitably applied to an immersive head mount display equipped with no camera. It is needless to say that the technology can also be applied to a video see-through head mount display equipped with a camera, a transmissive, that is, optical see-through head mount display, and, in addition, various types of image display apparatuses other than a head mount display in a similar manner.
- Further, the technology disclosed in the present specification can be suitably applied to both binocular and monocular head mount displays.
- In short, although the technology disclosed in the present specification has been described in the form of exemplification, description contents of the present specification should not be interpreted in a limited manner. In order to judge the gist of the technology disclosed in the present specification, the claims should be considered.
- Note that the technology disclosed in the present specification can employ the following configurations.
- (1) An information processing apparatus including:
- an image capturing unit;
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus; and
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information relating to the information processing apparatus in a screen of the image display apparatus.
- (2) An information processing apparatus including:
- an image capturing unit;
- a display unit;
- a position/posture calculation unit that calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus;
- a position/posture transform unit that transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus; and
- an embedding position calculation unit that calculates, on the basis of the position and posture of the information processing apparatus in the second reference coordinate system, a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
- (3) The information processing apparatus according to (2), further including a communication unit that communicates with the image display apparatus,
- wherein screen information of the display unit and information of the embedding position are transmitted to the image display apparatus through the communication unit.
- (4) The information processing apparatus according to (2) or (3), wherein the position/posture calculation unit calculates the position and posture of the image display apparatus in the first reference coordinate system on the basis of a reference index mounted on the image display apparatus.
(5) The information processing apparatus according to (4), further including a reference index arrangement information acquisition unit that acquires information relating to a relationship between posture of the image display apparatus and posture of the reference index.
(6) The information processing apparatus according to (5), - wherein the reference index arrangement information acquisition unit preloads design information relating to the relationship between the posture of the image display apparatus and the posture of the reference index, and
- the position/posture calculation unit calculates the position and posture of the image display apparatus in the first reference coordinate system on the basis of the design information from an image captured by the image capturing unit when the information processing apparatus is arranged in a predetermined direction with respect to the image display apparatus.
- (7) The information processing apparatus according to (5), wherein the reference index arrangement information acquisition unit receives information from the reference index.
(8) The information processing apparatus according to (7), - wherein three or more reference indexes whose arrangement information is known are arranged on the image display apparatus, and
- the reference index arrangement information acquisition unit makes discrimination between the reference indexes on the basis of at least one of color, texture, and shape and acquires axial information of the reference coordinate system of the image display apparatus.
- (9) The information processing apparatus according to (7), wherein the reference index arrangement information acquisition unit receives information encoded in a flashing pattern from the reference index.
(10) The information processing apparatus according to (9), wherein the reference index arrangement information acquisition unit makes an inquiry about information obtained by decoding the flashing pattern to a database to acquire coordinate information of the reference index, information of the reference index, or information of the image display apparatus.
(11) The information processing apparatus according to (2) or (3), wherein the position/posture calculation unit calculates the position and posture of the image display apparatus in the first reference coordinate system using visual simultaneous localization and mapping (SLAM).
(12) The information processing apparatus according to (11), - wherein position and posture of the image display apparatus calculated by the position/posture calculation unit in a reference state in which the information processing apparatus is held in an initial posture are stored, and
- the position/posture transform unit calculates a change in the posture of the image display apparatus from the reference state and transforms the position and posture of the image display apparatus in the first reference coordinate system to the position and posture of the information processing apparatus in the second reference coordinate system.
- (13) An image processing method including:
- a position/posture calculation step of calculating, on the basis of an image of another image display apparatus captured by an image capturing unit mounted on an information terminal, position and posture of the image display apparatus in a first reference coordinate system of the information terminal;
- a position/posture transform step of transforming the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information terminal in a second reference coordinate system of the image display apparatus; and
- an embedding position calculation step of calculating, on the basis of the position and posture of the information terminal in the second reference coordinate system, a position of embedding screen information of the information terminal with respect to a screen of the image display apparatus.
- (14) An image display system including:
- a display apparatus fixed to a head or face of an observer; and
- an information processing apparatus that includes an image capturing unit and a display unit, calculates, on the basis of an image of an external image display apparatus captured by the image capturing unit, position and posture of the image display apparatus in a first reference coordinate system of the information processing apparatus, transforms the position and posture of the image display apparatus in the first reference coordinate system to position and posture of the information processing apparatus in a second reference coordinate system of the image display apparatus, and calculates a position of embedding screen information of the display unit with respect to a screen of the image display apparatus.
-
- 100 Head mount display
- 101L, 101R Virtual image optical unit
- 103L, 103R Microphone
- 104L, 104R Display panel
- 105 Pupil distance adjustment mechanism
- 301 Control unit
- 301A ROM
- 301B RAM
- 302 Input operation unit
- 303 Remote-control reception unit
- 304 State information acquisition unit
- 305 Communication unit
- 306 Storage unit
- 307 Image processing unit
- 308 Display drive unit
- 309 Display unit
- 310 Virtual image optical unit
- 312 Outer camera
- 313 Audio processing unit
- 314 Audio input/output unit
- 315 Touch panel
- 316 Environment information acquisition unit
- 1610 Control unit
- 1611 CPU
- 1612 ROM
- 1613 RAM
- 1620 Display unit
- 1621 Display panel
- 1622 Display interface
- 1623 Touch panel
- 1624 Touch interface
- 1630 Audio processing unit
- 1631 Audio output unit
- 1632 Audio input unit
- 1634 Output terminal
- 1640 Communication unit
- 1650 Storage unit
- 1660 Camera unit
- 1661 Image sensor
- 1662 AFE processing unit
- 1663 Camera interface
- 1670 Sensor unit
Claims (14)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014159169 | 2014-08-05 | ||
JP2014-159169 | 2014-08-05 | ||
PCT/JP2015/062934 WO2016021252A1 (en) | 2014-08-05 | 2015-04-30 | Information processing device, information processing method, and image display system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170206673A1 true US20170206673A1 (en) | 2017-07-20 |
US10719950B2 US10719950B2 (en) | 2020-07-21 |
Family
ID=55263535
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/328,089 Active US10719950B2 (en) | 2014-08-05 | 2015-04-30 | Head mount display (HMD) operated with mobile device for transforming reference coordinate systems for providing screen information |
Country Status (3)
Country | Link |
---|---|
US (1) | US10719950B2 (en) |
JP (1) | JP6525010B2 (en) |
WO (1) | WO2016021252A1 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180107267A1 (en) * | 2015-04-17 | 2018-04-19 | Meta Technologies Inc. | Motion-sensing interactive system and motion-sensing interation method |
US20180164589A1 (en) * | 2015-05-29 | 2018-06-14 | Kyocera Corporation | Wearable device |
US20180176547A1 (en) * | 2016-12-19 | 2018-06-21 | Seiko Epson Corporation | Display apparatus and method for controlling display apparatus |
US20180181273A1 (en) * | 2013-05-15 | 2018-06-28 | Sony Corporation | Display control device, display control method, and recording medium |
EP3470966A1 (en) * | 2017-10-13 | 2019-04-17 | Nintendo Co., Ltd. | Orientation and/or position estimation system, orientation and/or position estimation method, and orientation and/or position estimation apparatus |
US20190147158A1 (en) * | 2017-11-13 | 2019-05-16 | International Business Machines Corporation | Secure password input in electronic devices |
US20190172334A1 (en) * | 2015-09-01 | 2019-06-06 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
CN109992100A (en) * | 2017-12-30 | 2019-07-09 | 深圳多哚新技术有限责任公司 | One kind wearing display system and its display methods |
US20200201542A1 (en) * | 2017-09-04 | 2020-06-25 | Wacom Co., Ltd. | Spatial position indication system |
US10835809B2 (en) | 2017-08-26 | 2020-11-17 | Kristina Contreras | Auditorium efficient tracking in auditory augmented reality |
US11153488B2 (en) * | 2019-09-26 | 2021-10-19 | United States Of America, As Represented By The Secretary Of The Army | Variable latency and frame rate camera |
US11169600B1 (en) * | 2019-12-06 | 2021-11-09 | Snap Inc. | Virtual object display interface between a wearable device and a mobile device |
US11393431B2 (en) * | 2019-02-21 | 2022-07-19 | Seiko Epson Corporation | Display system, control program for information processor, and control method for information processor that are configured to adjust display of a first image on a first display unit based on the position of a second display unit |
US20220414990A1 (en) * | 2021-06-25 | 2022-12-29 | Acer Incorporated | Augmented reality system and operation method thereof |
US20230067564A1 (en) * | 2021-08-31 | 2023-03-02 | Htc Corporation | Virtual image display system and calibration method for pointing direction of controller thereof |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6686319B2 (en) * | 2015-08-04 | 2020-04-22 | 富士通株式会社 | Image projection device and image display system |
JP6668170B2 (en) * | 2016-06-07 | 2020-03-18 | 株式会社ソニー・インタラクティブエンタテインメント | Information processing apparatus, information processing method, and computer program |
EP3545385B1 (en) * | 2016-11-25 | 2021-07-14 | Sensoryx AG | Wearable motion tracking system |
JP2020507221A (en) * | 2017-02-03 | 2020-03-05 | ベステル エレクトロニク サナイー ベ ティカレト エー.エス. | Improved method and system for video conferencing using HMD |
GB2555501B (en) * | 2017-05-04 | 2019-08-28 | Sony Interactive Entertainment Europe Ltd | Head mounted display and method |
JP7035510B2 (en) * | 2017-12-22 | 2022-03-15 | セイコーエプソン株式会社 | Display system and control method of display system |
CN110060202B (en) * | 2019-04-19 | 2021-06-08 | 湖北亿咖通科技有限公司 | Monocular SLAM algorithm initialization method and system |
JP2021077185A (en) | 2019-11-12 | 2021-05-20 | ソニー株式会社 | Information processing device, information processing method, and program |
JP7376616B2 (en) * | 2019-12-27 | 2023-11-08 | マクセル株式会社 | Head-mounted information output device |
CN113467602B (en) * | 2020-03-31 | 2024-03-19 | 中国移动通信集团浙江有限公司 | VR display method and system |
CN114827338A (en) * | 2021-01-29 | 2022-07-29 | 北京外号信息技术有限公司 | Method and electronic device for presenting virtual objects on a display medium of a device |
JP2022170264A (en) * | 2021-04-28 | 2022-11-10 | ソニーグループ株式会社 | Information processing device, information processing method, and program |
US20240244173A1 (en) * | 2021-05-17 | 2024-07-18 | Maxell, Ltd. | Head mounted display |
WO2022269753A1 (en) | 2021-06-22 | 2022-12-29 | マクセル株式会社 | Information processing system, information processing device, and image display device |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020106135A1 (en) * | 2000-06-26 | 2002-08-08 | Waro Iwane | Information converting system |
US20050068293A1 (en) * | 2003-09-30 | 2005-03-31 | Canon Kabushiki Kaisha | Data conversion method and apparatus, and orientation measurement apparatus |
US20060273984A1 (en) * | 2005-04-20 | 2006-12-07 | Canon Kabushiki Kaisha | Image processing method and image processing apparatus |
JP2011242591A (en) * | 2010-05-18 | 2011-12-01 | Fujifilm Corp | Head-mounted display device |
US20120327115A1 (en) * | 2011-06-21 | 2012-12-27 | Chhetri Amit S | Signal-enhancing Beamforming in an Augmented Reality Environment |
US20140002329A1 (en) * | 2012-06-29 | 2014-01-02 | Sony Computer Entertainment Inc. | Image processing device, image processing method, and image processing system |
US20140015736A1 (en) * | 2012-07-16 | 2014-01-16 | Lg Electronics Inc. | Head mounted display and method of outputting a content using the same in which the same identical content is displayed |
US20140168268A1 (en) * | 2011-08-24 | 2014-06-19 | Sony Corporation | Information processing device, information processing method, and program |
US20150015459A1 (en) * | 2013-07-10 | 2015-01-15 | Lg Electronics Inc. | Mobile device, head mounted display and method of controlling therefor |
US20150316767A1 (en) * | 2014-05-01 | 2015-11-05 | Michael John Ebstyne | 3d mapping with flexible camera rig |
US20160203642A1 (en) * | 2015-01-14 | 2016-07-14 | Oculus Vr, Llc | Passive locators for a virtual reality headset |
US20200057495A1 (en) * | 2015-12-17 | 2020-02-20 | Looxid Labs, Inc. | Eye-brain interface (ebi) system and method for controlling same |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005038321A (en) | 2003-07-18 | 2005-02-10 | Canon Inc | Head mount display device |
JP2010276792A (en) * | 2009-05-27 | 2010-12-09 | Ntt Docomo Inc | Image display system, image display method, and portable terminal |
JP5681850B2 (en) * | 2010-03-09 | 2015-03-11 | レノボ・イノベーションズ・リミテッド(香港) | A portable terminal using a head-mounted display as an external display device |
JP5434848B2 (en) | 2010-08-18 | 2014-03-05 | ソニー株式会社 | Display device |
JP2012141461A (en) | 2010-12-29 | 2012-07-26 | Sony Corp | Head mount display |
JP2012186660A (en) | 2011-03-06 | 2012-09-27 | Sony Corp | Head mounted display |
JP2013175208A (en) * | 2013-04-05 | 2013-09-05 | Nintendo Co Ltd | Information processing apparatus and information processing program |
-
2015
- 2015-04-30 WO PCT/JP2015/062934 patent/WO2016021252A1/en active Application Filing
- 2015-04-30 US US15/328,089 patent/US10719950B2/en active Active
- 2015-04-30 JP JP2016539870A patent/JP6525010B2/en active Active
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020106135A1 (en) * | 2000-06-26 | 2002-08-08 | Waro Iwane | Information converting system |
US20050068293A1 (en) * | 2003-09-30 | 2005-03-31 | Canon Kabushiki Kaisha | Data conversion method and apparatus, and orientation measurement apparatus |
US20060273984A1 (en) * | 2005-04-20 | 2006-12-07 | Canon Kabushiki Kaisha | Image processing method and image processing apparatus |
JP2011242591A (en) * | 2010-05-18 | 2011-12-01 | Fujifilm Corp | Head-mounted display device |
US20120327115A1 (en) * | 2011-06-21 | 2012-12-27 | Chhetri Amit S | Signal-enhancing Beamforming in an Augmented Reality Environment |
US20140168268A1 (en) * | 2011-08-24 | 2014-06-19 | Sony Corporation | Information processing device, information processing method, and program |
US20140002329A1 (en) * | 2012-06-29 | 2014-01-02 | Sony Computer Entertainment Inc. | Image processing device, image processing method, and image processing system |
US20140015736A1 (en) * | 2012-07-16 | 2014-01-16 | Lg Electronics Inc. | Head mounted display and method of outputting a content using the same in which the same identical content is displayed |
US20150015459A1 (en) * | 2013-07-10 | 2015-01-15 | Lg Electronics Inc. | Mobile device, head mounted display and method of controlling therefor |
US20150316767A1 (en) * | 2014-05-01 | 2015-11-05 | Michael John Ebstyne | 3d mapping with flexible camera rig |
US20160203642A1 (en) * | 2015-01-14 | 2016-07-14 | Oculus Vr, Llc | Passive locators for a virtual reality headset |
US20200057495A1 (en) * | 2015-12-17 | 2020-02-20 | Looxid Labs, Inc. | Eye-brain interface (ebi) system and method for controlling same |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180181273A1 (en) * | 2013-05-15 | 2018-06-28 | Sony Corporation | Display control device, display control method, and recording medium |
US20180107267A1 (en) * | 2015-04-17 | 2018-04-19 | Meta Technologies Inc. | Motion-sensing interactive system and motion-sensing interation method |
US10768690B2 (en) * | 2015-04-17 | 2020-09-08 | Meta Techologies Inc. | Motion-sensing interactive system and motion-sensing interation method |
US10591729B2 (en) * | 2015-05-29 | 2020-03-17 | Kyocera Corporation | Wearable device |
US20180164589A1 (en) * | 2015-05-29 | 2018-06-14 | Kyocera Corporation | Wearable device |
US11741811B2 (en) | 2015-09-01 | 2023-08-29 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
US11176797B2 (en) | 2015-09-01 | 2021-11-16 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
US20190172334A1 (en) * | 2015-09-01 | 2019-06-06 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
US10755545B2 (en) * | 2015-09-01 | 2020-08-25 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
US10785472B2 (en) * | 2016-12-19 | 2020-09-22 | Seiko Epson Corporation | Display apparatus and method for controlling display apparatus |
US20180176547A1 (en) * | 2016-12-19 | 2018-06-21 | Seiko Epson Corporation | Display apparatus and method for controlling display apparatus |
US11310483B2 (en) | 2016-12-19 | 2022-04-19 | Seiko Epson Corporation | Display apparatus and method for controlling display apparatus |
US10835809B2 (en) | 2017-08-26 | 2020-11-17 | Kristina Contreras | Auditorium efficient tracking in auditory augmented reality |
US11630570B2 (en) * | 2017-09-04 | 2023-04-18 | Wacom Co., Ltd. | Spatial position indication system |
US20200201542A1 (en) * | 2017-09-04 | 2020-06-25 | Wacom Co., Ltd. | Spatial position indication system |
US11272152B2 (en) * | 2017-10-13 | 2022-03-08 | Nintendo Co., Ltd. | Orientation and/or position estimation system, orientation and/or position estimation method, and orientation and/or position estimation apparatus |
EP3470966A1 (en) * | 2017-10-13 | 2019-04-17 | Nintendo Co., Ltd. | Orientation and/or position estimation system, orientation and/or position estimation method, and orientation and/or position estimation apparatus |
US20190116350A1 (en) * | 2017-10-13 | 2019-04-18 | Nintendo Co., Ltd. | Orientation and/or position estimation system, orientation and/or position estimation method, and orientation and/or position estimation apparatus |
US20190147158A1 (en) * | 2017-11-13 | 2019-05-16 | International Business Machines Corporation | Secure password input in electronic devices |
US10776475B2 (en) * | 2017-11-13 | 2020-09-15 | International Business Machines Corporation | Secure password input in electronic devices |
CN109992100A (en) * | 2017-12-30 | 2019-07-09 | 深圳多哚新技术有限责任公司 | One kind wearing display system and its display methods |
US11393431B2 (en) * | 2019-02-21 | 2022-07-19 | Seiko Epson Corporation | Display system, control program for information processor, and control method for information processor that are configured to adjust display of a first image on a first display unit based on the position of a second display unit |
US11153488B2 (en) * | 2019-09-26 | 2021-10-19 | United States Of America, As Represented By The Secretary Of The Army | Variable latency and frame rate camera |
US20220043508A1 (en) * | 2019-12-06 | 2022-02-10 | Ilteris Canberk | Virtual object display interface between a wearable device and a mobile device |
US11604508B2 (en) * | 2019-12-06 | 2023-03-14 | Snap Inc. | Virtual object display interface between a wearable device and a mobile device |
US11169600B1 (en) * | 2019-12-06 | 2021-11-09 | Snap Inc. | Virtual object display interface between a wearable device and a mobile device |
US11886633B2 (en) | 2019-12-06 | 2024-01-30 | Snap Inc. | Virtual object display interface between a wearable device and a mobile device |
US20220414990A1 (en) * | 2021-06-25 | 2022-12-29 | Acer Incorporated | Augmented reality system and operation method thereof |
US20230067564A1 (en) * | 2021-08-31 | 2023-03-02 | Htc Corporation | Virtual image display system and calibration method for pointing direction of controller thereof |
US12093470B2 (en) * | 2021-08-31 | 2024-09-17 | Htc Corporation | Virtual image display system and calibration method for pointing direction of controller thereof |
Also Published As
Publication number | Publication date |
---|---|
WO2016021252A1 (en) | 2016-02-11 |
JPWO2016021252A1 (en) | 2017-05-18 |
JP6525010B2 (en) | 2019-06-05 |
US10719950B2 (en) | 2020-07-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10719950B2 (en) | Head mount display (HMD) operated with mobile device for transforming reference coordinate systems for providing screen information | |
US20240296633A1 (en) | Augmented reality experiences using speech and text captions | |
CN111052043B (en) | Controlling external devices using a real-world interface | |
EP3098689B1 (en) | Image display device and image display method | |
US11127210B2 (en) | Touch and social cues as inputs into a computer | |
WO2016013269A1 (en) | Image display device, image display method, and computer program | |
JP6217747B2 (en) | Information processing apparatus and information processing method | |
US9753285B2 (en) | Information processing device, notification state control method, and program | |
US11869156B2 (en) | Augmented reality eyewear with speech bubbles and translation | |
US20150355463A1 (en) | Image display apparatus, image display method, and image display system | |
US20130174213A1 (en) | Implicit sharing and privacy control through physical behaviors using sensor-rich devices | |
US20160054565A1 (en) | Information processing device, presentation state control method, and program | |
EP3676745B1 (en) | Privacy screen for computer simulated reality | |
US11195341B1 (en) | Augmented reality eyewear with 3D costumes | |
JP6822410B2 (en) | Information processing system and information processing method | |
US10482670B2 (en) | Method for reproducing object in 3D scene and virtual reality head-mounted device | |
US20240256052A1 (en) | User interactions with remote devices | |
US20210406542A1 (en) | Augmented reality eyewear with mood sharing | |
JP2011152593A (en) | Robot operation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KAWAMOTO, KENTA;REEL/FRAME:041061/0037 Effective date: 20161124 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |