[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20150015611A1 - Method for representing virtual information in a real environment - Google Patents

Method for representing virtual information in a real environment Download PDF

Info

Publication number
US20150015611A1
US20150015611A1 US14/502,366 US201414502366A US2015015611A1 US 20150015611 A1 US20150015611 A1 US 20150015611A1 US 201414502366 A US201414502366 A US 201414502366A US 2015015611 A1 US2015015611 A1 US 2015015611A1
Authority
US
United States
Prior art keywords
region
real environment
virtual information
view
item
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/502,366
Inventor
Peter Meier
Frank Angermann
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Apple Inc
Original Assignee
Metaio GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Metaio GmbH filed Critical Metaio GmbH
Priority to US14/502,366 priority Critical patent/US20150015611A1/en
Publication of US20150015611A1 publication Critical patent/US20150015611A1/en
Assigned to APPLE INC. reassignment APPLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: METAIO GMBH
Priority to US16/716,169 priority patent/US20200193712A1/en
Priority to US16/843,462 priority patent/US11562540B2/en
Priority to US18/153,746 priority patent/US20230290083A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/147Digital output to display device ; Cooperation and interconnection of the display device with other functional units using display panels
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • G06T15/20Perspective computation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C23/00Combined instruments indicating more than one navigational value, e.g. for aircraft; Combined measuring devices for measuring two or more variables of movement, e.g. distance, speed or acceleration
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2354/00Aspects of interface with display user

Definitions

  • the present invention relates to a method for representing virtual information in a real environment.
  • the method is particularly suitable for ergonomic representation and marking of points of interest in the world by means of augmented reality technology.
  • Augmented Reality is a technology in which virtual data are overlaid with reality and which thus facilitates the association of data with reality.
  • the use of mobile AR systems is already known in the prior art.
  • high-performance mobile devices e.g. smartphones
  • These devices meanwhile have comparatively large color displays, installed cameras, good processors and additional sensors, such as e.g. orientation sensors and GPS.
  • additional sensors such as e.g. orientation sensors and GPS.
  • the position of the device can be approximated via radio networks.
  • a first aspect of the invention relates to a method for representing virtual information in a real environment, comprising the following steps: providing at least one view of a real environment and of a system setup for blending in or overlaying virtual information for superimposing with the real environment in at least part of the view, the system setup comprising at least one display device, determining a position and orientation of at least one part of the system setup relative to at least one component of the real environment, subdividing at least part of the view of the real environment into a plurality of regions comprising a first region and a second region, with objects of the real environment within the first region being placed closer to the system setup than objects of the real environment within the second region, blending in or overlaying at least one item of virtual information on the display device in at least part of the view of the real environment, considering the position and orientation of said at least one part of the system setup, wherein the virtual information is shown differently in the first region than in the second region with respect to the type of blending in or overlay in the view of the real environment.
  • the at least one part of the system setup may be a camera, for example, the position and orientation (pose) of which are determined, with the camera being not necessarily fixedly connected to the display device.
  • the camera being not necessarily fixedly connected to the display device.
  • there is no camera necessary at all for the overall system e.g. when the pose of at least one part of the system setup is determined via GPS and orientation sensors only.
  • the pose determination of any part of the system setup is suitable as a matter of principle, provided there are conclusions possible as to the viewing direction of the user.
  • the first region can be a near region, whereas the second region can be a far or remote region, for example. However, it is also possible that the first region represents a location region, whereas the second region represents a near region.
  • An embodiment with near region, far region and location region will be elucidated in more detail hereinafter with reference to the drawings.
  • the at least one item of virtual information represents a point of interest (generally also abbreviated as POI, especially in connection with navigation devices) with respect to the real environment.
  • a point of interest generally also abbreviated as POI, especially in connection with navigation devices
  • the invention offers the advantage that an ergonomic representation of virtual information, in particular of points of interest, in the real world is obtained without the field of view of the user being excessively restricted and without overstraining the user with too many items of information. At the same time, there can be displayed many different items of virtual information which, however, due to the ergonomic representation of the virtual information, does not result in overstraining the user. Moreover, the association of information can be enhanced by consideration of the human perception mechanisms.
  • FIG. 1 shows a plan view of a schematic arrangement of exemplary system setups with respect to a real environment that can be used for performing a method according to the invention.
  • FIG. 2 shows a schematic arrangement of an exemplary division of points of interest (surrounding POIs) to different radius regions.
  • FIG. 3 shows a schematic arrangement illustrating a possible approach for calculating the radii over POIs in sections, in which the radii per sector may differ from each other.
  • FIG. 4 shows possible elements of an exemplary view of the real environment with blended in virtual information (here: POI objects) for forming an embodiment of a user interface according to aspects of the invention.
  • POI objects blended in virtual information
  • FIG. 5 shows possible elements of the user interface according to FIG. 4 when the focus of operation is located in the location region.
  • FIG. 6 shows an exemplary preview for virtual information (here: POI) in the location region of the view according to FIG. 4 .
  • FIG. 7 shows an exemplary possibility of shielding surface elements from the view according to FIG. 4 , when these are not required.
  • FIGS. 8 to 11 show different representation possibilities for embodiments of items of virtual information (here: POI objects).
  • FIG. 12 shows a flowchart giving an overview of the overall process of an embodiment of a method according to the invention.
  • FIG. 13 shows an embodiment illustrating a possible course for calculating radius 0 (cp. FIG. 2 ) of the location region, as exemplified in the view according to FIG. 4 ,
  • FIG. 14 shows an embodiment illustrating a possible course for calculating radius 1 as exemplified in FIG. 2 .
  • FIG. 15 shows an embodiment illustrating a possible course for calculating radius 2 as exemplified in FIG. 2 .
  • FIG. 16 shows a schematic view for illustrating exemplary possibilities for the aggregation of items of virtual information (here: POI objects).
  • FIG. 17 shows an exemplary possible course for associating shadows and material effects (“shading”) with the items of virtual information (here: POI objects).
  • FIG. 18 shows exemplary possible interactions with items of virtual information (here: POI objects).
  • FIG. 1 shows a plan view illustrating a schematic arrangement of exemplary system setups with respect to a real environment, which can be used for performing a method according to the invention.
  • FIG. 1 illustrates various possibilities of a system setup.
  • the user wears, as display device, a head mounted display system (abbreviated to HMD) comprising a display 21 that is part of the system setup 20 .
  • the display 21 e.g. may be generally known semi-transparent data glasses (“optical see-through display”) in which virtual information provided by a computer 23 can be blended in.
  • the user then sees through the semi-permeable data glasses 21 , in a view of the real world 40 , objects of the real world augmented with blended in virtual information (such as e.g. POI objects related to the real world).
  • the system setup 20 constitutes an embodiment of a generally known Augmented Reality (AR) system.
  • AR Augmented Reality
  • the display 21 may have additional sensors 24 , such as rotation sensors, and a camera 22 for optical tracking mounted thereon.
  • Display 21 can be semi-transparent or may be fed with images of the reality by a camera image. With a semi-transparent display 21 , calibration between eye 25 and display 21 is necessary. To this end, there are various methods documented in the prior art and known to the expert.
  • position sensors such as e.g. GPS sensors (GPS: Global Positioning System) for rendering possible geographic position determination of the system setup 20 (e.g. in accordance with longitude and latitude) in the real world 40 .
  • GPS Global Positioning System
  • FIG. 1B there is shown another exemplary system setup 30 that can be found often e.g. in modern mobile telephones (so-called “smartphones”).
  • the display device 31 e.g. in the form of a display screen or display
  • computer 33 e.g. in the form of a display screen or display
  • sensors 34 and camera 32 constitute a system unit that is accommodated e.g. in a common housing of a mobile telephone.
  • the view of the real environment 40 is provided by display 31 showing a camera image of the real environment 40 captured by camera 32 .
  • the camera image can be shown on display 31 and augmented with additional virtual information (such as POI objects related to the real world).
  • the system setup 30 constitutes another embodiment of a generally known augmented reality (AR) system.
  • AR augmented reality
  • the present invention can be used expediently for all forms of AR.
  • the representation is implemented in the so-called optical see-through mode with semi-transparent HMD or in the video see-through mode with camera and display screen.
  • this may be a camera or the optical system resulting from the combination of eye and see-through display (cp. display 21 of FIG. 1A ). Both thereof comprise relevant camera properties for blending in virtual information, such as aperture angle and main point of image).
  • the invention basically can also be used in connection with stereoscopic displays, in which the video see-through approach advantageously uses two cameras each for recording one video stream per eye.
  • the items of virtual 3D information can be calculated individually for each eye.
  • the processing of the different partial steps described hereinafter basically can be distributed to various computers via a network.
  • a client/sever architecture or a pure client-based solution is possible.
  • the client could send an image to a server which, on the basis of the image, makes available to the client information on the 3D position and 3D orientation of the system setup (cp. FIG. 1 ) or part thereof in relation to the real world (referred to as pose in the following) and on the range of vision.
  • the client or the server may also comprise several computing units, such as several CPUs or specialized hardware components, such as generally known FPGAs, ASICs, GPUs or DSPs.
  • Several clients may also exchange information among each other which e.g. is generated with respect to the range of vision at this location, or in case a client generates a POI. This exchange of information can take place via a server, however, direct connections via Bluetooth or WLAN would be conceivable as well.
  • the pose (position and orientation) of the camera in space is necessary. This can be realized in variety of different ways. It is possible to determine the pose in the world e.g. by using merely GPS and an orientation sensor with electronic compass (as installed e.g. in some modem mobile telephones). However, the uncertainty of the pose then is very high. Thus, it is also possible to use other methods, such as e.g. optical initialization and tracking or the combination of optical methods using GPS and orientation sensors. WLAN locating can be used as well or RFIDs or optical markers can support the locating process. As mentioned hereinbefore. a client/server-based approach is possible here as well. In particular, the client can request from the server location-specific information needed for optical tracking. Such information may be e.g. reference images of the surrounding environment with pose information and depth information.
  • the invention can improve the information representation for the client. However, it can also be utilized in a remote scenario (remote view scenario).
  • a remote scenario e.g. a maintenance expert in a control room views the client's image transferred via the data network as well as the correspondingly processed items of information on his display screen. The expert then could give instructions to the client or just observe.
  • a person views recorded image or video material along with interactive additional information represented according to the invention, and, if possible, can navigate through the material similar to the Internet-based application “Google Streetview”.
  • the invention can also be installed, or carried along, in vehicles, aircraft or ships as a monitor, HMD or by way of a head-up display.
  • a point of interest can be set up for a large variety of different forms of information. Examples are given hereinafter. It is possible to represent images of places using GPS information. It is possible to automatically extract information from the Internet. For example, this may be company or restaurant websites with addresses or pages giving ratings. Users can deposit texts, images or 3D objects at locations and make the same available to others. Information pages, such as Wikipedia, can be searched for geo-information, and the pages can be made accessible as POI. POIs can be generated automatically from the search and browsing behavior of the users of mobile devices. It is possible to show other locations of interest, such as underground transportation or bus stations, hospitals, police stations, physicians, real estate ads or fitness clubs.
  • FIG. 2 shows a schematic arrangement illustrating an exemplary division of points of interests (surrounding POIs) to different radii regions
  • FIG. 3 shows a schematic arrangement illustrating a possible approach for calculating the radii over POIs in sectors, with the possibility that the radii per sector may be different
  • FIG. 4 shows possible elements of an exemplary view of the real environment with blended in items of virtual information (here: POI objects) for creating an embodiment of a user interface in accordance with aspects of the invention.
  • POI objects items of virtual information
  • At least a part of the view 1 of the real environment (which may be the view through a semi-transparent HMD 21 or on a screen HMD 21 according to FIG. 1A or a screen view of a camera image on display 31 according to FIG. 1B ) is subdivided into a plurality of regions.
  • these regions comprise a first region 3 , constituting a near region 3 in the present embodiment, a second region 4 , constituting a far region 4 in the present embodiment, and a third region 5 constituting a location region 5 in the present embodiment.
  • Objects of the real environment 40 within near region 3 , are placed closer to the system setup 20 and 30 , respectively, than objects of the real environment within far region 4 .
  • objects of the real environment 40 are placed closer to system setup 20 and 30 , respectively, than objects of the real environment within near region 3 .
  • the view 1 according to FIG. 4 has at least one item of virtual information blended in (in the present embodiment, several items of virtual information POI 1 -POI 4 , POI 11 -POI 14 and POI 21 -POI 23 are blended in). In the present embodiment these represent points of interest (POIs) in relation to the reality that can be seen in view 1 .
  • virtual information POI 1 is blended in relating to a building visible in view 1 (not shown in FIG. 4 ) that makes reference to this building and/or is optically associated with the same and possibly permits additional information to be retrieved with respect to the building.
  • the information POI 1 is blended in on the display device 21 and 31 , respectively, in at least part of the view 1 of the real environment, considering the position and orientation of the corresponding system setup 20 or 30 or part thereof, respectively.
  • the blended in virtual information is shown differently in the near region 3 than in the far region 4 , with respect to type of blending in view 1 .
  • the same virtual information (e.g. POI 1 ) in the near region 3 e.g. when it is associated with a real object in the near region 3
  • is shown differently from its blending in the far region 4 e.g. when it is associated with a real object in the far region 4 (a POI is present here at the same time in one region only).
  • the near region 3 is separated from the far region 4 and the location region 5 by a respective boundary which in the present embodiment each have a radius (radius 0 to radius 3) in the direction of a viewing angle of the display device ( FIG. 2 in addition shows further boundaries or radii, as will still be explained in more detail hereinafter).
  • the determination and calculation of the radii according to FIGS. 2 and 3 will still be described in more detail in the following.
  • the boundaries do not have a fixed shape, but preferably are rotation-symmetric with respect to the camera and viewer, respectively.
  • the boundaries in addition do not need to be circular, but may also have an e.g. elliptical or other shape.
  • FIG. 12 shows in a flowchart an overview of the process of an embodiment of a method according to the invention in conjunction with FIGS. 1 to 11 .
  • a first step 1 . 0 the system is initialized or, in case a pass has already taken place, prior values can be retained optionally.
  • data are loaded if the same are not yet present or if the situation has changed.
  • the data source may be contained in more or more servers or can be stored locally on the device or can be coded as information in the environment (e.g. via RFID or QR codes).
  • the number can be restricted via the position and the determination of a radius 3.
  • Such functionality is offered e.g.
  • POIs may already be filtered out in this step as well by way of filters defined by the user himself. For example, gastronomy only or no gastronomy. POIs carrying corresponding information are sorted out then. It is apparent to the expert here that a corresponding data structure can be deposited for a POI, which may contain e.g. its position, its 2D or 3D shape, an image and/or additional meta information, such as e.g. the category.
  • steps 3 . 0 - 6 . 0 the POIs then are organized according to the invention. Thereafter, they can be processed optically and displayed. Following this, interaction with the POIs shown is possible, as will be explained in more detail by way of an example.
  • FIGS. 13 to 18 relate to possible sub-steps of the general method steps illustrated in FIG. 12 . These steps basically can be implemented per pass or in the background, in certain intervals, depending on the capacity of the system.
  • a fundamental aspect of the invention consists in that items of blended in virtual information (here: POIs) will be assigned to fixedly configured or automatically calculated, different regions of a view of the real world.
  • FIG. 2 shows a view of a possible scenario from a bird's eye perspective.
  • Various POIs are shown in a spatial arrangement in the surroundings of the camera. If all of these POIs were simply displayed, the screen rapidly would be overfilled with numerous objects. For this reason, a subdivision is made to different regions.
  • the location region 5 is confined by radius 0. As regards the radii, it is to be noted that they are not necessarily constant over all angles, but may vary in accordance with the angle (cp. FIG. 3 ).
  • the location region 5 shows objects located in close proximity, so that it is difficult for the user to find the same with the device. The possibility of finding them is dependent in particular on the accuracy of the ascertainable pose. Accordingly, radius 0 may be set greater or smaller advantageously in accordance with the (uncertainty of the) pose determination ( FIG. 13 ).
  • FIG. 4 shows possible elements of an exemplary view of the real environment with blended in virtual information (here: POI objects) for forming an embodiment of a user interface according to aspects of the invention.
  • FIG. 4 in this regard shows an advantageous subdivision of the user interface.
  • POIs in the location region 5 are fixedly arranged in the lower portion of view 1 and can be selected e.g. by mouse click or touching in the event of a touchscreen.
  • FIG. 6 shows a possible view as a result of activation of an exemplary POI 23 . If provided, an image P 23 of the exact location of POI 23 can be displayed for orientation. If provided, a preview V 23 of the deposited information can be displayed in addition.
  • POI objects present outside the location region 5 are arranged in near region 3 .
  • the correct association of virtual information with reality is deemed to be particularly important. For example, it may be relevant to the personal planning of the path whether a POI object is present before or behind a road crossing.
  • Near region 3 is confined by radii 0 and 1 ( FIG. 2 ).
  • FIG. 14 illustrates a possibility for calculating radius 1 .
  • the expedient visibility of POI objects is of relevance. If the representation of a POI still would have a size of 2 mm only (exemplary value), it would not be of much use, but cause just disorder on the display.
  • the calculation in step 4 . 1 thus incorporates the resolution of the display, the size of the POI (the representation e.g. of a dinosaur permits a larger radius 1 than the representation of a sphere with a size of 1 m) and the aperture angle of the camera.
  • radius 1 can be matched further if there are too many POIs filling the screen (referred to as global approach).
  • the calculation can refer to the entire aperture angle of the camera or an even larger range. Smaller sectors can be adjusted individually as well.
  • the sectors may be fixed, for example, or can be formed on the basis of clusters, with clusters being of assistance to divide sectors (cp. step 6 . 2 in FIG. 16 ).
  • An exemplary method could be implemented as follows: the center of gravity is calculated for each cluster. A straight line is drawn through this center. Straight lines arranged at a small angle from each other are combined. Each resulting straight line is aligned with integral angles, and a sector is assigned to the same. The size of the sectors then is rounded an increased iteratively in integral angle steps until contact is established with a neighboring sector.
  • the sectors are defined by clusters or by fixed settings, the number of POIs of a sector is counted in step 4 . 2 . If the number exceeds a specific configurable threshold value, the radius is reduced until the value is below the threshold value (step 4 . 3 ).
  • step 4 . 4 the 2D position and extension of the POIs, in accordance with the rendering pipeline, is calculated.
  • the rendering pipeline is understood by the expert as the projection of objects in 3D space to a smaller-dimension display. In the near region, this is effected e.g. in accordance with position, orientation, camera aperture angle and resolution of the 2D display portion available for the near region. For the far region, the rendering pipeline performs projection to a line. This can be effected e.g.
  • the rear POI is marked for the far region 4 (step 4 . 5 ).
  • This marking may be retained for one display pass (step 1 . 0 to 9 . 0 of FIG. 12 ) or longer (e.g. 3 second or 5 display passes).
  • the system can perform steps 4 . 4 and 4 . 5 until there are no more overlappings present or these are below a specific value.
  • There may also be a tolerance as to when overlapping is rated as such e.g. more than 10% of the sum of the area of both objects intersecting each other).
  • the far region 4 contains POIs that were displaced from the near region or are located outside of radius 1 in the sector. But even here, it is not necessarily expedient to display all existing POI objects. Thus, it may be advantageous to calculate a radius 2, similar to radius 1, as shown in FIG. 15 . Due to the fact that radius 2 in particular still is to display POI objects located within the reach of the user, it is advantageously possible to employ for calculating radius 2 the current speed or the average speed of the user or the distance that can be covered by the user within a specific period of time, using public transport, a vehicle or bicycle or the like. The possibility of calculating, in accordance with these factors, which objects can be displayed at all by AR, can also be applied independently of the other claims of the invention.
  • FIG. 4 shows a possible exemplary form of representation for POIs in different regions.
  • POIs in the location region 5 below are represented in a uniform size and non-movably and for time being independently of the viewing direction.
  • POI objects in near region 3 are overlaid on the environment 40 three-dimensionally and with correct perspective in accordance with the pose, the camera parameters and the model properties.
  • POI objects in far region 4 advantageously are represented in a uniform size and advantageously move along in accordance with the orientation of the device (in particular the display device) such that a vertical line in downward direction would hit the corresponding associated real location. It is advantageously possible to additionally display, on the right and left margins 6 , symbols hinting at POI objects outside the aperture angle of the camera.
  • An arrow 8 can give a directional hint so that, when the device is moved around the viewer in the direction of the arrow, the POI object would enter the visible range.
  • These margins 6 may exist for every region ( FIG. 4 ) or e.g. for the near region 3 only ( FIG. 5 ).
  • Near region 3 , far region 4 and/or location region 5 can be hidden or shielded optionally, especially when there is no virtual information displayed therein. This can also be effected automatically when there is no POI present in these regions ( FIG. 7 ).
  • a point of interest can be represented by a representation, an image, a 3D object or the like.
  • a 3D rendering technology such as the known methods OpenGL or DirectX
  • 3D objects the rendering of so-called billboards (2D objects that always face the viewer) or so-called 2D overlays with 2D rendering techniques the projection of which is calculated autonomously.
  • the kind of representation may be based on the category of the POIs (e.g. a globe for representing a website) or can be determined by the user (e.g. placing a dinosaur with additional information).
  • the POIs can be shown in a contrasting color.
  • POIs can be provided with additional items of information detailed to different extents (exemplified in the following by way of POI 1 ).
  • the lowest detail level is the POI only.
  • labels cp. label L 1
  • the next stage is the so-called preview (preview cp. V 1 or image P 1 in FIG. 11 ) (e.g. an image or rendering of the website or an infotext).
  • preview preview cp. V 1 or image P 1 in FIG. 11
  • Some POI objects may be visualized still more closely in a subsequent stage (cp. information I 1 ).
  • an own program of the operating system such as starting of an Internet browser or a media player.
  • the POIs can be displayed simultaneously with their additional information, or upon activation only.
  • it can also be determined which POI object is located closest, and only the foremost POI objects indicate firstly labels as of a specific distance, and automatically indicate previews when coming closer.
  • the display can also be controlled by so-called eye tracking.
  • the POIs viewed by the user are displayed with additional information.
  • the additional information advantageously can be anchored in the view in non-movable manner and can be connected to the movable POI representation by way of a dynamic connection. This provides for enhanced legibility of the information.
  • eye tracking for activating additional information via a POI in AR can also be applied independently of the method according to claim 1 .
  • the following method steps basically can also be implemented independently of the idea described so far, of dividing the view of the real environment into a plurality of regions (such as near, far and location regions) and to represent the blended in virtual information differently in accordance with the respective region, as described hereinbefore.
  • the standard POI objects moreover have a fixed size to permit a continuous estimation of the distance to the same.
  • Various degrees of complexity are conceivable in this regard for processing.
  • On the basis of the weather situation e.g. in accordance with the Google weather service “mostly cloudy”, “occasional stormy”, “occasional rain” etc.) it is possible to assign fixed illumination models or materials (e.g. textures) matching the weather situation.
  • the easiest technical realization in addition to further known methods, consists in adjusting the fog settings in OpenGL.
  • the visibility distance for reality-conforming representation of the distance
  • this can be realized e.g. by a second rendering pass considering specific materials only, but displaying no fog.
  • the position of the sun and the moon respectively, can be calculated on the basis of the position, the date and the time, and can be used for adjusting the light sources. This has an effect in particular on the shadows (cp.
  • the shadows can be pre-calculated textures (advantageously with a transparency value) which, in accordance with the position of the sun or the moon, are located below the POI on ground level where the straight line between sun or moon and POI intersects the ground level (exceptions, if this is not so). In case sun and moon are visible at the same time, the sun will be used for calculation.
  • the shadows may also be calculated dynamically, as known in the prior art.
  • this can include the mutual shading of POIs.
  • a 3D model of the environment exists (cf. also step 7 . 4 in FIG. 17 , e.g. in the form of a model for occluding a real object; so-called “occlusion geometry”)
  • this can be used in addition for realistic calculation of the shadow situation e.g. in that it casts shadows on POIs (cf. also FIG. 9 ).
  • the reality degree of the blending in can be enhanced in addition by augmenting the materials by images of the environment of the POI.
  • the use of so-called environment maps is known to the expert. Novel is that these are dynamically extracted in location-specific manner e.g. from Google Streetview and considered.
  • step 7 . 4 there is taken a further step for enhancing the depth perception of the viewer.
  • occlusion geometry so-called “occlusion geometry”
  • POIs in case of occlusion of the same, can be shifted directly to the far region 4 or can be marked specifically.
  • the part occluded may be shown semi-transparent, in broken lines or in a different color.
  • the shadows are not calculated and displayed either.
  • the depth model can be deposited or generated dynamically by meas of SLAM algorithms, stereo cameras or a time-of-flight camera. In that event, depth information per image pixel is sufficient.
  • step 7 is taken a further step for enhancing the depth perception of the viewer.
  • the camera parameters are generated for correct superimposition or overlaying in the near region (this does not have to take place continuously). These parameters can be generated dynamically e.g by means of a SLAM mechanism or retrieved from the server in accordance with the device name or can be deposited in the program. In case of a see-through HMD or HUD, the results of see-through calibration or dynamic measuring of the position of the eye in relation to the display are used.
  • the camera image advantageously can be processed such that only the most important image components are shown in contrasting manner. This makes sense in the video see-through mode only, and should assist users in particular in case of very bright surrounding light. Processing of the video image can take place e.g. by means of the Sobel operator for edge extraction. This mode advantageously can be turned on and off in case of bright outside light, e.g. when the device contains a brightness sensor.
  • the parameters of the pose are made available to the rendering system in addition, if this was not already necessary for the calculations in steps 3 to 6 of FIG. 12 . Depending on the hardware properties of the system, everything can be displayed and calculated in step 7 . 8 .
  • step 7 . 8 B If the hardware of the system is weak, calculation of the correct material surfaces can also be effected on the part of the server (step 7 . 8 B), or the entire image can be calculated on the part of the server.
  • a modern GPU Graphic Processor Unit
  • step 7 . 8 A There are numerous possibilities known to the expert in this regard.
  • step 4 . 0 or after step 5 . 0 it is also possible to combine POIs, as shown in FIG. 16 , at the lower left.
  • the points of a cluster constitute a new, differently shaped POI. This can be effected individually in near region 3 , in location region 5 and in far region 4 , or can be effected already before the subdivision.
  • this can be effected with POIs of one category only (e.g. websites only). POIs in the most forward region advantageously can be excluded from this process.
  • FIG. 18 shows interaction possibilities with POIs according to aspects of the method of the invention.
  • the user can alter the segment of the items of virtual information and reality already by changing his location or the viewing direction. This can also trigger an automatic interaction, as already described in “representation of objects”.
  • the device is equipped with a touchscreen or a kind of mouse control (e.g. a trackball) it is possible in principle to address POIs directly.
  • the individual POIs now are arranged selectably, and the labels are displayed (step 9 . 8 ). Otherwise, the label is displayed if the same is not yet displayed (step 9 . 9 ).
  • the POI step 9 .
  • the infotext or preview is displayed if it is not yet displayed (step 9 . 11 ).
  • the detailed display is triggered for POIs having corresponding items of information deposited, step 9 . 13 (e.g. playing a movie, music or showing a website).
  • the current display can be closed. Alternatively, it could possibly be retained in operation in the background.
  • the user optionally can also provide a setting to the effect that the next audio or video file deposited is executed and run in the background.
  • the interaction can also be implemented by means of the target cross 7 (optional) shown in FIG. 7 , as described in FIG. 18 in steps 9 . 5 to 9 . 7 .
  • the target cross 7 optionally shown in FIG. 7 , as described in FIG. 18 in steps 9 . 5 to 9 . 7 .
  • the system can also activate the POI on the user interface that is closest to the crosshairs. By prolonged aiming at a POI or pressing a button, the next detail level can be activated.
  • the user For reaching the far region 4 , the user directs the camera upwardly until the highest point POI in the near region 3 has been exceeded by a threshold value (e.g. 10 degrees). The user then can navigate between the POIs by turning right and left. In similar manner, by directing the camera downwardly, the user can reach the location region 5 . By turning right and left, the user here too can navigate through the otherwise fixed POIs.
  • a special feature in the location region 5 resides in that a POI, if indicated or adapted to be generated automatically, in addition to POI detail information indicates an image of the location in order to facilitate finding the same (cp. preview V 23 and image P 23 for POI 23 in FIG. 6 ). The image can be introduced manually or can be generated automatically by means of an image database containing pose information.
  • the selection of a POI optionally can be effected by voice control as well.
  • the user activates the control by a specific sentence (e.g. “activate voice control”).
  • Each POI is then marked with a number. By pronouncing the number, the corresponding POI can then be activated.
  • the virtual information (e.g. POIs) can be blended in the view of the real environment in the far region 4 in a uniform size.
  • the virtual information in location region 5 is represented differently from its type of blending in the near region.
  • the virtual information can be overlaid in the view of the real environment in location region 5 in non-movable manner, in particular also with uniform size and/or irrespective of the orientation of the display device.
  • the several items of virtual information can be selected from a group of items of virtual information which together form a cluster, with the virtual object having a different shape as compared to the items of virtual information.
  • a geographic map and/or a bird's eye perspective view can be blended in the display device when the display device is held approximately horizontally with respect to the earth's surface or is held by more than a specific angle below the lowest POI in the near region or, in case this function is already taken for reaching the location region, is held still further below.
  • a respective boundary, in particular the radius thereof, between the regions can be altered when virtual information is transferred by the user from one of the regions, e.g. the near region, to another region, e.g. the far region, and/or vice versa.
  • the boundary in particular the radius therof, can also be calculated in accordance with the number of items of virtual information within a specific sector of the view.
  • the boundary, in particular the radius thereof can also be calculated in accordance with the two-dimensional density of several items of virtual information within a specific sector.
  • the boundary, in particular the radius thereof can also be calculated in accordance with several items of virtual information which together constitute a cluster.
  • a shadow can be indicated in the display device below the information, near a ground level shown in the display device, said shadow corresponding to the position of the virtual information.
  • the display device used may be a video display device in which the view of the real environment is augmented or replaced by an edge image for enhanced contrast.
  • the user may be given acoustic and/or haptic feedback at an input device used for making a selection.
  • depth information with respect to at least one real object contained in the view can be calculated or loaded, and the depth information can be used for blending in an occlusion model for occluding a real object in the display device in case virtual information is occluded in the view by the real object to be occluded.
  • Such depth information can also be used for calculating a boundary between regions, e.g. between near region and far region.
  • several items of blended in virtual information in the display device may have a respective number assigned thereto, and the corresponding virtual information can be selected by voice recognition of the number or selection of the number on a keyboard or a touch-sensitive input pad.
  • several items of virtual information can be associated with one of several categories each, and the items of virtual information can be blended in and/or masked in accordance with the category.
  • an edge 6 in the display device which indicates a range or coverage of the near region 3 , with a boundary of the near region being variable by a user action, in particular by dragging the boundary.
  • the virtual information can be represented in at least three stages.
  • a first stage comprises a body (e.g. a 2D body or a 3D body) as a just local hint to the virtual information (cp. POI 1 in FIGS. 10 , 11 )
  • a second stage comprises a hint to the virtual information in the form a label with inscription (cp. label L 1 in FIG. 11 )
  • a third stage comprises an extract-like preview of the virtual information (cp. preview V 1 and image P 1 in FIGS. 10 , 11 ) which is blended in particular when the dot-like representation or the label is selected by a user action.
  • a fourth stage shows the information in full length (cp. information I 1 from FIG. 10 ).
  • a development may provide that, in a first part of the far region 4 , virtual information is represented in the first stage only, in a second part of far region 4 with real objects placed closer to the display device than in the first part of the far region and in a first part of near region 3 , virtual information is represented in the second stage, and in a second part of the near region 3 with real objects placed closer to the display device than in the first part of the near region, the virtual information is represented in the third stage.
  • a boundary between near region and far region is calculated in accordance with a size of the blended in virtual information, a resolution of the display device and/or a resolution of a camera used for generating the view.
  • the limit of the location region can be increased with increased measurement uncertainty of position detection.
  • the boundary which in particular may be a radius and determines which objects are to be displayed at all, may be dependent on the current speed or the average speed of the user or the distance that can be covered by the user within a specific period of time, using public transport, a vehicle or bicycle or the like.
  • the system can superimpose virtual information on reality, and in blending in virtual information in a view of a real environment, weather data can be considered that are retrieved e.g. via the Internet (“online”) in order to increase the degree of reality of blended in virtual information with respect to the real environment and thereby improve the association therewith.
  • weather data can be considered that are retrieved e.g. via the Internet (“online”) in order to increase the degree of reality of blended in virtual information with respect to the real environment and thereby improve the association therewith.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Geometry (AREA)
  • Computing Systems (AREA)
  • Processing Or Creating Images (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention relates to a method for ergonomically representing virtual information in a real environment, including the following steps: providing at least one view of a real environment and of a system setup for blending in virtual information for superimposing with the real environment in at least part of the view, the system setup comprising at least one display device, ascertaining a position and orientation of at least one part of the system setup relative to at least one component of the real environment, subdividing at least part of the view of the real environment into a plurality of regions comprising a first region and a second region, with objects of the real environment within the first region being placed closer to the system setup than objects of the real environment within the second region, and blending in at least one item of virtual information on the display device in at least part of the view of the real environment, considering the position and orientation of said at least one part of the system setup, wherein the virtual information is shown differently in the first region than in the second region with respect to the type of blending in the view of the real environment.

Description

  • This application is a continuation of U.S. patent application Ser. No. 13/391,589 filed Mar. 22, 2012, which is a national phase application of PCT Application No. PCT/EP2010/061841 filed on Aug. 13, 2010, which claims priority to German Application No. 10 2009 037 835.9 filed Aug. 18, 2009.
  • BACKGROUND OF THE INVENTION
  • 1. Technical Field
  • The present invention relates to a method for representing virtual information in a real environment. The method is particularly suitable for ergonomic representation and marking of points of interest in the world by means of augmented reality technology.
  • 2. Background Information
  • Augmented Reality (AR) is a technology in which virtual data are overlaid with reality and which thus facilitates the association of data with reality. The use of mobile AR systems is already known in the prior art. In the past years, high-performance mobile devices (e.g. smartphones) turned out to be suitable for AR application. These devices meanwhile have comparatively large color displays, installed cameras, good processors and additional sensors, such as e.g. orientation sensors and GPS. In addition thereto, the position of the device can be approximated via radio networks.
  • In the past, there were various projects implemented on mobile devices using AR. At first, there were used special optical marks for ascertaining the position and orientation of the device. As regards AR, which is usable for large areas as well and thus is also referred to as large area AR, there have also been published hints for sensible representation of objects in connection with HMDs [2,4,5,6,7,8,9]. In more recent times, there are also approaches to utilize GPS and the orientation sensor systems of modern devices [1,3,10,15].
  • Especially as regards these approaches using video see-through AR on small mobile devices, however, there have been no innovative methods published for enhancing usability and user friendliness of the same, respectively.
  • The applicant has arrived at the finding that certain system properties should be fulfilled for using large-area AR: (1) Uncomplicated, rapid use and direct access to relevant information. (2) The accuracy of associating virtual information with reality is important. (3) The user interface should be clean and tidy.
  • Disadvantage of methods used so far: (1) There is often used a so-called “birdview” (a kind of bird's eye perspective) as overview on points in the surrounding environment. This view is used e.g. to show to the user, by blended in virtual information, where approximately points of interest are located in the real world. With a limited screen size, the user has a limited view of remote elements, or in other words the resolution becomes too small and the elements are not distinguishable/visible. The interface is rendered more complex by the simultaneous representation of two viewing directions. (2) If the size of the objects is not scaled properly, the users' distance perception and thus the capability of association will be impaired. (3) If the size of the objects is scaled, these become small and illegible with large distances. The interface has an unclean and untidy appearance.
  • SUMMARY OF THE INVENTION
  • It is the object of the invention to indicate a method for representing virtual information in a real environment, which is capable of achieving an ergonomic representation of points of interest in the real world without excessively restricting the user's field of view and without overstraining the user with an excess of information.
  • A first aspect of the invention relates to a method for representing virtual information in a real environment, comprising the following steps: providing at least one view of a real environment and of a system setup for blending in or overlaying virtual information for superimposing with the real environment in at least part of the view, the system setup comprising at least one display device, determining a position and orientation of at least one part of the system setup relative to at least one component of the real environment, subdividing at least part of the view of the real environment into a plurality of regions comprising a first region and a second region, with objects of the real environment within the first region being placed closer to the system setup than objects of the real environment within the second region, blending in or overlaying at least one item of virtual information on the display device in at least part of the view of the real environment, considering the position and orientation of said at least one part of the system setup, wherein the virtual information is shown differently in the first region than in the second region with respect to the type of blending in or overlay in the view of the real environment.
  • The at least one part of the system setup may be a camera, for example, the position and orientation (pose) of which are determined, with the camera being not necessarily fixedly connected to the display device. In specific cases, there is no camera necessary at all for the overall system, e.g. when the pose of at least one part of the system setup is determined via GPS and orientation sensors only. The pose determination of any part of the system setup is suitable as a matter of principle, provided there are conclusions possible as to the viewing direction of the user.
  • The first region can be a near region, whereas the second region can be a far or remote region, for example. However, it is also possible that the first region represents a location region, whereas the second region represents a near region. An embodiment with near region, far region and location region will be elucidated in more detail hereinafter with reference to the drawings.
  • In accordance with another aspect of the invention which can also be applied separately from the above first aspect of the invention involving the subdivision of the view into regions, it is also possible in blending in virtual information in a view of a real environment to consider weather data that are queried e.g. via the Internet (“online”) so as to enhance the degree of reality of blended in virtual information with respect to the real environment and to thus improve the association therewith. There are various degrees of complexity conceivable for processing. For example, e.g. on the basis of the weather situation, there may be assigned fixed illumination models or materials (e.g. textures) matching the weather situation. In addition or as an alternative, there can be calculated shadows cast or light conditions in accordance with weather data (such as e.g. clouds, solar radiation etc.) and/or other data (such as time of the day, time of the year etc.).
  • In an embodiment, the at least one item of virtual information represents a point of interest (generally also abbreviated as POI, especially in connection with navigation devices) with respect to the real environment.
  • The invention offers the advantage that an ergonomic representation of virtual information, in particular of points of interest, in the real world is obtained without the field of view of the user being excessively restricted and without overstraining the user with too many items of information. At the same time, there can be displayed many different items of virtual information which, however, due to the ergonomic representation of the virtual information, does not result in overstraining the user. Moreover, the association of information can be enhanced by consideration of the human perception mechanisms.
  • Further advantageous developments of the invention can be taken from the dependent claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The invention will now be elucidated in more detail by way of the figures shown in the drawings, in which
  • FIG. 1 shows a plan view of a schematic arrangement of exemplary system setups with respect to a real environment that can be used for performing a method according to the invention.
  • FIG. 2 shows a schematic arrangement of an exemplary division of points of interest (surrounding POIs) to different radius regions.
  • FIG. 3 shows a schematic arrangement illustrating a possible approach for calculating the radii over POIs in sections, in which the radii per sector may differ from each other.
  • FIG. 4 shows possible elements of an exemplary view of the real environment with blended in virtual information (here: POI objects) for forming an embodiment of a user interface according to aspects of the invention.
  • FIG. 5 shows possible elements of the user interface according to FIG. 4 when the focus of operation is located in the location region.
  • FIG. 6 shows an exemplary preview for virtual information (here: POI) in the location region of the view according to FIG. 4.
  • FIG. 7 shows an exemplary possibility of shielding surface elements from the view according to FIG. 4, when these are not required.
  • FIGS. 8 to 11 show different representation possibilities for embodiments of items of virtual information (here: POI objects).
  • FIG. 12 shows a flowchart giving an overview of the overall process of an embodiment of a method according to the invention.
  • FIG. 13 shows an embodiment illustrating a possible course for calculating radius 0 (cp. FIG. 2) of the location region, as exemplified in the view according to FIG. 4,
  • FIG. 14 shows an embodiment illustrating a possible course for calculating radius 1 as exemplified in FIG. 2.
  • FIG. 15 shows an embodiment illustrating a possible course for calculating radius 2 as exemplified in FIG. 2.
  • FIG. 16 shows a schematic view for illustrating exemplary possibilities for the aggregation of items of virtual information (here: POI objects).
  • FIG. 17 shows an exemplary possible course for associating shadows and material effects (“shading”) with the items of virtual information (here: POI objects).
  • FIG. 18 shows exemplary possible interactions with items of virtual information (here: POI objects).
  • DETAILED DESCRIPTION OF THE INVENTION
  • FIG. 1 shows a plan view illustrating a schematic arrangement of exemplary system setups with respect to a real environment, which can be used for performing a method according to the invention. In particular, FIG. 1 illustrates various possibilities of a system setup.
  • In the illustration of FIG. 1A, the user wears, as display device, a head mounted display system (abbreviated to HMD) comprising a display 21 that is part of the system setup 20. The display 21 e.g. may be generally known semi-transparent data glasses (“optical see-through display”) in which virtual information provided by a computer 23 can be blended in. The user then sees through the semi-permeable data glasses 21, in a view of the real world 40, objects of the real world augmented with blended in virtual information (such as e.g. POI objects related to the real world). In this manner, the system setup 20 constitutes an embodiment of a generally known Augmented Reality (AR) system.
  • The display 21 may have additional sensors 24, such as rotation sensors, and a camera 22 for optical tracking mounted thereon. Display 21 can be semi-transparent or may be fed with images of the reality by a camera image. With a semi-transparent display 21, calibration between eye 25 and display 21 is necessary. To this end, there are various methods documented in the prior art and known to the expert. Advantageously on display 21 or anywhere on the user's body or in computer unit 23, there may also be installed position sensors, such as e.g. GPS sensors (GPS: Global Positioning System) for rendering possible geographic position determination of the system setup 20 (e.g. in accordance with longitude and latitude) in the real world 40.
  • In the illustration of FIG. 1B, there is shown another exemplary system setup 30 that can be found often e.g. in modern mobile telephones (so-called “smartphones”). The display device 31 (e.g. in the form of a display screen or display), computer 33, sensors 34 and camera 32 constitute a system unit that is accommodated e.g. in a common housing of a mobile telephone. The view of the real environment 40 is provided by display 31 showing a camera image of the real environment 40 captured by camera 32. For augmented reality applications, the camera image can be shown on display 31 and augmented with additional virtual information (such as POI objects related to the real world). In this manner, the system setup 30 constitutes another embodiment of a generally known augmented reality (AR) system.
  • Basically, the present invention can be used expediently for all forms of AR. For example, it is of no relevance whether the representation is implemented in the so-called optical see-through mode with semi-transparent HMD or in the video see-through mode with camera and display screen. When reference is made hereinafter to the camera, this may be a camera or the optical system resulting from the combination of eye and see-through display (cp. display 21 of FIG. 1A). Both thereof comprise relevant camera properties for blending in virtual information, such as aperture angle and main point of image).
  • The invention basically can also be used in connection with stereoscopic displays, in which the video see-through approach advantageously uses two cameras each for recording one video stream per eye. In any situation, the items of virtual 3D information can be calculated individually for each eye.
  • The processing of the different partial steps described hereinafter basically can be distributed to various computers via a network. Thus, a client/sever architecture or a pure client-based solution is possible. For example, the client could send an image to a server which, on the basis of the image, makes available to the client information on the 3D position and 3D orientation of the system setup (cp. FIG. 1) or part thereof in relation to the real world (referred to as pose in the following) and on the range of vision. Moreover, the client or the server may also comprise several computing units, such as several CPUs or specialized hardware components, such as generally known FPGAs, ASICs, GPUs or DSPs. Several clients may also exchange information among each other which e.g. is generated with respect to the range of vision at this location, or in case a client generates a POI. This exchange of information can take place via a server, however, direct connections via Bluetooth or WLAN would be conceivable as well.
  • For permitting AR to be realized, the pose (position and orientation) of the camera in space is necessary. This can be realized in variety of different ways. It is possible to determine the pose in the world e.g. by using merely GPS and an orientation sensor with electronic compass (as installed e.g. in some modem mobile telephones). However, the uncertainty of the pose then is very high. Thus, it is also possible to use other methods, such as e.g. optical initialization and tracking or the combination of optical methods using GPS and orientation sensors. WLAN locating can be used as well or RFIDs or optical markers can support the locating process. As mentioned hereinbefore. a client/server-based approach is possible here as well. In particular, the client can request from the server location-specific information needed for optical tracking. Such information may be e.g. reference images of the surrounding environment with pose information and depth information.
  • The invention can improve the information representation for the client. However, it can also be utilized in a remote scenario (remote view scenario). In this scenario, e.g. a maintenance expert in a control room views the client's image transferred via the data network as well as the correspondingly processed items of information on his display screen. The expert then could give instructions to the client or just observe. In a similar scenario it is conceivable that a person views recorded image or video material along with interactive additional information represented according to the invention, and, if possible, can navigate through the material similar to the Internet-based application “Google Streetview”.
  • In addition thereto, the invention can also be installed, or carried along, in vehicles, aircraft or ships as a monitor, HMD or by way of a head-up display.
  • Basically, a point of interest (“POI”) can be set up for a large variety of different forms of information. Examples are given hereinafter. It is possible to represent images of places using GPS information. It is possible to automatically extract information from the Internet. For example, this may be company or restaurant websites with addresses or pages giving ratings. Users can deposit texts, images or 3D objects at locations and make the same available to others. Information pages, such as Wikipedia, can be searched for geo-information, and the pages can be made accessible as POI. POIs can be generated automatically from the search and browsing behavior of the users of mobile devices. It is possible to show other locations of interest, such as underground transportation or bus stations, hospitals, police stations, physicians, real estate ads or fitness clubs.
  • In the following, aspects and embodiments of the invention will be explained in more detail by way of the flowcharts shown beginning with FIG. 12 in conjunction with the remaining FIGS. 1 to 11.
  • In this regard, FIG. 2 shows a schematic arrangement illustrating an exemplary division of points of interests (surrounding POIs) to different radii regions, FIG. 3 shows a schematic arrangement illustrating a possible approach for calculating the radii over POIs in sectors, with the possibility that the radii per sector may be different, and FIG. 4 shows possible elements of an exemplary view of the real environment with blended in items of virtual information (here: POI objects) for creating an embodiment of a user interface in accordance with aspects of the invention.
  • As illustrated in more detail in FIG. 4, at least a part of the view 1 of the real environment (which may be the view through a semi-transparent HMD 21 or on a screen HMD 21 according to FIG. 1A or a screen view of a camera image on display 31 according to FIG. 1B) is subdivided into a plurality of regions. In the present embodiment, these regions comprise a first region 3, constituting a near region 3 in the present embodiment, a second region 4, constituting a far region 4 in the present embodiment, and a third region 5 constituting a location region 5 in the present embodiment. Objects of the real environment 40, within near region 3, are placed closer to the system setup 20 and 30, respectively, than objects of the real environment within far region 4. In the location region 5, in turn, objects of the real environment 40 are placed closer to system setup 20 and 30, respectively, than objects of the real environment within near region 3.
  • The view 1 according to FIG. 4 has at least one item of virtual information blended in (in the present embodiment, several items of virtual information POI1-POI4, POI11-POI14 and POI21-POI23 are blended in). In the present embodiment these represent points of interest (POIs) in relation to the reality that can be seen in view 1. For, example virtual information POI1 is blended in relating to a building visible in view 1 (not shown in FIG. 4) that makes reference to this building and/or is optically associated with the same and possibly permits additional information to be retrieved with respect to the building. The information POI1 is blended in on the display device 21 and 31, respectively, in at least part of the view 1 of the real environment, considering the position and orientation of the corresponding system setup 20 or 30 or part thereof, respectively. As will still be explained in more detail hereinafter, the blended in virtual information is shown differently in the near region 3 than in the far region 4, with respect to type of blending in view 1. In other words, the same virtual information (e.g. POI1) in the near region 3, e.g. when it is associated with a real object in the near region 3, is shown differently from its blending in the far region 4, e.g. when it is associated with a real object in the far region 4 (a POI is present here at the same time in one region only).
  • As illustrated in FIGS. 2 and 3, the near region 3 is separated from the far region 4 and the location region 5 by a respective boundary which in the present embodiment each have a radius (radius 0 to radius 3) in the direction of a viewing angle of the display device (FIG. 2 in addition shows further boundaries or radii, as will still be explained in more detail hereinafter). The determination and calculation of the radii according to FIGS. 2 and 3 will still be described in more detail in the following. The boundaries do not have a fixed shape, but preferably are rotation-symmetric with respect to the camera and viewer, respectively. The boundaries in addition do not need to be circular, but may also have an e.g. elliptical or other shape.
  • Overview
  • FIG. 12 shows in a flowchart an overview of the process of an embodiment of a method according to the invention in conjunction with FIGS. 1 to 11. In a first step 1.0, the system is initialized or, in case a pass has already taken place, prior values can be retained optionally. In a subsequent step, data are loaded if the same are not yet present or if the situation has changed. The data source may be contained in more or more servers or can be stored locally on the device or can be coded as information in the environment (e.g. via RFID or QR codes). In case a very large number of POIs is deposited in the data source, the number can be restricted via the position and the determination of a radius 3. Such functionality is offered e.g. by modem database systems with spatial features or properties. Basically, POIs may already be filtered out in this step as well by way of filters defined by the user himself. For example, gastronomy only or no gastronomy. POIs carrying corresponding information are sorted out then. It is apparent to the expert here that a corresponding data structure can be deposited for a POI, which may contain e.g. its position, its 2D or 3D shape, an image and/or additional meta information, such as e.g. the category.
  • In steps 3.0-6.0 the POIs then are organized according to the invention. Thereafter, they can be processed optically and displayed. Following this, interaction with the POIs shown is possible, as will be explained in more detail by way of an example.
  • Division Into Regions
  • In the following, steps 3.0-6.0 of FIG. 12 will be described in more detail in conjunction with subsequent FIGS. 13 to 18 and FIGS. 1 to 11. In this regard, FIGS. 13 to 18 relate to possible sub-steps of the general method steps illustrated in FIG. 12. These steps basically can be implemented per pass or in the background, in certain intervals, depending on the capacity of the system. A fundamental aspect of the invention consists in that items of blended in virtual information (here: POIs) will be assigned to fixedly configured or automatically calculated, different regions of a view of the real world.
  • FIG. 2 shows a view of a possible scenario from a bird's eye perspective. Various POIs are shown in a spatial arrangement in the surroundings of the camera. If all of these POIs were simply displayed, the screen rapidly would be overfilled with numerous objects. For this reason, a subdivision is made to different regions. The location region 5 is confined by radius 0. As regards the radii, it is to be noted that they are not necessarily constant over all angles, but may vary in accordance with the angle (cp. FIG. 3). The location region 5 shows objects located in close proximity, so that it is difficult for the user to find the same with the device. The possibility of finding them is dependent in particular on the accuracy of the ascertainable pose. Accordingly, radius 0 may be set greater or smaller advantageously in accordance with the (uncertainty of the) pose determination (FIG. 13).
  • FIG. 4 shows possible elements of an exemplary view of the real environment with blended in virtual information (here: POI objects) for forming an embodiment of a user interface according to aspects of the invention. FIG. 4 in this regard shows an advantageous subdivision of the user interface. POIs in the location region 5 are fixedly arranged in the lower portion of view 1 and can be selected e.g. by mouse click or touching in the event of a touchscreen. FIG. 6 shows a possible view as a result of activation of an exemplary POI23. If provided, an image P23 of the exact location of POI23 can be displayed for orientation. If provided, a preview V23 of the deposited information can be displayed in addition.
  • POI objects present outside the location region 5, but in near sight, are arranged in near region 3. In this region, the correct association of virtual information with reality is deemed to be particularly important. For example, it may be relevant to the personal planning of the path whether a POI object is present before or behind a road crossing. Near region 3 is confined by radii 0 and 1 (FIG. 2).
  • FIG. 14 illustrates a possibility for calculating radius 1. On the one hand, the expedient visibility of POI objects is of relevance. If the representation of a POI still would have a size of 2 mm only (exemplary value), it would not be of much use, but cause just disorder on the display. The calculation in step 4.1 thus incorporates the resolution of the display, the size of the POI (the representation e.g. of a dinosaur permits a larger radius 1 than the representation of a sphere with a size of 1 m) and the aperture angle of the camera. In the following steps, radius 1 can be matched further if there are too many POIs filling the screen (referred to as global approach). The calculation can refer to the entire aperture angle of the camera or an even larger range. Smaller sectors can be adjusted individually as well.
  • The sectors may be fixed, for example, or can be formed on the basis of clusters, with clusters being of assistance to divide sectors (cp. step 6.2 in FIG. 16). An exemplary method could be implemented as follows: the center of gravity is calculated for each cluster. A straight line is drawn through this center. Straight lines arranged at a small angle from each other are combined. Each resulting straight line is aligned with integral angles, and a sector is assigned to the same. The size of the sectors then is rounded an increased iteratively in integral angle steps until contact is established with a neighboring sector. When the sectors are defined by clusters or by fixed settings, the number of POIs of a sector is counted in step 4.2. If the number exceeds a specific configurable threshold value, the radius is reduced until the value is below the threshold value (step 4.3).
  • As an alternative, it is also possible to shift individual POIs without radius matching to the far region 4 (referred to as local approach) if there are POIs overlapping in the representation. To this end, in step 4.4, the 2D position and extension of the POIs, in accordance with the rendering pipeline, is calculated. The rendering pipeline is understood by the expert as the projection of objects in 3D space to a smaller-dimension display. In the near region, this is effected e.g. in accordance with position, orientation, camera aperture angle and resolution of the 2D display portion available for the near region. For the far region, the rendering pipeline performs projection to a line. This can be effected e.g. using a standard 3D-2D projection in which the objects are communicated for the renderer as being very far away, in order to determine the position of the objects on the possibly rotated display in the vanishing point of the same. In case of overlapping, the rear POI is marked for the far region 4 (step 4.5). This marking may be retained for one display pass (step 1.0 to 9.0 of FIG. 12) or longer (e.g. 3 second or 5 display passes). The system can perform steps 4.4 and 4.5 until there are no more overlappings present or these are below a specific value. There may also be a tolerance as to when overlapping is rated as such (e.g. more than 10% of the sum of the area of both objects intersecting each other). Advantageously, it is also possible to combine the global approach with the local approach.
  • Although the drawings, for reasons of visibility, illustrate the sectors and POIs from a bird's eye perspective, all of these methods can be applied selectively for 3D application as well.
  • The far region 4 contains POIs that were displaced from the near region or are located outside of radius 1 in the sector. But even here, it is not necessarily expedient to display all existing POI objects. Thus, it may be advantageous to calculate a radius 2, similar to radius 1, as shown in FIG. 15. Due to the fact that radius 2 in particular still is to display POI objects located within the reach of the user, it is advantageously possible to employ for calculating radius 2 the current speed or the average speed of the user or the distance that can be covered by the user within a specific period of time, using public transport, a vehicle or bicycle or the like. The possibility of calculating, in accordance with these factors, which objects can be displayed at all by AR, can also be applied independently of the other claims of the invention.
  • In an exemplary development of the invention, it is also possible to drag individual POIs from one region to another region (e.g. via a touchscreen) and thereby mark the same for display in the near region. The POI object then will be displayed permanently e.g. in near region 3.
  • For manual setting of the radii, it is also conceivable to display the same to the user in the correct position as blended in on reality. The user can then alter the radii manually by means of the touchscreen.
  • Representation of Objects
  • FIG. 4 shows a possible exemplary form of representation for POIs in different regions. POIs in the location region 5 below are represented in a uniform size and non-movably and for time being independently of the viewing direction. POI objects in near region 3 are overlaid on the environment 40 three-dimensionally and with correct perspective in accordance with the pose, the camera parameters and the model properties. POI objects in far region 4 advantageously are represented in a uniform size and advantageously move along in accordance with the orientation of the device (in particular the display device) such that a vertical line in downward direction would hit the corresponding associated real location. It is advantageously possible to additionally display, on the right and left margins 6, symbols hinting at POI objects outside the aperture angle of the camera. An arrow 8 can give a directional hint so that, when the device is moved around the viewer in the direction of the arrow, the POI object would enter the visible range. These margins 6 may exist for every region (FIG. 4) or e.g. for the near region 3 only (FIG. 5). Near region 3, far region 4 and/or location region 5 can be hidden or shielded optionally, especially when there is no virtual information displayed therein. This can also be effected automatically when there is no POI present in these regions (FIG. 7).
  • Basically, a point of interest (POI) can be represented by a simbol, an image, a 3D object or the like. For representation, it is possible to use a 3D rendering technology (such as the known methods OpenGL or DirectX) using 3D objects, the rendering of so-called billboards (2D objects that always face the viewer) or so-called 2D overlays with 2D rendering techniques the projection of which is calculated autonomously. The kind of representation may be based on the category of the POIs (e.g. a globe for representing a website) or can be determined by the user (e.g. placing a dinosaur with additional information). In particular, the POIs can be shown in a contrasting color.
  • As shown in FIGS. 10 and 11, POIs can be provided with additional items of information detailed to different extents (exemplified in the following by way of POI1). The lowest detail level is the POI only. In a next level or stage, labels (cp. label L1) can be displayed (FIG. 10, left) indicating e.g. a descriptive text for the associated POI object. The next stage is the so-called preview (preview cp. V1 or image P1 in FIG. 11) (e.g. an image or rendering of the website or an infotext). Some POI objects may be visualized still more closely in a subsequent stage (cp. information I1). This may also trigger activation of an own program of the operating system, such as starting of an Internet browser or a media player. Depending on the configuration or automatically exploitable criteria, such as the resolution of the display, the POIs can be displayed simultaneously with their additional information, or upon activation only. Advantageously, it can also be determined which POI object is located closest, and only the foremost POI objects indicate firstly labels as of a specific distance, and automatically indicate previews when coming closer.
  • As an alternative, the display can also be controlled by so-called eye tracking. The POIs viewed by the user are displayed with additional information. The additional information advantageously can be anchored in the view in non-movable manner and can be connected to the movable POI representation by way of a dynamic connection. This provides for enhanced legibility of the information. The use of eye tracking for activating additional information via a POI in AR can also be applied independently of the method according to claim 1.
  • For improving the association of the POI objects with real locations, it is possible to combine various method steps according to the invention, as illustrated in FIG. 17. The following method steps basically can also be implemented independently of the idea described so far, of dividing the view of the real environment into a plurality of regions (such as near, far and location regions) and to represent the blended in virtual information differently in accordance with the respective region, as described hereinbefore.
  • There is basically effected (especially in near region 3 when a division into regions is made) a 3D rendering of the POI object with correct perspective. The standard POI objects moreover have a fixed size to permit a continuous estimation of the distance to the same. For increasing the degree of reality and thus improving association, it is advantageously possible to query or retrieve weather data online. Various degrees of complexity are conceivable in this regard for processing. On the basis of the weather situation (e.g. in accordance with the Google weather service “mostly cloudy”, “occasional stormy”, “occasional rain” etc.) it is possible to assign fixed illumination models or materials (e.g. textures) matching the weather situation. However, at the highest complexity level, is would also be possible to make use of a current cloud or rain satellite or radar image for dynamically preparing an approximated model of the cloud cover and to calculate therefrom the shadows cast and optionally the detailed light conditions (cf. also FIG. 9). As was already mentioned, this can be carried out by a server making available the data to the client in location-specific form. As regards the perception of the distance, it is also useful to ascertain the visibility range in case of fog, rain or haze. This can be effected automatically (cp. “From video image e.g. (Automatic Fog Detection and Estimation of Visibility Distance through use of an Onboard Camera)”, Journal Machine Vision and Applications, Publisher Springer Berlin/Heidelberg ISSN 0932-8092 (Print) 1432-1769 (Online), Volume 17, Number 1/April 2006, pages 8-20) or also by means of current weather data.
  • The easiest technical realization, in addition to further known methods, consists in adjusting the fog settings in OpenGL. In particular for using the visibility distance for reality-conforming representation of the distance, it is also possible to represent part of a virtual object differently by fog, whereas another part is shown clearly. This is to prevent that e.g. when the technology is used in a vehicle, important items of POI information disappear completely due to fog. Technically, this can be realized e.g. by a second rendering pass considering specific materials only, but displaying no fog. As described in step 7.2 of FIG. 17, the position of the sun and the moon, respectively, can be calculated on the basis of the position, the date and the time, and can be used for adjusting the light sources. This has an effect in particular on the shadows (cp. shadows S1-S4 for POI1-POI4 in FIGS. 4, 10 and 11) assisting the user in better determining the position of the POI. The shadows can be pre-calculated textures (advantageously with a transparency value) which, in accordance with the position of the sun or the moon, are located below the POI on ground level where the straight line between sun or moon and POI intersects the ground level (exceptions, if this is not so). In case sun and moon are visible at the same time, the sun will be used for calculation.
  • However, the shadows may also be calculated dynamically, as known in the prior art. Advantageously, this can include the mutual shading of POIs. When a 3D model of the environment exists (cf. also step 7.4 in FIG. 17, e.g. in the form of a model for occluding a real object; so-called “occlusion geometry”), this can be used in addition for realistic calculation of the shadow situation e.g. in that it casts shadows on POIs (cf. also FIG. 9). In step 7.3, the reality degree of the blending in can be enhanced in addition by augmenting the materials by images of the environment of the POI. The use of so-called environment maps is known to the expert. Novel is that these are dynamically extracted in location-specific manner e.g. from Google Streetview and considered.
  • In step 7.4 there is taken a further step for enhancing the depth perception of the viewer. By loading occlusion models (so-called “occlusion geometry”), it can be determined whether a POI is visible for the viewer or disappears behind another building, for example. POIs, in case of occlusion of the same, can be shifted directly to the far region 4 or can be marked specifically. For example, the part occluded may be shown semi-transparent, in broken lines or in a different color. Advantageously, the shadows are not calculated and displayed either. The depth model can be deposited or generated dynamically by meas of SLAM algorithms, stereo cameras or a time-of-flight camera. In that event, depth information per image pixel is sufficient. In step 7.5, the camera parameters are generated for correct superimposition or overlaying in the near region (this does not have to take place continuously). These parameters can be generated dynamically e.g by means of a SLAM mechanism or retrieved from the server in accordance with the device name or can be deposited in the program. In case of a see-through HMD or HUD, the results of see-through calibration or dynamic measuring of the position of the eye in relation to the display are used.
  • In step 7.6, the camera image advantageously can be processed such that only the most important image components are shown in contrasting manner. This makes sense in the video see-through mode only, and should assist users in particular in case of very bright surrounding light. Processing of the video image can take place e.g. by means of the Sobel operator for edge extraction. This mode advantageously can be turned on and off in case of bright outside light, e.g. when the device contains a brightness sensor. In step 7.7, the parameters of the pose are made available to the rendering system in addition, if this was not already necessary for the calculations in steps 3 to 6 of FIG. 12. Depending on the hardware properties of the system, everything can be displayed and calculated in step 7.8. If the hardware of the system is weak, calculation of the correct material surfaces can also be effected on the part of the server (step 7.8B), or the entire image can be calculated on the part of the server. In case of a strong system, a modern GPU (Graphic Processor Unit) can take over a large share of the work (step 7.8A). There are numerous possibilities known to the expert in this regard.
  • Aggregation of POIs
  • At a time in the course (FIG. 12), advantageously before step 4.0 or after step 5.0, it is also possible to combine POIs, as shown in FIG. 16, at the lower left. The points of a cluster constitute a new, differently shaped POI. This can be effected individually in near region 3, in location region 5 and in far region 4, or can be effected already before the subdivision. Advantageously, this can be effected with POIs of one category only (e.g. websites only). POIs in the most forward region advantageously can be excluded from this process.
  • Interaction with POIs
  • FIG. 18 shows interaction possibilities with POIs according to aspects of the method of the invention. As is usual in augmented reality, the user can alter the segment of the items of virtual information and reality already by changing his location or the viewing direction. This can also trigger an automatic interaction, as already described in “representation of objects”. If the device is equipped with a touchscreen or a kind of mouse control (e.g. a trackball) it is possible in principle to address POIs directly. If several POIs have been combined to fowl a larger POI object, the individual POIs now are arranged selectably, and the labels are displayed (step 9.8). Otherwise, the label is displayed if the same is not yet displayed (step 9.9). By renewed activation of the POI (step 9.1 or key pressing) the infotext or preview is displayed if it is not yet displayed (step 9.11). By renewed activation, the detailed display is triggered for POIs having corresponding items of information deposited, step 9.13 (e.g. playing a movie, music or showing a website). By touching the closing button or advantageously by shaking the device and registration by movement sensors, the current display can be closed. Alternatively, it could possibly be retained in operation in the background. The user optionally can also provide a setting to the effect that the next audio or video file deposited is executed and run in the background.
  • Instead of a direct selection by touch, the interaction can also be implemented by means of the target cross 7 (optional) shown in FIG. 7, as described in FIG. 18 in steps 9.5 to 9.7. When the user directs the crosshairs 7 (or a similar target object) in the camera to a POI, the next detail level will be activated there (label in case POI only is visible, preview in case label is visible). Advantageously, the system can also activate the POI on the user interface that is closest to the crosshairs. By prolonged aiming at a POI or pressing a button, the next detail level can be activated. For reaching the far region 4, the user directs the camera upwardly until the highest point POI in the near region 3 has been exceeded by a threshold value (e.g. 10 degrees). The user then can navigate between the POIs by turning right and left. In similar manner, by directing the camera downwardly, the user can reach the location region 5. By turning right and left, the user here too can navigate through the otherwise fixed POIs. A special feature in the location region 5 resides in that a POI, if indicated or adapted to be generated automatically, in addition to POI detail information indicates an image of the location in order to facilitate finding the same (cp. preview V23 and image P23 for POI23 in FIG. 6). The image can be introduced manually or can be generated automatically by means of an image database containing pose information.
  • The selection of a POI optionally can be effected by voice control as well. To this end, the user activates the control by a specific sentence (e.g. “activate voice control”). Each POI is then marked with a number. By pronouncing the number, the corresponding POI can then be activated.
  • It is advantageously possible with the system to activate a map mode by directing the same downwards still further. When switching the POI or triggering an interaction, there is advantageously triggered an acoustic signal immediately. Moreover, a haptic signal (e.g. slight vibration) can be issued as well.
  • In case of a specific interaction with a POI (e.g. three clicks within a short period of time), it is advantageously possible to open a map mode centering the map at the location of the POI. As an alternative, navigation to this POI can be started by specific interaction.
  • In addition thereto, the invention contains the following aspects and embodiments that can be applied in connection with what has been described so far:
  • The virtual information (e.g. POIs) can be blended in the view of the real environment in the far region 4 in a uniform size. With respect to the type of blending in the view of the real environment, the virtual information in location region 5 is represented differently from its type of blending in the near region. In particular, the virtual information can be overlaid in the view of the real environment in location region 5 in non-movable manner, in particular also with uniform size and/or irrespective of the orientation of the display device.
  • It is also possible to combine several items of different virtual information in one virtual object and to display the virtual object in the view instead of the several items of virtual information. The several items of virtual information can be selected from a group of items of virtual information which together form a cluster, with the virtual object having a different shape as compared to the items of virtual information.
  • Moreover, a geographic map and/or a bird's eye perspective view can be blended in the display device when the display device is held approximately horizontally with respect to the earth's surface or is held by more than a specific angle below the lowest POI in the near region or, in case this function is already taken for reaching the location region, is held still further below.
  • A respective boundary, in particular the radius thereof, between the regions can be altered when virtual information is transferred by the user from one of the regions, e.g. the near region, to another region, e.g. the far region, and/or vice versa.
  • The boundary, in particular the radius therof, can also be calculated in accordance with the number of items of virtual information within a specific sector of the view. The boundary, in particular the radius thereof, can also be calculated in accordance with the two-dimensional density of several items of virtual information within a specific sector. Moreover, the boundary, in particular the radius thereof, can also be calculated in accordance with several items of virtual information which together constitute a cluster.
  • In the near region, a shadow can be indicated in the display device below the information, near a ground level shown in the display device, said shadow corresponding to the position of the virtual information.
  • The display device used may be a video display device in which the view of the real environment is augmented or replaced by an edge image for enhanced contrast.
  • In a user action for selecting virtual information of for switching between several items of virtual information, the user may be given acoustic and/or haptic feedback at an input device used for making a selection.
  • By determining the position and orientation of the at least one part of the system setup 20 and 30, respectively, in relation to the real environment, depth information with respect to at least one real object contained in the view can be calculated or loaded, and the depth information can be used for blending in an occlusion model for occluding a real object in the display device in case virtual information is occluded in the view by the real object to be occluded. Such depth information can also be used for calculating a boundary between regions, e.g. between near region and far region.
  • In addition thereto, several items of blended in virtual information in the display device may have a respective number assigned thereto, and the corresponding virtual information can be selected by voice recognition of the number or selection of the number on a keyboard or a touch-sensitive input pad.
  • Moreover, several items of virtual information can be associated with one of several categories each, and the items of virtual information can be blended in and/or masked in accordance with the category.
  • It is also possible to show an edge 6 in the display device, which indicates a range or coverage of the near region 3, with a boundary of the near region being variable by a user action, in particular by dragging the boundary.
  • For example, the virtual information can be represented in at least three stages. A first stage comprises a body (e.g. a 2D body or a 3D body) as a just local hint to the virtual information (cp. POI1 in FIGS. 10, 11), a second stage comprises a hint to the virtual information in the form a label with inscription (cp. label L1 in FIG. 11), and a third stage comprises an extract-like preview of the virtual information (cp. preview V1 and image P1 in FIGS. 10, 11) which is blended in particular when the dot-like representation or the label is selected by a user action. A fourth stage then shows the information in full length (cp. information I1 from FIG. 10).
  • A development may provide that, in a first part of the far region 4, virtual information is represented in the first stage only, in a second part of far region 4 with real objects placed closer to the display device than in the first part of the far region and in a first part of near region 3, virtual information is represented in the second stage, and in a second part of the near region 3 with real objects placed closer to the display device than in the first part of the near region, the virtual information is represented in the third stage.
  • Moreover, it can be provided that a boundary between near region and far region, in particular a radius of the boundary, is calculated in accordance with a size of the blended in virtual information, a resolution of the display device and/or a resolution of a camera used for generating the view.
  • Furthermore, the limit of the location region can be increased with increased measurement uncertainty of position detection.
  • Furthermore, the boundary, which in particular may be a radius and determines which objects are to be displayed at all, may be dependent on the current speed or the average speed of the user or the distance that can be covered by the user within a specific period of time, using public transport, a vehicle or bicycle or the like.
  • Furthermore, irrespective of the other disclosures, the system can superimpose virtual information on reality, and in blending in virtual information in a view of a real environment, weather data can be considered that are retrieved e.g. via the Internet (“online”) in order to increase the degree of reality of blended in virtual information with respect to the real environment and thereby improve the association therewith. The features and embodiments described hereinbefore in connection with weather data and the like can also be applied in connection with this aspect independently of other aspects described.
  • While the invention has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to the particular embodiment(s) disclosed herein as the best mode contemplated for carrying out this invention.

Claims (21)

What is claimed is:
1. A method for representing virtual virtual information in a real environment, comprising:
providing at least one view of a real environment and a system setup for blending in virtual information for superimposing with the real environment in at least part of the view, the system setup comprising at least one display device;
determining a position and orientation of at least one part of the system setup relative to at least one component of the real environment;
blending in at least one item of virtual information on the display device in at least part of the view of the real environment, considering the position and orientation of said at least one part of the system setup; and
wherein the at least one item of virtual information is configured to be interacted with a user according to at least one of eye tracking and voice control.
2. The method of claim 1, wherein the at least one item of virtual information is activated if it is determined to be viewed by the user according to said eye tracking.
3. The method of claim 1, further comprising:
displaying at least one sign for the at least one item of virtual information,
wherein the at least one item of virtual information is activated according to said voice control based on a voice indicative of the at least one sign.
4. The method of claim 1, wherein the at least one item of virtual information is configured to be viewed or activated by the user and is configured to be displayed with additional information.
5. The method of claim 1, further comprising:
subdividing at least part of the view of the real environment into a plurality of regions comprising a first region and a second region, with objects of the real environment within the first region being placed closer to the system setup than objects of the real environment within the second region; and
wherein the virtual information is configured to be shown differently in the first region than in the second region with respect to the type of blending in the view of the real environment.
6. The method of claim 5, wherein the at least one item of virtual information is blended in the view of the real environment in the first region with the correct perspective corresponding to the position and orientation of the display device relative to the real environment.
7. The method of claim 5, wherein the at least one virtual information is blended in the view of the real environment in the second region in a uniform size.
8. The method of claim 5, wherein the plurality of regions into which said at least part of the view of the real environment is subdivided, in addition to said first region and said second region, comprise a third region within which objects of the real environment are placed closer to the system setup than objects of the real environment within said first region, and wherein the at least one item of virtual information is shown differently in the third region than in the first region with respect to the type of blending in the view of the real environment.
9. The method of claim 5, wherein the at least one item of virtual information is blended in the view of the real environment in a location region in a non-movable manner.
10. The method of claim 1, wherein several items of different virtual information are combined to form a virtual object and the virtual object is displayed in the view instead of the several items of virtual information.
11. The method of claim 1, wherein the display device used is a video display device in which the view of the real environment is augmented or replaced by an edge image for enhanced contrast.
12. A method for representing virtual information in a real environment, comprising:
providing at least one view of a real environment and a system setup for blending in virtual information for superimposing with the real environment in at least part of the view, the system setup comprising at least one display device;
determining a position and orientation of at least one part of the system setup relative to at least one component of the real environment;
subdividing at least part of the view of the real environment into a plurality of regions comprising a first region and a second region, with objects of the real environment within the first region being placed closer to the system setup than objects of the real environment within the second region;
blending in at least one item of virtual information on the display device in at least part of the view of the real environment, considering the position and orientation of said at least one part of the system setup;
wherein the virtual information is shown differently in the first region than in the second region with respect to the type of blending in the view of the real environment;
wherein the first region is separated from the second region by a boundary, said boundary being calculated dynamically; and
wherein at least one of:
the boundary is altered when virtual information is transferred by a user from one of said regions to another one of said regions, or
the boundary is calculated in accordance with at least one of the number of items of virtual information within a particular sector of the view, the two-dimensional density of several items of virtual information within a particular sector of the view, or several items of virtual information which together constitute a cluster.
13. The method of claim 12, wherein the at least one item of virtual information is blended in the view of the real environment in the first region with the correct perspective corresponding to the position and orientation of the display device relative to the real environment.
14. The method of claim 12, wherein the at least one virtual information is blended in the view of the real environment in the second region in a uniform size.
15. The method of claim 12, wherein the plurality of regions into which said at least part of the view of the real environment is subdivided, in addition to said first region and said second region, comprise a third region within which objects of the real environment are placed closer to the system setup than objects of the real environment within said first region, and wherein the at least one item of virtual information is shown differently in the third region than in the first region with respect to the type of blending in the view of the real environment.
16. The method of claim 15, wherein the at least one item of virtual information is blended in the view of the real environment in a location region in a non-movable manner
17. The method of claim 12, wherein several items of different virtual information are combined to form a virtual object and the virtual object is displayed in the view instead of the several items of virtual information.
18. The method of claim 12, wherein said at least one item of virtual information is adapted to be displayed in at least three stages, a first stage comprising a body as a local hint to the virtual information, a second stage comprising a hint to the virtual information in the form of a label with inscription, and a third stage comprising an extract-like preview of the virtual information.
19. The method of claim 18, wherein in a first part of the second region virtual information is displayed in the first stage only, in a second part of the second region having real objects placed closer to the display device than in the first part of the second region and in a first part of the first region, virtual information is displayed in the second stage, and in a second part of the first region having real objects placed closer to the display device than in the first part of the first region, the virtual information is displayed in the third stage.
20. The method of claim 5, wherein the at least one item of virtual information is blended in the view of the real environment in the first region with the correct perspective corresponding to the position and orientation of the display device relative to the real environment in various sizes in accordance with the perspective positioning of the virtual information in the view.
21. The method of claim 12, wherein the at least one item of virtual information is blended in the view of the real environment in the first region with the correct perspective corresponding to the position and orientation of the display device relative to the real environment in various sizes in accordance with the perspective positioning of the virtual information in the view.
US14/502,366 2009-08-18 2014-09-30 Method for representing virtual information in a real environment Abandoned US20150015611A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US14/502,366 US20150015611A1 (en) 2009-08-18 2014-09-30 Method for representing virtual information in a real environment
US16/716,169 US20200193712A1 (en) 2009-08-18 2019-12-16 Method for Representing Virtual Information in a Real Environment
US16/843,462 US11562540B2 (en) 2009-08-18 2020-04-08 Method for representing virtual information in a real environment
US18/153,746 US20230290083A1 (en) 2009-08-18 2023-01-12 Method for representing virtual information in a real environment

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
DE102009037835A DE102009037835B4 (en) 2009-08-18 2009-08-18 Method for displaying virtual information in a real environment
DE102009037835.9 2009-08-18
PCT/EP2010/061841 WO2011020793A2 (en) 2009-08-18 2010-08-13 Method for representing virtual information in a real environment
US201213391589A 2012-03-22 2012-03-22
US14/502,366 US20150015611A1 (en) 2009-08-18 2014-09-30 Method for representing virtual information in a real environment

Related Parent Applications (3)

Application Number Title Priority Date Filing Date
PCT/EP2010/061841 Continuation WO2011020793A2 (en) 2009-08-18 2010-08-13 Method for representing virtual information in a real environment
US13/391,589 Continuation US8896629B2 (en) 2009-08-18 2010-08-13 Method for representing virtual information in a real environment
PCT/EP2010/006841 Continuation WO2011063894A2 (en) 2009-11-25 2010-11-10 Photovoltaic module structure and method of establishing an electrically conductive connection between contact layers spaced apart from one another, in particular in the photovoltaic module structure

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/716,169 Continuation US20200193712A1 (en) 2009-08-18 2019-12-16 Method for Representing Virtual Information in a Real Environment

Publications (1)

Publication Number Publication Date
US20150015611A1 true US20150015611A1 (en) 2015-01-15

Family

ID=43495370

Family Applications (5)

Application Number Title Priority Date Filing Date
US13/391,589 Active 2031-01-14 US8896629B2 (en) 2009-08-18 2010-08-13 Method for representing virtual information in a real environment
US14/502,366 Abandoned US20150015611A1 (en) 2009-08-18 2014-09-30 Method for representing virtual information in a real environment
US16/716,169 Pending US20200193712A1 (en) 2009-08-18 2019-12-16 Method for Representing Virtual Information in a Real Environment
US16/843,462 Active US11562540B2 (en) 2009-08-18 2020-04-08 Method for representing virtual information in a real environment
US18/153,746 Pending US20230290083A1 (en) 2009-08-18 2023-01-12 Method for representing virtual information in a real environment

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/391,589 Active 2031-01-14 US8896629B2 (en) 2009-08-18 2010-08-13 Method for representing virtual information in a real environment

Family Applications After (3)

Application Number Title Priority Date Filing Date
US16/716,169 Pending US20200193712A1 (en) 2009-08-18 2019-12-16 Method for Representing Virtual Information in a Real Environment
US16/843,462 Active US11562540B2 (en) 2009-08-18 2020-04-08 Method for representing virtual information in a real environment
US18/153,746 Pending US20230290083A1 (en) 2009-08-18 2023-01-12 Method for representing virtual information in a real environment

Country Status (5)

Country Link
US (5) US8896629B2 (en)
EP (2) EP3675067A1 (en)
CN (2) CN102473324B (en)
DE (1) DE102009037835B4 (en)
WO (1) WO2011020793A2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106020437A (en) * 2015-03-24 2016-10-12 诺基亚技术有限公司 Augmented reality
WO2016166138A1 (en) * 2015-04-17 2016-10-20 Thales Method for managing and displaying georeferenced graphic symbols and associated display device
EP3136204A3 (en) * 2015-08-27 2017-03-29 Fujitsu Limited Image processing device and image processing method
US9922446B2 (en) 2012-12-21 2018-03-20 Apple Inc. Method for representing virtual information in a real environment
US9928626B2 (en) 2011-11-11 2018-03-27 Sony Corporation Apparatus, method, and program for changing augmented-reality display in accordance with changed positional relationship between apparatus and object
CN108292448A (en) * 2015-12-10 2018-07-17 索尼公司 Information processing apparatus, information processing method, and program
WO2018227502A1 (en) 2017-06-15 2018-12-20 Tencent Technology (Shenzhen) Company Limited System and method of instantly previewing immersive content
US10325414B2 (en) 2017-05-08 2019-06-18 Microsoft Technology Licensing, Llc Application of edge effects to 3D virtual objects
US10983594B2 (en) * 2017-04-17 2021-04-20 Intel Corporation Sensory enhanced augmented reality and virtual reality device
TWI733245B (en) * 2019-11-07 2021-07-11 南開科技大學 System for switching between augmented reality and virtual reality based on interaction process and method thereof
DE112016005040B4 (en) 2015-11-03 2024-10-10 Motorola Solutions, Inc. Out-of-focus notifications displayed using a head-mounted display

Families Citing this family (116)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102009037835B4 (en) 2009-08-18 2012-12-06 Metaio Gmbh Method for displaying virtual information in a real environment
WO2011106520A1 (en) 2010-02-24 2011-09-01 Ipplex Holdings Corporation Augmented reality panorama supporting visually impaired individuals
US9057874B2 (en) * 2010-12-30 2015-06-16 GM Global Technology Operations LLC Virtual cursor for road scene object selection on full windshield head-up display
CN102176197A (en) * 2011-03-23 2011-09-07 上海那里网络科技有限公司 Method for performing real-time interaction by using virtual avatar and real-time image
US9727132B2 (en) * 2011-07-01 2017-08-08 Microsoft Technology Licensing, Llc Multi-visor: managing applications in augmented reality environments
DE102011084596A1 (en) 2011-10-17 2013-04-18 Robert Bosch Gmbh Method for assisting a driver in a foreign environment
DE102011084993A1 (en) 2011-10-21 2013-04-25 Robert Bosch Gmbh Transfer of data from image data-based map services to an assistance system
US9536251B2 (en) * 2011-11-15 2017-01-03 Excalibur Ip, Llc Providing advertisements in an augmented reality environment
US20130132959A1 (en) * 2011-11-23 2013-05-23 Yahoo! Inc. System for generating or using quests
US8872853B2 (en) 2011-12-01 2014-10-28 Microsoft Corporation Virtual light in augmented reality
US9311751B2 (en) * 2011-12-12 2016-04-12 Microsoft Technology Licensing, Llc Display of shadows via see-through display
US20150029214A1 (en) * 2012-01-19 2015-01-29 Pioneer Corporation Display device, control method, program and storage medium
WO2013145614A1 (en) * 2012-03-27 2013-10-03 パナソニック株式会社 Information processing device, server device, information processing method and program
DE102012006881A1 (en) 2012-04-03 2012-11-08 Daimler Ag Vehicle navigation system for motor car, has display unit and operating unit, which is provided for user input for virtual vehicle position, where processing unit is provided for displaying realistic imaging
US9183676B2 (en) * 2012-04-27 2015-11-10 Microsoft Technology Licensing, Llc Displaying a collision between real and virtual objects
WO2013169601A2 (en) * 2012-05-07 2013-11-14 Honda Motor Co., Ltd. Method to generate virtual display surfaces from video imagery of road based scenery
EP2672458A3 (en) * 2012-06-06 2017-04-26 Samsung Electronics Co., Ltd Terminal, apparatus and method for providing an augmented reality service
DE102012013503B4 (en) 2012-07-06 2014-10-09 Audi Ag Method and control system for operating a motor vehicle
US9819878B2 (en) 2012-07-20 2017-11-14 Rakuten, Inc. Moving-image processing device, moving-image processing method, and information recording medium
US9870642B2 (en) * 2012-08-10 2018-01-16 Here Global B.V. Method and apparatus for layout for augmented reality view
US8928695B2 (en) 2012-10-05 2015-01-06 Elwha Llc Formatting of one or more persistent augmentations in an augmented view in response to multiple input factors
US10269179B2 (en) 2012-10-05 2019-04-23 Elwha Llc Displaying second augmentations that are based on registered first augmentations
US9077647B2 (en) 2012-10-05 2015-07-07 Elwha Llc Correlating user reactions with augmentations displayed through augmented views
US9141188B2 (en) 2012-10-05 2015-09-22 Elwha Llc Presenting an augmented view in response to acquisition of data inferring user activity
US10713846B2 (en) 2012-10-05 2020-07-14 Elwha Llc Systems and methods for sharing augmentation data
US9111383B2 (en) 2012-10-05 2015-08-18 Elwha Llc Systems and methods for obtaining and using augmentation data and for sharing usage data
US10180715B2 (en) 2012-10-05 2019-01-15 Elwha Llc Correlating user reaction with at least an aspect associated with an augmentation of an augmented view
US9619911B2 (en) 2012-11-13 2017-04-11 Qualcomm Incorporated Modifying virtual object display properties
EP2936441B1 (en) * 2012-12-21 2020-09-30 Apple Inc. Method for representing virtual information in a real environment
US9395543B2 (en) * 2013-01-12 2016-07-19 Microsoft Technology Licensing, Llc Wearable behavior-based vision system
EP2956927A4 (en) * 2013-02-15 2016-10-12 Elwha Llc Displaying in response to detecting one or more user behaviors one or more second augmentations that are based on one or more registered first augmentations
US9959674B2 (en) 2013-02-26 2018-05-01 Qualcomm Incorporated Directional and X-ray view techniques for navigation using a mobile device
US9377519B2 (en) * 2013-03-12 2016-06-28 Qualcomm Incorporated Server-based mobile device regional candidate position fix mode selection
US9247518B2 (en) * 2013-03-12 2016-01-26 Qualcomm Incorporated Mobile device positioning responsive to externally generated regional candidate position fix mode selection
US9685001B2 (en) 2013-03-15 2017-06-20 Blackberry Limited System and method for indicating a presence of supplemental information in augmented reality
US9639964B2 (en) 2013-03-15 2017-05-02 Elwha Llc Dynamically preserving scene elements in augmented reality systems
US10025486B2 (en) 2013-03-15 2018-07-17 Elwha Llc Cross-reality select, drag, and drop for augmented reality systems
EP2778842A1 (en) * 2013-03-15 2014-09-17 BlackBerry Limited System and method for indicating a presence of supplemental information in augmented reality
US10109075B2 (en) 2013-03-15 2018-10-23 Elwha Llc Temporal element restoration in augmented reality systems
CN103220500B (en) * 2013-03-20 2015-12-02 积成电子股份有限公司 Grid equipment monitoring image superposes methods of exhibiting with business diagnosis image
DE102013206173A1 (en) * 2013-04-09 2014-10-09 Bayerische Motoren Werke Aktiengesellschaft Selection of individual elements for display on data glasses
US10509533B2 (en) * 2013-05-14 2019-12-17 Qualcomm Incorporated Systems and methods of generating augmented reality (AR) objects
US10139623B2 (en) * 2013-06-18 2018-11-27 Microsoft Technology Licensing, Llc Virtual object orientation and visualization
US9235051B2 (en) 2013-06-18 2016-01-12 Microsoft Technology Licensing, Llc Multi-space connected virtual data objects
DE102013013698B4 (en) 2013-08-16 2024-10-02 Audi Ag Method for operating electronic data glasses
CN105659295B (en) * 2013-09-24 2019-07-26 苹果公司 For indicating the method for point of interest in the view of true environment on the mobile apparatus and for the mobile device of the method
DE102013016246A1 (en) * 2013-10-01 2015-04-02 Daimler Ag Method and device for augmented presentation
DE102013016241A1 (en) * 2013-10-01 2015-04-02 Daimler Ag Method and device for augmented presentation
US9652892B2 (en) * 2013-10-29 2017-05-16 Microsoft Technology Licensing, Llc Mixed reality spotlight
FR3012618B1 (en) * 2013-10-30 2018-08-31 Thales OPERATOR TERMINAL WITH VIEWPOINT QUALITY ZONES DISPLAY
KR102138520B1 (en) * 2013-11-11 2020-08-11 엘지전자 주식회사 A head mounted display and the method of controlling the same
US9791700B2 (en) * 2013-11-27 2017-10-17 Magic Leap, Inc. Virtual and augmented reality systems and methods
US10186087B2 (en) * 2013-12-06 2019-01-22 Hewlett-Packard Development Company, L. P. Occluding augmented reality objects
CN110120072B (en) * 2013-12-19 2023-05-09 苹果公司 Method and system for tracking mobile devices
US9986225B2 (en) * 2014-02-14 2018-05-29 Autodesk, Inc. Techniques for cut-away stereo content in a stereoscopic display
DE102014003178B4 (en) * 2014-03-01 2021-09-16 Audi Ag Devices and methods for displaying an image by means of a display device which can be worn on the head of a user
US20150262428A1 (en) * 2014-03-17 2015-09-17 Qualcomm Incorporated Hierarchical clustering for view management augmented reality
US10417824B2 (en) * 2014-03-25 2019-09-17 Apple Inc. Method and system for representing a virtual object in a view of a real environment
CN104951059B (en) * 2014-03-31 2018-08-10 联想(北京)有限公司 A kind of data processing method, device and a kind of electronic equipment
US9392212B1 (en) 2014-04-17 2016-07-12 Visionary Vr, Inc. System and method for presenting virtual reality content to a user
KR102209511B1 (en) 2014-05-12 2021-01-29 엘지전자 주식회사 Wearable glass-type device and method of controlling the device
KR101888566B1 (en) 2014-06-03 2018-08-16 애플 인크. Method and system for presenting a digital information related to a real object
DE102014210481A1 (en) * 2014-06-03 2015-12-03 Siemens Aktiengesellschaft Information display on moving objects visible through windows
CN106462328B (en) * 2014-06-24 2020-09-04 索尼公司 Information processing apparatus, information processing method, and program
CN105227618B (en) * 2014-07-04 2018-12-04 中国移动通信集团广东有限公司 A kind of communication site's position information processing method and system
US9959681B2 (en) * 2014-07-09 2018-05-01 Ji-yeon Jeong Augmented reality contents generation and play system and method using the same
KR20160009879A (en) * 2014-07-17 2016-01-27 엘지전자 주식회사 Wearable display device and method for controlling the same
TWI533240B (en) * 2014-12-31 2016-05-11 拓邁科技股份有限公司 Methods and systems for displaying data, and related computer program prodcuts
US9965029B2 (en) * 2015-03-30 2018-05-08 Sony Corporation Information processing apparatus, information processing method, and program
DE102015006610B4 (en) 2015-05-21 2024-10-02 Audi Ag Method for operating data glasses in a motor vehicle and system with data glasses
JP6344311B2 (en) * 2015-05-26 2018-06-20 ソニー株式会社 Display device, information processing system, and control method
US9665170B1 (en) 2015-06-10 2017-05-30 Visionary Vr, Inc. System and method for presenting virtual reality content to a user based on body posture
US10169917B2 (en) * 2015-08-20 2019-01-01 Microsoft Technology Licensing, Llc Augmented reality
WO2017051564A1 (en) * 2015-09-25 2017-03-30 株式会社ソニー・インタラクティブエンタテインメント Information processing device
KR101718628B1 (en) * 2015-09-30 2017-03-21 라인 가부시키가이샤 Method and system for recognizing poi of outside of map screen
US9857874B2 (en) * 2015-11-03 2018-01-02 Chunghwa Picture Tubes, Ltd. Augmented reality system and augmented reality interaction method
CN106371571B (en) 2015-11-30 2019-12-13 北京智谷睿拓技术服务有限公司 Information processing method, information processing device and user equipment
US10096158B2 (en) * 2016-03-24 2018-10-09 Ford Global Technologies, Llc Method and system for virtual sensor data generation with depth ground truth annotation
CN105894584B (en) * 2016-04-15 2019-08-02 北京小鸟看看科技有限公司 The method and apparatus that are interacted with actual environment under a kind of three-dimensional immersive environment
JP2018005091A (en) * 2016-07-06 2018-01-11 富士通株式会社 Display control program, display control method and display controller
CA3032812A1 (en) 2016-08-04 2018-02-08 Reification Inc. Methods for simultaneous localization and mapping (slam) and related apparatus and systems
TWI585617B (en) * 2016-08-31 2017-06-01 宅妝股份有限公司 Interaction method and system thereof
EP3291531A1 (en) 2016-09-06 2018-03-07 Thomson Licensing Methods, devices and systems for automatic zoom when playing an augmented reality scene
JP2018077644A (en) 2016-11-08 2018-05-17 富士ゼロックス株式会社 Information processing system and program
US10373358B2 (en) * 2016-11-09 2019-08-06 Sony Corporation Edge user interface for augmenting camera viewfinder with information
EP3349099A1 (en) * 2017-01-13 2018-07-18 Kai-Holger Brassel Immersive information and communication medium for very large user numbers
CN108509824B (en) * 2017-02-24 2020-08-18 亮风台(上海)信息科技有限公司 Article feature identification method based on AR equipment and system for checking article
US10429926B2 (en) * 2017-03-15 2019-10-01 International Business Machines Corporation Physical object addition and removal based on affordance and view
GB2574780B (en) 2017-04-04 2022-09-21 Canon Kk Electronic device and control method thereof
US10264380B2 (en) * 2017-05-09 2019-04-16 Microsoft Technology Licensing, Llc Spatial audio for three-dimensional data sets
US10249096B2 (en) * 2017-05-17 2019-04-02 International Business Machines Corporation Mixing virtual image data and physical image data
CN107315915A (en) * 2017-06-28 2017-11-03 上海联影医疗科技有限公司 A kind of simulated medical surgery method and system
US10231167B2 (en) 2017-06-30 2019-03-12 Otis Elevator Company Building access zone specification for mobile applications
US11118930B2 (en) * 2017-07-14 2021-09-14 Lyft, Inc. Providing information to users of a transportation system using augmented reality elements
CN107656961B (en) * 2017-08-04 2020-03-27 阿里巴巴集团控股有限公司 Information display method and device
KR102422929B1 (en) * 2017-08-16 2022-07-20 삼성전자 주식회사 Display apparatus, server and control method thereof
US11455565B2 (en) * 2017-08-31 2022-09-27 Ford Global Technologies, Llc Augmenting real sensor recordings with simulated sensor data
US11164380B2 (en) * 2017-12-05 2021-11-02 Samsung Electronics Co., Ltd. System and method for transition boundaries and distance responsive interfaces in augmented and virtual reality
CN108182730B (en) 2018-01-12 2022-08-12 北京小米移动软件有限公司 Virtual and real object synthesis method and device
TWI707306B (en) * 2018-03-06 2020-10-11 國立臺灣大學 Method and device for enhancing the efficiency of searching regions of interest in a virtual environment
US10747312B2 (en) * 2018-03-14 2020-08-18 Apple Inc. Image enhancement devices with gaze tracking
CN108398787B (en) * 2018-03-20 2023-05-16 京东方科技集团股份有限公司 Augmented reality display device, method and augmented reality glasses
CN110536125A (en) * 2018-05-25 2019-12-03 光宝电子(广州)有限公司 Image processing system and image treatment method
CN112771473A (en) * 2018-09-07 2021-05-07 苹果公司 Inserting imagery from a real environment into a virtual environment
US11244483B2 (en) * 2018-10-31 2022-02-08 Facebook Technologies, Llc Systems and methods for modifying a safety boundary for virtual reality systems
US11087541B2 (en) * 2018-12-03 2021-08-10 Honeywell International Inc. Location-based identification of petrochemical assets in an industrial plant
US10839594B2 (en) 2018-12-11 2020-11-17 Canon Kabushiki Kaisha Method, system and apparatus for capture of image data for free viewpoint video
US10783714B2 (en) 2019-01-29 2020-09-22 Verizon Patent And Licensing Inc. Methods and systems for automatically tailoring a form of an extended reality overlay object
KR102627612B1 (en) * 2019-02-19 2024-01-22 삼성전자주식회사 Method for displaying nerby information using augmented reality and electonic device therof
EP3911992A4 (en) * 2019-04-11 2022-03-23 Samsung Electronics Co., Ltd. Head-mounted display device and operating method of the same
CN110248381B (en) * 2019-07-08 2021-04-02 腾讯科技(深圳)有限公司 Positioning data processing method and device, computer equipment and storage medium
US11263787B2 (en) * 2020-03-05 2022-03-01 Rivian Ip Holdings, Llc Augmented reality detection for locating autonomous vehicles
AU2021330970A1 (en) 2020-08-24 2023-03-23 Fd Ip & Licensing Llc Previsualization devices and systems for the film industry
CN113393516B (en) * 2021-06-17 2022-05-24 贝壳找房(北京)科技有限公司 Method and apparatus for breaking up virtual objects in an AR scene
CN113722937B (en) * 2021-10-29 2022-03-18 深圳市城市交通规划设计研究中心股份有限公司 Method for setting guide mark, computer and storage medium
US11875492B1 (en) 2023-05-01 2024-01-16 Fd Ip & Licensing Llc Systems and methods for digital compositing

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020044152A1 (en) * 2000-10-16 2002-04-18 Abbott Kenneth H. Dynamic integration of computer generated and real world images
US20040128070A1 (en) * 2002-12-31 2004-07-01 Hauke Schmidt System and method for advanced 3D visualization for mobile navigation units
US6847336B1 (en) * 1996-10-02 2005-01-25 Jerome H. Lemelson Selectively controllable heads-up display system
US20050046953A1 (en) * 2003-08-29 2005-03-03 C.R.F. Societa Consortile Per Azioni Virtual display device for a vehicle instrument panel
US20060028400A1 (en) * 2004-08-03 2006-02-09 Silverbrook Research Pty Ltd Head mounted display with wave front modulator
US20060241860A1 (en) * 2005-04-21 2006-10-26 Microsoft Corporation Virtual earth mapping
US20070162942A1 (en) * 2006-01-09 2007-07-12 Kimmo Hamynen Displaying network objects in mobile devices based on geolocation
US20080024484A1 (en) * 2006-06-26 2008-01-31 University Of Southern California Seamless Image Integration Into 3D Models
US20090157313A1 (en) * 2007-12-17 2009-06-18 Hyundai Motor Company Method of loading 3-d image data
US20100023878A1 (en) * 2008-07-23 2010-01-28 Yahoo! Inc. Virtual notes in a reality overlay
US20100053069A1 (en) * 2008-08-26 2010-03-04 Symbol Technologies, Inc. Mobile computing system facilitating adaptive display of content among a plurality of display components including at least one virtual image display component
US8331611B2 (en) * 2009-07-13 2012-12-11 Raytheon Company Overlay information over video
US8427508B2 (en) * 2009-06-25 2013-04-23 Nokia Corporation Method and apparatus for an augmented reality user interface
US8457991B2 (en) * 2005-02-04 2013-06-04 The Invention Science Fund I, Llc Virtual credit in simulated environments
US8467991B2 (en) * 2008-06-20 2013-06-18 Microsoft Corporation Data services based on gesture and location information of device
US8487957B1 (en) * 2007-05-29 2013-07-16 Google Inc. Displaying and navigating within photo placemarks in a geographic information system, and applications thereof
US8896629B2 (en) * 2009-08-18 2014-11-25 Metaio Gmbh Method for representing virtual information in a real environment
US9420251B2 (en) * 2010-02-08 2016-08-16 Nikon Corporation Imaging device and information acquisition system in which an acquired image and associated information are held on a display
US9710554B2 (en) * 2010-09-23 2017-07-18 Nokia Technologies Oy Methods, apparatuses and computer program products for grouping content in augmented reality

Family Cites Families (103)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5379215A (en) * 1991-02-25 1995-01-03 Douglas P. Kruhoeffer Method for creating a 3-D image of terrain and associated weather
US6037936A (en) * 1993-09-10 2000-03-14 Criticom Corp. Computer vision system with a graphic user interface and remote camera control
JP3461980B2 (en) 1995-08-25 2003-10-27 株式会社東芝 High-speed drawing method and apparatus
EP1426910A3 (en) * 1996-04-16 2006-11-02 Xanavi Informatics Corporation Map display device, navigation device and map display method
US6720949B1 (en) 1997-08-22 2004-04-13 Timothy R. Pryor Man machine interfaces and applications
FR2775814B1 (en) 1998-03-06 2001-01-19 Rasterland Sa SYSTEM FOR VIEWING REALISTIC VIRTUAL THREE-DIMENSIONAL IMAGES IN REAL TIME
US6064354A (en) 1998-07-01 2000-05-16 Deluca; Michael Joseph Stereoscopic user interface method and apparatus
US6983203B1 (en) * 2000-07-28 2006-01-03 Alpine Electronics, Inc. POI icon display method and navigation system
US6944607B1 (en) * 2000-10-04 2005-09-13 Hewlett-Packard Development Compnay, L.P. Aggregated clustering method and system
US6405129B1 (en) 2000-11-29 2002-06-11 Alpine Electronics, Inc. Method of displaying POI icons for navigation apparatus
US6992659B2 (en) 2001-05-22 2006-01-31 Palmone, Inc. High transparency integrated enclosure touch screen assembly for a portable hand held device
US7082365B2 (en) * 2001-08-16 2006-07-25 Networks In Motion, Inc. Point of interest spatial rating search method and system
US6968511B1 (en) * 2002-03-07 2005-11-22 Microsoft Corporation Graphical user interface, data structure and associated method for cluster-based document management
US7203384B2 (en) 2003-02-24 2007-04-10 Electronic Scripting Products, Inc. Implement for optically inferring information from a planar jotting surface
JP4203354B2 (en) * 2003-05-19 2008-12-24 パナソニック株式会社 Content distribution apparatus and content reception apparatus
US6856901B2 (en) 2003-06-02 2005-02-15 Alpine Electronics, Inc. Display method and apparatus for navigation system
JP4532856B2 (en) 2003-07-08 2010-08-25 キヤノン株式会社 Position and orientation measurement method and apparatus
US7110100B2 (en) 2003-11-04 2006-09-19 Electronic Scripting Products, Inc. Apparatus and method for determining an inclination of an elongate object contacting a plane surface
US7038846B2 (en) 2003-11-24 2006-05-02 Electronic Scripting Products, Inc. Solid catadioptric lens with a single viewpoint
US7268956B2 (en) 2003-11-24 2007-09-11 Electronic Scripting Products, Inc. Solid catadioptric lens with two viewpoints
US7088440B2 (en) 2003-12-22 2006-08-08 Electronic Scripting Products, Inc. Method and apparatus for determining absolute position of a tip of an elongate object on a plane surface with invariant features
US7826641B2 (en) 2004-01-30 2010-11-02 Electronic Scripting Products, Inc. Apparatus and method for determining an absolute pose of a manipulated object in a real three-dimensional environment with invariant features
US7729515B2 (en) 2006-03-08 2010-06-01 Electronic Scripting Products, Inc. Optical navigation apparatus using fixed beacons and a centroid sensing device
US9229540B2 (en) 2004-01-30 2016-01-05 Electronic Scripting Products, Inc. Deriving input from six degrees of freedom interfaces
US8542219B2 (en) 2004-01-30 2013-09-24 Electronic Scripting Products, Inc. Processing pose data derived from the pose of an elongate object
US7961909B2 (en) 2006-03-08 2011-06-14 Electronic Scripting Products, Inc. Computer interface employing a manipulated object with absolute pose detection component and a display
US7353109B2 (en) * 2004-02-05 2008-04-01 Alpine Electronics, Inc. Display method and apparatus for navigation system for performing cluster search of objects
US7023536B2 (en) 2004-03-08 2006-04-04 Electronic Scripting Products, Inc. Apparatus and method for determining orientation parameters of an elongate object
US7161664B2 (en) 2004-04-13 2007-01-09 Electronic Scripting Products, Inc. Apparatus and method for optical determination of intermediate distances
US7295220B2 (en) * 2004-05-28 2007-11-13 National University Of Singapore Interactive system and method
JP4688577B2 (en) * 2004-06-07 2011-05-25 パナソニック株式会社 Content display device and content display method
US7113270B2 (en) 2004-06-18 2006-09-26 Electronics Scripting Products, Inc. Determination of an orientation parameter of an elongate object with a scan beam apparatus
US7460953B2 (en) * 2004-06-30 2008-12-02 Navteq North America, Llc Method of operating a navigation system using images
EP1866043A1 (en) * 2005-04-06 2007-12-19 Eidgenössische Technische Hochschule Zürich (ETH) Method of executing an application in a mobile device
US8040361B2 (en) * 2005-04-11 2011-10-18 Systems Technology, Inc. Systems and methods for combining virtual and real-time physical environments
WO2006121986A2 (en) * 2005-05-06 2006-11-16 Facet Technology Corp. Network-based navigation system having virtual drive-thru advertisements integrated with actual imagery from along a physical route
IL169934A (en) * 2005-07-27 2013-02-28 Rafael Advanced Defense Sys Real-time geographic information system and method
US7634354B2 (en) * 2005-08-31 2009-12-15 Microsoft Corporation Location signposting and orientation
US7904483B2 (en) * 2005-12-23 2011-03-08 Geopeg, Inc. System and method for presenting geo-located objects
DE102005061952B4 (en) * 2005-12-23 2008-09-11 Metaio Gmbh Method and system for determining inaccuracy information in an augmented reality system
US8280405B2 (en) 2005-12-29 2012-10-02 Aechelon Technology, Inc. Location based wireless collaborative environment with a visual user interface
US7913192B2 (en) * 2006-03-31 2011-03-22 Research In Motion Limited Methods and apparatus for retrieving and displaying map-related data for visually displayed maps of mobile communication devices
ATE409307T1 (en) * 2006-03-31 2008-10-15 Research In Motion Ltd USER INTERFACE METHOD AND APPARATUS FOR CONTROLLING THE VISUAL DISPLAY OF MAPS WITH SELECTABLE MAP ELEMENTS IN MOBILE COMMUNICATION DEVICES
KR101285360B1 (en) * 2007-01-25 2013-07-11 삼성전자주식회사 Point of interest displaying apparatus and method for using augmented reality
KR101387479B1 (en) 2007-03-16 2014-04-21 엘지전자 주식회사 A method for processing image in mobile communication terminal and the mobile communication terminal
JP4412342B2 (en) * 2007-03-30 2010-02-10 ソニー株式会社 CONTENT MANAGEMENT DEVICE, IMAGE DISPLAY DEVICE, IMAGING DEVICE, PROCESSING METHOD IN THEM, AND PROGRAM FOR CAUSING COMPUTER TO EXECUTE THE METHOD
US7689457B2 (en) * 2007-03-30 2010-03-30 Amazon Technologies, Inc. Cluster-based assessment of user interests
EP2143086A4 (en) * 2007-04-03 2010-11-10 Human Network Labs Inc Method and apparatus for acquiring local position and overlaying information
US20080268876A1 (en) 2007-04-24 2008-10-30 Natasha Gelfand Method, Device, Mobile Terminal, and Computer Program Product for a Point of Interest Based Scheme for Improving Mobile Visual Searching Functionalities
US8165808B2 (en) 2007-07-17 2012-04-24 Yahoo! Inc. Techniques for representing location information
DE102007033486B4 (en) * 2007-07-18 2010-06-17 Metaio Gmbh Method and system for mixing a virtual data model with an image generated by a camera or a presentation device
US8994851B2 (en) * 2007-08-07 2015-03-31 Qualcomm Incorporated Displaying image data and geographic element data
US9329052B2 (en) * 2007-08-07 2016-05-03 Qualcomm Incorporated Displaying image data and geographic element data
US8671355B2 (en) 2007-10-05 2014-03-11 Mapquest, Inc. Methods and systems for decluttering icons representing points of interest on a map
US8339399B2 (en) * 2007-10-31 2012-12-25 Microsoft Corporation Declustering point-of-interest icons
US20100214111A1 (en) 2007-12-21 2010-08-26 Motorola, Inc. Mobile virtual and augmented reality system
US20090169060A1 (en) * 2007-12-26 2009-07-02 Robert Bosch Gmbh Method and apparatus for spatial display and selection
JP4502005B2 (en) 2007-12-27 2010-07-14 アイシン・エィ・ダブリュ株式会社 Navigation device and computer program
US9582937B2 (en) * 2008-01-02 2017-02-28 Nokia Technologies Oy Method, apparatus and computer program product for displaying an indication of an object within a current field of view
US8504945B2 (en) * 2008-02-01 2013-08-06 Gabriel Jakobson Method and system for associating content with map zoom function
US8428873B2 (en) * 2008-03-24 2013-04-23 Google Inc. Panoramic images within driving directions
US8711176B2 (en) * 2008-05-22 2014-04-29 Yahoo! Inc. Virtual billboards
KR101526970B1 (en) 2008-05-29 2015-06-16 엘지전자 주식회사 Terminal and method for controlling the same
US20090319166A1 (en) * 2008-06-20 2009-12-24 Microsoft Corporation Mobile computing services based on devices with dynamic direction information
US8401771B2 (en) 2008-07-22 2013-03-19 Microsoft Corporation Discovering points of interest from users map annotations
KR101561913B1 (en) 2009-04-17 2015-10-20 엘지전자 주식회사 Method for displaying image for mobile terminal and apparatus thereof
US20100332539A1 (en) * 2009-06-30 2010-12-30 Sunil Mohan Presenting a related item using a cluster
US20110010650A1 (en) 2009-07-09 2011-01-13 Mapquest, Inc. Systems and methods for decluttering electronic map displays
US9488488B2 (en) * 2010-02-12 2016-11-08 Apple Inc. Augmented reality maps
KR101643869B1 (en) 2010-05-06 2016-07-29 엘지전자 주식회사 Operating a Mobile Termianl with a Vibration Module
KR101572892B1 (en) 2010-05-06 2015-11-30 엘지전자 주식회사 Mobile terminal and Method for displying image thereof
EP2482561A4 (en) 2010-05-27 2013-09-11 Nintendo Co Ltd Portable electronic device
US20120050144A1 (en) 2010-08-26 2012-03-01 Clayton Richard Morlock Wearable augmented reality computing apparatus
US8944745B2 (en) 2010-11-05 2015-02-03 Honda Motor Co., Ltd. Carrier device, transfer method and carrier hand
US9514717B2 (en) 2011-09-26 2016-12-06 Nokia Technology Oy Method and apparatus for rendering items in a user interface
US9087412B2 (en) * 2011-09-26 2015-07-21 Nokia Technologies Oy Method and apparatus for grouping and de-overlapping items in a user interface
CN102629183B (en) 2012-02-29 2014-12-03 华为终端有限公司 Terminal screen control method and device
CN102695034A (en) 2012-05-30 2012-09-26 青岛海信移动通信技术股份有限公司 Method and device for regulating end display of video image during video call
EP2936443A1 (en) * 2012-12-21 2015-10-28 Metaio GmbH Method for representing virtual information in a real environment
CN103034416B (en) 2012-12-25 2015-09-09 珠海金山办公软件有限公司 A kind of by shaking the method making device screen display forward gravity direction to
CN103092344A (en) 2013-01-11 2013-05-08 深圳市金立通信设备有限公司 Method of controlling terminal screen to rotate and terminal
CN103139354B (en) 2013-01-31 2016-03-02 广东欧珀移动通信有限公司 A kind of mobile phone unlock method and device
US8970709B2 (en) 2013-03-13 2015-03-03 Electronic Scripting Products, Inc. Reduced homography for recovery of pose parameters of an optical apparatus producing image data with structural uncertainty
CN103246431B (en) 2013-04-27 2016-12-28 深圳市金立通信设备有限公司 A kind of regulation screen display direction method, Apparatus and system
US9552675B2 (en) 2013-06-03 2017-01-24 Time Traveler App Llc Display application and perspective views of virtual space
US9129430B2 (en) 2013-06-25 2015-09-08 Microsoft Technology Licensing, Llc Indicating out-of-view augmented reality images
US9355123B2 (en) 2013-07-19 2016-05-31 Nant Holdings Ip, Llc Fast recognition algorithm processing, systems and methods
KR102065408B1 (en) 2013-08-16 2020-01-13 엘지전자 주식회사 Mobile terminal
CN105659295B (en) * 2013-09-24 2019-07-26 苹果公司 For indicating the method for point of interest in the view of true environment on the mobile apparatus and for the mobile device of the method
US10126822B2 (en) 2013-12-16 2018-11-13 Leap Motion, Inc. User-defined virtual interaction space and manipulation of virtual configuration
US20160307374A1 (en) 2013-12-19 2016-10-20 Metaio Gmbh Method and system for providing information associated with a view of a real environment superimposed with a virtual object
KR20150092560A (en) 2014-02-05 2015-08-13 엘지전자 주식회사 Mobile terminal and method for controlling the same
US9685005B2 (en) 2015-01-02 2017-06-20 Eon Reality, Inc. Virtual lasers for interacting with augmented reality environments
US9767613B1 (en) 2015-01-23 2017-09-19 Leap Motion, Inc. Systems and method of interacting with a virtual object
US9696795B2 (en) 2015-02-13 2017-07-04 Leap Motion, Inc. Systems and methods of creating a realistic grab experience in virtual reality/augmented reality environments
US10354441B2 (en) 2016-05-20 2019-07-16 ETAK Systems, LLC Augmented reality systems and methods for telecommunications site modeling
US10728767B2 (en) 2015-04-14 2020-07-28 ETAK Systems, LLC Systems and methods for augmented reality add-in of equipment and structures at a telecommunications site
US10149958B1 (en) 2015-07-17 2018-12-11 Bao Tran Systems and methods for computer assisted operation
WO2017201568A1 (en) 2016-05-23 2017-11-30 tagSpace Pty Ltd Media tags location-anchored digital media for augmented reality and virtual reality environments
US20180095636A1 (en) 2016-10-04 2018-04-05 Facebook, Inc. Controls and Interfaces for User Interactions in Virtual Spaces
GB2554914B (en) 2016-10-14 2022-07-20 Vr Chitect Ltd Virtual reality system and method
US10527849B2 (en) 2017-07-18 2020-01-07 Toyota Jidosha Kabushiki Kaisha Augmented reality vehicular assistance for color blindness
DK180842B1 (en) 2018-01-24 2022-05-12 Apple Inc Devices, procedures, and graphical user interfaces for System-Wide behavior for 3D models

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6847336B1 (en) * 1996-10-02 2005-01-25 Jerome H. Lemelson Selectively controllable heads-up display system
US20020044152A1 (en) * 2000-10-16 2002-04-18 Abbott Kenneth H. Dynamic integration of computer generated and real world images
US20040128070A1 (en) * 2002-12-31 2004-07-01 Hauke Schmidt System and method for advanced 3D visualization for mobile navigation units
US20050046953A1 (en) * 2003-08-29 2005-03-03 C.R.F. Societa Consortile Per Azioni Virtual display device for a vehicle instrument panel
US20060028400A1 (en) * 2004-08-03 2006-02-09 Silverbrook Research Pty Ltd Head mounted display with wave front modulator
US8457991B2 (en) * 2005-02-04 2013-06-04 The Invention Science Fund I, Llc Virtual credit in simulated environments
US20060241860A1 (en) * 2005-04-21 2006-10-26 Microsoft Corporation Virtual earth mapping
US20070162942A1 (en) * 2006-01-09 2007-07-12 Kimmo Hamynen Displaying network objects in mobile devices based on geolocation
US20080024484A1 (en) * 2006-06-26 2008-01-31 University Of Southern California Seamless Image Integration Into 3D Models
US8487957B1 (en) * 2007-05-29 2013-07-16 Google Inc. Displaying and navigating within photo placemarks in a geographic information system, and applications thereof
US20090157313A1 (en) * 2007-12-17 2009-06-18 Hyundai Motor Company Method of loading 3-d image data
US8467991B2 (en) * 2008-06-20 2013-06-18 Microsoft Corporation Data services based on gesture and location information of device
US20100023878A1 (en) * 2008-07-23 2010-01-28 Yahoo! Inc. Virtual notes in a reality overlay
US20100053069A1 (en) * 2008-08-26 2010-03-04 Symbol Technologies, Inc. Mobile computing system facilitating adaptive display of content among a plurality of display components including at least one virtual image display component
US8427508B2 (en) * 2009-06-25 2013-04-23 Nokia Corporation Method and apparatus for an augmented reality user interface
US8331611B2 (en) * 2009-07-13 2012-12-11 Raytheon Company Overlay information over video
US8896629B2 (en) * 2009-08-18 2014-11-25 Metaio Gmbh Method for representing virtual information in a real environment
US9420251B2 (en) * 2010-02-08 2016-08-16 Nikon Corporation Imaging device and information acquisition system in which an acquired image and associated information are held on a display
US9710554B2 (en) * 2010-09-23 2017-07-18 Nokia Technologies Oy Methods, apparatuses and computer program products for grouping content in augmented reality

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9928626B2 (en) 2011-11-11 2018-03-27 Sony Corporation Apparatus, method, and program for changing augmented-reality display in accordance with changed positional relationship between apparatus and object
US10614605B2 (en) 2011-11-11 2020-04-07 Sony Corporation Information processing apparatus, information processing method, and program for displaying a virtual object on a display
US9922446B2 (en) 2012-12-21 2018-03-20 Apple Inc. Method for representing virtual information in a real environment
US10878617B2 (en) 2012-12-21 2020-12-29 Apple Inc. Method for representing virtual information in a real environment
CN106020437A (en) * 2015-03-24 2016-10-12 诺基亚技术有限公司 Augmented reality
EP3073352A3 (en) * 2015-03-24 2016-12-21 Nokia Technologies Oy Augmented reality
EP4130938A1 (en) * 2015-03-24 2023-02-08 Nokia Technologies Oy Augmented reality
US10235812B2 (en) 2015-03-24 2019-03-19 Nokia Technologies Oy Augmented reality
WO2016166138A1 (en) * 2015-04-17 2016-10-20 Thales Method for managing and displaying georeferenced graphic symbols and associated display device
FR3035208A1 (en) * 2015-04-17 2016-10-21 Thales Sa METHOD FOR MANAGING AND DISPLAYING GEO-REFERENCE GRAPHIC SYMBOLS AND ASSOCIATED VISUALIZATION SYSTEM
CN107466360A (en) * 2015-04-17 2017-12-12 泰勒斯公司 Method for managing and displaying geographic coordinate reference graphic symbols and related display device
US20180128642A1 (en) * 2015-04-17 2018-05-10 Thales Method for managing and displaying georeferenced graphic symbols and associated display device
US10235118B2 (en) 2015-08-27 2019-03-19 Fujitsu Limited Augmented reality device and method for providing assistance to a worker at a remote site
EP3136204A3 (en) * 2015-08-27 2017-03-29 Fujitsu Limited Image processing device and image processing method
DE112016005040B4 (en) 2015-11-03 2024-10-10 Motorola Solutions, Inc. Out-of-focus notifications displayed using a head-mounted display
US10489981B2 (en) * 2015-12-10 2019-11-26 Sony Corporation Information processing device, information processing method, and program for controlling display of a virtual object
US20180315246A1 (en) * 2015-12-10 2018-11-01 Sony Corporation Information processing device, information processing method, and program
CN108292448A (en) * 2015-12-10 2018-07-17 索尼公司 Information processing apparatus, information processing method, and program
US10983594B2 (en) * 2017-04-17 2021-04-20 Intel Corporation Sensory enhanced augmented reality and virtual reality device
US20210382548A1 (en) * 2017-04-17 2021-12-09 Intel Corporation Sensory enhanced augemented reality and virtual reality device
US11829525B2 (en) * 2017-04-17 2023-11-28 Intel Corporation Sensory enhanced augmented reality and virtual reality device
US10325414B2 (en) 2017-05-08 2019-06-18 Microsoft Technology Licensing, Llc Application of edge effects to 3D virtual objects
WO2018227502A1 (en) 2017-06-15 2018-12-20 Tencent Technology (Shenzhen) Company Limited System and method of instantly previewing immersive content
EP3639114A4 (en) * 2017-06-15 2021-01-20 Tencent Technology (Shenzhen) Company Limited System and method of instantly previewing immersive content
TWI733245B (en) * 2019-11-07 2021-07-11 南開科技大學 System for switching between augmented reality and virtual reality based on interaction process and method thereof

Also Published As

Publication number Publication date
CN102473324A (en) 2012-05-23
US20200193712A1 (en) 2020-06-18
CN104731337A (en) 2015-06-24
WO2011020793A2 (en) 2011-02-24
EP3675067A1 (en) 2020-07-01
US20200242846A1 (en) 2020-07-30
US20120176410A1 (en) 2012-07-12
US8896629B2 (en) 2014-11-25
WO2011020793A3 (en) 2011-10-27
CN104731337B (en) 2017-12-15
US20230290083A1 (en) 2023-09-14
EP2467833A2 (en) 2012-06-27
US11562540B2 (en) 2023-01-24
CN102473324B (en) 2015-04-29
DE102009037835B4 (en) 2012-12-06
DE102009037835A1 (en) 2011-02-24

Similar Documents

Publication Publication Date Title
US11562540B2 (en) Method for representing virtual information in a real environment
US20210407160A1 (en) Method and sytem for presenting a digital information related to a real object
US10580162B2 (en) Method for determining the pose of a camera and for recognizing an object of a real environment
US10347046B2 (en) Augmented reality transportation notification system
EP3237845B1 (en) System and methods for interactive hybrid-dimension map visualization
CN110084878B (en) Method for representing virtual information in a real environment
US20100287500A1 (en) Method and system for displaying conformal symbology on a see-through display
JP6050518B2 (en) How to represent virtual information in the real environment
US20150310667A1 (en) Systems and methods for context based information delivery using augmented reality
US10602052B2 (en) Transition from display of first camera information to display of second camera information
US20190364224A1 (en) Information processing apparatus, information processing method, and program
US20180290057A1 (en) Dynamic scaling of augmented reality-based game
US11151882B2 (en) Providing a model of a vehicle to a rider at an accurate orientation
US11568616B1 (en) Display apparatuses and methods for facilitating location-based virtual content
KR20240019421A (en) Display device for vehicle and method for controlling the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: APPLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:METAIO GMBH;REEL/FRAME:040821/0462

Effective date: 20161118

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION