[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

WO2012015956A2 - Augmented reality and location determination methods and apparatus - Google Patents

Augmented reality and location determination methods and apparatus Download PDF

Info

Publication number
WO2012015956A2
WO2012015956A2 PCT/US2011/045586 US2011045586W WO2012015956A2 WO 2012015956 A2 WO2012015956 A2 WO 2012015956A2 US 2011045586 W US2011045586 W US 2011045586W WO 2012015956 A2 WO2012015956 A2 WO 2012015956A2
Authority
WO
WIPO (PCT)
Prior art keywords
user interaction
wireless communications
location information
interaction device
communications
Prior art date
Application number
PCT/US2011/045586
Other languages
French (fr)
Other versions
WO2012015956A3 (en
Inventor
Luke Richey
Allen Greaves
Original Assignee
Gravity Jack, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US12/847,771 external-priority patent/US8493206B2/en
Priority claimed from US12/847,754 external-priority patent/US8502659B2/en
Priority claimed from US12/847,790 external-priority patent/US8519844B2/en
Application filed by Gravity Jack, Inc. filed Critical Gravity Jack, Inc.
Publication of WO2012015956A2 publication Critical patent/WO2012015956A2/en
Publication of WO2012015956A3 publication Critical patent/WO2012015956A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/002Specific input/output arrangements not covered by G06F3/01 - G06F3/16
    • G06F3/005Input arrangements through a video camera

Definitions

  • This disclosure relates to augmented reality and location determination methods and apparatus.
  • Computing systems have continually evolved and the popularity of computing systems continues to increase.
  • the advancement of computing systems creates new uses and applications for the computing systems.
  • the processing speeds, storage capacities and network communication speeds are constantly increasing enabling the use of computing systems in increasing numbers of applications.
  • computing systems have evolved from typical office or desk systems to smaller devices, some of which have increased portability, which further expands the possible applications of the computing systems. More specifically, notebook computers have evolved from desktop computers, and more recently, handheld portable devices have also advanced significantly. Personal digital assistants, media players, cellular telephones, smartphones, and other portable devices have increased processing power and storage capacities while communications networks have also been improved allowing greater rates of data transfer between the computing systems.
  • Some computing systems and networks have evolved to a sufficient extent to perform augmented reality operations which augment the physical world with virtual computer-generated imagery in one example.
  • some portable computing systems have sufficient processing, storage and communications capabilities to provide real-time augmented reality data for mobile users.
  • At least some aspects of the disclosure are directed to improved methods, apparatus and programming for implementing augmented reality operations.
  • FIG. 1 is an illustrative representation of a user interaction device implementing augmented reality operations according to one embodiment.
  • Fig. 2 is a functional block diagram of a media system according to one embodiment.
  • Fig. 3 is a functional block diagram of a computing system according to one embodiment.
  • FIG. 4 is a functional block diagram of communications circuitry of a user interaction device according to one embodiment.
  • Fig. 5 is a functional block diagram of a user interface of a user interaction device according to one embodiment.
  • Fig. 6 is a flow chart of a method implemented by a user interaction device to implement augmented reality operations according to one embodiment.
  • Fig. 7 is a flow chart of a method implemented by a management device with respect to a plurality of user interaction devices according to one embodiment.
  • Fig. 8 is a flow chart of a method of outputting wireless communications signals according to one embodiment.
  • Fig. 9 is a flow chart of a method of receiving wireless communications signals according to one embodiment.
  • Fig. 11 is a flow chart of a method implemented by a management device with respect to image recognition operations according to one embodiment.
  • augmented reality operations where the physical world is augmented with additional information, such as virtual objects.
  • images of the physical world observed through user interaction devices may be augmented or enhanced with augmented reality representations, for example in the form of visual and/or audio data which may be experienced by users.
  • augmented reality representations may include virtual objects which augment a user's experience of the physical world.
  • the virtual objects may be associated with physical world objects which may be static or dynamically moving.
  • Some of the described embodiments include a media system configured to implement and co-ordinate or manage augmented reality operations of one user interaction device or a plurality of user interaction devices which may be interacting in a collaborative augmented reality session in one arrangement.
  • Some augmented reality systems use location information regarding locations of the user interaction devices and locations of physical objects in the physical world to accurately augment the physical world with the augmented reality representations.
  • the location information may be used to associate virtual objects with respective objects of the physical world in one illustrative example.
  • At least some aspects of the disclosure are directed towards increasing the accuracy of generated location information regarding the locations of user interaction devices and which location information may be used to implement augmented reality operations.
  • a plurality of different techniques may be available to determine the location information of user interaction devices.
  • information from the different techniques may be utilized and/or combined to provide location information of the user interaction devices of increased accuracy compared with other available location information of the user interaction devices, perhaps obtained from a single source. Additional aspects are described in the following disclosure.
  • a location determination method includes accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, identifying a plurality of wireless communication devices which are proximately located with respect to the user interaction device, initiating wireless communications between the user interaction device and the wireless communications devices, after the initiating, accessing information regarding the wireless communications of the user interaction device and the wireless communication devices, and using the information regarding the wireless communications, determining second location information regarding the location of the user interaction device, and wherein the second location information has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information.
  • a location determination method comprises accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, identifying a marker which is proximately located with respect to the location of the user interaction device, accessing an image generated by the user interaction device which includes the marker, and processing the image to determine second location information regarding the location of the user interaction device, and wherein the second location information has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information.
  • an augmented reality method comprises accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, generating second location information which has increased accuracy regarding the location of the user interaction device in the physical world, and communicating augmented data to the user interaction device, and wherein the augmented data comprises the augmented reality representation.
  • a location determination method comprises using a user interaction device, emitting a wireless communications signal at a first moment in time, using a plurality of wireless communication devices, receiving the wireless communications signal emitted by the user interaction device at a plurality of second moments in time, and using the first and second moments in time, determining information regarding a location of the user interaction device.
  • a computing system comprises communications circuitry configured to implement communications externally of the computing system, and processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information regarding a location of a user interaction device in a physical world, to use the first location information to identify a plurality of wireless communications devices which are proximately located with respect to the location of the user interaction device, to control the communications circuitry to output a control signal which is configured to initiate wireless communications between the user interaction device and the wireless communications devices, to access information regarding the wireless communications between the user interaction device and the wireless communications devices, and to use the information regarding the wireless communications to determine second location information regarding the location of the user interaction device and which has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information.
  • a computing system comprises communications circuitry configured to implement communications externally of the computing system and processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information received by the communications circuitry regarding a location of a user interaction device in a physical world, to identify a marker which is proximately located with respect to the location of the user interaction device, to access a plurality of images generated by the user interaction device, and to process the images with respect to the marker to determine second location information regarding the location of the user interaction device and which has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information
  • a computing system comprises communications circuitry configured to implement communications externally of the computing system and processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information received by the communications circuitry regarding a location of a user interaction device in a physical world, to generate second location information which has increased accuracy regarding the location of the user interaction device in the physical world, and to control the communications circuitry to communicate augmented data comprising an augmented reality representation to the user interaction device.
  • an augmented reality user interaction device comprises a camera, a display system, communications circuitry configured to implement wireless communications externally of the user interaction device, and processing circuitry coupled with the camera, the display system, and the communications circuitry, wherein the processing circuitry is configured to control the display system to generate a plurality of images which comprise image data generated by the camera and augmented data which augments the image data with an augmented reality representation, wherein the processing circuitry is further configured to control the communications circuitry to communicate first location information regarding a location of the user interaction device in a physical world externally of the user interaction device, and wherein the processing circuitry is further configured to access second location information regarding the location of the user interaction device after the outputting of the first location information and to use the second location information to generate the plurality of images, and wherein the second location information has increased accuracy regarding the location of the user interaction device in the physical world compared with the first location information.
  • FIG. 1 illustrates a user interaction device 10 which is used to generate an image of the physical world and which is augmented by an augmented reality representation. More specifically, in the example of Fig. 1 , the user interaction device 10 includes a camera (not shown) which is configured to capture images of the physical world and which may depicted using a display 12. As a user moves the user interaction device 10, a plurality of images are captured of different scenes viewed by the camera of the device 10.
  • the scene viewed by the device 10 includes a marker 14 on a wall of the physical world.
  • the generated image depicted using the display 12 includes an augmented reality representation 18 which augments a user's experience of the physical world by replacing the physical world marker 14 with the representation 18.
  • the augmented reality representation 18 is a virtual 3D object in the form of a puppy, which may be selected by another user to be associated with the marker 14.
  • marker 14 is one example of augmented reality operations which may be implemented using the user interaction device 10 and other augmented reality operations may be implemented in other embodiments.
  • virtual objects may be associated with other physical objects of the physical world, such as other user interaction devices 10 (not shown), in images generated by device 10.
  • augmented reality representations 18 may entirely replace physical objects of the physical world.
  • the augmented reality representations 18 may include advertising objects (e.g., banner with a product name) and the representations 18 may be associated with famous physical structures of the physical world when observed through a user interaction device 10. For example, a user at a significant football game may view a virtual object banner draped between the physical world goalposts when a user of a device 10 captures images of the end zone during a football game. Companies may pay advertising fees to have augmented reality representations of advertisements of their products associated with physical world objects and which may be viewed by users using their user interaction devices 10 who are proximately located to the physical world objects in one embodiment.
  • advertising objects e.g., banner with a product name
  • Location information regarding the locations of the user interaction device 10 and other physical objects in the physical world may be used to generate augmented reality representations 18 in captured images.
  • the location information may be used to depict the augmented reality representations 18 accurately associated with content (e.g., objects) of the physical world (e.g., other user interaction devices, buildings, structures, mountains, etc.).
  • location information may be included with the augmented data which determines where the augmented reality representations 18 are to be displayed with respect to content of the physical world when the static physical object is within the field of view of the camera.
  • the augmented data may be associated with an identifier of the portable physical object. Identification information of the portable physical object and location information of the portable physical object may be used to determine when the portable physical object is present within the field of view of the camera and where augmented reality representations 18 associated with the portable physical object should be shown in generated images. Location information regarding the user interaction devices 10 and/or physical objects may be used to accurately show the augmented reality representations 18 associated with the user interaction devices 10 and/or physical world objects in images generated by the user interaction devices 10 in one embodiment.
  • one user interaction device 10 may be present and the user may be experiencing augmented reality representations with respect to physical objects of the physical world.
  • a plurality of users having user interaction devices 10 may be present and proximately located to one another and experiencing augmented reality representations with respect to one another in a collaborative session and/or physical world objects.
  • the augmented reality representations may be associated with user interaction devices 10 of users and/or with physical objects. Users and user interaction devices 10 may be free to enter and leave interactive augmented reality collaborative sessions in some embodiments.
  • the user interaction devices 10 may be configured to communicate (e.g., wirelessly) with one another as well as with external devices (e.g., a management device, Wi-Fi communications devices) to implement augmented reality operations including operations with respect to determining location information of the user interaction devices 10 of increased accuracy.
  • external devices e.g., a management device, Wi-Fi communications devices
  • Media system 20 is configured to implement operations with respect augmenting the physical world with augmented reality representations.
  • media system 20 is configured to assist user interaction devices 10 with the generation of augmented reality representations.
  • media system 20 is configured to perform operations with respect to determining locations of user interaction devices 10 (which may be portable) for use in accurately associating augmented reality representations with the physical world.
  • media system 20 may communicate augmented data with respect to the user interaction devices 10 which may be used by the devices 10 to generate augmented reality representations.
  • media system 20 includes a plurality of use interaction devices 10 and a management device 22.
  • User interaction devices 10 may be configured to communicate with one another as well as with management device 22.
  • the user interaction devices 10 may communicate with management device 22 via a network 24.
  • Network 24 may be considered to be a part of media system 20 or may be external of media system 20 in different embodiments.
  • the user interaction devices 10 may also implement wireless communications with respect to other wireless communications devices (e.g., Wi-Fi communications devices) which may be within the communications ranges of the devices 10 (the Wi-Fi communications devices are not shown in Fig. 2).
  • the user interaction devices 10 may be proximately located with respect to one another in a group (e.g., within communications ranges of the devices 10 to implement communications with respect to one another) or in different geographical locations and not proximately located to one another. For example, different groups of user interaction devices 10 may exist in different geographical locations. User interaction devices 10 which are proximately located to one another may participate in a collaborative augmented reality session where augmented reality representations may be associated with the devices 10 in one embodiment. Additionally, only one user interaction devices 10 may be present in a given geographical location and may be implementing augmented reality operations with respect to static physical world objects.
  • User interaction devices 10 may be computing systems (e.g., one example is described with respect to Fig. 3) in one embodiment.
  • the user interaction devices 10 may have substantially the same configurations or have different configurations in example embodiments.
  • user interaction devices 10 may be configured as portable media devices, personal digital assistants, cellular telephones, smartphones, personal computers, notebook computers, glasses worn by a user including a camera and display system capable of generating images, or any other device capable of capturing images of the physical world and generating images and/or other media content for consumption by a user which include visual images of the physical world which are augmented by one or more augmented reality representations (e.g., additional virtual image content and/or audible content which augments physical world content).
  • augmented reality representations e.g., additional virtual image content and/or audible content which augments physical world content
  • management device 22 may be a server which is configured as a computing system, for example as described below with respect to Fig. 3.
  • Management device 22 is configured to implement communications with respect to user interaction devices 10 in the described embodiment.
  • Management device 22 may be configured to perform a plurality of operations with respect to the generation of augmented reality representations by the user interaction devices.
  • Example operations performed include operations with respect to co-ordination and management of user interaction devices 10, co-ordination and management of communications between user interaction devices 10 and a plurality of other wireless communications devices, determining locations of user interaction devices 10 (which may be portable), and storing and communicating augmented data for use by the user interaction devices 10 to generate augmented reality representations.
  • the user interaction devices 10 may communicate augmented data of their respective augmented reality representations (e.g., the above-described puppy) to the management device 22, perhaps for storage, and the management device 22 may thereafter provide the augmented data to others of the user interaction devices 10 for use in generating the augmented reality representations with respect to the devices 10 which provided the augmented data.
  • the management device 22 may communicate augmented data which includes the puppy representation to user interaction device 10 which uses the augmented data to generate the augmented reality representation 18.
  • identification data may be used to associate augmented data with respective appropriate objects of the physical world, such as user interaction devices 10 or other physical world objects.
  • initial location information regarding the location of a user interaction device 10 may be used to search a database of the management device 22 to identify other wireless communications devices (e.g., user interaction devices, Wi-Fi communications devices) within a communications range of the user interaction device 10.
  • the database may include location information regarding the devices 10 which the initial location information is searched against, and identification information which uniquely identifies each of the wireless communications devices.
  • the database may also include identification information which identifies the augmented data associated with the user interaction devices 10.
  • the management device 22 may communicate the augmented data associated with one of the devices to the other device and which includes an identifier of the device to which the augmented reality representation of the augmented data is to be associated with.
  • the receiving user interaction device may access location information regarding its present location and the location of the other device (e.g., access the location information from management device 22) and use the location information of the devices, the identification information of the other device, and the respective augmented data of the other device to generate the augmented reality representation 18 associated with the other device 10 in images captured by the device which include the other device.
  • the device 10 may also depict other augmented reality representations associated with other proximately located user interaction devices 10 and perhaps other objects of the physical world which may be present and have associated augmented data. More specifically, some objects of the physical world are static and do not change locations (e.g., marker 14 of Fig. 1 ).
  • management system 22 may store augmented data to be used to generate augmented reality representations with respect to static physical objects and the management system 22 may use location information regarding the device 10 to search the database to identify augmented data for static physical objects which are proximately located to the location of the device 10.
  • the management device 22 may communicate augmented data for a static physical object which is proximately located to the device 10 to the device 10 as well as location information where the static physical object exists in the physical world. Thereafter, the device 10 may use the location information to determine when the static physical object is within the field of view of the camera of device 10 and the device 10 may augment captured images of the static physical object with augmented reality representations using the appropriate augmented data.
  • Additional operations of the user interaction devices 10 with respect to generation of augmented reality representations may be managed by management device 22 in one embodiment.
  • user interaction devices 10 which attempt to access and utilize media system 20 to experience or implement operations with respect to augmented reality may be managed by management device 22.
  • management device 22 may manage an augmented reality session where a plurality of user interaction devices 10 are present and performing augmented reality operations with respect to one another.
  • media system 20 may be a closed system and user interaction devices 10 may first be registered and/or otherwise authorized prior to having permission to gain access to media system 20.
  • management device 22 may assign passwords to the devices 10 and use the passwords and identification information of the devices 10 to verify that the user interaction devices 10 are authorized to gain access to the media system 20.
  • users may initially sign-up to gain access to the media system 20, and after receiving a password, may subsequently participate in augmented reality sessions or otherwise access the media system 20 to implement augmented reality operations. Users may pay fees to participate in services provided by media system 20, for example, including services of accessing management device 22 and implementing augmented reality operations with respect to other user interaction devices 10 managed by management device 22. Access to or participation in augmented reality operations provided by media system 20 and management device 22 may be denied to users who cannot provide appropriate identification and password information in one embodiment.
  • management device 22 may perform operations with respect to determination of location information of the user interaction devices 10 which may be used to implement augmented reality operations with respect to the user interaction devices 10. Some of the described location determination embodiments include implementing wireless communications between the user interaction devices 10 and/or other wireless communications devices which may be present to determine the location information. Management device 22 may be configured to manage and/or coordinate at least some of the communications of the user interaction devices 10 in one example. In addition, management device 22 may be configured to implement operations with respect to image processing to determine the location information in additional embodiments described below. Other operations may be performed by management device 22 with respect to augmented reality and some of the operations performed by the management device 22 may be performed by other devices (e.g., the user interaction devices themselves) in other embodiments.
  • Network 24 which may be implemented as any suitable network configuration for implementing wired and/or wireless communications between user interaction devices 10, management device 22 or other wireless communications devices.
  • network 24 may include a wireless network, local or wide area networks, Internet, cellular network, and/or other suitable infrastructure for implementing communications of digital information.
  • user interaction devices 10 may have cellular communications capabilities to communicate with network 24 and management device 22.
  • Network 24 may implement packet-switched communications between user interaction devices 10 and management device 22 in one embodiment.
  • user interaction devices 10 may also communicate directly with one another or other wireless communications devices without use of network 24 in one embodiment.
  • FIG. 3 one example embodiment of a computing system 30 is shown.
  • the illustrated system 30 includes communications circuitry 32, processing circuitry 34, storage circuitry 36, a user interface 38, and a camera 40.
  • Other embodiments of computing system 30 are possible including more, less and/or alternative components.
  • the illustrated configuration of computing system 30 may correspond to one of user interaction devices 10 while some of the illustrated components (e.g., camera 40) may be omitted in one arrangement of computing system 30 which is implemented as management device 22.
  • Communications circuitry 32 is arranged to implement communications of computing system 30 with respect to external devices or systems (e.g., network 24 and other computing systems 30 implemented as other user interaction devices 10, Wi-Fi communications devices, or management device 22). Communications circuitry 32 may be configured to implement wired and/or wireless communications. Additional details of one example of communications circuitry 32 which may be utilized is discussed below with respect to Fig. 4. [0052] In one embodiment, processing circuitry 34 is arranged to process data, control data access and storage, issue control signals or commands, and control other augmented reality operations. Processing circuitry 34 may comprise circuitry configured to implement desired programming provided by appropriate computer-readable storage media in at least one embodiment.
  • processing circuitry 34 may be implemented as one or more processor(s) and/or other structure configured to execute executable instructions including, for example, software and/or firmware instructions.
  • Other exemplary embodiments of processing circuitry 34 include hardware logic, PGA, FPGA, ASIC, state machines, and/or other structures alone or in combination with one or more processor(s). These examples of processing circuitry 34 are for illustration and other configurations are possible.
  • Storage circuitry 36 is configured to store programming such as executable code or instructions (e.g., software and/or firmware), electronic data, databases, image data, augmented data, identifiers, location information and/or other digital information and the storage circuitry 46 may include computer-readable storage media. At least some embodiments or aspects described herein may be implemented using programming stored within one or more computer-readable storage medium of storage circuitry 36 and configured to control appropriate processing circuitry 34.
  • programming such as executable code or instructions (e.g., software and/or firmware), electronic data, databases, image data, augmented data, identifiers, location information and/or other digital information
  • the storage circuitry 46 may include computer-readable storage media. At least some embodiments or aspects described herein may be implemented using programming stored within one or more computer-readable storage medium of storage circuitry 36 and configured to control appropriate processing circuitry 34.
  • the computer-readable storage medium may be embodied in one or more articles of manufacture which can contain, store, or maintain programming, data and/or digital information for use by or in connection with an instruction execution system including processing circuitry 34 in the exemplary embodiment.
  • exemplary computer-readable storage media may include any one of physical media such as electronic, magnetic, optical, electromagnetic, infrared or semiconductor media.
  • Some more specific examples of computer-readable storage media include, but are not limited to, a portable magnetic computer diskette, such as a floppy diskette, a zip disk, a hard drive, random access memory, read only memory, flash memory, cache memory, and/or other configurations capable of storing programming, data, or other digital information.
  • User interface 38 is configured to interact with a user including conveying data to a user (e.g., displaying visual images for observation by the user) as well as receiving inputs from the user, for example, via a graphical user interface (GUI).
  • GUI graphical user interface
  • User interface 38 may be configured differently in different embodiments. One example embodiment of user interface 38 is discussed below with respect to Fig. 5.
  • Camera 40 is configured to capture images within its field of view and generate image data of scenes of the physical world viewed by the computing system 30 in one embodiment.
  • An example camera 40 includes an appropriate imaging sensor configured to generate digital image data responsive to received light in one implementation.
  • Movement/orientation circuitry 42 is configured to provide information regarding movement and orientation of the computing system 30 in the described embodiment.
  • circuitry 42 may include an accelerometer arranged to provide information regarding forces which the computing system is subjected to.
  • Circuitry 42 may also include a compass and inclinometer configured to provide information regarding an orientation of the computing system 30 in the physical world.
  • communications circuitry 32 of computing system 30 which is implemented as one of user interaction devices 10 is shown according to one possible embodiment.
  • communications circuitry 32 is configured to implement wired and/or wireless communications.
  • communications circuitry 32 includes network communications circuitry 50, GPS communications circuitry 52, and local communications circuitry 54.
  • Other configurations of communications circuitry 32 are possible including more, less and/or alternative components.
  • a communications interface e.g., USB port, NIC interface
  • implementing wired communications may also be provided.
  • network communications circuitry 50 is configured to implement wireless communications with respect to network 24.
  • network communications circuitry 50 is configured to communicate with cell towers of network 24 which includes a wireless cellular network. Accordingly, network communications circuitry 50 may implement wireless communications over relatively large distances (e.g., miles) in at least one embodiment.
  • GPS communications circuitry 52 is configured to implement wireless communications with external GPS transmitters (e.g., satellites).
  • GPS communications circuitry 52 may also include appropriate circuitry to provide Assisted GPS (AGPS) or Wireless Assisted GPS (WAGPS) which utilizes additional location information to provide location information of increased accuracy compared with use of GPS alone.
  • AGPS Assisted GPS
  • WAGPS Wireless Assisted GPS
  • GPS communications circuitry 52 may provide location information of computing system 30 directly in one embodiment or processing circuitry 34 may be configured to process signals received by GPS communications circuitry 52 to provide the location information of the computing system 30 in another embodiment.
  • Local communications circuitry 54 is configured to implement wireless communications with respect to local wireless communications devices (e.g., other user interaction devices 10, Wi-Fi communications devices) which are proximately located to the computing system 30.
  • local communications circuitry 54 is configured to implement Bluetooth, ad hoc or other suitable wireless communications with respect to the local wireless communications devices.
  • FIG. 5 one example embodiment of a user interface 38 of a user interaction device 10 is shown.
  • the example user interface 38 includes an input device 60, display system 62, speaker 64 and microphone 66.
  • Other arrangements of user interface 38 are possible including more, less and/or additional components.
  • Input device 60 is arranged to receive inputs of a user interacting with computing system 30.
  • Input device 60 may include a graphical user interface, keyboard, pointing device (e.g., mouse) or other suitable apparatus to permit a user to input information and interact with the computing system 30.
  • Display system 62 is configured to generate visual images which may be viewed by the user.
  • Display system 62 may depict images captured by camera 40, display digital content (e.g., windows, email, etc.) and display augmented reality representations in some examples.
  • Speaker 64 is configured to emit sound waves in one embodiment.
  • the sound waves may include voice information, audio information and other sound waves which are discussed further below.
  • Microphone 66 is a configured to receive sound waves in one embodiment. Received sound waves may include voices and sound waves emitted by other computing systems 30 in some illustrative examples.
  • user interaction devices 10 are configured to implement operations to enable users of the devices 10 to experience augmented reality.
  • the user interaction devices 10 may participate in augmented reality sessions which include a plurality of the devices 10.
  • a single user interaction device 10 may implement operations for its user to experience augmented reality without the presence of other user interaction devices 10 (e.g., associating augmented reality representations with physical objects present in the physical world).
  • the user interaction devices 10 may access management device 22 which may implement, manage and/or co-ordinate augmented reality operations for a single user interaction device 10 or a plurality of user interaction devices 10 (which may be participating in an augmented reality collaborative session with one another).
  • Accurate location information regarding locations of one or more user interaction device 10 is desired to improve the experience of augmented reality by the users.
  • the location information may be used by the user interaction device 10 to correctly and accurately associate augmented reality representations with respect to physical objects of the physical world.
  • the augmented reality representations may be correctly associated with physical objects of the physical world when sufficiently accurate location information regarding the locations of the user interaction devices 10 and physical objects are used to implement augmented reality operations.
  • Less accurate information regarding the locations of user interaction devices 10 may result in less than desirable augmented reality experiences.
  • an augmented reality symbol such as a puppy, which is to be associated with a respective user and user interaction device 10 may be associated with improper user interaction devices, improper physical world objects, or otherwise not correctly depicted as being associated with the appropriate user interaction device 10.
  • the physical objects which augmented reality representations are to be associated with may be portable (e.g., user interaction devices 10) whose locations may be constantly changing while augmented reality operations are being implemented.
  • the locations of the portable physical objects are determined in sufficient real time to enable the augmented reality representations to be appropriate associated with the portable physical objects (i.e., enable the augmented reality representations to be depicted near or upon the appropriate physical objects in real time).
  • portable physical objects may move indoors and it is also desirable to enable augmented reality operations to be implemented indoors or at other locations where location information of the portable physical objects obtained via conventional sources (e.g., GPS) may not be sufficiently accurate.
  • media system 20 including user interaction devices 10 and management device 22 may be arranged to implement location determination operations to provide sufficiently accurate location information regarding the user interaction devices 10 and other physical objects of the physical world to provide satisfactory experiences of augmented reality to users of the user interaction devices 10.
  • a plurality of different methods for providing location information of the physical objects may be available for use.
  • the location results of the different methods may be combined to further improve the accuracy of the location information.
  • less than all of the methods may be available for use, but the available methods may provide improved results of increased accuracy of the locations of the user interaction devices 10 or other physical objects compared with reliance upon conventional methodologies (e.g., GPS).
  • FIG. 6 one example process of determining and using location information of one or more user interaction devices 10 is shown.
  • the method may be implemented by processing circuitry of a user interaction device which has been authorized to participate in augmented reality operations of the media system. Additional methods are possible including more, less and/or alternative acts.
  • a user interaction device may determine initial location information regarding its present location at an initial moment in time, for example when the user interaction device accesses the media system to implement augmented reality operations.
  • the user interaction device may use location determination circuitry to provide the location information using aGPS.
  • the initial location information determined by the user interaction device may have unacceptable error to implement accurate augmented reality operations with respect to the physical world depending upon the location of the user interaction device (e.g., indoors or otherwise located where reception of GPS communications may be relatively poor).
  • Other techniques may also be utilized, for example, monitoring movements of the user interaction device from a last known position, for example, using output of movement/orientation circuitry 42.
  • the user interaction device may communicate the initial location information to the management device of the media system.
  • the initial location information may be aGPS data with an associated error reading.
  • the user interaction device may also include a timestamp corresponding to the time when the location information was obtained and identification information which uniquely identifies the user interaction device.
  • the management device of the media system may utilize the initial location information to perform operations to attempt to determine refined location information which has increased accuracy with respect to the actual location of the user interaction device in the physical world compared with the initial location information.
  • the management device may manage communications of the user interaction device with respect to other wireless communications devices (e.g., other user interaction devices and Wi-Fi communications devices) to provide refined location information according to one method and/or implement image recognition operations to provide the refined location information according to another method in illustrative example embodiments.
  • the user interaction device monitors for the reception of a communication from the management device.
  • the communication indicates whether a sufficient number of wireless communications devices are proximately located to the user interaction device to implement operations with respect to the devices to attempt to provide the refined location information or the communication indicates that an insufficient number of wireless communications devices are present to implement the operations.
  • a threshold distance e.g., corresponding to a local wireless communications range of the user interaction device which provided the initial location information in one example
  • the management device may maintain a database of locations of the wireless communications devices and the management device may use the initial location information to search the database to identify the presence of other wireless communications devices which may be proximately located to the user interaction device to implement wireless communications with the user interaction device.
  • the user interaction device may enter a device waiting state or mode as a waiter as described further below.
  • the method may proceed to an act A20 to implement additional operations with respect to image recognition as discussed in detail below.
  • act A16 the user interaction device implements localized communications with respect to the other proximately located wireless communications devices which were identified to be sufficiently close to the user interaction device for wireless communications.
  • the user interaction device may be controlled to be in different operational states as a transmitter and receiver of wireless communications signals at different moments in time to implement the wireless communications. Additional details of these communications and different operational states of the user interaction devices are discussed below with respect to the example embodiments of Figs. 7-9.
  • information regarding the wireless communications between the proximately-located wireless communications devices may be communicated by the wireless communications devices participating in the communications to the management device.
  • the management device may process the information regarding the wireless communications in an attempt to provide refined location information for the wireless communications devices including the user interaction device which provided the initial location information at act A10. Additional details regarding the processing to determine the refined location information using the information regarding the wireless communications is described in further detail below with respect to one example embodiment of Figs. 10a-10c.
  • the locations of the one or more Wi-Fi communications devices may be fixed and the known locations of these static communications devices may also be used to determine the location information of the user interaction device.
  • the usage of the location information of these devices may provide increased accuracy compared to communications which do not utilize a static device since the location information of the static device may be known with a relatively high accuracy compared with locations of portable devices.
  • the user interaction device which provided the initial location information may proceed to implement operations with respect to images in an attempt to determine additional information regarding its location as described further below.
  • the user interaction device monitors for the reception of a communication from the management device indicating that one or more markers are proximately located to the user interaction device (e.g., within visible range of the optics of the camera user interaction device).
  • the management device may maintain a database of visual markers and their respective locations in the physical world and the management device may use the initial location information (as well as other location information resulting from the wireless communications) to search the database to identify markers which are located sufficiently close to the user interaction device to be captured by the camera of the user interaction device.
  • the process proceeds to an act A26 where the location information for the user interaction device may be updated, for example, using refined location information from the management device (e.g., based upon the wireless communications with other wireless communications devices).
  • the refined location information may be stored as the location of the respective user interaction device in a database maintained by the management device and the refined location information may be communicated to one or more of the user interaction devices.
  • the process proceeds to an act A22 where the user interaction device which provided the initial location information captures images of the physical world about the user interaction device.
  • the images may be processed by the user interaction device in attempt to provide the refined location information.
  • the management device may communicate images of the markers (e.g., as well as information regarding locations in the physical world from which the images were captured) which are proximately located to the user interaction device to the user interaction device.
  • the markers may be displayed to a user to inform the user of the presence and types of nearby markers.
  • the user interaction device may implement image processing techniques to analyze images captured by the user interaction device to identify whether the markers are present in the images. For example, object recognition processing may be used in one arrangement to identify the markers. Once the markers are identified, the images captured by the user interaction device may be processed with respect to the images of the markers received by the management device to identify location information regarding the locations of the user interaction device from which the images of the markers were captured by the user interaction device. This analysis may compare differences of the images and use the comparison with the known location information from which the images of the markers provided by the management device were captured. This determined location information may have increased accuracy of the location of the user interaction device compared with other available location information regarding the location of the user interaction device. Additional details regarding processing of images captured by the user interaction device are described below with respect to Fig. 11 in one embodiment.
  • image data of images captured by the user interaction device may be communicated to the management device for processing to provide the determined location information.
  • the markers which are in the vicinity of the user interaction device may be communicated to the user interaction device to inform the user of available markers and thereafter the user may capture images of the markers and communicate the captured images to the management device for processing as described above in act A24.
  • the markers may not be communicated to the user interaction device and the user interaction device may upload captured images to the management device for processing to recognize the presence of any markers and determine the location information. Other methods are possible.
  • the refined location information determined by the wireless communications with other wireless communications devices and/or information obtained by image processing techniques may be used to update the location information of the user interaction device.
  • the refined location information may have increased accuracy of the location of the user interaction device in the physical world compared with other available location information (e.g., GPS signals received by the user interaction device located in an indoors environment).
  • the user interaction device and/or management device may use the refined location information for various purposes.
  • the user interaction device and/or management device may utilize the refined location information to implement augmented reality operations.
  • the user interaction device may utilize the refined location information to generate images which include augmented reality representations associated with physical world content.
  • the refined location information may be used to determine where the augmented reality representations will be depicted in the generated images of the physical world and to be viewed in real time by the user of the user interaction device in one embodiment.
  • the management device may use the refined location information of a user interaction device 10 to identify markers or other user interaction devices which may be proximately located to the refined location information of the user interaction device 10.
  • a user interaction device 10 may wirelessly communicate with other wireless communications devices which may be proximately located to the device 10 in attempts to determine refined location information regarding the actual location of the device 10 in the physical world with increased accuracy compared with other available location determination methods.
  • An example implementation of using wireless communications to determine refined location information is described in additional detail with respect to Figs. 7-9.
  • the methods may be performed to determine refined location information using wireless communications between the wireless communications devices (e.g., user interaction devices, Wi-Fi communications devices) which are within wireless communications range of one another. More specifically, the method of Fig. 7 controls the states or modes of operation of the wireless communications devices as waiters, senders and receivers in one embodiment.
  • the wireless communications devices may individually operate as a waiter, sender and receiver at different moments in time. At one moment in time, one of the wireless communications devices may operate as a sender to transmit a wireless communications signal and the other devices operate as receivers of the signal. At subsequent moments in time, the other wireless communications devices may individually operate as senders while the non-transmitting ones of the devices operate as receivers of the transmissions.
  • a method of controlling the states or modes of operation of the wireless communications devices is described according to one embodiment. Other methods are possible including more, less and/or alternative acts. The method may be executed by processing circuitry of the management device in one implementation.
  • the management device manages the wireless communications in one embodiment.
  • a plurality of wireless communications devices may be within a wireless communications range of one another and the devices may implement wireless communications with respect to one another.
  • the wireless communications may be used to determine refined location information of the devices with increased accuracy compared with the initial location information (e.g., GPS data) of the devices.
  • the wireless communications devices may communicate initial location information and time information to the management device 22 and the management device 22 may use the location information and time information to search a database to identify the wireless communications devices which are within range of one another and capable of communicating with one another.
  • the management device determines the states or modes of operations wireless communications devices at different moments in time during the wireless communications.
  • the management device may instruct one of the wireless communications to operate as a sender of a wireless communications signal at one moment in time and instruct others of the wireless communications devices to operate as receivers to receive the wireless communications signal.
  • the management device may instruct others of the wireless communications devices to individually operate as the senders and the non- transmitting devices may be instructed to operate as receivers.
  • the results of the wireless communications may be analyzed to provide refined location information regarding the locations of the wireless communications devices in the physical world.
  • the management device may instruct the wireless communications devices to operate as waiters where no wireless communications are implemented until a sufficient number of wireless communications (e.g., three) are within communications range of one another for the location determination operations to occur.
  • a sufficient number of wireless communications e.g., three
  • initial location information is accessed from a user interaction device.
  • the user interaction device communicates GPS location coordinates, accuracy information regarding the GPS data regarding the location of the user interaction device and a unique identifier to the management device.
  • the user interaction device may also communicate accelerometer information (i.e., if the user interaction device is configured with an on-board accelerometer) which may be used to modify the initial location information by an offset according to movement detected by the accelerometer.
  • accelerometer information i.e., if the user interaction device is configured with an on-board accelerometer
  • a timestamp regarding the initial location information may also be recorded in one embodiment.
  • the timestamp and initial location information for the user interaction device may be stored within a database of the management device as discussed below in one embodiment.
  • the processing circuitry accesses the database of the management device including information regarding wireless communications devices.
  • the database may include respective location information for a plurality of user interaction devices which are interacting with the management device, and perhaps implementing augmented reality operations. Furthermore, the database may also maintain location information for a plurality of additional wireless communications devices, such as Wi-Fi communications devices.
  • act A110 it is determined whether the length of time from the last entry to the new information exceeds a threshold. [0099] If the condition of act A110 is affirmative, the mapped entry for the user interaction device is reset with the received initial location information at an act A112.
  • the received initial location information may be disregarded and the stored location information for the user interaction device may be used.
  • the management device may maintain a database including location information for a plurality of wireless communication devices (e.g., user interaction devices, Wi-Fi communications devices).
  • the management device may search the database to determine whether a sufficient number of wireless communications devices are within a threshold distance (e.g., wireless communications range) of the user interaction device. It is desired that at least three wireless communications devices be able to send and receive wireless communications with respect to one another in one embodiment. The accuracy of the location information determined from wireless communications of the devices increases as the number of communicating devices within communications range of one another increases.
  • the accuracy of the information is increased if one or more of the wireless communications devices have a known static location in the physical world. If a sufficient number of devices were not located in act A114, the management device may output a control signal to instruct the user interaction device at an act A116 to become a waiter until a sufficient number of devices are proximately located to the user interaction device, for example, as determined by the management device.
  • the management device determines whether the user interaction device has recently operated as a sender at an act A118. For example, the management device may determine whether the user interaction device has previously operated as a sender with the other wireless communications devices which are currently within the communications range of the user interaction device.
  • the management device proceeds to an act A122 to determine whether any of the other wireless communications devices which are proximately located to the user interaction device may become a sender.
  • the management device outputs a control signal instructing the user interaction device which provided the initial location information to become a sender at an act A124.
  • the user interaction device may output a wireless communications signal which may be received by other wireless communications devices operating as receivers and the results of the communication may be used to provide refined location information regarding the location of the user interaction device in one embodiment.
  • the management device may output a control signal at an act A126 instructing the user interaction device to be a receiver which receives wireless communications signals from senders as discussed in detail below.
  • the management device may also output a common timing reference to the senders and receivers which the senders and receivers may use to timestamp moments in time when the wireless communications signal is transmitted and received.
  • the timing reference may be the current time as determined by the management device and the senders and receivers may align their internal timing references to this time in one embodiment.
  • the management device may implement operations with respect to the user interaction devices.
  • the management device may communicate with the user interaction devices operating as senders and receiver in the example embodiments of Figs. 8 and 9.
  • a wireless communications device e.g., user interaction device
  • Other methods are possible including more, less and/or alternative acts.
  • wireless communications between a plurality of wireless communications devices may be used to provide refined location information regarding the wireless communications devices.
  • one of the wireless communications devices operating as a sender emits a wireless communications signal which may be received by other wireless communications devices within a wireless communications range of the sender.
  • the sender may output a plurality of different types of wireless communications signals in the described arrangement.
  • the sender emits an electromagnetic wireless communications signal (e.g., Bluetooth, ad hoc wireless communications, or other signal).
  • the sender emits a sound wave (e.g., a sound wave having a frequency which may be outside of the audible range of humans).
  • the communicated signals may be received by the receivers and used to provide refined location information in one embodiment.
  • the wireless communications devices may communicate plural different types of signals (e.g., electromagnetic signals as well as sound waves).
  • a user interaction device which has been selected to be a sender communicates initial location information regarding its current location which is received by the management device.
  • the management device accesses the initial location information.
  • the management device creates a unique identifier for the user interaction device which acts as the sender. If sound waves are to be communicated, the management device may select (e.g., randomly in one embodiment) a unique frequency from a range of possible frequencies as the identifier which may be used. In other examples, the identifier may be an oscillation of sound frequencies or a Bluetooth server name which the device uses to transmit a Bluetooth signal.
  • the management device searches for other wireless communications devices which are proximately located to the user interaction device to act as receivers (if the receivers are not known) or the management device accesses a list of the proximately located devices which may act as receivers (if known).
  • the management device communicates the unique identifier of the sender to the receivers.
  • the unique identifier may also include signal identification information which enables the receivers to identify the communications from the sender (e.g., identifies the sender of a Bluetooth communications signal, identifies the frequency of a sound wave from the sender, etc.).
  • the management device receives a plurality of responses from the receivers indicating that they have received the unique identifier and signal identification information and they are ready to receive the wireless communications signal to be emitted from the sender.
  • the management device communicates the unique identifier to the sender indicating that the receivers are ready and the management device may communicate a control signal to initiate wireless communications.
  • the management device may also specify the frequency of a sound wave to be emitted if the devices are using sound waves.
  • the sender outputs the wireless communications signal (e.g., electromagnetic wave or sound wave) and records the time the signal is outputted.
  • the wireless communications signal e.g., electromagnetic wave or sound wave
  • the sender may use appropriate communications circuitry to output the signal.
  • a speaker of the sender may be used to output the signal.
  • the timestamp indicative of the time that the signal was outputted by the sender is communicated to the management device.
  • the management device accesses the timestamp regarding the outputting of the signal.
  • the management device accesses a plurality of communications from the receivers indicating the respective times the wireless communications signal which was outputted by the sender was received by the respective receivers.
  • the wireless communications signal outputted by the sender may be received by the receivers at different times depending upon their distances from the sender.
  • the management device uses the information regarding the lengths of time between the outputting of the wireless communications signal and the reception of the signal by the plural receivers to determine linear distances between the sender and the receivers using the respective lengths of time of communications with respect to the receivers which may be multiplied by the velocity of the wireless signals which were communicated (e.g., sound, electromagnetic).
  • the determined linear distances may be used to determine refined location information of the user interaction devices as discussed in one embodiment below with respect to Figs. 10a-10c.
  • the above-described process may be repeated a number of times between two devices and the resultant distances may be averaged in one embodiment.
  • the management device may also output a completion signal to the wireless communications devices indicating that it has obtained all information regarding the communication (i.e., transmission and reception) of the wireless communications signal.
  • the completion signal may instruct the sender to cease sending the wireless communications signal.
  • the sender receives the completion signal from the management device and ceases outputting of the wireless communications signal.
  • the sender outputs a query to determine the next desired state of operation of the wireless communications device as a receiver or waiter.
  • the sender receives a response to the query from the management device in the form of a control signal to enter an operational state as a receiver or a waiter.
  • the sender changes its operational state to be a receiver for subsequent communications from other proximately located wireless communications devices or a waiter in accordance with the received response to the query.
  • a wireless communications device e.g., user interaction device
  • Other methods are possible including more, less and/or alternative acts.
  • the user interaction device receives a unique identifier from the management device which identifies wireless communications to be received from the sender.
  • the unique identifier may identify the sender of Bluetooth signal or may identify a frequency of a sound wave and the sender of the sound wave.
  • the receiver prepares to receive the signal and communicates a ready signal to the management device.
  • the receiver waits for reception of the wireless communications signal to be communicated by the sender.
  • the receiver compares an amount of time waiting for the signal with a time interval to determine whether the time interval has passed.
  • the user interaction device continues to wait for reception of the wireless communications signal if the time interval has not passed.
  • act A206 If the result of act A206 is affirmative, the receiver proceeds to an act A208 to output a notification to the management device that the device failed to receive the wireless communications signal.
  • the receiver queries the management device of whether to operate as a receiver, sender or waiter.
  • the receiver During the waiting for reception of the wireless communications signal, the receiver continually monitors for reception of the signal. If the wireless communications signal is detected at an act A212, the receiver proceeds to an act A214 to timestamp the time when the wireless communications signal was received. The receiver returns to act A204 to wait for the signal if the result of act A212 is negative.
  • the receiver communicates the timestamp of the reception of the wireless communications signal to the management device.
  • the receiver may also include its respective identifier which allows the management device to identify the receiver which received the signal.
  • the receiver receives instructions from the management device which instructs the device to operate as a receiver, sender or waiter during subsequent communications of additional wireless communications signals. [0137] At an act A220, the receiver enters the specified state of operation as a receiver, sender or waiter.
  • the communicated wireless communications signals used to determine refined location information may include different types of signals (e.g., electromagnetic waves, sound waves).
  • signals e.g., electromagnetic waves, sound waves.
  • the above-recited methods of Figs. 8 and 9 with respect to transmission and reception of signals may be implemented for the different types of signals. More specifically, the methods of Figs. 8 and 9 may be performed to implement communications of electromagnetic signals at one moment in time and the methods may again be performed at another moment in time for communications of sound waves.
  • some of the wireless communications may be implemented automatically without control of a user while others of the wireless communications may be implemented as a result of user control.
  • the user interaction devices may be configured to initiate and implement Bluetooth communications automatically without user control while sound wave communications may be initiated as a result of user control or instruction.
  • the use of sound waves may provide refined location information of improved accuracy in some embodiments compared with communications of electromagnetic waves since sound waves travel slower than electromagnetic waves.
  • the duration of the times of sound wave communications between the sending wireless communications device and the receiving wireless communications devices may be multiplied by the velocity of sound (e.g., 340.29 m/s) to determine the distances of the respective receiving wireless communications devices with respect to the sending wireless communications device.
  • the speed of light may be used in calculations for the wireless communications of electromagnetic waves to determine the distances of the receivers with respect to the sender. The determined distances may be used as discussed below to provide refined location information.
  • the wireless communications devices may be Wi-Fi communications devices which are arranged to wirelessly communicate with other wireless communications devices including the user interaction devices.
  • Typical Wi-Fi communications devices are not portable but are rather implemented in fixed static locations (e.g., hot spots).
  • the management device may store the locations of Wi-Fi communications devices and may search for the presence of these devices which may be within a wireless communications range of one of more user interaction device which is implementing operations with respect to augmented reality described herein.
  • Wi-Fi communications devices which are in static locations is advantageous since accurate information regarding these devices may be stored in the management device and used to accurately locate other wireless communications devices (e.g., user interaction devices). More specifically, since the location information of these devices may include reduced error compared with available location information of portable devices, the location information of the static devices may be used to provide refined location information which may have increased accuracy with respect to the portable devices compared with operations implemented solely between portable devices to provide the refined location information.
  • portable user interaction devices may calculate their respective distances to static wireless communications devices which may result in reduced error in the determined locations of the portable user interaction devices since the error of the locations of the static wireless communications device in the physical world may be less compared with error present in available location information of the portable communications devices which would otherwise be used.
  • the location of one static device may be known with an increased degree of accuracy and accordingly the processing of information regarding communications with this device may provide location information with an increased degree of accuracy compared with processing of communications with devices which may all be dynamically moving.
  • the database of the management device may be populated with identification and location information regarding a plurality of Wi-Fi communications devices.
  • users of an augmented reality community may upload location information to the management device of the Wi-Fi communications devices.
  • Wi-Fi communications devices may be programmed with their static locations upon installation in a facility and the information regarding the static locations may be provided to the management device.
  • user interaction devices having accurate information regarding their present location may be positioned adjacent to the Wi-Fi communications devices and the location of the user interaction device may be uploaded with the identification information of the Wi-Fi communications devices to the management device. Any suitable arrangement may be used to provide the location information regarding the Wi-Fi communications devices to the management device and which may be stored therein for use in subsequent searching operations.
  • These Wi-Fi communications devices populated in the database of the management device may be subsequently searched at a later moment in time using initial location information of another user interaction device.
  • the identified Wi-Fi communications devices which are proximately located to one or more user interaction device may be used to implement communications with respect to the user interaction devices as discussed above to provide refined location information of the user interaction devices and which may have increased accuracy compared with location information obtained by other methods, such as GPS.
  • the management device determines the refined location information using information regarding the wireless communications which are implemented between the wireless communications devices.
  • the receiving wireless communications devices e.g., user interaction devices, Wi-Fi communications devices
  • the sender may use the received information to determine its refined location information using techniques described in further detail below (as opposed to having management device perform the calculations to determine the refined location information).
  • wireless communications signals including sound waves may be utilized in attempts to improve the accuracy of the refined location information of a plurality of wireless communications devices.
  • the following is a discussion of processing of sound waves received by a wireless communications device to identify wireless communications signals which were emitted by the sending wireless communications device according to one embodiment.
  • the receivers may receive a predefined frequency (i.e., the frequency is defined before the communication of the signal) from the management device and the receivers may process received sound waves in attempts to determine whether the predefined frequency is present in the received sounds waves.
  • the described processing may be performed upon output of a microphone of a receiving user interaction device in one embodiment.
  • the received sound may be converted into a function which can be tested to determine if the specified frequency of the communicated wireless communications signal is present in the received domain.
  • f(x) is the received sound at time x and f(cp) is the Fourier Transform for which ⁇ is the frequency in hertz being checked against. If a strong frequency ⁇ exists within the ambient noise, then f(c ) will present a discernable maxima within the graphed f(cp).
  • f(cp) uses a complex integration to determine the entire domain.
  • a Riemann's Sum may be used to estimate this value as closely to f(cp) as possible. This is known as a Discrete Fourier Transform where samples will be taken from the noise in the room at a given interval. These samples may be reduced to:
  • x n would represent the amplitude of the ambient sound at a sample (time) n, and x k represents the magnitude at frequency k.
  • received sound may be sampled at a rate of approximately 44100 samples per second.
  • the samples may be partitioned into different groups and the processing of Eqn. 2 may be performed on the groups separately.
  • a resolution may be computed by dividing the sampling rate of 44100 samples per second by the number of samples obtained (e.g., a number of obtained samples of 32768 provides a resolution of 1.34).
  • the quotient of the frequency of the wireless communications signal may be divided by the resolution to provide the value of the result of the discrete Fourier Transform if the sound wave is present in the sound received by the wireless communications device.
  • a threshold for the magnitude of frequency may be specified and the output of the processing for the frequency of interest may be compared with this threshold. As an example, if the frequency of interest with the ambience is
  • the operations of determining location using sound waves may be implemented at times specified by the user for example when the presence of interfering sounds is low.
  • a user interaction device is configured to receive a GPS signal which provides the device's Cartesian coordinates as well as an accuracy radius which defines an accuracy circle about the coordinates where the user interaction device may exist.
  • the example method described below reduces the size of the accuracy circle (e.g., the area of possible locations the user interaction device may be with respect to the provided coordinates) to provide refined location information regarding the location of the user interaction device with increased accuracy.
  • a plurality of user interaction devices 100, 102 are shown at different locations having respective coordinates as well as respective accuracy circles 101 , 103 based upon GPS signals received by the user interaction devices 100, 102.
  • a line 106 is illustrated which corresponds to a linear distance between the user interaction devices 100, 102 which was determined from processing of wireless communications between user interaction devices 100, 102 as described above.
  • the endpoints of the line 106 are processed with respect to the accuracy circles 101 , 103 to reduce the areas of the accuracy circles 101 , 103 where the user interaction devices 100, 102 may be located and to provide the refined location information regarding the locations of devices 100, 102 of increased accuracy compared with accuracy circles 101 , 103.
  • the user interaction devices 100, 102 exist within their respective accuracy circles 101 , 103 and the devices 100, 102 are spaced a distance of the line 106 apart from one another.
  • the method reduces the size of one of the accuracy circles 101 , 103 using the line 106 and the other of the accuracy circles 101 , 103.
  • a circle 107 is generated having a radius equal to the length of line 106.
  • the accuracy of circle 103 may be reduced by identifying the union or overlap of circle 107 with circle 103 for all possible locations of device 100 within circle 101 .
  • an end point of line 106 may be placed on the edge of circle 101 and the intersection or overlap of circles 103, 107 may be determined. Next, the intersections or overlap may be repeatedly determined for all locations of an end point of line 106 about the edge of circle 101 . After the intersections have been determined for circles 103, 107 for all locations of the end point of line 106 on the edge of circle 101 , a union of all the intersections may be determined which indicates all possible locations of device 102 within accuracy circle 103.
  • the above-method may be repeated to identify all possible locations of device 100 within circle 101 by using line 106 to create circle 108 and placing an endpoint of line 106 on different locations of the edge of circle 103 to identify all interactions of circles 101 , 108 indicating all possible locations of device 100 within accuracy circle 101 .
  • the above-described processing provides two new areas corresponding to the intersections or overlaps of circles 101 , 107 and 103, 108 and which include all possible coordinate locations of respective devices 100, 102. These two new areas may thereafter be used as updated accuracy areas for the devices 100, 102 for subsequent processing and which replace the initial accuracy circles 101 , 103. These areas include smaller numbers of possible locations of the user interaction devices 100, 102 compared with the numbers of the possible locations of the devices 100, 102 within accuracy circles 101 , 103 providing refined location information regarding the devices 100, 102.
  • ellipsoids may be fitted to the new accuracy areas and the midpoints of the respective ellipsoids may be determined and used to indicate the new coordinates of the user interaction devices 100, 102.
  • the updated accuracy areas are smaller than the areas of the initial accuracy circles 101 , 103 providing refined location information which may be used in combination with the newly determined respective coordinates of the user interaction devices 100, 102 from the midpoints of the ellipsoids.
  • the original accuracy circles and coordinates of the devices 100, 102 from the GPS signals may be disregarded and the newly determined accuracy areas and coordinate locations of the devices 100, 102 may be stored within the management device and used for subsequent operations, such as determining further refined location information using other user interaction devices, implementing augmented reality operations, or other uses.
  • Other embodiments and methods are possible for reducing the accuracy circles and providing new coordinates of the devices 100, 102 in other embodiments.
  • the above-described example embodiment may also be implemented in arrangements where more than two user interaction devices are present.
  • the above-described processing may initially be implemented with respect to two of the devices to refine their accuracy circles/areas and determine new coordinates.
  • the above-described processing may be implemented with respect to one of the two devices (e.g., device 100) and an additional third device (not shown in Figs. 10a-10c).
  • This processing was use the refined accuracy area and new coordinates of device 100 with the accuracy area (e.g., circle) of the third device and a linear distance between the device 100 and the third device which may be determined by processing wireless communications between device 100 and the third device.
  • the new accuracy area and coordinates for device 100 and the third device may be stored and thereafter used for subsequent processing operations with respect to other devices.
  • the above-discussion is an example embodiment of utilization of wireless communications to implement operations with respect to determining refined location information which may have increased accuracy regarding the locations of a plurality of user interaction devices compared with other available location information.
  • other methods or techniques may also be utilized to also provide refined location information regarding the locations of the user interaction devices in the physical world.
  • One method uses image recognition operations to provide refined location information as discussed below in one embodiment.
  • processing circuitry of the management device may be arranged to implement the operations of the described example method.
  • processing circuitry of the user interaction devices may be configured to implement one or more of the operations of the described method (e.g., processing of captured images to locate markers or to compare images of a marker as discussed below).
  • other methods are possible including more, less and/or alternative acts.
  • the processing circuitry accesses initial location information regarding a location of a user interaction device.
  • the initial location information may be obtained using aGPS circuitry of the user interaction device and which is communicated to the management device in one embodiment.
  • the processing circuitry may search a database of markers to locate one or more stored markers which may be viewable by the user interaction device using the initial location information.
  • initial location information of a user interaction device may be unavailable and the images captured by the user interaction device may be communicated to the management device and analyzed by the management device with respect to the stored images of the database. The use of initial location information improves the speed of the processing compared with some arrangements where initial location information is not available or is otherwise not used.
  • the identified markers may be communicated from the management device to the user interaction device.
  • the user interaction device may display one or more images of the markers to inform the user of the possible presence of the markers in their environment and the user may capture images of the markers if the user find the markers in his environment.
  • the management device accesses images obtained from the user interaction device.
  • the user of the user interaction device may use the identified markers to capture one or more images of the markers which are observable by the users.
  • the images may be communicated to and accessed by the management device in act A306.
  • the user may capture a plurality of images of their environment without knowledge of existing markers and the management device may process the images in attempts to locate the presence of the markers.
  • the processing circuitry performs image recognition operations to locate markers in the received images.
  • a database of known markers may be accessed and the processing circuitry may search images captured by the user interaction device for the presence of one or more of the known markers.
  • the processing circuitry may utilize location information regarding the user interaction device to identify known markers which are proximately located to the user interaction device.
  • the known markers may include images of physical objects in the physical world or symbols (e.g., geometric shapes such as a black outline of a bow tie on a white background).
  • the images or symbols may be pre-processed before use as markers and the pre-processed images and symbols may be stored in the database for subsequent comparison operations in some embodiments.
  • the markers are pre-processed through image recognition (computer vision).
  • image recognition computer vision
  • key features of the markers are determined using a multi-scale Hessian detector algorithm which is referenced in Herbert Bay, Andreas Ess, Tinne Tuytelaars, Luc Van Gool, called “Speeded-Up Robust Features (SURF),” Computer Vision and Image Understanding (CVIU), Vol. 110, No. 3, pp. 346-- 359, 2008, the teachings of which are incorporated herein by reference.
  • This processing provides descriptors around keypoints which describe features of the markers for subsequent comparison operations to indicate whether a match is present.
  • the keypoints and descriptors may be stored within a database of the management device.
  • the images received from the user interaction device may also be processed using the multi-scale Hessian detector and SURF for determining features through keypoints and descriptors of the images.
  • the keypoints and descriptors resulting from the processing of the images may be compared with the keypoints and descriptors of the known markers which are stored in the database to determine whether the images contain one or more of the known markers.
  • the keypoints and descriptors of the features of the images and the known markers are compared with one another to determine whether there is a sufficient match.
  • a method for determining closely matched features is to compare the descriptors along with the edges of the keypoint to keypoint within the received images of a scene or marker.
  • One example method is referred to as FLANN and is described in Marius Muja, and David Lowe, "Fast Approximate Nearest Neighbors with Automatic Algorithm Configuration," International Conference on Computer Vision Theory, 2009, the teachings of which are incorporated herein by reference.
  • the comparison will locate known markers present in the received images including location, pose, and scale of the markers.
  • the processing circuitry may process images captured by the user interaction device and which include one or more markers with respect to stored images of the markers to provide refined location information regarding the location of the user interaction device in one embodiment.
  • the management device may include a database of images of the markers.
  • the database may also include location information which indicates various locations in the physical world from which the images of the markers in the database were captured from at previous moments in time.
  • the images containing the markers captured by the user interaction device may be compared to the stored images of the markers which were captured from these known locations. The comparison of the images received from the user interaction device with the stored images and the usage of the known locations from which the stored images were captured provides refined location information regarding the locations of the user interaction device when the images of the markers were captured by the user interaction device.
  • the refined location information may have increased accuracy regarding the locations of the user interaction device compared with the initial location information.
  • a size of the marker of an image captured from a known location may be compared with a size of the marker in an image captured by the user interaction device and the results of the comparison may provide information regarding the location of the user interaction device.
  • the comparison of the images from the user interaction device with the stored images of the markers determines whether any of the images of the markers match. If one of the images from the user interaction device contains a marker which matches a marker in one of the stored images, then the refined location information of the user is set to the location from which the matching stored image was captured from.
  • an image from the user interaction device may contain a stored marker which may be detected by FLANN as being a match even though the image from the user interaction device may have been captured from a different location than the image of the stored marker.
  • the features of the stored marker may be modified using a perspective transform to match the features of the marker within the image from the user interaction device.
  • the perspective transform modifies the features of the stored marker to appear as if the image was captured from different locations.
  • the offset of the user's location from which the user captured the image is known with respect to the location from which the image of the stored marker was obtained and may be used to provide refined location information of the user interaction device. More specifically, in this example, the refined location information is set to the location from which the image of the stored marker was captured as modified by the offset which was determined by the perspective transform processing.
  • the images received from the user interaction devices and which include the markers may be stored in the database and used for future image processing operations of subsequently captured images to provide location information of devices which captured the subsequent images.
  • it is desired to have accurate information regarding the locations from which the images were captured and information regarding the accuracy of the location information may be used to determine whether images are stored in the database and used for subsequent image processing operations to provide location information of other users. Furthermore, accuracy resulting from the processing of images may be improved when a plurality of markers are present in the images for processing or a plurality of different images of a single marker which were obtained from different locations are available for processing. In some embodiments, a user may upload a plurality of images of a marker which may be slightly rotated from one another or observed from slightly different locations.
  • the refined location information may be communicated to one or more of the user interaction devices and used by the devices for implementing operations with respect to experiencing augmented reality or other uses.
  • the refined location information may also be stored within the database which includes the wireless communications devices and their respective locations.
  • the markers may not be communicated to the user interaction device, but rather the user interaction device may capture images of its environment and the images may be uploaded to the management device for processing to locate markers and provide refined location information using the located markers by comparison with existing images in a database of the management device.
  • images of markers from the database of the management device may be communicated to the user interaction devices and used by the devices to calculate refined location information.
  • the user interaction device may be configured to process images captured by the user interaction device including performing object recognition operations to identify the presence of markers and to compare the captured images with the images from the database to provide the refined location information as discussed above with respect to operations of the management device.
  • the management device may have greater processing capacity than the processing capacities of the user interaction devices and implementing processing of images using the management device may be faster than processing of the images using the user interaction devices.
  • the above discussion describes various methods for determining refined location information regarding locations of user interaction devices in the physical world.
  • the refined location information may be used to accurately associate augmented reality representations at appropriate locations of the physical world as captured by the user interaction devices.
  • the refined location information of the plurality of user interaction devices may be used to associate augmented reality representations with physical objects of the physical world.
  • a plurality of markers may be associated with a plurality of respective virtual objects. Accordingly, when a marker is located in an image of a scene, the user interaction device may depict the virtual object in the image by replacing the marker 14 with the virtual object representation in images depicted using the display (e.g., the puppy representation 18 replaces the marker 14 shown in Fig. 1 ).
  • Accurate location information of the augmented reality user interaction devices enables augmented reality representations to be correctly referenced with respect to the physical world such that different user interaction devices experiencing the augmented reality representations will portray the augmented reality representations correctly with respect to the physical world (i.e., at appropriate locations with respect to physical world locations and objects).
  • GPS Some conventional location technologies including GPS have shortcomings with respect to use in augmented reality systems as discussed above. More specifically, civilian entities are currently precluded from obtaining the highest accuracy possible from GPS. In addition, the accuracy of GPS is further degraded at indoor locations or at other locations where the signals from the satellites of the GPS system may encounter obstacles or may not be clearly received.
  • Cell tower triangulation has been used in conventional arrangements to augment GPS readings but this methodology also has shortcomings of being of limited use indoors or in areas which are lacking a sufficient number of cell towers.
  • At least some aspects of the disclosure include obtaining and combining location information from different techniques in attempts to accurately locate the positions of the user interaction devices. Location information obtained from aspects of the disclosure may be used separately or in combination with conventional location determination arrangements in some embodiments.
  • aspects herein have been presented for guidance in construction and/or operation of illustrative embodiments of the disclosure. Applicant(s) hereof consider these described illustrative embodiments to also include, disclose and describe further inventive aspects in addition to those explicitly disclosed. For example, the additional inventive aspects may include less, more and/or alternative features than those described in the illustrative embodiments. In more specific examples, Applicants consider the disclosure to include, disclose and describe methods which include less, more and/or alternative steps than those methods explicitly disclosed as well as apparatus which includes less, more and/or alternative structure than the explicitly disclosed structure.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Augmented reality and location determination methods and apparatus are disclosed according to some aspects of the description. In one aspect, an augmented reality method includes accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, generating second location information which has increased accuracy regarding the location of the user interaction device in the physical world, and communicating augmented data to the user interaction device, and wherein the augmented data comprises the augmented reality representation.

Description

AUGMENTED REALITY AND LOCATION DETERMINATION METHODS AND APPARATUS
TECHNICAL FIELD
[0001] This disclosure relates to augmented reality and location determination methods and apparatus.
BACKGROUND
[0002] Computing systems have continually evolved and the popularity of computing systems continues to increase. The advancement of computing systems creates new uses and applications for the computing systems. For example, the processing speeds, storage capacities and network communication speeds are constantly increasing enabling the use of computing systems in increasing numbers of applications.
[0003] Furthermore, computing systems have evolved from typical office or desk systems to smaller devices, some of which have increased portability, which further expands the possible applications of the computing systems. More specifically, notebook computers have evolved from desktop computers, and more recently, handheld portable devices have also advanced significantly. Personal digital assistants, media players, cellular telephones, smartphones, and other portable devices have increased processing power and storage capacities while communications networks have also been improved allowing greater rates of data transfer between the computing systems.
[0004] Some computing systems and networks have evolved to a sufficient extent to perform augmented reality operations which augment the physical world with virtual computer-generated imagery in one example. In addition, some portable computing systems have sufficient processing, storage and communications capabilities to provide real-time augmented reality data for mobile users.
[0005] At least some aspects of the disclosure are directed to improved methods, apparatus and programming for implementing augmented reality operations.
BRIEF DESCRIPTION OF THE DRAWINGS
[0006] Fig. 1 is an illustrative representation of a user interaction device implementing augmented reality operations according to one embodiment.
[0007] Fig. 2 is a functional block diagram of a media system according to one embodiment. [0008] Fig. 3 is a functional block diagram of a computing system according to one embodiment.
[0009] Fig. 4 is a functional block diagram of communications circuitry of a user interaction device according to one embodiment.
[0010] Fig. 5 is a functional block diagram of a user interface of a user interaction device according to one embodiment.
[0011] Fig. 6 is a flow chart of a method implemented by a user interaction device to implement augmented reality operations according to one embodiment.
[0012] Fig. 7 is a flow chart of a method implemented by a management device with respect to a plurality of user interaction devices according to one embodiment.
[0013] Fig. 8 is a flow chart of a method of outputting wireless communications signals according to one embodiment.
[0014] Fig. 9 is a flow chart of a method of receiving wireless communications signals according to one embodiment.
[0015] Figs. 10a-10c are illustrative representations of a method of determining refined location information of a user interactive device according to one embodiment.
[0016] Fig. 11 is a flow chart of a method implemented by a management device with respect to image recognition operations according to one embodiment.
DETAILED DESCRIPTION
[0017] Attention is directed to the following commonly assigned applications, which are incorporated herein by reference:
[0018] U.S. Patent Applications having Attorney Docket Nos. GR71 -001 , 003, and 004, all entitled "Augmented Reality and Location Determination Methods and Apparatus" by inventors Luke Richey and Allen Greaves, and having U.S. Patent Application Serial Nos. 12/847754, 12/847771 , and 12/847790, all of which were filed July 30,2010.
[0019] Some aspects of the disclosure described herein are directed towards apparatus, methods and programming for implementing augmented reality operations where the physical world is augmented with additional information, such as virtual objects. For example, images of the physical world observed through user interaction devices may be augmented or enhanced with augmented reality representations, for example in the form of visual and/or audio data which may be experienced by users. In one example embodiment, augmented reality representations may include virtual objects which augment a user's experience of the physical world. The virtual objects may be associated with physical world objects which may be static or dynamically moving. Some of the described embodiments include a media system configured to implement and co-ordinate or manage augmented reality operations of one user interaction device or a plurality of user interaction devices which may be interacting in a collaborative augmented reality session in one arrangement.
[0020] Some augmented reality systems use location information regarding locations of the user interaction devices and locations of physical objects in the physical world to accurately augment the physical world with the augmented reality representations. The location information may be used to associate virtual objects with respective objects of the physical world in one illustrative example. At least some aspects of the disclosure are directed towards increasing the accuracy of generated location information regarding the locations of user interaction devices and which location information may be used to implement augmented reality operations. At different times, a plurality of different techniques may be available to determine the location information of user interaction devices. In some embodiments, information from the different techniques may be utilized and/or combined to provide location information of the user interaction devices of increased accuracy compared with other available location information of the user interaction devices, perhaps obtained from a single source. Additional aspects are described in the following disclosure.
[0021] According to one embodiment, a location determination method includes accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, identifying a plurality of wireless communication devices which are proximately located with respect to the user interaction device, initiating wireless communications between the user interaction device and the wireless communications devices, after the initiating, accessing information regarding the wireless communications of the user interaction device and the wireless communication devices, and using the information regarding the wireless communications, determining second location information regarding the location of the user interaction device, and wherein the second location information has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information.
[0022] According to another embodiment, A location determination method comprises accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, identifying a marker which is proximately located with respect to the location of the user interaction device, accessing an image generated by the user interaction device which includes the marker, and processing the image to determine second location information regarding the location of the user interaction device, and wherein the second location information has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information.
[0023] According to yet another embodiment, an augmented reality method comprises accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world, using the first location information, generating second location information which has increased accuracy regarding the location of the user interaction device in the physical world, and communicating augmented data to the user interaction device, and wherein the augmented data comprises the augmented reality representation.
[0024] According to another embodiment, a location determination method comprises using a user interaction device, emitting a wireless communications signal at a first moment in time, using a plurality of wireless communication devices, receiving the wireless communications signal emitted by the user interaction device at a plurality of second moments in time, and using the first and second moments in time, determining information regarding a location of the user interaction device.
[0025] According to another embodiment, a computing system comprises communications circuitry configured to implement communications externally of the computing system, and processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information regarding a location of a user interaction device in a physical world, to use the first location information to identify a plurality of wireless communications devices which are proximately located with respect to the location of the user interaction device, to control the communications circuitry to output a control signal which is configured to initiate wireless communications between the user interaction device and the wireless communications devices, to access information regarding the wireless communications between the user interaction device and the wireless communications devices, and to use the information regarding the wireless communications to determine second location information regarding the location of the user interaction device and which has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information.
[0026] According to another embodiment, a computing system comprises communications circuitry configured to implement communications externally of the computing system and processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information received by the communications circuitry regarding a location of a user interaction device in a physical world, to identify a marker which is proximately located with respect to the location of the user interaction device, to access a plurality of images generated by the user interaction device, and to process the images with respect to the marker to determine second location information regarding the location of the user interaction device and which has increased accuracy with respect to the location of the user interaction device in the physical world compared with the first location information
[0027] According to another embodiment, a computing system comprises communications circuitry configured to implement communications externally of the computing system and processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information received by the communications circuitry regarding a location of a user interaction device in a physical world, to generate second location information which has increased accuracy regarding the location of the user interaction device in the physical world, and to control the communications circuitry to communicate augmented data comprising an augmented reality representation to the user interaction device.
[0028] According to another embodiment, an augmented reality user interaction device comprises a camera, a display system, communications circuitry configured to implement wireless communications externally of the user interaction device, and processing circuitry coupled with the camera, the display system, and the communications circuitry, wherein the processing circuitry is configured to control the display system to generate a plurality of images which comprise image data generated by the camera and augmented data which augments the image data with an augmented reality representation, wherein the processing circuitry is further configured to control the communications circuitry to communicate first location information regarding a location of the user interaction device in a physical world externally of the user interaction device, and wherein the processing circuitry is further configured to access second location information regarding the location of the user interaction device after the outputting of the first location information and to use the second location information to generate the plurality of images, and wherein the second location information has increased accuracy regarding the location of the user interaction device in the physical world compared with the first location information.
[0029] Referring to Fig. 1 , one example of augmented reality aspects of the disclosure is described. Fig. 1 illustrates a user interaction device 10 which is used to generate an image of the physical world and which is augmented by an augmented reality representation. More specifically, in the example of Fig. 1 , the user interaction device 10 includes a camera (not shown) which is configured to capture images of the physical world and which may depicted using a display 12. As a user moves the user interaction device 10, a plurality of images are captured of different scenes viewed by the camera of the device 10.
[0030] In the illustrated example, the scene viewed by the device 10 includes a marker 14 on a wall of the physical world. The generated image depicted using the display 12 includes an augmented reality representation 18 which augments a user's experience of the physical world by replacing the physical world marker 14 with the representation 18. In the illustrated example, the augmented reality representation 18 is a virtual 3D object in the form of a puppy, which may be selected by another user to be associated with the marker 14.
[0031] The use of marker 14 is one example of augmented reality operations which may be implemented using the user interaction device 10 and other augmented reality operations may be implemented in other embodiments. For example, virtual objects may be associated with other physical objects of the physical world, such as other user interaction devices 10 (not shown), in images generated by device 10. In some embodiments, augmented reality representations 18 may entirely replace physical objects of the physical world.
[0032] In one more specific example, the augmented reality representations 18 may include advertising objects (e.g., banner with a product name) and the representations 18 may be associated with famous physical structures of the physical world when observed through a user interaction device 10. For example, a user at a significant football game may view a virtual object banner draped between the physical world goalposts when a user of a device 10 captures images of the end zone during a football game. Companies may pay advertising fees to have augmented reality representations of advertisements of their products associated with physical world objects and which may be viewed by users using their user interaction devices 10 who are proximately located to the physical world objects in one embodiment.
[0033] Location information regarding the locations of the user interaction device 10 and other physical objects in the physical world may be used to generate augmented reality representations 18 in captured images. In one example, the location information may be used to depict the augmented reality representations 18 accurately associated with content (e.g., objects) of the physical world (e.g., other user interaction devices, buildings, structures, mountains, etc.).
[0034] For a static physical object which does not move (e.g., marker 14), location information may be included with the augmented data which determines where the augmented reality representations 18 are to be displayed with respect to content of the physical world when the static physical object is within the field of view of the camera. For portable physical objects (e.g., user interaction devices), the augmented data may be associated with an identifier of the portable physical object. Identification information of the portable physical object and location information of the portable physical object may be used to determine when the portable physical object is present within the field of view of the camera and where augmented reality representations 18 associated with the portable physical object should be shown in generated images. Location information regarding the user interaction devices 10 and/or physical objects may be used to accurately show the augmented reality representations 18 associated with the user interaction devices 10 and/or physical world objects in images generated by the user interaction devices 10 in one embodiment.
[0035] In some embodiments, one user interaction device 10 may be present and the user may be experiencing augmented reality representations with respect to physical objects of the physical world. In other examples (e.g., Fig. 1 ), a plurality of users having user interaction devices 10 may be present and proximately located to one another and experiencing augmented reality representations with respect to one another in a collaborative session and/or physical world objects. The augmented reality representations may be associated with user interaction devices 10 of users and/or with physical objects. Users and user interaction devices 10 may be free to enter and leave interactive augmented reality collaborative sessions in some embodiments.
[0036] It is desired to provide accurate information regarding the locations of user interaction devices 10 to correctly associate augmented reality representations with respect to physical world objects. As described further below, methods and apparatus are described which enable location information of the user interaction devices 10 to be determined with increased accuracy compared with, for example, arrangements which use conventional location determination methods, such as a global positioning system (GPS). In some embodiments disclosed below, the user interaction devices 10 may be configured to communicate (e.g., wirelessly) with one another as well as with external devices (e.g., a management device, Wi-Fi communications devices) to implement augmented reality operations including operations with respect to determining location information of the user interaction devices 10 of increased accuracy.
[0037] Referring to Fig. 2, one example of a media system 20 is shown. Media system 20 is configured to implement operations with respect augmenting the physical world with augmented reality representations. For example, media system 20 is configured to assist user interaction devices 10 with the generation of augmented reality representations. In a more specific example, media system 20 is configured to perform operations with respect to determining locations of user interaction devices 10 (which may be portable) for use in accurately associating augmented reality representations with the physical world. In one embodiment, media system 20 may communicate augmented data with respect to the user interaction devices 10 which may be used by the devices 10 to generate augmented reality representations.
[0038] In the illustrated example configuration of Fig. 2, media system 20 includes a plurality of use interaction devices 10 and a management device 22. User interaction devices 10 may be configured to communicate with one another as well as with management device 22. For example, the user interaction devices 10 may communicate with management device 22 via a network 24. Network 24 may be considered to be a part of media system 20 or may be external of media system 20 in different embodiments. In some embodiments, the user interaction devices 10 may also implement wireless communications with respect to other wireless communications devices (e.g., Wi-Fi communications devices) which may be within the communications ranges of the devices 10 (the Wi-Fi communications devices are not shown in Fig. 2).
[0039] The user interaction devices 10 may be proximately located with respect to one another in a group (e.g., within communications ranges of the devices 10 to implement communications with respect to one another) or in different geographical locations and not proximately located to one another. For example, different groups of user interaction devices 10 may exist in different geographical locations. User interaction devices 10 which are proximately located to one another may participate in a collaborative augmented reality session where augmented reality representations may be associated with the devices 10 in one embodiment. Additionally, only one user interaction devices 10 may be present in a given geographical location and may be implementing augmented reality operations with respect to static physical world objects.
[0040] User interaction devices 10 may be computing systems (e.g., one example is described with respect to Fig. 3) in one embodiment. The user interaction devices 10 may have substantially the same configurations or have different configurations in example embodiments. In some examples, user interaction devices 10 may be configured as portable media devices, personal digital assistants, cellular telephones, smartphones, personal computers, notebook computers, glasses worn by a user including a camera and display system capable of generating images, or any other device capable of capturing images of the physical world and generating images and/or other media content for consumption by a user which include visual images of the physical world which are augmented by one or more augmented reality representations (e.g., additional virtual image content and/or audible content which augments physical world content).
[0041] In one embodiment, management device 22 may be a server which is configured as a computing system, for example as described below with respect to Fig. 3. Management device 22 is configured to implement communications with respect to user interaction devices 10 in the described embodiment. Management device 22 may be configured to perform a plurality of operations with respect to the generation of augmented reality representations by the user interaction devices. Example operations performed include operations with respect to co-ordination and management of user interaction devices 10, co-ordination and management of communications between user interaction devices 10 and a plurality of other wireless communications devices, determining locations of user interaction devices 10 (which may be portable), and storing and communicating augmented data for use by the user interaction devices 10 to generate augmented reality representations.
[0042] In one example embodiment, the user interaction devices 10 may communicate augmented data of their respective augmented reality representations (e.g., the above-described puppy) to the management device 22, perhaps for storage, and the management device 22 may thereafter provide the augmented data to others of the user interaction devices 10 for use in generating the augmented reality representations with respect to the devices 10 which provided the augmented data. For example, in Fig. 1 , the management device 22 may communicate augmented data which includes the puppy representation to user interaction device 10 which uses the augmented data to generate the augmented reality representation 18. [0043] In one embodiment, identification data may be used to associate augmented data with respective appropriate objects of the physical world, such as user interaction devices 10 or other physical world objects. With respect to user interaction devices as described further below in one embodiment with respect to Fig. 1 , initial location information regarding the location of a user interaction device 10 may be used to search a database of the management device 22 to identify other wireless communications devices (e.g., user interaction devices, Wi-Fi communications devices) within a communications range of the user interaction device 10. The database may include location information regarding the devices 10 which the initial location information is searched against, and identification information which uniquely identifies each of the wireless communications devices.
[0044] The database may also include identification information which identifies the augmented data associated with the user interaction devices 10. In one example, once the management device 22 determines that another user interaction device is proximately located to another user interaction device 10, the management device 22 may communicate the augmented data associated with one of the devices to the other device and which includes an identifier of the device to which the augmented reality representation of the augmented data is to be associated with. The receiving user interaction device may access location information regarding its present location and the location of the other device (e.g., access the location information from management device 22) and use the location information of the devices, the identification information of the other device, and the respective augmented data of the other device to generate the augmented reality representation 18 associated with the other device 10 in images captured by the device which include the other device.
[0045] In one embodiment, the device 10 may also depict other augmented reality representations associated with other proximately located user interaction devices 10 and perhaps other objects of the physical world which may be present and have associated augmented data. More specifically, some objects of the physical world are static and do not change locations (e.g., marker 14 of Fig. 1 ). In one embodiment, management system 22 may store augmented data to be used to generate augmented reality representations with respect to static physical objects and the management system 22 may use location information regarding the device 10 to search the database to identify augmented data for static physical objects which are proximately located to the location of the device 10. The management device 22 may communicate augmented data for a static physical object which is proximately located to the device 10 to the device 10 as well as location information where the static physical object exists in the physical world. Thereafter, the device 10 may use the location information to determine when the static physical object is within the field of view of the camera of device 10 and the device 10 may augment captured images of the static physical object with augmented reality representations using the appropriate augmented data.
[0046] Additional operations of the user interaction devices 10 with respect to generation of augmented reality representations may be managed by management device 22 in one embodiment. In one example management embodiment, user interaction devices 10 which attempt to access and utilize media system 20 to experience or implement operations with respect to augmented reality may be managed by management device 22. In one embodiment, management device 22 may manage an augmented reality session where a plurality of user interaction devices 10 are present and performing augmented reality operations with respect to one another.
[0047] In one example, media system 20 may be a closed system and user interaction devices 10 may first be registered and/or otherwise authorized prior to having permission to gain access to media system 20. In one embodiment, management device 22 may assign passwords to the devices 10 and use the passwords and identification information of the devices 10 to verify that the user interaction devices 10 are authorized to gain access to the media system 20. In some arrangements, users may initially sign-up to gain access to the media system 20, and after receiving a password, may subsequently participate in augmented reality sessions or otherwise access the media system 20 to implement augmented reality operations. Users may pay fees to participate in services provided by media system 20, for example, including services of accessing management device 22 and implementing augmented reality operations with respect to other user interaction devices 10 managed by management device 22. Access to or participation in augmented reality operations provided by media system 20 and management device 22 may be denied to users who cannot provide appropriate identification and password information in one embodiment.
[0048] As described in detail below in some embodiments, management device 22 may perform operations with respect to determination of location information of the user interaction devices 10 which may be used to implement augmented reality operations with respect to the user interaction devices 10. Some of the described location determination embodiments include implementing wireless communications between the user interaction devices 10 and/or other wireless communications devices which may be present to determine the location information. Management device 22 may be configured to manage and/or coordinate at least some of the communications of the user interaction devices 10 in one example. In addition, management device 22 may be configured to implement operations with respect to image processing to determine the location information in additional embodiments described below. Other operations may be performed by management device 22 with respect to augmented reality and some of the operations performed by the management device 22 may be performed by other devices (e.g., the user interaction devices themselves) in other embodiments.
[0049] Network 24 which may be implemented as any suitable network configuration for implementing wired and/or wireless communications between user interaction devices 10, management device 22 or other wireless communications devices. For example, network 24 may include a wireless network, local or wide area networks, Internet, cellular network, and/or other suitable infrastructure for implementing communications of digital information. In one embodiment, user interaction devices 10 may have cellular communications capabilities to communicate with network 24 and management device 22. Network 24 may implement packet-switched communications between user interaction devices 10 and management device 22 in one embodiment. As mentioned above, user interaction devices 10 may also communicate directly with one another or other wireless communications devices without use of network 24 in one embodiment.
[0050] Referring to Fig. 3, one example embodiment of a computing system 30 is shown. The illustrated system 30 includes communications circuitry 32, processing circuitry 34, storage circuitry 36, a user interface 38, and a camera 40. Other embodiments of computing system 30 are possible including more, less and/or alternative components. For example, the illustrated configuration of computing system 30 may correspond to one of user interaction devices 10 while some of the illustrated components (e.g., camera 40) may be omitted in one arrangement of computing system 30 which is implemented as management device 22.
[0051] Communications circuitry 32 is arranged to implement communications of computing system 30 with respect to external devices or systems (e.g., network 24 and other computing systems 30 implemented as other user interaction devices 10, Wi-Fi communications devices, or management device 22). Communications circuitry 32 may be configured to implement wired and/or wireless communications. Additional details of one example of communications circuitry 32 which may be utilized is discussed below with respect to Fig. 4. [0052] In one embodiment, processing circuitry 34 is arranged to process data, control data access and storage, issue control signals or commands, and control other augmented reality operations. Processing circuitry 34 may comprise circuitry configured to implement desired programming provided by appropriate computer-readable storage media in at least one embodiment. For example, the processing circuitry 34 may be implemented as one or more processor(s) and/or other structure configured to execute executable instructions including, for example, software and/or firmware instructions. Other exemplary embodiments of processing circuitry 34 include hardware logic, PGA, FPGA, ASIC, state machines, and/or other structures alone or in combination with one or more processor(s). These examples of processing circuitry 34 are for illustration and other configurations are possible.
[0053] Storage circuitry 36 is configured to store programming such as executable code or instructions (e.g., software and/or firmware), electronic data, databases, image data, augmented data, identifiers, location information and/or other digital information and the storage circuitry 46 may include computer-readable storage media. At least some embodiments or aspects described herein may be implemented using programming stored within one or more computer-readable storage medium of storage circuitry 36 and configured to control appropriate processing circuitry 34.
[0054] The computer-readable storage medium may be embodied in one or more articles of manufacture which can contain, store, or maintain programming, data and/or digital information for use by or in connection with an instruction execution system including processing circuitry 34 in the exemplary embodiment. For example, exemplary computer-readable storage media may include any one of physical media such as electronic, magnetic, optical, electromagnetic, infrared or semiconductor media. Some more specific examples of computer-readable storage media include, but are not limited to, a portable magnetic computer diskette, such as a floppy diskette, a zip disk, a hard drive, random access memory, read only memory, flash memory, cache memory, and/or other configurations capable of storing programming, data, or other digital information.
[0055] User interface 38 is configured to interact with a user including conveying data to a user (e.g., displaying visual images for observation by the user) as well as receiving inputs from the user, for example, via a graphical user interface (GUI). User interface 38 may be configured differently in different embodiments. One example embodiment of user interface 38 is discussed below with respect to Fig. 5.
[0056] Camera 40 is configured to capture images within its field of view and generate image data of scenes of the physical world viewed by the computing system 30 in one embodiment. An example camera 40 includes an appropriate imaging sensor configured to generate digital image data responsive to received light in one implementation.
[0057] Movement/orientation circuitry 42 is configured to provide information regarding movement and orientation of the computing system 30 in the described embodiment. For example, circuitry 42 may include an accelerometer arranged to provide information regarding forces which the computing system is subjected to. Circuitry 42 may also include a compass and inclinometer configured to provide information regarding an orientation of the computing system 30 in the physical world.
[0058] Referring to Fig. 4, communications circuitry 32 of computing system 30 which is implemented as one of user interaction devices 10 is shown according to one possible embodiment. In one embodiment, communications circuitry 32 is configured to implement wired and/or wireless communications. In the illustrated configuration, communications circuitry 32 includes network communications circuitry 50, GPS communications circuitry 52, and local communications circuitry 54. Other configurations of communications circuitry 32 are possible including more, less and/or alternative components. For example, a communications interface (e.g., USB port, NIC interface) for implementing wired communications may also be provided.
[0059] In one embodiment, network communications circuitry 50 is configured to implement wireless communications with respect to network 24. In one more specific embodiment, network communications circuitry 50 is configured to communicate with cell towers of network 24 which includes a wireless cellular network. Accordingly, network communications circuitry 50 may implement wireless communications over relatively large distances (e.g., miles) in at least one embodiment.
[0060] GPS communications circuitry 52 is configured to implement wireless communications with external GPS transmitters (e.g., satellites). In one embodiment, GPS communications circuitry 52 may also include appropriate circuitry to provide Assisted GPS (AGPS) or Wireless Assisted GPS (WAGPS) which utilizes additional location information to provide location information of increased accuracy compared with use of GPS alone. GPS communications circuitry 52 may provide location information of computing system 30 directly in one embodiment or processing circuitry 34 may be configured to process signals received by GPS communications circuitry 52 to provide the location information of the computing system 30 in another embodiment.
[0061] Local communications circuitry 54 is configured to implement wireless communications with respect to local wireless communications devices (e.g., other user interaction devices 10, Wi-Fi communications devices) which are proximately located to the computing system 30. In one embodiment, local communications circuitry 54 is configured to implement Bluetooth, ad hoc or other suitable wireless communications with respect to the local wireless communications devices.
[0062] Referring to Fig. 5, one example embodiment of a user interface 38 of a user interaction device 10 is shown. The example user interface 38 includes an input device 60, display system 62, speaker 64 and microphone 66. Other arrangements of user interface 38 are possible including more, less and/or additional components.
[0063] Input device 60 is arranged to receive inputs of a user interacting with computing system 30. Input device 60 may include a graphical user interface, keyboard, pointing device (e.g., mouse) or other suitable apparatus to permit a user to input information and interact with the computing system 30.
[0064] Display system 62 is configured to generate visual images which may be viewed by the user. Display system 62 may depict images captured by camera 40, display digital content (e.g., windows, email, etc.) and display augmented reality representations in some examples.
[0065] Speaker 64 is configured to emit sound waves in one embodiment. The sound waves may include voice information, audio information and other sound waves which are discussed further below.
[0066] Microphone 66 is a configured to receive sound waves in one embodiment. Received sound waves may include voices and sound waves emitted by other computing systems 30 in some illustrative examples.
[0067] As discussed previously, user interaction devices 10 are configured to implement operations to enable users of the devices 10 to experience augmented reality. In one embodiment, the user interaction devices 10 may participate in augmented reality sessions which include a plurality of the devices 10. In another embodiment, a single user interaction device 10 may implement operations for its user to experience augmented reality without the presence of other user interaction devices 10 (e.g., associating augmented reality representations with physical objects present in the physical world). In some embodiments, the user interaction devices 10 may access management device 22 which may implement, manage and/or co-ordinate augmented reality operations for a single user interaction device 10 or a plurality of user interaction devices 10 (which may be participating in an augmented reality collaborative session with one another). [0068] Accurate location information regarding locations of one or more user interaction device 10 is desired to improve the experience of augmented reality by the users. For example, the location information may be used by the user interaction device 10 to correctly and accurately associate augmented reality representations with respect to physical objects of the physical world. The augmented reality representations may be correctly associated with physical objects of the physical world when sufficiently accurate location information regarding the locations of the user interaction devices 10 and physical objects are used to implement augmented reality operations. Less accurate information regarding the locations of user interaction devices 10 may result in less than desirable augmented reality experiences. For example, an augmented reality symbol, such as a puppy, which is to be associated with a respective user and user interaction device 10 may be associated with improper user interaction devices, improper physical world objects, or otherwise not correctly depicted as being associated with the appropriate user interaction device 10.
[0069] Furthermore, in some embodiments, the physical objects which augmented reality representations are to be associated with may be portable (e.g., user interaction devices 10) whose locations may be constantly changing while augmented reality operations are being implemented. The locations of the portable physical objects are determined in sufficient real time to enable the augmented reality representations to be appropriate associated with the portable physical objects (i.e., enable the augmented reality representations to be depicted near or upon the appropriate physical objects in real time). Additionally, portable physical objects may move indoors and it is also desirable to enable augmented reality operations to be implemented indoors or at other locations where location information of the portable physical objects obtained via conventional sources (e.g., GPS) may not be sufficiently accurate.
[0070] In one embodiment, media system 20 including user interaction devices 10 and management device 22 may be arranged to implement location determination operations to provide sufficiently accurate location information regarding the user interaction devices 10 and other physical objects of the physical world to provide satisfactory experiences of augmented reality to users of the user interaction devices 10. A plurality of different methods for providing location information of the physical objects may be available for use. In some arrangements, the location results of the different methods may be combined to further improve the accuracy of the location information. In other arrangements, less than all of the methods may be available for use, but the available methods may provide improved results of increased accuracy of the locations of the user interaction devices 10 or other physical objects compared with reliance upon conventional methodologies (e.g., GPS).
[0071] Referring to Fig. 6, one example process of determining and using location information of one or more user interaction devices 10 is shown. In one embodiment, the method may be implemented by processing circuitry of a user interaction device which has been authorized to participate in augmented reality operations of the media system. Additional methods are possible including more, less and/or alternative acts.
[0072] At an act A10, a user interaction device may determine initial location information regarding its present location at an initial moment in time, for example when the user interaction device accesses the media system to implement augmented reality operations. In one example, the user interaction device may use location determination circuitry to provide the location information using aGPS. The initial location information determined by the user interaction device may have unacceptable error to implement accurate augmented reality operations with respect to the physical world depending upon the location of the user interaction device (e.g., indoors or otherwise located where reception of GPS communications may be relatively poor). Other techniques may also be utilized, for example, monitoring movements of the user interaction device from a last known position, for example, using output of movement/orientation circuitry 42.
[0073] At an act A12, the user interaction device may communicate the initial location information to the management device of the media system. In one embodiment, the initial location information may be aGPS data with an associated error reading. The user interaction device may also include a timestamp corresponding to the time when the location information was obtained and identification information which uniquely identifies the user interaction device.
[0074] The management device of the media system may utilize the initial location information to perform operations to attempt to determine refined location information which has increased accuracy with respect to the actual location of the user interaction device in the physical world compared with the initial location information. As discussed below, the management device may manage communications of the user interaction device with respect to other wireless communications devices (e.g., other user interaction devices and Wi-Fi communications devices) to provide refined location information according to one method and/or implement image recognition operations to provide the refined location information according to another method in illustrative example embodiments. [0075] At an act A14, the user interaction device monitors for the reception of a communication from the management device. The communication indicates whether a sufficient number of wireless communications devices are proximately located to the user interaction device to implement operations with respect to the devices to attempt to provide the refined location information or the communication indicates that an insufficient number of wireless communications devices are present to implement the operations. In one embodiment, a threshold distance (e.g., corresponding to a local wireless communications range of the user interaction device which provided the initial location information in one example) may be used to determine whether other wireless communications devices are located sufficiently close to the communicating user interaction device. In one embodiment, the management device may maintain a database of locations of the wireless communications devices and the management device may use the initial location information to search the database to identify the presence of other wireless communications devices which may be proximately located to the user interaction device to implement wireless communications with the user interaction device.
[0076] If the result of act A14 is negative, the user interaction device may enter a device waiting state or mode as a waiter as described further below. In addition, the method may proceed to an act A20 to implement additional operations with respect to image recognition as discussed in detail below.
[0077] If the result of act A14 is affirmative, the process proceeds to an act A16 where the user interaction device implements localized communications with respect to the other proximately located wireless communications devices which were identified to be sufficiently close to the user interaction device for wireless communications. According to one example embodiment, the user interaction device may be controlled to be in different operational states as a transmitter and receiver of wireless communications signals at different moments in time to implement the wireless communications. Additional details of these communications and different operational states of the user interaction devices are discussed below with respect to the example embodiments of Figs. 7-9.
[0078] At an act A18, information regarding the wireless communications between the proximately-located wireless communications devices may be communicated by the wireless communications devices participating in the communications to the management device. In one embodiment, the management device may process the information regarding the wireless communications in an attempt to provide refined location information for the wireless communications devices including the user interaction device which provided the initial location information at act A10. Additional details regarding the processing to determine the refined location information using the information regarding the wireless communications is described in further detail below with respect to one example embodiment of Figs. 10a-10c.
[0079] In some embodiments where communications are implemented with respect to one or more Wi-Fi communications devices which are proximately located with respect to the user interaction device, the locations of the one or more Wi-Fi communications devices may be fixed and the known locations of these static communications devices may also be used to determine the location information of the user interaction device. The usage of the location information of these devices may provide increased accuracy compared to communications which do not utilize a static device since the location information of the static device may be known with a relatively high accuracy compared with locations of portable devices.
[0080] Following the communication of the information regarding the wireless communications, the user interaction device which provided the initial location information may proceed to implement operations with respect to images in an attempt to determine additional information regarding its location as described further below.
[0081] At an act A20, the user interaction device monitors for the reception of a communication from the management device indicating that one or more markers are proximately located to the user interaction device (e.g., within visible range of the optics of the camera user interaction device). For example, the management device may maintain a database of visual markers and their respective locations in the physical world and the management device may use the initial location information (as well as other location information resulting from the wireless communications) to search the database to identify markers which are located sufficiently close to the user interaction device to be captured by the camera of the user interaction device.
[0082] If the result of act A20 is negative, the process proceeds to an act A26 where the location information for the user interaction device may be updated, for example, using refined location information from the management device (e.g., based upon the wireless communications with other wireless communications devices). The refined location information may be stored as the location of the respective user interaction device in a database maintained by the management device and the refined location information may be communicated to one or more of the user interaction devices. [0083] If the result of act A22 is affirmative, the process proceeds to an act A22 where the user interaction device which provided the initial location information captures images of the physical world about the user interaction device.
[0084] At an act A24, the images may be processed by the user interaction device in attempt to provide the refined location information. In one embodiment, the management device may communicate images of the markers (e.g., as well as information regarding locations in the physical world from which the images were captured) which are proximately located to the user interaction device to the user interaction device. The markers may be displayed to a user to inform the user of the presence and types of nearby markers.
[0085] The user interaction device may implement image processing techniques to analyze images captured by the user interaction device to identify whether the markers are present in the images. For example, object recognition processing may be used in one arrangement to identify the markers. Once the markers are identified, the images captured by the user interaction device may be processed with respect to the images of the markers received by the management device to identify location information regarding the locations of the user interaction device from which the images of the markers were captured by the user interaction device. This analysis may compare differences of the images and use the comparison with the known location information from which the images of the markers provided by the management device were captured. This determined location information may have increased accuracy of the location of the user interaction device compared with other available location information regarding the location of the user interaction device. Additional details regarding processing of images captured by the user interaction device are described below with respect to Fig. 11 in one embodiment.
[0086] In another embodiment, image data of images captured by the user interaction device may be communicated to the management device for processing to provide the determined location information. In some embodiments, the markers which are in the vicinity of the user interaction device may be communicated to the user interaction device to inform the user of available markers and thereafter the user may capture images of the markers and communicate the captured images to the management device for processing as described above in act A24. In other arrangements, the markers may not be communicated to the user interaction device and the user interaction device may upload captured images to the management device for processing to recognize the presence of any markers and determine the location information. Other methods are possible.
[0087] At an act A26, the refined location information determined by the wireless communications with other wireless communications devices and/or information obtained by image processing techniques may be used to update the location information of the user interaction device. The refined location information may have increased accuracy of the location of the user interaction device in the physical world compared with other available location information (e.g., GPS signals received by the user interaction device located in an indoors environment).
[0088] At an act A28, the user interaction device and/or management device may use the refined location information for various purposes. In one example, the user interaction device and/or management device may utilize the refined location information to implement augmented reality operations. In one more specific example, the user interaction device may utilize the refined location information to generate images which include augmented reality representations associated with physical world content. For example, the refined location information may be used to determine where the augmented reality representations will be depicted in the generated images of the physical world and to be viewed in real time by the user of the user interaction device in one embodiment. The management device may use the refined location information of a user interaction device 10 to identify markers or other user interaction devices which may be proximately located to the refined location information of the user interaction device 10.
[0089] As mentioned above, a user interaction device 10 may wirelessly communicate with other wireless communications devices which may be proximately located to the device 10 in attempts to determine refined location information regarding the actual location of the device 10 in the physical world with increased accuracy compared with other available location determination methods. An example implementation of using wireless communications to determine refined location information is described in additional detail with respect to Figs. 7-9. The methods may be performed to determine refined location information using wireless communications between the wireless communications devices (e.g., user interaction devices, Wi-Fi communications devices) which are within wireless communications range of one another. More specifically, the method of Fig. 7 controls the states or modes of operation of the wireless communications devices as waiters, senders and receivers in one embodiment. The method of Fig. 8 discloses one method of operations with respect to one of the devices operating as a sender and the method of Fig. 9 discloses one method of operations with respect to one or more of the devices operating as receivers. The wireless communications devices may individually operate as a waiter, sender and receiver at different moments in time. At one moment in time, one of the wireless communications devices may operate as a sender to transmit a wireless communications signal and the other devices operate as receivers of the signal. At subsequent moments in time, the other wireless communications devices may individually operate as senders while the non-transmitting ones of the devices operate as receivers of the transmissions.
[0090] Referring to Fig. 7, a method of controlling the states or modes of operation of the wireless communications devices is described according to one embodiment. Other methods are possible including more, less and/or alternative acts. The method may be executed by processing circuitry of the management device in one implementation.
[0091] As mentioned above, the management device manages the wireless communications in one embodiment. In one illustrative example, a plurality of wireless communications devices may be within a wireless communications range of one another and the devices may implement wireless communications with respect to one another. The wireless communications may be used to determine refined location information of the devices with increased accuracy compared with the initial location information (e.g., GPS data) of the devices. The wireless communications devices may communicate initial location information and time information to the management device 22 and the management device 22 may use the location information and time information to search a database to identify the wireless communications devices which are within range of one another and capable of communicating with one another.
[0092] In one example communications arrangement discussed below, the management device determines the states or modes of operations wireless communications devices at different moments in time during the wireless communications. As discussed above in one example, the management device may instruct one of the wireless communications to operate as a sender of a wireless communications signal at one moment in time and instruct others of the wireless communications devices to operate as receivers to receive the wireless communications signal. At other moments in time, the management device may instruct others of the wireless communications devices to individually operate as the senders and the non- transmitting devices may be instructed to operate as receivers. After each of the wireless communications devices has operated as a sender, the results of the wireless communications may be analyzed to provide refined location information regarding the locations of the wireless communications devices in the physical world. When an insufficient number of wireless communications devices are present (or at other appropriate moments in time), the management device may instruct the wireless communications devices to operate as waiters where no wireless communications are implemented until a sufficient number of wireless communications (e.g., three) are within communications range of one another for the location determination operations to occur.
[0093] At an act A100, initial location information is accessed from a user interaction device. In one embodiment, the user interaction device communicates GPS location coordinates, accuracy information regarding the GPS data regarding the location of the user interaction device and a unique identifier to the management device. The user interaction device may also communicate accelerometer information (i.e., if the user interaction device is configured with an on-board accelerometer) which may be used to modify the initial location information by an offset according to movement detected by the accelerometer. A timestamp regarding the initial location information may also be recorded in one embodiment. The timestamp and initial location information for the user interaction device may be stored within a database of the management device as discussed below in one embodiment.
[0094] At an act A102, the processing circuitry accesses the database of the management device including information regarding wireless communications devices. The database may include respective location information for a plurality of user interaction devices which are interacting with the management device, and perhaps implementing augmented reality operations. Furthermore, the database may also maintain location information for a plurality of additional wireless communications devices, such as Wi-Fi communications devices.
[0095] At an act A104, it is determined whether the mapping is empty for the user interaction device which communicated the initial location information.
[0096] If the condition of act A104 is affirmative, a new mapping entry is created for the user interaction device and the received initial location information is stored as the location of the device at an act A106.
[0097] If the condition of act A104 is negative, the timestamp of the initial location information is compared with the last modification of the entry for the user interaction device at an act A108.
[0098] At an act A110, it is determined whether the length of time from the last entry to the new information exceeds a threshold. [0099] If the condition of act A110 is affirmative, the mapped entry for the user interaction device is reset with the received initial location information at an act A112.
[0100] If the condition of act A110 is negative, the received initial location information may be disregarded and the stored location information for the user interaction device may be used.
[0101] At an act A114, it is determined whether any other wireless communications devices are proximately located to the location of the user interaction device. As mentioned above, the management device may maintain a database including location information for a plurality of wireless communication devices (e.g., user interaction devices, Wi-Fi communications devices). The management device may search the database to determine whether a sufficient number of wireless communications devices are within a threshold distance (e.g., wireless communications range) of the user interaction device. It is desired that at least three wireless communications devices be able to send and receive wireless communications with respect to one another in one embodiment. The accuracy of the location information determined from wireless communications of the devices increases as the number of communicating devices within communications range of one another increases. Furthermore, the accuracy of the information is increased if one or more of the wireless communications devices have a known static location in the physical world. If a sufficient number of devices were not located in act A114, the management device may output a control signal to instruct the user interaction device at an act A116 to become a waiter until a sufficient number of devices are proximately located to the user interaction device, for example, as determined by the management device.
[0102] If a sufficient number of devices were located in act A114, the management device determines whether the user interaction device has recently operated as a sender at an act A118. For example, the management device may determine whether the user interaction device has previously operated as a sender with the other wireless communications devices which are currently within the communications range of the user interaction device.
[0103] If the result of act A118 is affirmative, the management device proceeds to an act A122 to determine whether any of the other wireless communications devices which are proximately located to the user interaction device may become a sender.
[0104] If the result of act A118 or act A122 is negative, the management device outputs a control signal instructing the user interaction device which provided the initial location information to become a sender at an act A124. As discussed further below, the user interaction device may output a wireless communications signal which may be received by other wireless communications devices operating as receivers and the results of the communication may be used to provide refined location information regarding the location of the user interaction device in one embodiment.
[0105] If the result of act A122 is affirmative, the management device may output a control signal at an act A126 instructing the user interaction device to be a receiver which receives wireless communications signals from senders as discussed in detail below. In addition to outputting the control signals to the sender and receivers, the management device may also output a common timing reference to the senders and receivers which the senders and receivers may use to timestamp moments in time when the wireless communications signal is transmitted and received. For example, the timing reference may be the current time as determined by the management device and the senders and receivers may align their internal timing references to this time in one embodiment.
[0106] At an act A128, the management device may implement operations with respect to the user interaction devices. For example, the management device may communicate with the user interaction devices operating as senders and receiver in the example embodiments of Figs. 8 and 9.
[0107] Referring to Fig. 8, an example method of operations implemented by a wireless communications device (e.g., user interaction device) operating as a sender is described. Other methods are possible including more, less and/or alternative acts.
[0108] As mentioned above, wireless communications between a plurality of wireless communications devices may be used to provide refined location information regarding the wireless communications devices. In one embodiment, one of the wireless communications devices operating as a sender emits a wireless communications signal which may be received by other wireless communications devices within a wireless communications range of the sender. The sender may output a plurality of different types of wireless communications signals in the described arrangement.
[0109] More specifically, in one example embodiment, the sender emits an electromagnetic wireless communications signal (e.g., Bluetooth, ad hoc wireless communications, or other signal). In another example embodiment, the sender emits a sound wave (e.g., a sound wave having a frequency which may be outside of the audible range of humans). The communicated signals may be received by the receivers and used to provide refined location information in one embodiment. In some embodiments, the wireless communications devices may communicate plural different types of signals (e.g., electromagnetic signals as well as sound waves).
[0110] At an act A150, a user interaction device which has been selected to be a sender communicates initial location information regarding its current location which is received by the management device.
[0111] At an act A152, the management device accesses the initial location information.
[0112] At an act A154, the management device creates a unique identifier for the user interaction device which acts as the sender. If sound waves are to be communicated, the management device may select (e.g., randomly in one embodiment) a unique frequency from a range of possible frequencies as the identifier which may be used. In other examples, the identifier may be an oscillation of sound frequencies or a Bluetooth server name which the device uses to transmit a Bluetooth signal.
[0113] At an act A156, the management device searches for other wireless communications devices which are proximately located to the user interaction device to act as receivers (if the receivers are not known) or the management device accesses a list of the proximately located devices which may act as receivers (if known).
[0114] At an act A158, the management device communicates the unique identifier of the sender to the receivers. The unique identifier may also include signal identification information which enables the receivers to identify the communications from the sender (e.g., identifies the sender of a Bluetooth communications signal, identifies the frequency of a sound wave from the sender, etc.).
[0115] At an act A160, the management device receives a plurality of responses from the receivers indicating that they have received the unique identifier and signal identification information and they are ready to receive the wireless communications signal to be emitted from the sender.
[0116] At an act A162, the management device communicates the unique identifier to the sender indicating that the receivers are ready and the management device may communicate a control signal to initiate wireless communications. The management device may also specify the frequency of a sound wave to be emitted if the devices are using sound waves.
[0117] At an act A164, the sender outputs the wireless communications signal (e.g., electromagnetic wave or sound wave) and records the time the signal is outputted. For electromagnetic signals, the sender may use appropriate communications circuitry to output the signal. For sound waves, a speaker of the sender may be used to output the signal.
[0118] At an act A166, the timestamp indicative of the time that the signal was outputted by the sender is communicated to the management device.
[0119] At an act A168, the management device accesses the timestamp regarding the outputting of the signal.
[0120] At an act A170, the management device accesses a plurality of communications from the receivers indicating the respective times the wireless communications signal which was outputted by the sender was received by the respective receivers. The wireless communications signal outputted by the sender may be received by the receivers at different times depending upon their distances from the sender.
[0121] At an act A172, the management device uses the information regarding the lengths of time between the outputting of the wireless communications signal and the reception of the signal by the plural receivers to determine linear distances between the sender and the receivers using the respective lengths of time of communications with respect to the receivers which may be multiplied by the velocity of the wireless signals which were communicated (e.g., sound, electromagnetic). The determined linear distances may be used to determine refined location information of the user interaction devices as discussed in one embodiment below with respect to Figs. 10a-10c. The above-described process may be repeated a number of times between two devices and the resultant distances may be averaged in one embodiment.
[0122] The management device may also output a completion signal to the wireless communications devices indicating that it has obtained all information regarding the communication (i.e., transmission and reception) of the wireless communications signal. The completion signal may instruct the sender to cease sending the wireless communications signal.
[0123] At an act A174, the sender receives the completion signal from the management device and ceases outputting of the wireless communications signal.
[0124] At an act A176, the sender outputs a query to determine the next desired state of operation of the wireless communications device as a receiver or waiter.
[0125] At an act A178, the sender receives a response to the query from the management device in the form of a control signal to enter an operational state as a receiver or a waiter. [0126] At an act A180, the sender changes its operational state to be a receiver for subsequent communications from other proximately located wireless communications devices or a waiter in accordance with the received response to the query.
[0127] Referring to Fig. 9, an example method of operations implemented by a wireless communications device (e.g., user interaction device) operating as a receiver is described. Other methods are possible including more, less and/or alternative acts.
[0128] At an act A200, the user interaction device receives a unique identifier from the management device which identifies wireless communications to be received from the sender. The unique identifier may identify the sender of Bluetooth signal or may identify a frequency of a sound wave and the sender of the sound wave.
[0129] At an act A202, the receiver prepares to receive the signal and communicates a ready signal to the management device.
[0130] At an act A204, the receiver waits for reception of the wireless communications signal to be communicated by the sender.
[0131] At an act A206, the receiver compares an amount of time waiting for the signal with a time interval to determine whether the time interval has passed. The user interaction device continues to wait for reception of the wireless communications signal if the time interval has not passed.
[0132] If the result of act A206 is affirmative, the receiver proceeds to an act A208 to output a notification to the management device that the device failed to receive the wireless communications signal.
[0133] At an act A210, the receiver queries the management device of whether to operate as a receiver, sender or waiter.
[0134] During the waiting for reception of the wireless communications signal, the receiver continually monitors for reception of the signal. If the wireless communications signal is detected at an act A212, the receiver proceeds to an act A214 to timestamp the time when the wireless communications signal was received. The receiver returns to act A204 to wait for the signal if the result of act A212 is negative.
[0135] At an act A216, the receiver communicates the timestamp of the reception of the wireless communications signal to the management device. The receiver may also include its respective identifier which allows the management device to identify the receiver which received the signal.
[0136] At an act A218, the receiver receives instructions from the management device which instructs the device to operate as a receiver, sender or waiter during subsequent communications of additional wireless communications signals. [0137] At an act A220, the receiver enters the specified state of operation as a receiver, sender or waiter.
[0138] As mentioned above, the communicated wireless communications signals used to determine refined location information may include different types of signals (e.g., electromagnetic waves, sound waves). The above-recited methods of Figs. 8 and 9 with respect to transmission and reception of signals may be implemented for the different types of signals. More specifically, the methods of Figs. 8 and 9 may be performed to implement communications of electromagnetic signals at one moment in time and the methods may again be performed at another moment in time for communications of sound waves.
[0139] In one implementation, some of the wireless communications may be implemented automatically without control of a user while others of the wireless communications may be implemented as a result of user control. In one more specific example, the user interaction devices may be configured to initiate and implement Bluetooth communications automatically without user control while sound wave communications may be initiated as a result of user control or instruction. The use of sound waves may provide refined location information of improved accuracy in some embodiments compared with communications of electromagnetic waves since sound waves travel slower than electromagnetic waves.
[0140] The duration of the times of sound wave communications between the sending wireless communications device and the receiving wireless communications devices may be multiplied by the velocity of sound (e.g., 340.29 m/s) to determine the distances of the respective receiving wireless communications devices with respect to the sending wireless communications device. The speed of light may be used in calculations for the wireless communications of electromagnetic waves to determine the distances of the receivers with respect to the sender. The determined distances may be used as discussed below to provide refined location information.
[0141] As mentioned above, some of the wireless communications devices may be Wi-Fi communications devices which are arranged to wirelessly communicate with other wireless communications devices including the user interaction devices. Typical Wi-Fi communications devices are not portable but are rather implemented in fixed static locations (e.g., hot spots). In one embodiment, the management device may store the locations of Wi-Fi communications devices and may search for the presence of these devices which may be within a wireless communications range of one of more user interaction device which is implementing operations with respect to augmented reality described herein.
[0142] The use of Wi-Fi communications devices which are in static locations is advantageous since accurate information regarding these devices may be stored in the management device and used to accurately locate other wireless communications devices (e.g., user interaction devices). More specifically, since the location information of these devices may include reduced error compared with available location information of portable devices, the location information of the static devices may be used to provide refined location information which may have increased accuracy with respect to the portable devices compared with operations implemented solely between portable devices to provide the refined location information.
[0143] In one embodiment, portable user interaction devices may calculate their respective distances to static wireless communications devices which may result in reduced error in the determined locations of the portable user interaction devices since the error of the locations of the static wireless communications device in the physical world may be less compared with error present in available location information of the portable communications devices which would otherwise be used. For example, the location of one static device may be known with an increased degree of accuracy and accordingly the processing of information regarding communications with this device may provide location information with an increased degree of accuracy compared with processing of communications with devices which may all be dynamically moving.
[0144] In one embodiment, the database of the management device may be populated with identification and location information regarding a plurality of Wi-Fi communications devices. In one embodiment, users of an augmented reality community may upload location information to the management device of the Wi-Fi communications devices. For example, Wi-Fi communications devices may be programmed with their static locations upon installation in a facility and the information regarding the static locations may be provided to the management device. In another example, user interaction devices having accurate information regarding their present location may be positioned adjacent to the Wi-Fi communications devices and the location of the user interaction device may be uploaded with the identification information of the Wi-Fi communications devices to the management device. Any suitable arrangement may be used to provide the location information regarding the Wi-Fi communications devices to the management device and which may be stored therein for use in subsequent searching operations. [0145] These Wi-Fi communications devices populated in the database of the management device may be subsequently searched at a later moment in time using initial location information of another user interaction device. The identified Wi-Fi communications devices which are proximately located to one or more user interaction device may be used to implement communications with respect to the user interaction devices as discussed above to provide refined location information of the user interaction devices and which may have increased accuracy compared with location information obtained by other methods, such as GPS.
[0146] In the example embodiment discussed above, the management device determines the refined location information using information regarding the wireless communications which are implemented between the wireless communications devices. In another embodiment, the receiving wireless communications devices (e.g., user interaction devices, Wi-Fi communications devices) may communicate information regarding moments in time when the wireless communications devices received a wireless communications signal which was emitted by a sender to the sender. The sender may use the received information to determine its refined location information using techniques described in further detail below (as opposed to having management device perform the calculations to determine the refined location information).
[0147] As mentioned above, wireless communications signals including sound waves may be utilized in attempts to improve the accuracy of the refined location information of a plurality of wireless communications devices. The following is a discussion of processing of sound waves received by a wireless communications device to identify wireless communications signals which were emitted by the sending wireless communications device according to one embodiment. As mentioned above, the receivers may receive a predefined frequency (i.e., the frequency is defined before the communication of the signal) from the management device and the receivers may process received sound waves in attempts to determine whether the predefined frequency is present in the received sounds waves.
[0148] The described processing may be performed upon output of a microphone of a receiving user interaction device in one embodiment. The received sound may be converted into a function which can be tested to determine if the specified frequency of the communicated wireless communications signal is present in the received domain. Received sound is a sinusoidal wave and a Fourier Transform of Eqn. 1 may be used to convert the received sound into a desired domain in one embodiment: f(<p) = f(x)e-2→dx EQN.1
Essentially f(x) is the received sound at time x and f(cp) is the Fourier Transform for which φ is the frequency in hertz being checked against. If a strong frequency φ exists within the ambient noise, then f(c ) will present a discernable maxima within the graphed f(cp).
[0149] f(cp) uses a complex integration to determine the entire domain. A Riemann's Sum may be used to estimate this value as closely to f(cp) as possible. This is known as a Discrete Fourier Transform where samples will be taken from the noise in the room at a given interval. These samples may be reduced to:
- k =∑n=o *ne ~ fc = 0, ... , iV - l EQN. 2 xn is a sequence of complex numbers that go from n = 0, ... , n = N - 1; i is the imaginary unit ^T, and k is the frequency to being checked against. For sound recognition, xn would represent the amplitude of the ambient sound at a sample (time) n, and xk represents the magnitude at frequency k.
[0150] In one example, received sound may be sampled at a rate of approximately 44100 samples per second. The samples may be partitioned into different groups and the processing of Eqn. 2 may be performed on the groups separately. A resolution may be computed by dividing the sampling rate of 44100 samples per second by the number of samples obtained (e.g., a number of obtained samples of 32768 provides a resolution of 1.34). The quotient of the frequency of the wireless communications signal may be divided by the resolution to provide the value of the result of the discrete Fourier Transform if the sound wave is present in the sound received by the wireless communications device.
[0151 ] In one embodiment, a threshold for the magnitude of frequency may be specified and the output of the processing for the frequency of interest may be compared with this threshold. As an example, if the frequency of interest with the ambience is
20Hz then Eqn. 2 becomes X20hz =∑n=o ½e N and X20hz is compared with the threshold. If the threshold is met, then it can be assumed that the sound frequency of interest is being communicated within the environment of the user interaction device. In one embodiment, the operations of determining location using sound waves may be implemented at times specified by the user for example when the presence of interfering sounds is low.
[0152] Once it is determined that the specified frequency of the wireless communications signal has been received, further operations may be performed in one embodiment to refine the processing of the time of reception of the signal by the receiver. The group of samples in which the appropriate sound wave was detected may be split into halves and the transform may be implemented on each half. The output of the Discrete Fourier Transform of the predefined frequency of the sound wave being analyzed should be smaller on the first half (if the frequency was present during the first and second halves) or only present in the second half (if the frequency was only present in the second half and not the first half). The half indicating where the frequency was received may be split into additional halves which may also be similarly processed until the amount of error of the time when the frequency was received is sufficiently small.
[0153] Referring to Figs. 10a-10c, an example embodiment of a method for increasing the accuracy of a user is shown. In one embodiment, a user interaction device is configured to receive a GPS signal which provides the device's Cartesian coordinates as well as an accuracy radius which defines an accuracy circle about the coordinates where the user interaction device may exist. The example method described below reduces the size of the accuracy circle (e.g., the area of possible locations the user interaction device may be with respect to the provided coordinates) to provide refined location information regarding the location of the user interaction device with increased accuracy.
[0154] Referring to Fig. 10a, a plurality of user interaction devices 100, 102 are shown at different locations having respective coordinates as well as respective accuracy circles 101 , 103 based upon GPS signals received by the user interaction devices 100, 102. In addition, a line 106 is illustrated which corresponds to a linear distance between the user interaction devices 100, 102 which was determined from processing of wireless communications between user interaction devices 100, 102 as described above. In one embodiment, the endpoints of the line 106 are processed with respect to the accuracy circles 101 , 103 to reduce the areas of the accuracy circles 101 , 103 where the user interaction devices 100, 102 may be located and to provide the refined location information regarding the locations of devices 100, 102 of increased accuracy compared with accuracy circles 101 , 103. [0155] In the described example, the user interaction devices 100, 102 exist within their respective accuracy circles 101 , 103 and the devices 100, 102 are spaced a distance of the line 106 apart from one another. In one embodiment, the method reduces the size of one of the accuracy circles 101 , 103 using the line 106 and the other of the accuracy circles 101 , 103.
[0156] Referring to Fig. 10b, one example is described where a circle 107 is generated having a radius equal to the length of line 106. The accuracy of circle 103 may be reduced by identifying the union or overlap of circle 107 with circle 103 for all possible locations of device 100 within circle 101 .
[0157] More specifically, an end point of line 106 may be placed on the edge of circle 101 and the intersection or overlap of circles 103, 107 may be determined. Next, the intersections or overlap may be repeatedly determined for all locations of an end point of line 106 about the edge of circle 101 . After the intersections have been determined for circles 103, 107 for all locations of the end point of line 106 on the edge of circle 101 , a union of all the intersections may be determined which indicates all possible locations of device 102 within accuracy circle 103.
[0158] Referring to Fig. 10c, the above-method may be repeated to identify all possible locations of device 100 within circle 101 by using line 106 to create circle 108 and placing an endpoint of line 106 on different locations of the edge of circle 103 to identify all interactions of circles 101 , 108 indicating all possible locations of device 100 within accuracy circle 101 .
[0159] The above-described processing provides two new areas corresponding to the intersections or overlaps of circles 101 , 107 and 103, 108 and which include all possible coordinate locations of respective devices 100, 102. These two new areas may thereafter be used as updated accuracy areas for the devices 100, 102 for subsequent processing and which replace the initial accuracy circles 101 , 103. These areas include smaller numbers of possible locations of the user interaction devices 100, 102 compared with the numbers of the possible locations of the devices 100, 102 within accuracy circles 101 , 103 providing refined location information regarding the devices 100, 102.
[0160] In one embodiment, ellipsoids may be fitted to the new accuracy areas and the midpoints of the respective ellipsoids may be determined and used to indicate the new coordinates of the user interaction devices 100, 102. The updated accuracy areas are smaller than the areas of the initial accuracy circles 101 , 103 providing refined location information which may be used in combination with the newly determined respective coordinates of the user interaction devices 100, 102 from the midpoints of the ellipsoids. The original accuracy circles and coordinates of the devices 100, 102 from the GPS signals may be disregarded and the newly determined accuracy areas and coordinate locations of the devices 100, 102 may be stored within the management device and used for subsequent operations, such as determining further refined location information using other user interaction devices, implementing augmented reality operations, or other uses. Other embodiments and methods are possible for reducing the accuracy circles and providing new coordinates of the devices 100, 102 in other embodiments.
[0161] The above-described example embodiment may also be implemented in arrangements where more than two user interaction devices are present. In one example method where three user interaction devices are present, the above-described processing may initially be implemented with respect to two of the devices to refine their accuracy circles/areas and determine new coordinates. After the processing with respect to the devices, the above-described processing may be implemented with respect to one of the two devices (e.g., device 100) and an additional third device (not shown in Figs. 10a-10c). This processing was use the refined accuracy area and new coordinates of device 100 with the accuracy area (e.g., circle) of the third device and a linear distance between the device 100 and the third device which may be determined by processing wireless communications between device 100 and the third device. The new accuracy area and coordinates for device 100 and the third device may be stored and thereafter used for subsequent processing operations with respect to other devices.
[0162] The determined refined location information may be communicated to the user interaction devices for use by the devices to accurately associate augmented reality representations with respect to appropriate locations of the physical world. As mentioned above, the augmented reality representations may be associated with physical objects of the physical world. Additional details regarding usage of the refined location information is discussed below. The refined location information may also be stored in a database of the management device along with timestamp information which indicates a moment in time to which refined location information pertains.
[0163] The above-discussion is an example embodiment of utilization of wireless communications to implement operations with respect to determining refined location information which may have increased accuracy regarding the locations of a plurality of user interaction devices compared with other available location information. As mentioned above, other methods or techniques may also be utilized to also provide refined location information regarding the locations of the user interaction devices in the physical world. One method uses image recognition operations to provide refined location information as discussed below in one embodiment.
[0164] Referring to Fig. 11 , one method is described which may be executed to calculate refined location information regarding one or more user interaction device using image recognition techniques according to one embodiment. Processing circuitry of the management device may be arranged to implement the operations of the described example method. In other arrangements, processing circuitry of the user interaction devices may be configured to implement one or more of the operations of the described method (e.g., processing of captured images to locate markers or to compare images of a marker as discussed below). Furthermore, other methods are possible including more, less and/or alternative acts.
[0165] At an act A300, the processing circuitry accesses initial location information regarding a location of a user interaction device. The initial location information may be obtained using aGPS circuitry of the user interaction device and which is communicated to the management device in one embodiment.
[0166] At an act A302, the processing circuitry may search a database of markers to locate one or more stored markers which may be viewable by the user interaction device using the initial location information. In other embodiments, initial location information of a user interaction device may be unavailable and the images captured by the user interaction device may be communicated to the management device and analyzed by the management device with respect to the stored images of the database. The use of initial location information improves the speed of the processing compared with some arrangements where initial location information is not available or is otherwise not used.
[0167] At an act A304, the identified markers may be communicated from the management device to the user interaction device. The user interaction device may display one or more images of the markers to inform the user of the possible presence of the markers in their environment and the user may capture images of the markers if the user find the markers in his environment.
[0168] At an act A306, the management device accesses images obtained from the user interaction device. In one example, the user of the user interaction device may use the identified markers to capture one or more images of the markers which are observable by the users. The images may be communicated to and accessed by the management device in act A306. As mentioned above, in another embodiment, the user may capture a plurality of images of their environment without knowledge of existing markers and the management device may process the images in attempts to locate the presence of the markers.
[0169] At an act A308, the processing circuitry performs image recognition operations to locate markers in the received images. In one embodiment, a database of known markers may be accessed and the processing circuitry may search images captured by the user interaction device for the presence of one or more of the known markers. The processing circuitry may utilize location information regarding the user interaction device to identify known markers which are proximately located to the user interaction device. The known markers may include images of physical objects in the physical world or symbols (e.g., geometric shapes such as a black outline of a bow tie on a white background). The images or symbols may be pre-processed before use as markers and the pre-processed images and symbols may be stored in the database for subsequent comparison operations in some embodiments.
[0170] In one preprocessing example, the markers are pre-processed through image recognition (computer vision). First, key features of the markers are determined using a multi-scale Hessian detector algorithm which is referenced in Herbert Bay, Andreas Ess, Tinne Tuytelaars, Luc Van Gool, called "Speeded-Up Robust Features (SURF)," Computer Vision and Image Understanding (CVIU), Vol. 110, No. 3, pp. 346-- 359, 2008, the teachings of which are incorporated herein by reference. This processing provides descriptors around keypoints which describe features of the markers for subsequent comparison operations to indicate whether a match is present. In one embodiment, the keypoints and descriptors may be stored within a database of the management device.
[0171] Thereafter, the images received from the user interaction device may also be processed using the multi-scale Hessian detector and SURF for determining features through keypoints and descriptors of the images. The keypoints and descriptors resulting from the processing of the images may be compared with the keypoints and descriptors of the known markers which are stored in the database to determine whether the images contain one or more of the known markers.
[0172] In one comparison embodiment, the keypoints and descriptors of the features of the images and the known markers are compared with one another to determine whether there is a sufficient match. A method for determining closely matched features is to compare the descriptors along with the edges of the keypoint to keypoint within the received images of a scene or marker. One example method is referred to as FLANN and is described in Marius Muja, and David Lowe, "Fast Approximate Nearest Neighbors with Automatic Algorithm Configuration," International Conference on Computer Vision Theory, 2009, the teachings of which are incorporated herein by reference. The comparison will locate known markers present in the received images including location, pose, and scale of the markers.
[0173] At an act A310, the processing circuitry may process images captured by the user interaction device and which include one or more markers with respect to stored images of the markers to provide refined location information regarding the location of the user interaction device in one embodiment. The management device may include a database of images of the markers. The database may also include location information which indicates various locations in the physical world from which the images of the markers in the database were captured from at previous moments in time. The images containing the markers captured by the user interaction device may be compared to the stored images of the markers which were captured from these known locations. The comparison of the images received from the user interaction device with the stored images and the usage of the known locations from which the stored images were captured provides refined location information regarding the locations of the user interaction device when the images of the markers were captured by the user interaction device. The refined location information may have increased accuracy regarding the locations of the user interaction device compared with the initial location information. In another example, a size of the marker of an image captured from a known location may be compared with a size of the marker in an image captured by the user interaction device and the results of the comparison may provide information regarding the location of the user interaction device.
[0174] In one embodiment using FLANN, the comparison of the images from the user interaction device with the stored images of the markers determines whether any of the images of the markers match. If one of the images from the user interaction device contains a marker which matches a marker in one of the stored images, then the refined location information of the user is set to the location from which the matching stored image was captured from.
[0175] In other examples, an image from the user interaction device may contain a stored marker which may be detected by FLANN as being a match even though the image from the user interaction device may have been captured from a different location than the image of the stored marker. The features of the stored marker may be modified using a perspective transform to match the features of the marker within the image from the user interaction device. The perspective transform modifies the features of the stored marker to appear as if the image was captured from different locations. Once the modifications by the perspective transform identify a match of the modified features of the stored marker with the features of the marker in the image from the user interaction device, the offset of the user's location from which the user captured the image is known with respect to the location from which the image of the stored marker was obtained and may be used to provide refined location information of the user interaction device. More specifically, in this example, the refined location information is set to the location from which the image of the stored marker was captured as modified by the offset which was determined by the perspective transform processing.
[0176] At an act A312, the images received from the user interaction devices and which include the markers (and the determined location information of the user interaction device when the images were captured) may be stored in the database and used for future image processing operations of subsequently captured images to provide location information of devices which captured the subsequent images.
[0177] In one embodiment, it is desired to have accurate information regarding the locations from which the images were captured and information regarding the accuracy of the location information may be used to determine whether images are stored in the database and used for subsequent image processing operations to provide location information of other users. Furthermore, accuracy resulting from the processing of images may be improved when a plurality of markers are present in the images for processing or a plurality of different images of a single marker which were obtained from different locations are available for processing. In some embodiments, a user may upload a plurality of images of a marker which may be slightly rotated from one another or observed from slightly different locations.
[0178] At an act A314, the refined location information may be communicated to one or more of the user interaction devices and used by the devices for implementing operations with respect to experiencing augmented reality or other uses. The refined location information may also be stored within the database which includes the wireless communications devices and their respective locations.
[0179] As mentioned above, other methods are possible and some of the acts may be performed by different entities or omitted. In another example, the markers may not be communicated to the user interaction device, but rather the user interaction device may capture images of its environment and the images may be uploaded to the management device for processing to locate markers and provide refined location information using the located markers by comparison with existing images in a database of the management device.
[0180] In another embodiment, images of markers from the database of the management device may be communicated to the user interaction devices and used by the devices to calculate refined location information. In one embodiment, the user interaction device may be configured to process images captured by the user interaction device including performing object recognition operations to identify the presence of markers and to compare the captured images with the images from the database to provide the refined location information as discussed above with respect to operations of the management device.
[0181] The management device may have greater processing capacity than the processing capacities of the user interaction devices and implementing processing of images using the management device may be faster than processing of the images using the user interaction devices.
[0182] The above discussion describes various methods for determining refined location information regarding locations of user interaction devices in the physical world. The refined location information may be used to accurately associate augmented reality representations at appropriate locations of the physical world as captured by the user interaction devices. For example, the refined location information of the plurality of user interaction devices may be used to associate augmented reality representations with physical objects of the physical world.
[0183] In one example, a plurality of markers may be associated with a plurality of respective virtual objects. Accordingly, when a marker is located in an image of a scene, the user interaction device may depict the virtual object in the image by replacing the marker 14 with the virtual object representation in images depicted using the display (e.g., the puppy representation 18 replaces the marker 14 shown in Fig. 1 ).
[0184] Improvements in processing capabilities and communications bandwidth have led to new implementations and applications of augmented reality systems. Provision of accurate location information utilizing apparatus and methods described herein according to example embodiments of the disclosure enable computing systems, including portable devices, to implement improved augmented reality operations. At least some aspects of the disclosure provide systems, apparatus and methods which enable locations of user interaction devices to be determined with increased accuracy permitting augmented reality representations to be referenced with respect to the physical world with increased accuracy compared with usage of location information determined by conventional methods such as GPS which has inherent limited accuracy and has yet larger error in some areas (e.g., indoors). Accurate location information of the augmented reality user interaction devices enables augmented reality representations to be correctly referenced with respect to the physical world such that different user interaction devices experiencing the augmented reality representations will portray the augmented reality representations correctly with respect to the physical world (i.e., at appropriate locations with respect to physical world locations and objects).
[0185] Some conventional location technologies including GPS have shortcomings with respect to use in augmented reality systems as discussed above. More specifically, civilian entities are currently precluded from obtaining the highest accuracy possible from GPS. In addition, the accuracy of GPS is further degraded at indoor locations or at other locations where the signals from the satellites of the GPS system may encounter obstacles or may not be clearly received. Cell tower triangulation has been used in conventional arrangements to augment GPS readings but this methodology also has shortcomings of being of limited use indoors or in areas which are lacking a sufficient number of cell towers.
[0186] At least some aspects of the disclosure include obtaining and combining location information from different techniques in attempts to accurately locate the positions of the user interaction devices. Location information obtained from aspects of the disclosure may be used separately or in combination with conventional location determination arrangements in some embodiments.
[0187] The protection sought is not to be limited to the disclosed embodiments, which are given by way of example only, but instead is to be limited only by the scope of the appended claims.
[0188] Further, aspects herein have been presented for guidance in construction and/or operation of illustrative embodiments of the disclosure. Applicant(s) hereof consider these described illustrative embodiments to also include, disclose and describe further inventive aspects in addition to those explicitly disclosed. For example, the additional inventive aspects may include less, more and/or alternative features than those described in the illustrative embodiments. In more specific examples, Applicants consider the disclosure to include, disclose and describe methods which include less, more and/or alternative steps than those methods explicitly disclosed as well as apparatus which includes less, more and/or alternative structure than the explicitly disclosed structure.

Claims

CLAIMS: The invention claimed is:
1. An augmented reality method comprising:
accessing first location information regarding a location of a user interaction device in a physical world, wherein the user interaction device is configured to generate an augmented reality representation with respect to the physical world;
using the first location information, generating second location information which has increased accuracy regarding the location of the user interaction device in the physical world; and
communicating augmented data to the user interaction device, and wherein the augmented data comprises the augmented reality representation.
2. The method of claim 1 wherein the communicating comprises communicating the second location information to the user interaction device.
3. The method of claim 2 wherein the augmented reality representation comprises a virtual object, and further comprising associating the virtual object with the physical world.
4. The method of claim 1 further comprising using the first location information, identifying an image of a marker which is proximately located to the user interaction device, and wherein the generating comprises generating using the image of the marker.
5. The method of claim 4 further comprising receiving an image from the user interaction device, and wherein the generating comprises processing the received image with respect to the image of the marker to generate the second location information.
6. The method of claim 1 further comprising receiving information regarding wireless communications of the user interaction device with a plurality of wireless communications devices, and wherein the generating comprises generating the second location information using the information regarding the wireless communications.
7. The method of claim 6 wherein the accessing, the generating and the communicating comprise acts implemented by a management device, and further comprising managing the wireless communications of the user interaction device with the wireless communications devices using the management device.
8. The method of claim 6 wherein the managing comprises communicating control signals from the management device to the user interaction device and the wireless communications devices to control the wireless communications.
9. A computing system comprising:
communications circuitry configured to implement communications externally of the computing system; and
processing circuitry coupled with the communications circuitry, and wherein the processing circuitry is configured to access first location information received by the communications circuitry regarding a location of a user interaction device in a physical world, to generate second location information which has increased accuracy regarding the location of the user interaction device in the physical world, and to control the communications circuitry to communicate augmented data comprising an augmented reality representation to the user interaction device.
10. The system of claim 9 wherein the processing circuitry is configured to control the communications circuitry to communicate the second location information to the user interaction device.
1 1 . The system of claim 10 wherein the augmented reality representation comprises a virtual object.
12. The system of claim 9 wherein the processing circuitry is configured to use the first location information to identify an image of a marker which is proximately located to the user interaction device, and wherein the processing circuitry is configured to generate the second location information using the image of the marker.
13. The system of claim 12 wherein the communications circuitry is configured to receive an image from the user interaction device, and the processing circuitry is configured to process the image with respect to the image of the marker to generate the second location information.
14. The system of claim 9 wherein the communications circuitry is configured to receive information regarding wireless communications of the user interaction device with a plurality of wireless communications devices, and the processing circuitry is configured to use the information regarding the wireless communications to generate the second location information.
15. The system of claim 14 wherein the processing circuitry is configured to manage the wireless communications of the user interaction device with the wireless communications devices.
16. The system of claim 15 wherein the processing circuitry is configured to output control signals to the user interaction device and the other wireless communications devices to control the wireless communications.
17. An augmented reality user interaction device comprising:
a camera;
a display system;
communications circuitry configured to implement wireless communications externally of the user interaction device; and
processing circuitry coupled with the camera, the display system, and the communications circuitry, wherein the processing circuitry is configured to control the display system to generate a plurality of images which comprise image data generated by the camera and augmented data which augments the image data with an augmented reality representation, wherein the processing circuitry is further configured to control the communications circuitry to communicate first location information regarding a location of the user interaction device in a physical world externally of the user interaction device, and wherein the processing circuitry is further configured to access second location information regarding the location of the user interaction device after the outputting of the first location information and wherein the second location information has increased accuracy regarding the location of the user interaction device in the physical world compared with the first location information.
18. The device of claim 17 further comprising location determination circuitry configured to generate the first location information.
19. The device of claim 17 wherein the communications circuitry receives an image of a marker after the outputting the first location information, and the processing circuitry is configured to process a plurality of images captured by the image capture system with respect to the image of the marker, and to generate the second location information using the processing of the images.
20. The device of claim 19 wherein the processing circuitry is configured to receive information regarding a location from which the image of the marker was captured, and to generate the second location information using the information regarding the location from which the image of the marker was captured.
21 . The device of claim 19 wherein the processing circuitry is configured to control the display system to generate the image of the marker.
22. The device of claim 17 wherein the processing circuitry is configured to control the communications circuitry to output a plurality of images captured by the image capture system, and the communications circuitry receives the second location information as a result of the outputting of the images.
23. The device of claim 17 wherein the communications circuitry is configured to implement wireless communications with respect to a plurality of wireless communications devices which are within a communications range of the user interaction device, and wherein the processing circuitry is configured to control the communications circuitry to output information regarding the wireless communications with respect to the wireless communications devices, and the communications circuitry receives the second location information as a result of the outputting of the information regarding the wireless communications.
24. The device of claim 23 wherein the communications circuitry is configured to output a wireless communications signal to the other wireless communications devices, and the communications circuitry is configured to output the information regarding the wireless communications comprising information regarding a moment in time when the wireless communications signal was outputted by the communications circuitry.
25. The device of claim 24 wherein the communications circuitry comprises a speaker configured to output the wireless communications signal comprising a sound wave.
26. The device of claim 25 wherein the processing circuitry is configured to control the speaker to output the wireless communications signal comprising a sound wave having a frequency defined before the outputting.
27. The device of claim 24 wherein the communications circuitry is configured to output the wireless communications signal comprising an electromagnetic wave.
28. The device of claim 24 wherein the communications circuitry is configured to receive a control signal instructing the user interaction device to operate as a sender, and the processing circuitry is configured to control the communications circuitry to output the wireless communications signal as a result of the reception of the control signal.
29. The device of claim 23 wherein the communications circuitry is configured to receive a wireless communications signal from one of the wireless communications devices, and the processing circuitry is configured to control the communications circuitry to output the information regarding the wireless communications comprising information regarding a moment in time when the wireless communications signal was received by the communications circuitry.
30. The device of claim 29 wherein the communications circuitry is configured to receive a control signal instructing the user interaction device to operate as a receiver, and the processing circuitry is configured to monitor the communications circuitry for reception of the wireless communications signal to determine the moment in time when the wireless communications signal was received.
31 . The device of claim 30 wherein the communications circuitry comprises a microphone configured to receive the wireless communications signal comprising a sound wave.
32. The device of claim 31 wherein the processing circuitry is configured to monitor the microphone for reception of the sound wave comprising a frequency defined before the reception.
33. The device of claim 1 7 wherein the processing circuitry is configured to control the communications circuitry to output an image externally of user interaction device which was captured by the camera and which includes a marker.
PCT/US2011/045586 2010-07-30 2011-07-27 Augmented reality and location determination methods and apparatus WO2012015956A2 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US12/847,771 US8493206B2 (en) 2010-07-30 2010-07-30 Augmented reality and location determination methods and apparatus
US12/847,754 US8502659B2 (en) 2010-07-30 2010-07-30 Augmented reality and location determination methods and apparatus
US12/847,754 2010-07-30
US12/847,790 US8519844B2 (en) 2010-07-30 2010-07-30 Augmented reality and location determination methods and apparatus
US12/847,771 2010-07-30
US12/847,790 2010-07-30

Publications (2)

Publication Number Publication Date
WO2012015956A2 true WO2012015956A2 (en) 2012-02-02
WO2012015956A3 WO2012015956A3 (en) 2012-05-03

Family

ID=45530696

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/045586 WO2012015956A2 (en) 2010-07-30 2011-07-27 Augmented reality and location determination methods and apparatus

Country Status (1)

Country Link
WO (1) WO2012015956A2 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8519844B2 (en) 2010-07-30 2013-08-27 Gravity Jack, Inc. Augmented reality and location determination methods and apparatus
WO2013184629A2 (en) * 2012-06-04 2013-12-12 Fluor Technologies Corporation Mobile device for monitoring and controlling facility systems
WO2015142334A1 (en) * 2014-03-20 2015-09-24 Hewlett-Packard Development Company, L.P. Identifying electronic components for augmented reality
WO2017039911A1 (en) * 2015-09-02 2017-03-09 Microsoft Technology Licensing, Llc Localizing devices in an augmented reality environment
US11210705B1 (en) * 2013-10-18 2021-12-28 United Services Automobile Association (Usaa) System and method for transmitting direct advertising information to an augmented reality device
US12141840B1 (en) 2021-11-18 2024-11-12 United Services Automobile Association (Usaa) System and method for transmitting direct advertising information to an augmented reality device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6064335A (en) * 1997-07-21 2000-05-16 Trimble Navigation Limited GPS based augmented reality collision avoidance system
US20040239756A1 (en) * 2003-05-30 2004-12-02 Aliaga Daniel G. Method and apparatus for computing error-bounded position and orientation of panoramic cameras in real-world environments
US20090244097A1 (en) * 2008-03-25 2009-10-01 Leonardo William Estevez System and Method for Providing Augmented Reality
US20100315418A1 (en) * 2008-02-12 2010-12-16 Gwangju Institute Of Science And Technology Tabletop, mobile augmented reality system for personalization and cooperation, and interaction method using augmented reality

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6064335A (en) * 1997-07-21 2000-05-16 Trimble Navigation Limited GPS based augmented reality collision avoidance system
US20040239756A1 (en) * 2003-05-30 2004-12-02 Aliaga Daniel G. Method and apparatus for computing error-bounded position and orientation of panoramic cameras in real-world environments
US20100315418A1 (en) * 2008-02-12 2010-12-16 Gwangju Institute Of Science And Technology Tabletop, mobile augmented reality system for personalization and cooperation, and interaction method using augmented reality
US20090244097A1 (en) * 2008-03-25 2009-10-01 Leonardo William Estevez System and Method for Providing Augmented Reality

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8519844B2 (en) 2010-07-30 2013-08-27 Gravity Jack, Inc. Augmented reality and location determination methods and apparatus
WO2013184629A2 (en) * 2012-06-04 2013-12-12 Fluor Technologies Corporation Mobile device for monitoring and controlling facility systems
WO2013184629A3 (en) * 2012-06-04 2014-02-06 Fluor Technologies Corporation Mobile device for monitoring and controlling facility systems
US11210705B1 (en) * 2013-10-18 2021-12-28 United Services Automobile Association (Usaa) System and method for transmitting direct advertising information to an augmented reality device
WO2015142334A1 (en) * 2014-03-20 2015-09-24 Hewlett-Packard Development Company, L.P. Identifying electronic components for augmented reality
WO2017039911A1 (en) * 2015-09-02 2017-03-09 Microsoft Technology Licensing, Llc Localizing devices in an augmented reality environment
US9865091B2 (en) 2015-09-02 2018-01-09 Microsoft Technology Licensing, Llc Localizing devices in augmented reality environment
CN108027649A (en) * 2015-09-02 2018-05-11 微软技术许可有限责任公司 Locating devices in an augmented reality environment
US12141840B1 (en) 2021-11-18 2024-11-12 United Services Automobile Association (Usaa) System and method for transmitting direct advertising information to an augmented reality device

Also Published As

Publication number Publication date
WO2012015956A3 (en) 2012-05-03

Similar Documents

Publication Publication Date Title
US8519844B2 (en) Augmented reality and location determination methods and apparatus
US8493206B2 (en) Augmented reality and location determination methods and apparatus
US8502659B2 (en) Augmented reality and location determination methods and apparatus
Vo et al. A survey of fingerprint-based outdoor localization
US9869748B2 (en) Locating a mobile device
US20240085189A1 (en) System For Determining Position Both Indoor and Outdoor
AU2012262822B2 (en) Monitoring geofence exit
US9717065B2 (en) Indoor remote triggered location scanning
US11243288B2 (en) Location error radius determination
EP2761896B1 (en) Utilizing relationships between places of relevance
CN105100390A (en) Mobile terminal and method for controlling the mobile terminal
WO2012015956A2 (en) Augmented reality and location determination methods and apparatus
WO2017070969A1 (en) Floor positioning method, network device and mobile terminal
US20200090405A1 (en) Geophysical sensor positioning system
WO2023005482A1 (en) Team recommendation method, team recommendation apparatus, electronic device, and storage medium
Hu Wi-Fi based indoor positioning system using smartphones
Dellosa et al. Modified fingerprinting localization technique of indoor positioning system based on coordinates
US20160198300A1 (en) Mobile Device Distance Measurement and Object Identification by Utilizing the Rotation Vector and Accelerometer
US10452150B2 (en) Electronic map augmentation through pointing gestures background
CN114356182B (en) Article positioning method, device, equipment and storage medium
CN110095792A (en) The method and device of positioning terminal
CN115175747A (en) Slicing storage of geo-located data with predictable query response time
CN113329333A (en) Indoor positioning method and device, computer equipment and storage medium
CN116664812B (en) Visual positioning method, visual positioning system and electronic equipment
US20230366973A1 (en) Map generation system and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11813132

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11813132

Country of ref document: EP

Kind code of ref document: A2