US20140036025A1 - Video conference using an external video stream - Google Patents
Video conference using an external video stream Download PDFInfo
- Publication number
- US20140036025A1 US20140036025A1 US14/050,673 US201314050673A US2014036025A1 US 20140036025 A1 US20140036025 A1 US 20140036025A1 US 201314050673 A US201314050673 A US 201314050673A US 2014036025 A1 US2014036025 A1 US 2014036025A1
- Authority
- US
- United States
- Prior art keywords
- communication device
- portable communication
- video
- cameras
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000004891 communication Methods 0.000 claims abstract description 149
- 238000000034 method Methods 0.000 claims description 53
- 230000000007 visual effect Effects 0.000 claims description 7
- 230000006870 function Effects 0.000 description 14
- 238000010586 diagram Methods 0.000 description 13
- 230000004913 activation Effects 0.000 description 6
- 230000001413 cellular effect Effects 0.000 description 6
- 230000000977 initiatory effect Effects 0.000 description 5
- 230000008569 process Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 238000001514 detection method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000002463 transducing effect Effects 0.000 description 2
- 230000003213 activating effect Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/56—Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities
- H04M3/567—Multimedia conference systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/414—Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
- H04N21/41407—Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance embedded in a portable device, e.g. video client on a mobile phone, PDA, laptop
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/4223—Cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
- H04N21/4363—Adapting the video stream to a specific local network, e.g. a Bluetooth® network
- H04N21/43637—Adapting the video stream to a specific local network, e.g. a Bluetooth® network involving a wireless protocol, e.g. Bluetooth, RF or wireless LAN [IEEE 802.11]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4788—Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/147—Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2203/00—Aspects of automatic or semi-automatic exchanges
- H04M2203/20—Aspects of automatic or semi-automatic exchanges related to features of supplementary services
- H04M2203/2094—Proximity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2242/00—Special services or facilities
- H04M2242/30—Determination of the location of a subscriber
Definitions
- portable communication devices such as mobile phones
- portable communication devices offer a number of additional features to their users, such as text messaging, multimedia playback, and video conferencing.
- the portable device may be used to send/receive a video stream (i.e., both audio and image information) with another party.
- the portable communication device may include a camera or video camera that is used to obtain the images for the video stream.
- the user may need to hold the portable device in an orientation that points the camera of the portable communication device at the user. In this manner, the portable communication device may enable a video conference. Participating in a video conference by holding the portable device in an orientation that allows the device to focus on the user, however, can be awkward and may not be generally conducive to a good user experience.
- a method may include determining, by a portable communication device, whether at least one video camera external to the portable communication device is in a vicinity of the portable communication device; and using video taken by the at least one video camera in a video conference conducted via the portable communication device.
- the method may further include presenting a list of video cameras that are determined to be in the vicinity of the portable communication device to a user of the portable communication device.
- the determining in the method may further include obtaining a location of the portable communication device; and identifying the video cameras in the vicinity of the portable communication device based on the location information. Additionally, obtaining the location of the portable communication device may include obtaining the location using a GPS signal.
- the determining in the method may further include determining the video cameras in a vicinity of the portable communication device based on face recognition techniques used to recognize a user of the portable communication device in video taken by the video cameras.
- the determining in the method may further include determining video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
- the video cameras external to the portable communication device may include public surveillance cameras.
- the method may include using the video taken by the at least one video camera in a video conference by: receiving the video from the at least one video camera at the portable communication device; and transmitting the received video to a party with whom the video conference is to be conducted.
- using the video taken by the at least one video camera in a video conference may further include causing the video from the at least one video camera to be transmitted to a party with whom the video conference is to be conducted.
- Another aspect is directed to a method that may include receiving, at a portable communication device, a request for video cameras that are external to the portable communication device; identifying, by the portable communication device, video cameras that are external to the portable communication device and that are in a vicinity of the portable communication device; and presenting, via a visual display on the portable communication device, the identified video cameras.
- the identifying may further include obtaining a location of the portable communication device; and determining the video cameras in the vicinity of the portable communication device based on the location.
- the identifying may further include determining the video cameras in the vicinity of the portable communication device based on face recognition techniques used to recognize the user in video taken by the video cameras.
- the identifying may further include determining the video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
- the method may further include enabling a video conference in which video of a user of the portable communication device corresponds to video taken by one of the video cameras external to the portable communication device and in which audio input to the portable communication device corresponds to audio recorded by the portable communication device.
- the method may further include presenting a graphical interface configured to facilitate selection of one of the identified video cameras for use in a video conference.
- the portable communication device may include a memory to store instructions and a processor to execute the instructions.
- the device may determine whether video cameras external to the portable communication device are in a vicinity of the portable communication device and enable a video conference in which video taken by at least one of the determined video cameras is used as a video source for the video conference.
- the portable communication device may further include a display, wherein the processor further executes instructions to present a list of the video cameras determined to be in the vicinity of the portable communication device to the user of the portable communication device.
- the processor of the portable communication may further execute instructions to obtain a location of the portable communication device and determine the video cameras in a vicinity of the portable communication device based on the location.
- the processor of the portable communication device may further execute instructions to obtain the location of the portable communication device by obtaining the location using a GPS signal.
- the processor of the portable communication device may further execute instructions to determine the video cameras in a vicinity of the portable communication device based on face recognition techniques used to recognize a user of the portable communication device in video taken by the video cameras.
- the processor of the portable communication device may further execute instructions to determine the video cameras in a vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
- the video cameras external to the portable communication device may include public surveillance cameras.
- the computer-readable medium includes instructions to determine, by a portable communication device, whether a first video camera external to the portable communication device is in a vicinity of the portable communication device; and instructions to conduct a video conference using video taken by the first video camera.
- Another aspect is directed to a device that may comprise means for identifying, by a portable communication device, whether video cameras external to the portable communication device are in a vicinity of the portable communication device; and means for conducting a video conference from the portable communication device using video taken by at least one of the identified video cameras.
- FIG. 1 is an exemplary diagram of an environment or system in which concepts described herein may be implemented
- FIG. 2 is a diagram illustrating a front view of exemplary external components of a portable communication device
- FIG. 3 is a diagram illustrating a rear view of exemplary external components of the portable communication device
- FIG. 4 is a diagram illustrating exemplary internal components of the device shown in FIGS. 2 and 3 ;
- FIG. 5 is a flow chart illustrating exemplary operations for providing a video conference
- FIG. 6 is a diagram of an exemplary system that conceptually illustrates the determination of external cameras that are watching a user
- FIG. 7 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user based on the system shown in FIG. 6 ;
- FIG. 8 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user consistent with another implementation
- FIG. 9 is a diagram of an exemplary system that conceptually illustrates the determination of external cameras that are in the vicinity of the user based on detection using local wireless signals;
- FIG. 10 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user based on the system shown in FIG. 9 ;
- FIG. 11 is an exemplary graphical interface that a device may present to a user.
- a portable communication device may determine whether an external camera, such as a surveillance camera, is currently viewing the user of the portable communication device.
- the image or video feed from the external camera may be used as the image/video source for a video conference.
- FIG. 1 is an exemplary diagram of an environment or system in which concepts described herein may be implemented.
- the environment may include a portable communication device 100 that may be carried by a user 105 .
- Portable communication device 100 may be a portable device such as a cellular phone, a personal digital assistant (PDA), a laptop, or another type of computation or communication device, a thread or process running on one of these devices, and/or an object executable by one of these devices.
- Portable communication device 100 may be a device that provides connectivity to network 110 .
- portable communication device 100 may connect wirelessly to network 110 to communicate voice, video, and/or data through network 110 .
- Network 110 may include a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), a telephone network, such as the Public Switched Telephone Network (PSTN) or a cellular telephone network, an intranet, the Internet, or a combination of networks.
- PSTN Public Switched Telephone Network
- User 105 of portable communication device 100 may connect to network 110 via wired, wireless, and/or optical connections.
- network 110 may represent a number of different types of networks.
- network 110 may represent a traditional cellular network through which a user may place wireless calls and a data network, such as the Internet, over which the user may browse various web pages.
- Camera 120 may be a public surveillance camera (e.g., a camera operated by police or other law enforcement personnel), a private camera, or any other type of camera.
- Structure 125 may be any structure, such as a building or utility pole, appropriate for mounting camera 120 .
- cameras 120 are commonly used by governments and/or businesses to reduce crime and maintain order. In some metropolitan areas, security cameras are so prevalent that an average citizen may be filmed numerous times each day. Security cameras can also provide a sense of security for citizens. Thus, it is reasonable to assume that the coverage of security cameras will continue to increase.
- FIG. 2 is a diagram illustrating a front view of exemplary external components of portable communication device 100 .
- device 100 may include a housing 205 , a microphone 210 , a speaker 220 , a keypad 230 , function keys 240 , a display 250 , and a camera/video activation button 260 .
- Housing 205 may include a structure configured to contain components of device 100 .
- housing 205 may be formed from plastic and may be configured to support microphone 210 , speaker 220 , keypad 230 , function keys 240 , display 250 , and camera/video activation button 260 .
- Microphone 210 may include any component capable of transducing air pressure waves to a corresponding electrical signal. For example, a user may speak into microphone 210 during a telephone call.
- Speaker 220 may include any component capable of transducing an electrical signal to a corresponding sound wave. For example, a user may listen to music through speaker 220 .
- Keypad 230 may include any component capable of providing input to device 100 .
- Keypad 230 may include a standard telephone keypad.
- Keypad 230 may also include one or more special purpose keys.
- each key of keypad 230 may be, for example, a pushbutton.
- a user may utilize keypad 230 for entering information, such as text or a phone number, or activating a special function.
- Function keys 240 may include any component capable of providing input to device 100 .
- Function keys 240 may include a key that permits a user to cause device 100 to perform one or more operations.
- the functionality associated with a key of function keys 240 may change depending on the mode of device 100 .
- function keys 240 may perform a variety of operations, such as placing a telephone call, playing various media, setting various camera features (e.g., focus, zoom, etc.) or accessing an application.
- Function keys 240 may include a key that provides a cursor function and a select function. In one implementation, each key of function keys 240 may be, for example, a pushbutton.
- Display 250 may include any component capable of providing visual information.
- display 250 may be a liquid crystal display (LCD).
- display 250 may be any one of other display technologies, such as a plasma display panel (PDP), a field emission display (FED), a thin film transistor (TFT) display, etc.
- Display 250 may display, for example, text, image, and/or video information. Display 250 may also operate as a view finder.
- Camera/video activation button 260 may be a pushbutton that enables a user to initiate taking an image or to initiate taking a series of images (i.e., video). Camera/video activation button 260 may activate a camera located at, for example, the back of device 100 .
- FIG. 3 is a diagram illustrating a rear view of exemplary external components of the device depicted in FIG. 2 .
- device 100 may include a camera 370 .
- Camera 370 may include any component capable of capturing an image or a series of images. Camera 370 may be a digital camera. Display 250 may operate as a view finder when a user of device 100 operates camera 370 . Camera 370 may provide for automatic and/or manual adjustment of a camera setting. In one implementation, device 100 may include functionality that is displayable on display 250 to allow a user to adjust a camera setting. For example, a user may be able to adjust a setting of camera 370 by operating one or more of function keys 240 .
- Camera 370 may be used to take a picture, a series of pictures, and/or real-time video that may be used in a video conference. As is described in more detail below, a user of device 100 may, in certain situations, participate in video conferences through device 100 using images or video from a camera external to device 100 . Accordingly, in some situations camera 370 is not necessary and may thus not be provided on device 100 .
- Device 100 is intended to be broadly interpreted to include any type of communication device.
- device 100 may include a handheld device, such as a wireless telephone, a personal digital assistant (PDA), or a computer.
- PDA personal digital assistant
- FIGS. 2 and 3 illustrate exemplary external components of device 100
- device 100 may contain fewer, different, or additional external components than the external components depicted in FIG. 1 .
- one or more external components of device 100 may include the capabilities of one or more other external components of device 100 .
- display 250 may be an input component (e.g., a touch screen). Additionally, or alternatively, the external components may be arranged differently than the external components depicted in FIG. 1 .
- FIG. 4 is a diagram illustrating exemplary internal components of device 100 .
- device 100 may include microphone 210 , speaker 220 , keypad 230 , function keys 240 , display 250 , camera/video activation button 260 , camera 370 , a memory 400 , a transceiver 440 , a control unit 450 , a GPS component 480 , and a directional component 490 .
- Microphone 210 , speaker 220 , keypad 230 , function keys 240 , display 250 , camera/video activation button 260 , and camera 370 may represent the circuitry or other processing elements associated with the corresponding elements, as shown in FIGS. 2 and 3 . Accordingly, no further description of these elements will be provided with respect to FIG. 4 .
- Memory 400 may include any type of computer-readable medium to store data and instructions related to the operation and use of device 100 .
- memory 400 may include a memory component, such as a random access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), synchronous dynamic random access memory (SRAM), ferroelectric random access memory, (FRAM), a read only memory (ROM), a programmable read only memory (PROM), an erasable programmable read only memory (EPROM), electrically erasable read only memory (EEPROM), and/or flash memory.
- RAM random access memory
- DRAM dynamic random access memory
- SRAM static random access memory
- SRAM synchronous dynamic random access memory
- FRAM ferroelectric random access memory
- ROM read only memory
- PROM programmable read only memory
- EPROM erasable programmable read only memory
- EEPROM electrically erasable read only memory
- memory 400 may include a storage component, such as a magnetic storage component (e.g., a hard drive), a compact disc (CD) drive, a digital versatile disc (DVD) drive, or another type of computer-readable medium.
- a storage component such as a magnetic storage component (e.g., a hard drive), a compact disc (CD) drive, a digital versatile disc (DVD) drive, or another type of computer-readable medium.
- Memory 400 may also include an external storing component, such as a Universal Serial Bus (USB) memory stick, a digital camera memory card, and/or a Subscriber Identity Module (SIM) card.
- USB Universal Serial Bus
- SIM Subscriber Identity Module
- Transceiver 440 may include any component capable of transmitting and receiving information.
- transceiver 440 may include a radio circuit that provides wireless communication with a network or another device.
- GPS (Global Positioning System) component 480 may include logic to obtain position information of device 100 using the Global Navigation Satellite System (GNSS).
- GNSS Global Navigation Satellite System
- other technologies may alternatively be used to obtain positional information of device 100 instead of GPS component 480 .
- the location of device 100 may be determined through contact with local short range wireless networks or via triangulation with cellular towers.
- Device 100 may optionally include directional component 490 , which may include, for example, gyroscope, a compass or accelerometer logic that may be used to ascertain the directional orientation of device 100 and potentially of the user holding device 100 .
- directional component 490 may include, for example, gyroscope, a compass or accelerometer logic that may be used to ascertain the directional orientation of device 100 and potentially of the user holding device 100 .
- Control unit 450 may include any logic that may interpret and execute instructions, and may control the overall operation of device 100 .
- Logic as used herein, may include hardware, software, and/or a combination of hardware and software.
- Control unit 450 may include, for example, a general-purpose processor, a microprocessor, a data processor, a co-processor, application specific integrated circuit (ASIC), field programmable gate array (FPGA), and/or a network processor.
- Control unit 450 may access instructions from memory 400 , from other components of device 100 , and/or from a source external to device 100 (e.g., a network or another device).
- Control unit 450 may provide for different operational modes associated with device 100 . Additionally, control unit 450 may operate in multiple modes simultaneously. For example, control unit 450 may operate in a camera mode, a video conference mode, and/or a telephone mode. The operation of control unit 450 and device 100 in videoconference mode will be particularly described in more detail below.
- FIGS. 2-4 illustrate exemplary internal and external components
- device 100 may include fewer, additional, and/or different components than the exemplary components depicted in FIGS. 2-4 .
- one or more components of device 100 may include the capabilities of one or more other components of device 100 .
- transceiver 440 and/or control unit 450 may include their own on-board memory 400 .
- hardwired circuitry may be used in place of or in combination with software instructions to implement operations of device 100 .
- implementations described herein are not limited to any specific combination of hardware circuitry and software.
- device 100 may be used to facilitate a video conference.
- the images for the video conference may be obtained from a camera external to device 100 .
- a video conference is to be broadly interpreted as communication between two or more parties in which both audio and visual information is transferred.
- the visual information may be video (i.e., a sequence of images that are updated fast enough to appear to fluidly capture movement) or a series of images in which the images are separated by longer time periods. Further, the visual information for a video conference does not need to be transmitted in both directions. That is, a video conference may be conducted in which only one party receives visual information.
- a user of a portable communication device such as a user of device 100 , may wish to establish a video conference with another party.
- the user may, for example, be speaking to the other party via device 100 and during the course of the conversation, the users may decide that they would like to view images of each other.
- the user when initially dialing the called party, may initiate the conversation as a video conference.
- FIG. 5 is a flow chart illustrating exemplary operations for providing a video conference.
- the external camera may be any camera that is not part of portable communication device 100 and that is focused on the user or is located in the vicinity of device 100 .
- public or private surveillance cameras or security cameras may be used.
- GPS component 480 and/or directional component 490 may be used to locate the user and the location of the user may then be compared to a list of known available external cameras.
- external cameras may use face recognition technologies to identify people and then proactively “publish” or notify users when they are in focus. These techniques are discussed in more detail below.
- the user may use camera 370 on portable communication device 100 to generate video for the video conference (acts 502 and 503 ). Alternatively, or if device 100 does not have a camera, the user may decide to forgo transmitting video to the other party.
- video from the external video camera may be sent to the other party as part of the video conference (acts 502 and 504 ). If, for example, an available public surveillance camera is focusing on the user, the other party may view video from the surveillance camera as the video that is included in the video conference.
- the other party may receive the video in a number of possible ways. For example, portable communication device 100 may receive video from the nearby external camera via a short range wireless signal (e.g., a Bluetooth or WiFi signal), and may then forward the video to the other party over network 110 . Alternatively, the other party may receive video directly from the external camera over network 110 . In this case, the other party to the video conference may be forwarded a link or address associated with the external camera so that the other party may begin to receive the video from the external camera.
- a short range wireless signal e.g., a Bluetooth or WiFi signal
- the determination of whether an external camera is available for the video conference can be performed using a number of possible techniques.
- a user may, at any particular time, wish to know which, if any, external cameras are viewing or “watching” the user.
- the user may, through portable communication device 100 , request that the device 100 determine which external cameras are watching the user.
- the results of this request may then be used in initiating a video conference, as shown in FIG. 5 .
- the list of external cameras that are watching the user may simply be shown to the user for amusement or informational purposes.
- FIG. 6 is a diagram of a system 600 that conceptually illustrates the determination of external cameras that are watching the user using information from GPS component 480 and/or a directional component 490 . Assume that user 105 would like to use portable communication device 100 to determine whether any external cameras are available and watching the user.
- device 100 includes a data structure, labeled as external camera list 601 .
- system 600 additionally includes external cameras 610 and 615 . External cameras 610 and 615 may be, for example, public surveillance or security cameras.
- External camera list 601 may be, for example, a list stored in memory 400 of portable communication device 100 .
- External camera list 601 may generally describe the external cameras that are available for possible use as a video source.
- external camera list 601 includes a camera identification (ID) field 620 , camera location field 625 , and camera orientation field 630 .
- Camera identification field 620 may store a name or identification coded associated with the external cameras 610 and 615
- camera location field 625 may store an indication of the location of cameras 610 and 615 (e.g., as an address or a latitude and longitude value)
- camera orientation field 630 may store an indication of the orientation of cameras 610 and 615 (e.g., the direction the camera is pointing, whether the camera is panning, etc.).
- external camera list 601 instead of being stored locally in device 100 , may be stored at, for example, a database located externally from device 100 and that may be accessed remotely over network 110 .
- Public surveillance and security cameras are becoming increasingly prevalent.
- the video from at least some of these cameras may be made publicly available.
- cameras owned by government organizations may be required by law to make the location and video from the cameras available.
- private companies that own surveillance cameras may, for free or for a fee, make the video taken by the cameras available.
- the location and optionally, the orientation of the video cameras may be published by the entity that operates the camera. This information may form the basis for external camera list 601 .
- FIG. 7 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user 105 based on the system shown in FIG. 6 .
- the user 105 may indicate that the user 105 would like to view a list of available external cameras (act 701 ).
- portable communication device 100 may obtain location information of the user 105 (act 702 ). The location information may be obtained using GPS location via GPS component 480 . Other techniques for obtaining the location of the user 105 , such as triangulation using cellular phone towers may alternately be used to obtain the position of the user 105 .
- Portable communication device 100 may optionally obtain the orientation of the device 100 (act 703 ). The orientation (e.g., the direction the user is facing) may be obtained via directional component 490 . To obtain the orientation of the user 105 , it may be assumed that the user 105 is carrying or holding portable communication device 100 in a particular orientation when the user 105 requests the list of available external cameras.
- portable communication device 100 may search external camera list 601 for external cameras that are likely to be focused on the user 105 . This list of external cameras may then be displayed or otherwise presented to the user 105 (act 704 ). In situations in which the user 105 is initiating a video conference, the returned list of video cameras may be used to obtain video for the video conference.
- FIG. 8 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user 105 consistent with another implementation.
- automated face recognition techniques may be used to identify when a user is in view of a camera.
- Video from a network of external video cameras such as the cameras operated by a police force in a particular jurisdiction, may be continuously monitored and analyzed to recognize faces in the cameras (act 801 ). Automated facial recognition techniques are known in the art and will not be described further herein. In one implementation, users that wish to use the service may be required to pre-register images of their face with the operator of the video cameras.
- the user 105 may indicate, via portable electronic device 100 , that the user 105 would like to view a list of available external cameras (act 802 ).
- portable communication device 100 may obtain a list of any cameras in which the user's face is currently (or recently) visible (act 803 ). This list may be obtained, for example, by querying a database or other server device over network 110 .
- the database or server device may be dynamically updated as users move in and out of a network of available external cameras as their faces are recognized.
- This list of external cameras that are currently viewing the user 105 may then be displayed or otherwise presented to the user 105 (act 804 ). In situations in which the user is initiating a video conference, the returned list of video cameras may be used to obtain video for the video conference.
- FIGS. 9 and 10 are diagrams illustrating exemplary operations for determining which external cameras are watching the user consistent with another implementation.
- detection via local wireless signals may be used to identify when a user is near an external camera.
- FIG. 9 is a diagram of a system 900 that conceptually illustrates the determination of external cameras that are in the vicinity of the user based on detection using local wireless signals.
- System 900 includes external cameras 910 and 915 .
- External cameras 910 and 915 may be, for example, public surveillance or security cameras.
- External cameras 910 and 915 may include respective wireless transceivers 911 and 916 through which the cameras may communicate with local devices.
- external cameras 910 and 915 may continuously broadcast, via radio or light signals, information that indicates the presence of the broadcasting camera, called “presence information” herein.
- wireless transceivers 911 and 916 may be additionally used by external cameras 910 and 915 to receive requests for and transmit video information to device 100 .
- FIG. 10 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user based on the system shown in FIG. 9 .
- the user may indicate that the user would like to view a list of available external cameras (act 1001 ).
- portable communication device 100 may determine if it is currently receiving or has recently received wireless presence information from a local external camera (act 1002 ).
- a list of determined external cameras (if any) may then be displayed or otherwise presented to the user (act 1003 ).
- device 100 may additionally obtain and display an image or video stream currently or recently taken by the external camera.
- a user such as user 105 shown in FIG. 6 or 9 , decides to initiate a video conference with another party.
- the user 105 may first control device 100 to determine whether any external cameras are available.
- the user 105 may first dial the called party and then, during the course of the call, check to determine whether an external camera is available.
- device 100 may check whether any external cameras are available. Device 100 may perform this check using one or more of the techniques discussed with respect to FIGS. 7 , 8 , and 10 .
- Device 100 may allow the user 105 to choose which, if any, of these two available cameras to use in the video conference.
- FIG. 11 is an exemplary graphical interface 1100 that device 100 may present to user 105 in display 250 .
- Interface 1100 allows the user to select an external camera.
- labels 1110 and 1120 are shown corresponding to the two available cameras.
- images 1115 and 1120 which may represent current or recently taken images provided by each of the available cameras, are shown in interface 1100 .
- the user may select one of the external cameras to use in the video conference.
- device 100 may automatically determine which of multiple external cameras are taking the most relevant images of the user and automatically select this external camera. This determination may be based on, for example, the presence or size of the face of the user 105 in the image, whether the camera is focused on user 105 , whether the orientation of the camera is appropriate, etc.
- audio spoken by the user 105 and received by microphone 210 may be transmitted over network 110 to the called party.
- Video from the selected camera may also be transmitted to the called party.
- the video may be transmitted directly to the called party (i.e., without first being transmitted through device 100 ) and the audio from user 105 may be transmitted as a separate stream to the called party from user 105 .
- the video may be first sent to device 100 , such as via a short range local wireless network, and then forwarded, with the audio of the user, to the called party.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Telephonic Communication Services (AREA)
- Telephone Function (AREA)
Abstract
A portable communication device may determine whether a video camera is in a vicinity of the portable communication device. If so, video taken by the video camera may be used in a video conference conducted from the portable communication device.
Description
- This is a continuation of and claims priority to U.S. patent application Ser. No. 11/971,493 filed Jan. 9, 2008, titled “Video Conference Using An External Video Stream,” the entirety of which is incorporated by reference herein.
- The proliferation of portable communication devices, such as mobile phones, has grown tremendously within the past decade. In addition to voice conversations, portable communication devices offer a number of additional features to their users, such as text messaging, multimedia playback, and video conferencing. When conducting a video conferencing using a portable communication device, the portable device may be used to send/receive a video stream (i.e., both audio and image information) with another party. The portable communication device may include a camera or video camera that is used to obtain the images for the video stream. The user may need to hold the portable device in an orientation that points the camera of the portable communication device at the user. In this manner, the portable communication device may enable a video conference. Participating in a video conference by holding the portable device in an orientation that allows the device to focus on the user, however, can be awkward and may not be generally conducive to a good user experience.
- According to one aspect, a method may include determining, by a portable communication device, whether at least one video camera external to the portable communication device is in a vicinity of the portable communication device; and using video taken by the at least one video camera in a video conference conducted via the portable communication device.
- The method may further include presenting a list of video cameras that are determined to be in the vicinity of the portable communication device to a user of the portable communication device.
- The determining in the method may further include obtaining a location of the portable communication device; and identifying the video cameras in the vicinity of the portable communication device based on the location information. Additionally, obtaining the location of the portable communication device may include obtaining the location using a GPS signal.
- The determining in the method may further include determining the video cameras in a vicinity of the portable communication device based on face recognition techniques used to recognize a user of the portable communication device in video taken by the video cameras.
- The determining in the method may further include determining video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
- Further, the video cameras external to the portable communication device may include public surveillance cameras.
- Further, the method may include using the video taken by the at least one video camera in a video conference by: receiving the video from the at least one video camera at the portable communication device; and transmitting the received video to a party with whom the video conference is to be conducted.
- Further, in the method, using the video taken by the at least one video camera in a video conference, may further include causing the video from the at least one video camera to be transmitted to a party with whom the video conference is to be conducted.
- Another aspect is directed to a method that may include receiving, at a portable communication device, a request for video cameras that are external to the portable communication device; identifying, by the portable communication device, video cameras that are external to the portable communication device and that are in a vicinity of the portable communication device; and presenting, via a visual display on the portable communication device, the identified video cameras.
- Further, in the method, the identifying may further include obtaining a location of the portable communication device; and determining the video cameras in the vicinity of the portable communication device based on the location.
- Further, in the method, the identifying may further include determining the video cameras in the vicinity of the portable communication device based on face recognition techniques used to recognize the user in video taken by the video cameras.
- Further, in the method, the identifying may further include determining the video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
- Further, the method may further include enabling a video conference in which video of a user of the portable communication device corresponds to video taken by one of the video cameras external to the portable communication device and in which audio input to the portable communication device corresponds to audio recorded by the portable communication device.
- Further, the method may further include presenting a graphical interface configured to facilitate selection of one of the identified video cameras for use in a video conference.
- Another aspect is directed to a portable communication device. The portable communication device may include a memory to store instructions and a processor to execute the instructions. The device may determine whether video cameras external to the portable communication device are in a vicinity of the portable communication device and enable a video conference in which video taken by at least one of the determined video cameras is used as a video source for the video conference.
- The portable communication device may further include a display, wherein the processor further executes instructions to present a list of the video cameras determined to be in the vicinity of the portable communication device to the user of the portable communication device.
- Further, the processor of the portable communication may further execute instructions to obtain a location of the portable communication device and determine the video cameras in a vicinity of the portable communication device based on the location.
- Further, the processor of the portable communication device may further execute instructions to obtain the location of the portable communication device by obtaining the location using a GPS signal.
- Further, the processor of the portable communication device may further execute instructions to determine the video cameras in a vicinity of the portable communication device based on face recognition techniques used to recognize a user of the portable communication device in video taken by the video cameras.
- Further, the processor of the portable communication device may further execute instructions to determine the video cameras in a vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
- Further, the video cameras external to the portable communication device may include public surveillance cameras.
- Another aspect is directed to a computer-readable medium containing instructions executable by at least one processor. The computer-readable medium includes instructions to determine, by a portable communication device, whether a first video camera external to the portable communication device is in a vicinity of the portable communication device; and instructions to conduct a video conference using video taken by the first video camera.
- Another aspect is directed to a device that may comprise means for identifying, by a portable communication device, whether video cameras external to the portable communication device are in a vicinity of the portable communication device; and means for conducting a video conference from the portable communication device using video taken by at least one of the identified video cameras.
- The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate exemplary embodiments described herein and, together with the description, explain these exemplary embodiments. In the drawings:
-
FIG. 1 is an exemplary diagram of an environment or system in which concepts described herein may be implemented; -
FIG. 2 is a diagram illustrating a front view of exemplary external components of a portable communication device; -
FIG. 3 is a diagram illustrating a rear view of exemplary external components of the portable communication device; -
FIG. 4 is a diagram illustrating exemplary internal components of the device shown inFIGS. 2 and 3 ; -
FIG. 5 is a flow chart illustrating exemplary operations for providing a video conference; -
FIG. 6 is a diagram of an exemplary system that conceptually illustrates the determination of external cameras that are watching a user; -
FIG. 7 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user based on the system shown inFIG. 6 ; -
FIG. 8 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user consistent with another implementation; -
FIG. 9 is a diagram of an exemplary system that conceptually illustrates the determination of external cameras that are in the vicinity of the user based on detection using local wireless signals; -
FIG. 10 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user based on the system shown inFIG. 9 ; and -
FIG. 11 is an exemplary graphical interface that a device may present to a user. - The following detailed description refers to the accompanying drawings. The same reference numbers in different drawings may identify the same or similar elements. Also, the following description does not limit the invention.
- Surveillance cameras for law enforcement or other agencies can be prevalent in certain areas. Consistent with aspects described herein, a portable communication device may determine whether an external camera, such as a surveillance camera, is currently viewing the user of the portable communication device. The image or video feed from the external camera may be used as the image/video source for a video conference.
-
FIG. 1 is an exemplary diagram of an environment or system in which concepts described herein may be implemented. The environment may include aportable communication device 100 that may be carried by auser 105.Portable communication device 100 may be a portable device such as a cellular phone, a personal digital assistant (PDA), a laptop, or another type of computation or communication device, a thread or process running on one of these devices, and/or an object executable by one of these devices.Portable communication device 100 may be a device that provides connectivity to network 110. For example,portable communication device 100 may connect wirelessly to network 110 to communicate voice, video, and/or data throughnetwork 110. -
Network 110 may include a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), a telephone network, such as the Public Switched Telephone Network (PSTN) or a cellular telephone network, an intranet, the Internet, or a combination of networks.User 105 ofportable communication device 100 may connect to network 110 via wired, wireless, and/or optical connections. Although shown as asingle network 110,network 110 may represent a number of different types of networks. For example, whendevice 100 is a cellular phone with web browsing capabilities,network 110 may represent a traditional cellular network through which a user may place wireless calls and a data network, such as the Internet, over which the user may browse various web pages. - Assume that the environment shown in
FIG. 1 includes acamera 120 mounted on astructure 125.Camera 120 may be a public surveillance camera (e.g., a camera operated by police or other law enforcement personnel), a private camera, or any other type of camera.Structure 125 may be any structure, such as a building or utility pole, appropriate for mountingcamera 120. - In general, in some locations,
cameras 120 are commonly used by governments and/or businesses to reduce crime and maintain order. In some metropolitan areas, security cameras are so prevalent that an average citizen may be filmed numerous times each day. Security cameras can also provide a sense of security for citizens. Thus, it is reasonable to assume that the coverage of security cameras will continue to increase. -
FIG. 2 is a diagram illustrating a front view of exemplary external components ofportable communication device 100. As illustrated,device 100 may include ahousing 205, amicrophone 210, aspeaker 220, akeypad 230,function keys 240, adisplay 250, and a camera/video activation button 260. -
Housing 205 may include a structure configured to contain components ofdevice 100. For example,housing 205 may be formed from plastic and may be configured to supportmicrophone 210,speaker 220,keypad 230,function keys 240,display 250, and camera/video activation button 260. -
Microphone 210 may include any component capable of transducing air pressure waves to a corresponding electrical signal. For example, a user may speak intomicrophone 210 during a telephone call.Speaker 220 may include any component capable of transducing an electrical signal to a corresponding sound wave. For example, a user may listen to music throughspeaker 220. -
Keypad 230 may include any component capable of providing input todevice 100.Keypad 230 may include a standard telephone keypad.Keypad 230 may also include one or more special purpose keys. In one implementation, each key ofkeypad 230 may be, for example, a pushbutton. A user may utilizekeypad 230 for entering information, such as text or a phone number, or activating a special function. -
Function keys 240 may include any component capable of providing input todevice 100.Function keys 240 may include a key that permits a user to causedevice 100 to perform one or more operations. The functionality associated with a key offunction keys 240 may change depending on the mode ofdevice 100. For example,function keys 240 may perform a variety of operations, such as placing a telephone call, playing various media, setting various camera features (e.g., focus, zoom, etc.) or accessing an application.Function keys 240 may include a key that provides a cursor function and a select function. In one implementation, each key offunction keys 240 may be, for example, a pushbutton. -
Display 250 may include any component capable of providing visual information. For example, in one implementation,display 250 may be a liquid crystal display (LCD). In another implementation,display 250 may be any one of other display technologies, such as a plasma display panel (PDP), a field emission display (FED), a thin film transistor (TFT) display, etc.Display 250 may display, for example, text, image, and/or video information.Display 250 may also operate as a view finder. - Camera/
video activation button 260 may be a pushbutton that enables a user to initiate taking an image or to initiate taking a series of images (i.e., video). Camera/video activation button 260 may activate a camera located at, for example, the back ofdevice 100. -
FIG. 3 is a diagram illustrating a rear view of exemplary external components of the device depicted inFIG. 2 . As illustrated, in addition to the components previously described,device 100 may include acamera 370. -
Camera 370 may include any component capable of capturing an image or a series of images.Camera 370 may be a digital camera.Display 250 may operate as a view finder when a user ofdevice 100 operatescamera 370.Camera 370 may provide for automatic and/or manual adjustment of a camera setting. In one implementation,device 100 may include functionality that is displayable ondisplay 250 to allow a user to adjust a camera setting. For example, a user may be able to adjust a setting ofcamera 370 by operating one or more offunction keys 240. -
Camera 370 may be used to take a picture, a series of pictures, and/or real-time video that may be used in a video conference. As is described in more detail below, a user ofdevice 100 may, in certain situations, participate in video conferences throughdevice 100 using images or video from a camera external todevice 100. Accordingly, in somesituations camera 370 is not necessary and may thus not be provided ondevice 100. -
Device 100 is intended to be broadly interpreted to include any type of communication device. For example,device 100 may include a handheld device, such as a wireless telephone, a personal digital assistant (PDA), or a computer. Accordingly, althoughFIGS. 2 and 3 illustrate exemplary external components ofdevice 100, in other implementations,device 100 may contain fewer, different, or additional external components than the external components depicted inFIG. 1 . Additionally, or alternatively, one or more external components ofdevice 100 may include the capabilities of one or more other external components ofdevice 100. For example,display 250 may be an input component (e.g., a touch screen). Additionally, or alternatively, the external components may be arranged differently than the external components depicted inFIG. 1 . -
FIG. 4 is a diagram illustrating exemplary internal components ofdevice 100. As illustrated,device 100 may includemicrophone 210,speaker 220,keypad 230,function keys 240,display 250, camera/video activation button 260,camera 370, amemory 400, atransceiver 440, acontrol unit 450, aGPS component 480, and adirectional component 490.Microphone 210,speaker 220,keypad 230,function keys 240,display 250, camera/video activation button 260, andcamera 370 may represent the circuitry or other processing elements associated with the corresponding elements, as shown inFIGS. 2 and 3 . Accordingly, no further description of these elements will be provided with respect toFIG. 4 . -
Memory 400 may include any type of computer-readable medium to store data and instructions related to the operation and use ofdevice 100. For example,memory 400 may include a memory component, such as a random access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), synchronous dynamic random access memory (SRAM), ferroelectric random access memory, (FRAM), a read only memory (ROM), a programmable read only memory (PROM), an erasable programmable read only memory (EPROM), electrically erasable read only memory (EEPROM), and/or flash memory. Additionally,memory 400 may include a storage component, such as a magnetic storage component (e.g., a hard drive), a compact disc (CD) drive, a digital versatile disc (DVD) drive, or another type of computer-readable medium.Memory 400 may also include an external storing component, such as a Universal Serial Bus (USB) memory stick, a digital camera memory card, and/or a Subscriber Identity Module (SIM) card. -
Transceiver 440 may include any component capable of transmitting and receiving information. For example,transceiver 440 may include a radio circuit that provides wireless communication with a network or another device. - GPS (Global Positioning System)
component 480 may include logic to obtain position information ofdevice 100 using the Global Navigation Satellite System (GNSS). In some implementations, other technologies may alternatively be used to obtain positional information ofdevice 100 instead ofGPS component 480. For example, the location ofdevice 100 may be determined through contact with local short range wireless networks or via triangulation with cellular towers. -
Device 100 may optionally includedirectional component 490, which may include, for example, gyroscope, a compass or accelerometer logic that may be used to ascertain the directional orientation ofdevice 100 and potentially of theuser holding device 100. -
Control unit 450 may include any logic that may interpret and execute instructions, and may control the overall operation ofdevice 100. Logic, as used herein, may include hardware, software, and/or a combination of hardware and software.Control unit 450 may include, for example, a general-purpose processor, a microprocessor, a data processor, a co-processor, application specific integrated circuit (ASIC), field programmable gate array (FPGA), and/or a network processor.Control unit 450 may access instructions frommemory 400, from other components ofdevice 100, and/or from a source external to device 100 (e.g., a network or another device). -
Control unit 450 may provide for different operational modes associated withdevice 100. Additionally,control unit 450 may operate in multiple modes simultaneously. For example,control unit 450 may operate in a camera mode, a video conference mode, and/or a telephone mode. The operation ofcontrol unit 450 anddevice 100 in videoconference mode will be particularly described in more detail below. - Although
FIGS. 2-4 illustrate exemplary internal and external components, in other implementations,device 100 may include fewer, additional, and/or different components than the exemplary components depicted inFIGS. 2-4 . In still other implementations, one or more components ofdevice 100 may include the capabilities of one or more other components ofdevice 100. For example,transceiver 440 and/orcontrol unit 450 may include their own on-board memory 400. Additionally, hardwired circuitry may be used in place of or in combination with software instructions to implement operations ofdevice 100. Thus, implementations described herein are not limited to any specific combination of hardware circuitry and software. - As previously discussed,
device 100 may be used to facilitate a video conference. Consistent with aspects described herein, the images for the video conference may be obtained from a camera external todevice 100. - A video conference, as this term is used herein, is to be broadly interpreted as communication between two or more parties in which both audio and visual information is transferred. The visual information may be video (i.e., a sequence of images that are updated fast enough to appear to fluidly capture movement) or a series of images in which the images are separated by longer time periods. Further, the visual information for a video conference does not need to be transmitted in both directions. That is, a video conference may be conducted in which only one party receives visual information.
- A user of a portable communication device, such as a user of
device 100, may wish to establish a video conference with another party. The user may, for example, be speaking to the other party viadevice 100 and during the course of the conversation, the users may decide that they would like to view images of each other. Alternatively, the user, when initially dialing the called party, may initiate the conversation as a video conference. -
FIG. 5 is a flow chart illustrating exemplary operations for providing a video conference. - Initially, it may be determined whether an external camera is available for the video conference (act 501). The external camera may be any camera that is not part of
portable communication device 100 and that is focused on the user or is located in the vicinity ofdevice 100. For example, public or private surveillance cameras or security cameras may be used. A number of different possible techniques can be used to determine or to inform the user when the user is in view of or near an external camera. For example,GPS component 480 and/ordirectional component 490 may be used to locate the user and the location of the user may then be compared to a list of known available external cameras. Alternatively, external cameras may use face recognition technologies to identify people and then proactively “publish” or notify users when they are in focus. These techniques are discussed in more detail below. - If an external camera is not available, the user may use
camera 370 onportable communication device 100 to generate video for the video conference (acts 502 and 503). Alternatively, or ifdevice 100 does not have a camera, the user may decide to forgo transmitting video to the other party. - If an external video camera is determined to be available, video from the external video camera may be sent to the other party as part of the video conference (
acts 502 and 504). If, for example, an available public surveillance camera is focusing on the user, the other party may view video from the surveillance camera as the video that is included in the video conference. The other party may receive the video in a number of possible ways. For example,portable communication device 100 may receive video from the nearby external camera via a short range wireless signal (e.g., a Bluetooth or WiFi signal), and may then forward the video to the other party overnetwork 110. Alternatively, the other party may receive video directly from the external camera overnetwork 110. In this case, the other party to the video conference may be forwarded a link or address associated with the external camera so that the other party may begin to receive the video from the external camera. - As previously mentioned, the determination of whether an external camera is available for the video conference (act 501) can be performed using a number of possible techniques. In general, a user may, at any particular time, wish to know which, if any, external cameras are viewing or “watching” the user. The user may, through
portable communication device 100, request that thedevice 100 determine which external cameras are watching the user. The results of this request may then be used in initiating a video conference, as shown inFIG. 5 . Alternatively, in some implementations, the list of external cameras that are watching the user may simply be shown to the user for amusement or informational purposes. - Various techniques for performing
act 501 will now be described in additional detail with reference toFIGS. 6-10 . -
FIG. 6 is a diagram of a system 600 that conceptually illustrates the determination of external cameras that are watching the user using information fromGPS component 480 and/or adirectional component 490. Assume thatuser 105 would like to useportable communication device 100 to determine whether any external cameras are available and watching the user. In this example,device 100 includes a data structure, labeled asexternal camera list 601. As shown, system 600 additionally includesexternal cameras External cameras -
External camera list 601 may be, for example, a list stored inmemory 400 ofportable communication device 100.External camera list 601 may generally describe the external cameras that are available for possible use as a video source. For example, as shown inFIG. 6 ,external camera list 601 includes a camera identification (ID)field 620,camera location field 625, andcamera orientation field 630.Camera identification field 620 may store a name or identification coded associated with theexternal cameras camera location field 625 may store an indication of the location ofcameras 610 and 615 (e.g., as an address or a latitude and longitude value), andcamera orientation field 630 may store an indication of the orientation ofcameras 610 and 615 (e.g., the direction the camera is pointing, whether the camera is panning, etc.). - In an alternative possible implementation,
external camera list 601, instead of being stored locally indevice 100, may be stored at, for example, a database located externally fromdevice 100 and that may be accessed remotely overnetwork 110. - Public surveillance and security cameras, as discussed above, are becoming increasingly prevalent. The video from at least some of these cameras may be made publicly available. For example, cameras owned by government organizations may be required by law to make the location and video from the cameras available. Alternatively, private companies that own surveillance cameras may, for free or for a fee, make the video taken by the cameras available. In these situations, the location and optionally, the orientation of the video cameras may be published by the entity that operates the camera. This information may form the basis for
external camera list 601. -
FIG. 7 is a flow chart illustrating exemplary operations for determining which external cameras are watching theuser 105 based on the system shown inFIG. 6 . Theuser 105 may indicate that theuser 105 would like to view a list of available external cameras (act 701). In response,portable communication device 100 may obtain location information of the user 105 (act 702). The location information may be obtained using GPS location viaGPS component 480. Other techniques for obtaining the location of theuser 105, such as triangulation using cellular phone towers may alternately be used to obtain the position of theuser 105.Portable communication device 100 may optionally obtain the orientation of the device 100 (act 703). The orientation (e.g., the direction the user is facing) may be obtained viadirectional component 490. To obtain the orientation of theuser 105, it may be assumed that theuser 105 is carrying or holdingportable communication device 100 in a particular orientation when theuser 105 requests the list of available external cameras. - Using the position information and possibly the orientation information of the
user 105,portable communication device 100 may searchexternal camera list 601 for external cameras that are likely to be focused on theuser 105. This list of external cameras may then be displayed or otherwise presented to the user 105 (act 704). In situations in which theuser 105 is initiating a video conference, the returned list of video cameras may be used to obtain video for the video conference. -
FIG. 8 is a flow chart illustrating exemplary operations for determining which external cameras are watching theuser 105 consistent with another implementation. In general, as described inFIG. 8 , automated face recognition techniques may be used to identify when a user is in view of a camera. - Video from a network of external video cameras, such as the cameras operated by a police force in a particular jurisdiction, may be continuously monitored and analyzed to recognize faces in the cameras (act 801). Automated facial recognition techniques are known in the art and will not be described further herein. In one implementation, users that wish to use the service may be required to pre-register images of their face with the operator of the video cameras.
- The
user 105 may indicate, via portableelectronic device 100, that theuser 105 would like to view a list of available external cameras (act 802). In response,portable communication device 100 may obtain a list of any cameras in which the user's face is currently (or recently) visible (act 803). This list may be obtained, for example, by querying a database or other server device overnetwork 110. The database or server device may be dynamically updated as users move in and out of a network of available external cameras as their faces are recognized. This list of external cameras that are currently viewing theuser 105 may then be displayed or otherwise presented to the user 105 (act 804). In situations in which the user is initiating a video conference, the returned list of video cameras may be used to obtain video for the video conference. -
FIGS. 9 and 10 are diagrams illustrating exemplary operations for determining which external cameras are watching the user consistent with another implementation. In general, as described with respect to the description ofFIGS. 9 and 10 , detection via local wireless signals may be used to identify when a user is near an external camera. -
FIG. 9 is a diagram of a system 900 that conceptually illustrates the determination of external cameras that are in the vicinity of the user based on detection using local wireless signals. System 900, as shown, includesexternal cameras External cameras External cameras respective wireless transceivers external cameras wireless transceivers external cameras device 100. -
FIG. 10 is a flow chart illustrating exemplary operations for determining which external cameras are watching the user based on the system shown inFIG. 9 . The user may indicate that the user would like to view a list of available external cameras (act 1001). In response,portable communication device 100 may determine if it is currently receiving or has recently received wireless presence information from a local external camera (act 1002). A list of determined external cameras (if any) may then be displayed or otherwise presented to the user (act 1003). In some implementations, in addition to simply displaying whether an external camera is present,device 100 may additionally obtain and display an image or video stream currently or recently taken by the external camera. - Initiation of an exemplary video
conference using device 100 will now be described. Assume that a user, such asuser 105 shown inFIG. 6 or 9, decides to initiate a video conference with another party. Before dialing the number of the party, theuser 105 may first controldevice 100 to determine whether any external cameras are available. Alternatively, in other implementations, theuser 105 may first dial the called party and then, during the course of the call, check to determine whether an external camera is available. - In response to the user's command,
device 100 may check whether any external cameras are available.Device 100 may perform this check using one or more of the techniques discussed with respect toFIGS. 7 , 8, and 10. - Assume that two external cameras are determined to be available to capture images for a video conference. That is, assume that two external cameras are determined to be in the vicinity of the
user 105 and may be used to take videos of theuser 105.Device 100 may allow theuser 105 to choose which, if any, of these two available cameras to use in the video conference. -
FIG. 11 is an exemplarygraphical interface 1100 thatdevice 100 may present touser 105 indisplay 250.Interface 1100 allows the user to select an external camera. Ininterface 1100,labels images interface 1100. The user may select one of the external cameras to use in the video conference. In an alternative implementation,device 100 may automatically determine which of multiple external cameras are taking the most relevant images of the user and automatically select this external camera. This determination may be based on, for example, the presence or size of the face of theuser 105 in the image, whether the camera is focused onuser 105, whether the orientation of the camera is appropriate, etc. - During the course of the video conference, audio spoken by the
user 105 and received bymicrophone 210 may be transmitted overnetwork 110 to the called party. Video from the selected camera may also be transmitted to the called party. In one implementation, the video may be transmitted directly to the called party (i.e., without first being transmitted through device 100) and the audio fromuser 105 may be transmitted as a separate stream to the called party fromuser 105. In another implementation, the video may be first sent todevice 100, such as via a short range local wireless network, and then forwarded, with the audio of the user, to the called party. - The foregoing description of implementations provides illustration, but is not intended to be exhaustive or to limit the implementations to the precise form disclosed. Modifications and variations are possible in light of the above teachings or may be acquired from practice of the teachings.
- It should be emphasized that the term “comprises” or “comprising” when used in the specification is taken to specify the presence of stated features, integers, steps, or components but does not preclude the presence or addition of one or more other features, integers, steps, components, or groups thereof.
- In addition, while series of processes have been described herein, the order of the processes may be modified in other implementations. Further, non-dependent processes may be performed in parallel.
- It will be apparent that aspects described herein may be implemented in many different forms of software, firmware, and hardware in the implementations illustrated in the figures. The actual software code or specialized control hardware used to implement aspects does not limit the invention. Thus, the operation and behavior of the aspects were described without reference to the specific software code—it being understood that software and control hardware can be designed to implement the aspects based on the description herein.
- Even though particular combinations of features are recited in the claims and/or disclosed in the specification, these combinations are not intended to limit the invention. In fact, many of these features may be combined in ways not specifically recited in the claims and/or disclosed in the specification.
- No element, act, or instruction used in the present application should be construed as critical or essential to the implementations described herein unless explicitly described as such. Also, as used herein, the article “a” and “an” are intended to include one or more items. Where only one item is intended, the term “one” or similar language is used. Further, the phrase “based on” is intended to mean “based, at least in part, on” unless explicitly stated otherwise. As used herein, the term “and/or” includes any and all combinations of one or more of the associated list items.
Claims (23)
1. A method for conducting a video conference, the method comprising:
obtaining a location of a portable communication device;
determining, by the portable communication device and based on the location, whether at least one video camera external to the portable communication device is in a vicinity of the portable communication device; and
using video taken by the at least one video camera in a video conference conducted via the portable communication device, wherein the video taken by the at least one video camera displays a user of the portable communication device.
2. The method of claim 1 , further comprising:
presenting a list of video cameras that are determined to be in the vicinity of the portable communication device.
3. The method of claim 1 , wherein the determining step includes:
determining video cameras in the vicinity of the portable communication device based on face recognition techniques used to recognize a user of the portable communication device in video taken by the video cameras.
4. The method of claim 1 , wherein the determining step includes: determining video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
5. The method of claim 1 , wherein the video cameras external to the portable communication device include public surveillance cameras.
6. The method of claim 1 , wherein using the video taken by the at least one video camera in a video conference further includes:
receiving the video from the at least one video camera at the portable communication device; and
transmitting the received video to a party with whom the video conference is to be conducted.
7. The method of claim 1 , wherein using the video taken by the at least one video camera in a video conference further includes:
causing the video from the at least one video camera to be transmitted to a party with whom the video conference is to be conducted.
8. The method of claim 1 , wherein the determining step further comprises:
determining an orientation of the portable communication device; and
determining, based on the orientation of the portable communication device, whether the at least one video camera external to the portable communication device is in the vicinity of the portable communication device.
9. A method for conducting a video conference, the method comprising:
receiving, at a portable communication device, a request for identification of video cameras that are external to the portable communication device;
identifying, by the portable communication device, video cameras that are external to the portable communication device and that are in a vicinity of the portable communication device;
presenting, via a visual display on the portable communication device, the identified video cameras; and
using the portable communication device, conducting a video conference using video associated with at least one of the identified video cameras, wherein a user of the portable communication device is viewable in the video.
10. The method of claim 9 , wherein the identifying step includes:
obtaining a location of the portable communication device; and
determining the video cameras in the vicinity of the portable communication device based on the location.
11. The method of claim 9 , wherein the identifying includes:
determining the video cameras in the vicinity of the portable communication device based on face recognition techniques used to recognize the user of the portable communication device in video taken by the video cameras.
12. The method of claim 9 , wherein the identifying step includes:
determining the video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
13. The method of claim 9 , further comprising:
presenting a graphical interface configured to facilitate selection of one of the identified video cameras for use in the video conference.
14. A method for conducting a video conference, the method comprising:
determining at least one video camera in a vicinity of a portable communication device based on at least one of presence information that is wirelessly transmitted by the at least one video camera, a face recognition technique to recognize a user of the portable communication device in video taken by the at least one video camera, or a location of the portable communication device;
using video taken by the at least one video camera in a video conference conducted via the portable communication device; and
wherein the video captured by the at least one video camera displays the user of the portable communication device.
15. The method of claim 14 , further comprising:
determining an orientation of the portable communication device; and
determining, based on the orientation of the portable communication device, whether the at least one video camera external to the portable communication device is in the vicinity of the portable communication device.
16. A method for conducting a video conference, the method comprising:
determining a location and an orientation of a portable communication device;
determining, based on the location and the orientation of the portable communication device, whether at least one video camera external to the portable communication device is in a vicinity of the portable communication device; and
using video taken by the at least one video camera in a video conference conducted via the portable communication device.
17. The method of claim 16 , wherein the video captured by the at least one video camera displays a user of the portable communication device.
18. A portable communication device, comprising:
a memory to store instructions; and
a processor to execute the instructions to:
obtain a location of the portable communication device;
determine, based on the location, video cameras external to the portable communication device that are in a vicinity of the portable communication device; and
enable a video conference in which video taken by at least one of the determined video cameras is used as a video source for the video conference, wherein a user of the portable communication device is viewable in the video.
19. The portable communication device of claim 18 , further comprising: a display, and wherein the processor further executes instructions to present a list of the video cameras determined to be in the vicinity of the portable communication device.
20. The portable communication device of claim 18 , wherein the processor further executes instructions to obtain the location of the portable communication device using a GPS signal.
21. The portable communication device of claim 18 , wherein the processor further executes instructions to determine the video cameras in the vicinity of the portable communication device based on face recognition techniques used to recognize the user of the portable communication device in video taken by the video cameras.
22. The portable communication device of claim 18 , wherein the processor further executes instructions to determine the video cameras in the vicinity of the portable communication device based on presence information that is wirelessly transmitted by the video cameras.
23. The portable communication device of claim 18 , wherein the video cameras external to the portable communication device include public surveillance cameras.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/050,673 US20140036025A1 (en) | 2008-01-09 | 2013-10-10 | Video conference using an external video stream |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/971,493 US8581957B2 (en) | 2008-01-09 | 2008-01-09 | Video conference using an external video stream |
US14/050,673 US20140036025A1 (en) | 2008-01-09 | 2013-10-10 | Video conference using an external video stream |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/971,493 Continuation US8581957B2 (en) | 2008-01-09 | 2008-01-09 | Video conference using an external video stream |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140036025A1 true US20140036025A1 (en) | 2014-02-06 |
Family
ID=40427253
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/971,493 Active 2031-06-04 US8581957B2 (en) | 2008-01-09 | 2008-01-09 | Video conference using an external video stream |
US14/050,673 Abandoned US20140036025A1 (en) | 2008-01-09 | 2013-10-10 | Video conference using an external video stream |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/971,493 Active 2031-06-04 US8581957B2 (en) | 2008-01-09 | 2008-01-09 | Video conference using an external video stream |
Country Status (2)
Country | Link |
---|---|
US (2) | US8581957B2 (en) |
WO (1) | WO2009087500A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104702437A (en) * | 2015-02-28 | 2015-06-10 | 深圳市宝德软件开发有限公司 | Device for using two-dimensional code to perform network configuration for equipment |
WO2016026314A1 (en) * | 2014-08-22 | 2016-02-25 | 中兴通讯股份有限公司 | Method,device and system for configuring wireless camera |
EP4250720A4 (en) * | 2021-01-13 | 2024-05-22 | Samsung Electronics Co., Ltd. | Electronic device and image transmitting/receiving method |
Families Citing this family (53)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8024007B2 (en) * | 2008-07-28 | 2011-09-20 | Embarq Holdings Company, Llc | System and method for a projection enabled VoIP phone |
US8285256B2 (en) | 2008-07-28 | 2012-10-09 | Embarq Holdings Company, Llc | System and method for projecting information from a wireless device |
US20100113068A1 (en) * | 2008-11-06 | 2010-05-06 | Lmr Inventions, Llc | Hosted imagery capture in an ad hoc for mobile computing |
US8577000B1 (en) * | 2009-04-06 | 2013-11-05 | Wendell Brown | Method and apparatus for content presentation in association with a telephone call |
US20110109742A1 (en) * | 2009-10-07 | 2011-05-12 | Robert Laganiere | Broker mediated video analytics method and system |
CA2716637A1 (en) | 2009-10-07 | 2011-04-07 | Telewatch Inc. | Video analytics with pre-processing at the source end |
WO2011041904A1 (en) | 2009-10-07 | 2011-04-14 | Telewatch Inc. | Video analytics method and system |
WO2011068464A1 (en) * | 2009-12-01 | 2011-06-09 | T-Data Systems (S) Pte Ltd | Memory card and method for storage and wireless transceiving of data |
TWI400940B (en) * | 2010-03-24 | 2013-07-01 | Hon Hai Prec Ind Co Ltd | Handheld device and method for controlling orbit cameras remotely |
US9143739B2 (en) | 2010-05-07 | 2015-09-22 | Iwatchlife, Inc. | Video analytics with burst-like transmission of video data |
CA2748059A1 (en) | 2010-08-04 | 2012-02-04 | Iwatchlife Inc. | Method and system for initiating communication via a communication network |
CA2748199A1 (en) * | 2010-08-04 | 2012-02-04 | Iwatchlife Inc. | Method and system for making video calls |
CA2748065A1 (en) | 2010-08-04 | 2012-02-04 | Iwatchlife Inc. | Method and system for locating an individual |
CA2748060A1 (en) | 2010-08-04 | 2012-02-04 | Iwatchlife Inc. | Method and system for making video calls |
US9485461B2 (en) * | 2012-03-19 | 2016-11-01 | Sony Corporation | Video conferencing using wireless peripheral video conferencing device |
CA2822217A1 (en) | 2012-08-02 | 2014-02-02 | Iwatchlife Inc. | Method and system for anonymous video analytics processing |
EP2696591B1 (en) * | 2012-08-09 | 2019-04-10 | Samsung Electronics Co., Ltd | Video calling using a remote camera device to stream video to a local endpoint host acting as a proxy |
US9024995B2 (en) * | 2012-08-09 | 2015-05-05 | Samsung Electronics Co., Ltd. | Video calling using a remote camera device to stream video to a local endpoint host acting as a proxy |
US9621780B2 (en) | 2012-10-04 | 2017-04-11 | Nvidia Corporation | Method and system of curve fitting for common focus measures |
US9392158B2 (en) | 2012-10-04 | 2016-07-12 | Nvidia Corporation | Method and system for intelligent dynamic autofocus search |
US9020109B2 (en) * | 2013-02-26 | 2015-04-28 | Cisco Technology, Inc. | Location aware conferencing system and method |
US9942515B2 (en) | 2013-03-14 | 2018-04-10 | Microsoft Technology Licensing, Llc | Smart device pairing and configuration for meeting spaces |
US9031393B2 (en) * | 2013-06-12 | 2015-05-12 | Nvidia Corporation | Methods for enhancing camera focusing performance using camera orientation |
US9113036B2 (en) * | 2013-07-17 | 2015-08-18 | Ebay Inc. | Methods, systems, and apparatus for providing video communications |
US9672649B2 (en) * | 2013-11-04 | 2017-06-06 | At&T Intellectual Property I, Lp | System and method for enabling mirror video chat using a wearable display device |
WO2015183367A1 (en) | 2014-05-30 | 2015-12-03 | Apple Inc. | Continuity |
US10306125B2 (en) | 2014-10-09 | 2019-05-28 | Belkin International, Inc. | Video camera with privacy |
US9179105B1 (en) * | 2014-09-15 | 2015-11-03 | Belkin International, Inc. | Control of video camera with privacy feedback |
JP5830729B1 (en) | 2014-11-07 | 2015-12-09 | 株式会社ビアンコネロ | Shooting and video distribution system that automatically or manually captures a user with a portable communication terminal |
US20160330411A1 (en) * | 2015-05-04 | 2016-11-10 | Alexander Favors | Camera assembly and method for sharing public surveillance data. |
CN104853136A (en) * | 2015-05-29 | 2015-08-19 | 小米科技有限责任公司 | Video communication method and device |
CN105120099A (en) * | 2015-08-31 | 2015-12-02 | 小米科技有限责任公司 | Shooting control method and device |
US9787940B2 (en) * | 2015-10-05 | 2017-10-10 | Mutualink, Inc. | Video management defined embedded voice communication groups |
DK201670622A1 (en) | 2016-06-12 | 2018-02-12 | Apple Inc | User interfaces for transactions |
CN106375703A (en) * | 2016-09-14 | 2017-02-01 | 北京小米移动软件有限公司 | Video communication method and device |
WO2018076357A1 (en) | 2016-10-31 | 2018-05-03 | 北京小米移动软件有限公司 | Multimedia information playback method and system, collection device, and standardised server |
JP6934623B2 (en) * | 2017-01-20 | 2021-09-15 | パナソニックIpマネジメント株式会社 | Communication control method, telepresence robot, and communication control program |
CN108326875B (en) * | 2017-01-20 | 2023-03-24 | 松下知识产权经营株式会社 | Communication control method and apparatus, telepresence robot, and storage medium |
CN111343060B (en) | 2017-05-16 | 2022-02-11 | 苹果公司 | Method and interface for home media control |
US10402056B2 (en) * | 2017-06-28 | 2019-09-03 | Blue Jeans Network, Inc. | Selecting and managing devices to use for video conferencing |
US10511810B2 (en) * | 2017-07-06 | 2019-12-17 | Amazon Technologies, Inc. | Accessing cameras of audio/video recording and communication devices based on location |
DK201870364A1 (en) | 2018-05-07 | 2019-12-03 | Apple Inc. | Multi-participant live communication user interface |
US11128792B2 (en) | 2018-09-28 | 2021-09-21 | Apple Inc. | Capturing and displaying images with multiple focal planes |
US11620103B2 (en) | 2019-05-31 | 2023-04-04 | Apple Inc. | User interfaces for audio media control |
US10996917B2 (en) | 2019-05-31 | 2021-05-04 | Apple Inc. | User interfaces for audio media control |
CN113168246B (en) | 2019-10-10 | 2024-09-27 | 微软技术许可有限责任公司 | Configuring a mouse device by pressure detection |
US11392291B2 (en) | 2020-09-25 | 2022-07-19 | Apple Inc. | Methods and interfaces for media control with dynamic feedback |
US20220368548A1 (en) | 2021-05-15 | 2022-11-17 | Apple Inc. | Shared-content session user interfaces |
US11907605B2 (en) | 2021-05-15 | 2024-02-20 | Apple Inc. | Shared-content session user interfaces |
US11893214B2 (en) | 2021-05-15 | 2024-02-06 | Apple Inc. | Real-time communication user interface |
US11847378B2 (en) | 2021-06-06 | 2023-12-19 | Apple Inc. | User interfaces for audio routing |
US11770600B2 (en) | 2021-09-24 | 2023-09-26 | Apple Inc. | Wide angle video conference |
WO2023196231A1 (en) * | 2022-04-04 | 2023-10-12 | Apple Inc. | User interfaces for camera sharing |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6839079B2 (en) * | 2001-10-31 | 2005-01-04 | Alphamosaic Limited | Video-telephony system |
US20060017692A1 (en) * | 2000-10-02 | 2006-01-26 | Wehrenberg Paul J | Methods and apparatuses for operating a portable device based on an accelerometer |
EP1659794A1 (en) * | 2004-11-23 | 2006-05-24 | Sony Ericsson Mobile Communications AB | System and method for phone video conferencing |
US7133062B2 (en) * | 2003-07-31 | 2006-11-07 | Polycom, Inc. | Graphical user interface for video feed on videoconference terminal |
US20070047479A1 (en) * | 2005-08-29 | 2007-03-01 | Cisco Technology, Inc. | Method and system for conveying media source location information |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040080624A1 (en) * | 2002-10-29 | 2004-04-29 | Yuen Siltex Peter | Universal dynamic video on demand surveillance system |
KR100493858B1 (en) * | 2002-12-16 | 2005-06-10 | 삼성전자주식회사 | Multi ear jack and his control method |
US20050021620A1 (en) * | 2003-05-30 | 2005-01-27 | Todd Simon | Web data conferencing system and method with full motion interactive video |
DE602005011320D1 (en) | 2004-10-27 | 2009-01-08 | Intellprop Ltd | DEVICE AND METHOD FOR TELECOMMUNICATIONS SERVICES |
-
2008
- 2008-01-09 US US11/971,493 patent/US8581957B2/en active Active
- 2008-07-08 WO PCT/IB2008/052745 patent/WO2009087500A1/en active Application Filing
-
2013
- 2013-10-10 US US14/050,673 patent/US20140036025A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060017692A1 (en) * | 2000-10-02 | 2006-01-26 | Wehrenberg Paul J | Methods and apparatuses for operating a portable device based on an accelerometer |
US6839079B2 (en) * | 2001-10-31 | 2005-01-04 | Alphamosaic Limited | Video-telephony system |
US7133062B2 (en) * | 2003-07-31 | 2006-11-07 | Polycom, Inc. | Graphical user interface for video feed on videoconference terminal |
EP1659794A1 (en) * | 2004-11-23 | 2006-05-24 | Sony Ericsson Mobile Communications AB | System and method for phone video conferencing |
US20070047479A1 (en) * | 2005-08-29 | 2007-03-01 | Cisco Technology, Inc. | Method and system for conveying media source location information |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016026314A1 (en) * | 2014-08-22 | 2016-02-25 | 中兴通讯股份有限公司 | Method,device and system for configuring wireless camera |
CN104702437A (en) * | 2015-02-28 | 2015-06-10 | 深圳市宝德软件开发有限公司 | Device for using two-dimensional code to perform network configuration for equipment |
EP4250720A4 (en) * | 2021-01-13 | 2024-05-22 | Samsung Electronics Co., Ltd. | Electronic device and image transmitting/receiving method |
Also Published As
Publication number | Publication date |
---|---|
US20090174763A1 (en) | 2009-07-09 |
WO2009087500A1 (en) | 2009-07-16 |
US8581957B2 (en) | 2013-11-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8581957B2 (en) | Video conference using an external video stream | |
CN106453924B (en) | A kind of image capturing method and device | |
US9674489B2 (en) | Systems and methods for location-based multimedia | |
CN106488251B (en) | Realize the method and device, main broadcaster's client and user client for connecting wheat in live streaming | |
US20170272916A1 (en) | System and method for monitoring locations of mobile devices | |
US20110093266A1 (en) | Voice pattern tagged contacts | |
US20100235768A1 (en) | Personalized user interface based on picture analysis | |
US8682374B2 (en) | Mobile terminal and controlling method thereof | |
KR20090104484A (en) | Terminal and method of performing instant messaging service therein | |
EP3147802B1 (en) | Method and apparatus for processing information | |
CN107423386B (en) | Method and device for generating electronic card | |
CN104158985A (en) | Communication method, device and system | |
CN113365153B (en) | Data sharing method and device, storage medium and electronic equipment | |
CN105898573B (en) | Multimedia file playing method and device | |
EP3291508B1 (en) | Message pushing method and device, and terminal equipment | |
CN114009003A (en) | Image acquisition method, device, equipment and storage medium | |
CN112291631A (en) | Information acquisition method, device, terminal and storage medium | |
CN105549300A (en) | Automatic focusing method and device | |
CN110636318A (en) | Message display method, message display device, client device, server and storage medium | |
CN112146676B (en) | Information navigation method, device, equipment and storage medium | |
WO2021237592A1 (en) | Anchor point information processing method, apparatus and device and storage medium | |
CN105260088B (en) | Information classification display processing method and device | |
CN105682187B (en) | Icon display method and device | |
CN105338482A (en) | Method and device for emergency call | |
CN109391944B (en) | Wireless network remarking method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BENGTSSON, HENRIK;BACKLUND, ERIK JOHAN VENDEL;SASSI, JARI AARNI;AND OTHERS;SIGNING DATES FROM 20131002 TO 20131010;REEL/FRAME:031584/0598 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |