US20140055401A1 - Method and apparatus for controlling contents in realtime fashion - Google Patents
Method and apparatus for controlling contents in realtime fashion Download PDFInfo
- Publication number
- US20140055401A1 US20140055401A1 US14/069,384 US201314069384A US2014055401A1 US 20140055401 A1 US20140055401 A1 US 20140055401A1 US 201314069384 A US201314069384 A US 201314069384A US 2014055401 A1 US2014055401 A1 US 2014055401A1
- Authority
- US
- United States
- Prior art keywords
- screen
- touch panel
- position information
- touch
- network communication
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1454—Digital output to display device ; Cooperation and interconnection of the display device with other functional units involving copying of the display data of a local workstation or window to a remote workstation or window so that an actual copy of the data is displayed simultaneously on two or more displays, e.g. teledisplay
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2330/00—Aspects of power supply; Aspects of display protection and defect management
- G09G2330/02—Details of power systems and of start or stop of display operation
- G09G2330/021—Power management, e.g. power saving
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/02—Handling of images in compressed format, e.g. JPEG, MPEG
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2350/00—Solving problems of bandwidth in display systems
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2354/00—Aspects of interface with display user
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/16—Use of wireless transmission of display information
Definitions
- the present invention relates to a method and an apparatus for controlling contents in realtime fashion, and more particularly, to a method and an apparatus for controlling multimedia contents in realtime fashion by means of using wireless touch, motion sensing and battery charging.
- each one of such devices may be operated under a specific platform or OS that may be unique from each other, which may be good in providing varieties to the market in general but may too raise a preference issue to the user depending on his/her own demands in terms of such as function, design or performance of different operating systems and devices.
- the inventor realizes an user may advantageously be provided with an universal apparatus capable of communicating with the any example of new breed of electronic devices running any certain type of operating systems and may be advantageously free from the preference selection issue while still being able to utilize or interact with the any one such new breed of electronic devices running different type of operating systems.
- portable devices are integrated with more functions to be easily accessed by and interacted with users for more dedicated purposes or services including such as gaming, web browsing, teleconferencing and navigation.
- the inventor realizes that in order to provide an electronic apparatus capable of communicating with external devices of various of different operating systems offering more dedicated services or applications, the most accessible functions or main services of the communicated external devices shall be preferably preserved, controlled and/or interacted by users while preserving a relatively less hardware structures to provide a highly portable and interactive devices to users.
- an electronic apparatus capable of communicating with external device(s) of various of different types of operating systems such that applications and services of the connected/communicated external device of different types of operating systems in relation to such as interactive gaming, teleconferencing, web browsing, media playing, navigation, image displaying or taking may all be preserved or advantageously made available, and preferably with enhanced data transmission, while realizing or having a relatively “light” hardware structures with great portability and interactive feasibility to users.
- the present invention provides a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time, in which the first device is coupled to the second device via a network communication, the first screen is configured to display a mirrored content of the second screen, and the first screen includes touch panel capability.
- the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time includes the following steps: detecting a first touch signal on the first screen of the first device; converting the first touch signal into a first touch data associated with a first position information with respect to a first coordinate system of the first screen; transmitting the first touch data to the second device via the network communication, so that the actual operation is performed on the second device.
- the actual operation performed on the second device can be overridden if a detecting event is detected on the second screen of the second device.
- the transmit (TX) side is also referred to as a second device
- the receive (RX) side is also referred to a first device.
- optional steps of the embodiment include of the following: detecting a motion event on the first device and generating a sensor data, and sending the sensor data to the second device via the network communication for enabling a corresponding operation on the second device, in which the corresponding operation includes changing orientation of the content displayed on the second screen.
- one of a plurality of sensing ranges can be selected for detecting the motion event, and a threshold value can be provided for detecting the motion event on the first device.
- the method further includes charging capability to charge the first device with the second device or to charge the second device with the first device by detecting a charging connection between the first device and the second device, retrieving remaining battery amounts from the first device and the second device respectively, and providing an option to charge the first device with the second device, or to charge the second device with the first device.
- a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time further includes calculating a second position information with respect to a second coordinate system of the second screen based on the first position information.
- the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time of the embodiment further includes additional steps which include: receiving compressed multimedia content compressed with a compression ratio from the second device via the network communication, and decompressing and displaying the multimedia content on the first screen of the first device.
- the compression ratio is determined based on a bandwidth of the network communication.
- the decompressed multimedia content can be scaled based on a resolution of the first screen.
- the present invention provides a method for performing an actual operation on a second device having a second touch panel from a first device having a first touch panel in real time, and the first device is coupled to the second device via a network communication.
- the method for performing the actual operation on the second device having the second touch panel from the first device having the first touch panel in real time includes steps as follow: receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel; calculating a second position information with respect to a second coordinate system of the second touch panel based on the first position information; and performing the actual operation on the second device based on the second position information.
- the transmit (TX) side is also referred to as a second device
- the receive (RX) side is also referred to a first device.
- the performing of the actual operation is overridden if a detecting event is detected on the second touch panel of the second device.
- the method of this embodiment may further include steps for compressing and decompressing of multimedia content from the second device to the first device as follow: compressing a multimedia content displayed on the second touch panel, and sending the compressed multimedia content from the second device to the first device for decompressing and displaying on the first touch panel via the network communication.
- This embodiment may further include steps for performing operations on the second device in response to sensor data indicating a motion event on the first device as follow: receiving a sensor data from the first device, the sensor data indicating a motion event on the first device, and performing a corresponding operation on the second device in response to the sensor data.
- the corresponding operation may include changing orientation of the multimedia content displayed on the second display.
- the present invention provides a second device configured to be controlled by a first device with a first touch panel, and the first device is coupled to the second device by a network communication.
- the second device includes a second touch panel, a compression unit for compressing multimedia content with a compression ratio, a transceiver for transmitting the compressed multimedia content to the first device for decompressing and displaying on the first touch panel and for receiving a touch data associated with a first position information with respect to a first coordinate system of the first touch panel, a coordinate calculating unit for calculating a second position information with respect to a second coordinate system of the second touch panel, and a processor for performing an actual operation on the second device according to the touch data and the second position information.
- the actual operation can be overridden if the second touch panel detects a detecting event or if the second device receives an incoming call.
- the compression ratio is determined based on a bandwidth of the network communication.
- the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device; in addition, the first touch data is referred to the touch data being formed after converting of the at least one touch signal into touch data.
- FIG. 1A is a schematic diagram showing wireless control from a RX side over a TX side according to an embodiment of the present invention.
- FIG. 1B and FIG. 1C are schematic diagrams showing touch control operations of from the RX side over the TX side shown in FIG. 1A according to an embodiment of the present invention.
- FIG. 2 is a schematic diagram showing a compression ratio in relation to an available bandwidth between the RX side and the TX side according to an alteration of the present invention.
- FIG. 3A and FIG. 3B are schematic diagrams showing an image orientation on the TX side in response to a motion detection on the RX side where FIG. 3B is further provided with a switch element for adjusting motion sensitivity according to another alteration of the present invention.
- FIG. 3C is a schematic diagram showing motion control operations of the RX side over the TX side corresponding to FIGS. 3A and 3B according to an embodiment of the present invention.
- FIG. 3D and FIG. 3E are schematic diagrams showing adjustment of motion sensitivity on the RX side implemented by an application program on the TX side according to a further alteration of the present invention.
- FIG. 3F and FIG. 3G are schematic diagrams showing adjustment of magnitude and/or threshold values of application programs on the TX side after motion data is received according to a further alteration of the present invention.
- FIG. 4 is a flow diagram showing a process for controlling multimedia contents in realtime fashion according to an embodiment of the present invention.
- FIG. 5 is a flow diagram showing a process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention.
- FIG. 6 is a flow diagram showing another process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention.
- FIG. 7A and FIG. 7B are schematic diagrams showing battery charging between the TX side and the RX side according to a further alteration of the present invention.
- FIG. 8 is a flow diagram of a battery charging process according to an embodiment of the present invention.
- FIG. 9 is a flow diagram of a battery charging process according to another embodiment of the present invention.
- FIG. 1A is a schematic diagram of a RX side 10 (e.g. pad-like device having a touch-control display) and a TX side 12 (e.g. phone-like device having a computing power & operating system therein) according to an embodiment of the present invention.
- the RX side 10 is provided for controlling in realtime fashion multimedia contents that are originated from the TX side 12 .
- the multimedia contents can include audio and video streams and be processed to present at the TX side 12 (such as cellphone, smartphone, PDA, and GPS navigation, or any other portable wireless device) having a phone-like screen 1202 and the TX side 12 is coupled to the RX side 10 via a network communication NC, which is preferably a wireless communication and a two-way communication.
- NC is preferably a wireless communication and a two-way communication.
- the multimedia contents may include video and audio data, i.e. the video data can be displayed on the screen 1202 while the audio data can be played via a speaker (not shown) in the TX side 12 .
- the RX side 10 may be advantageously provided in a CPU-less (without central processing unit) and/or GPU-less (without graphical processing unit) environment for the sake of less power consumption and/or simplified circuit complexity.
- the TX side 12 may be a computing system or mobile device equipped with CPU and/or GPU computing power, such as a PC, laptop, tablet PC, mobile phone, smart-phone, and may include a processor (not shown) for executing an application program (not shown) of an operation system (OS) 1204 installed in a storage medium (not shown) and a transceiver 1206 coupled to the processor and capable of transmitting and receiving the multimedia contents generated by the application program of OS 1204 executed by the processor.
- OS operation system
- the OS 1204 may be any one of existing or known OS, such as the abovementioned Windows®, iOS®, Android®, Linux®, Blackberry®, Symbian®, Palm®, installed on the TX side 12 such that an application program may be performed or run thereon.
- the network communication NC between the RX side 10 and the TX side 12 may preferably be any one of the network communication standards or profiles of: IEEE 802.11, IEEE 802.16, CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth, Zigbee and combination thereof, or wired communication such as ADSL network, Optical/Fiber network or Cable network.
- the RX side 10 (refer back to FIG. 1A ) includes a pad-like screen 1002 (e.g. touch panel), a coordinate mapping unit 1004 and a transceiver 1006 .
- the screen 1002 is provided for detecting at least one touch signal, and is preferably larger than the screen 1202 .
- the coordinate mapping unit 1004 can be controlled and/or controllable by an operating system (not shown).
- the coordinate mapping unit 1004 is provided for converting the at least one touch signal (from the screen 1002 ) into touch data associated with a first position information that indicates a virtual operation on the first screen 1002 corresponding to an actual operation on the second screen 1202 , wherein the first position information is with respect to a first coordinate system of the first screen.
- a pad-like screen 1002 e.g. touch panel
- the coordinate mapping unit 1004 can be controlled and/or controllable by an operating system (not shown).
- the coordinate mapping unit 1004 is provided for converting the at least one touch signal (from the screen 100
- the first position information represents a current cursor position (right side of FIG. 1B ) on the pad-like screen 1202 moved by the user's finger (not shown), and thus the mirrored cursor on the phone-like screen 1202 will be moved to the corresponding position (left side of FIG. 1B ) such that the user can do any operation on the phone-like screen by touching the pad-like screen 1202 while without touching the phone-like screen 1202 .
- the transceiver 1006 is provided for transmitting the touch data (e.g. the first position information indicative of the current position of cursor on screen 1002 ) sent from the coordinate mapping unit 1004 to the TX side 12 via the network communication NC.
- the TX side 12 further includes a coordinate calculating unit 1208 for calculating a second position information (e.g. new cursor position on phone-like screen) with respect to a second coordinate system of the screen 1202 based on the first position information.
- a second position information e.g. new cursor position on phone-like screen
- mapping movement (delta_x′, delta_y′) on the phone-like screen can be obtained as following expression:
- the first position information may further include the current position of the cursor on the pad-like screen 1002 of the RX side 10 , and the resolution data of the RX side 10 , e.g. 1920 ⁇ 1080, and the second position information can be calculated so as to obtain the mapping position of the cursor on the phone-like screen 1202 , e.g. scaling down to a resolution 480 ⁇ 320.
- the actual action can be “play a video movie” when the mapping position of the cursor on the screen 1202 is pointed on a “play” icon.
- the user can “touch” on the pad-like screen 1002 of the RX side 10 to implement operations without access to the phone-like screen 1202 as if what he/she usually does on the phone-like screen 1202 of the TX side 12 .
- all the tasks requiring extensive computing power/resources are handled by the TX side 12 , and only the display task is duplicated (mirrored) to the RX side 10 from the TX side 12 such that the multimedia contents can be presented in realtime fashion between the pad-like screen 1002 of the RX side and the phone-like screen 1202 of the TX side 12 .
- FIG. 1B and FIG. 1C are schematic diagrams of touch control operations of the RX side 10 and the TX side 12 shown in FIG. 1A according to an embodiment of the present invention.
- a user can do the “open” operation to open a file or folder on the larger screen 1002 of the RX side 10 (right side) and quasi-simultaneously the smaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in the RX side 10 , and thus the user can do the “double-click” operation to open some object like file/folder on the larger screen 1002 while utilize computing power from the TX side 12 , i.e.
- a user can do the “key-in” operation on the larger RX side 10 (right side) to type words and simultaneously the smaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in the RX side 10 , and thus the user can do the “key-in” operation on the larger screen 1002 while utilize computing power from the TX side 12 , i.e. the operating system 1204 , to run the text-editing application.
- the TX side 12 may further include a compressing unit 1210 , for compressing the multimedia contents with a compression ratio, e.g. compressing RGB data into MJPEG data, which requires less time for compression and thus facilitates operating in realtime fashion, and the transceiver 1206 can be further utilized for transmitting the compressed multimedia contents to the screen 1002 for displaying.
- a compressing unit 1210 for compressing the multimedia contents with a compression ratio, e.g. compressing RGB data into MJPEG data, which requires less time for compression and thus facilitates operating in realtime fashion
- the transceiver 1206 can be further utilized for transmitting the compressed multimedia contents to the screen 1002 for displaying.
- the transceiver 1006 receives the compressed multimedia contents from the transceiver 1206 via the network communication NC, and then compressed multimedia contents are decompressed by a decompressing unit 1010 with the compression ratio, such that the video data of the multimedia contents can be displayed on the screen 1002 while the audio data of the multimedia contents can be played by a speaker 1008 .
- the multimedia contents can be displayed and controlled in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 .
- FIG. 2 is a schematic diagram of the RX side 10 and the TX side 12 according to an alteration of the present invention.
- the transceiver 1206 can further detect a bandwidth of the network communication NC
- the TX side 12 further includes a dynamic compression unit 1212 , for determining the compression ratio of the multimedia content based on the bandwidth of the network communication NC, such that the compressing unit 1210 can compress the multimedia contents at the compression ratio, which is determined by the bandwidth of the network communication NC.
- the transceiver 1006 also can detect the bandwidth of the network communication NC, and then the decompressing unit 1010 decompresses the compressed multimedia contents by the compression ratio provided by a bandwidth calculating unit 1012 .
- the multimedia contents can be compressed with a compression ratio according to a bandwidth of the network communication NC, such that the multimedia contents can be smoothly displayed in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 by facilitating the realtime codec function (e.g. the video compression ratio is determined by the actual available bandwidth).
- the transceiver 1006 also can directly receive the compression ratio from the dynamic compression unit 1212 (TX side) rather than the bandwidth calculating unit 1012 , and then to pass the compression ratio into the decompressing unit 1010 so that the bandwidth calculating unit 1012 can be removed as well.
- the present invention discloses another method for controlling in realtime fashion multimedia contents between a RX side having a first touch panel and a TX side having a second touch panel via a network communication.
- the multimedia contents comprise at least a screen frame data (not shown) where the screen frame data includes the screen resolution data, image data displayed on the screen and audio data.
- the method comprising the steps of:
- FIGS. 3A-3B are schematic diagrams of the TX side 12 and the RX side 10 according to another alteration of the present invention.
- the RX side 10 may further include a motion sensor 1014 (including at least one of an accelerometer, a gyroscope and a magnetic compass) for detecting at least one motion event.
- the receiver 1006 transmits the sensor data (RX) from the motion sensor 1014 after the motion sensor 1014 has detected the motion event on the RX side, and then the sensor data (RX) is sent to the TX side 12 via the network communication NC.
- a sensor decision unit may disable sensor data (TX) generated from a motion sensor 1216 on the TX side 12 only if the detecting exists (i.e. the sensor data (RX) from the RX side has being received on the TX side), and then send the sensor data (RX) to a motion process unit 1214 .
- the motion process unit 1214 can convert the sensor data (RX) into motion data (indicating a translation representative of linear acceleration in three-axis directions, and/or an angular rotation representative of Yaw, Roll and Pitch in three-axis directions) on the TX side 12 , so as to present the multimedia contents on the screen 1202 and subsequently on the screen 1002 in a same particular orientation (e.g. portrait orientation or landscape orientation).
- the particular orientation is determined by the operating system 1204 according to the motion data (RX) such that the image at the screen 1202 will be presented at the landscape or portrait orientation.
- RX motion data
- the multimedia contents can be displayed and controlled in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 through motion operations.
- FIG. 3C is a schematic diagrams of motion control operations between the RX side 10 and the TX side 12 shown in FIGS. 3A-3B according to one embodiment of the present invention.
- a user can do the “game play” driving motion, e.g. moving the image (RX) of the larger pad-like screen 1002 from Landscape to Portrait, and thus the mirrored image (TX) on the smaller phone-like screen 1002 (shown on left side of FIG. 3C ) is quasi-simultaneously adjusted from Landscape to Portrait on the TX side 12 performed by the operating system 1204 based on the motion data converted from the sensor data (RX) generated by the motion sensor 1014 (RX).
- the user can play the game on the pad-like screen 1002 while utilize computing power from the TX side 12 , i.e. the operating system 1204 , to run the game application actually on the phone-like screen 1202 .
- the motion sensor 1014 may have a plurality of sensing ranges, i.e. different sensing sensitivities, or a particular sensing range for generating the sensor data.
- one of the multiple sensing ranges can be selected by a control code determined by a switch signal so as to enable a motion chip to generate one motion data based on the selected sensing range determined by the control code.
- the motion sensor 1014 has a plurality of sensing ranges. As shown in FIG. 3A-3B , the RX side 10 detects whether a switch signal is invoked by a switch element 1016 operative to a user, and the switch element 1016 may issue an adjustment signal to a motion chip (i.e. motion sensor 1014 ) after the switch signal was invoked (e.g. user turns the switch element 1016 on). Then, the motion sensor 1014 selects one of the plurality of sensing ranges of the motion chip based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between the RX side 10 and the TX side 12 .
- a motion chip i.e. motion sensor 1014
- the motion chip has a sensing range of ⁇ 1 G at its “off” (original) state and ⁇ 2 G at “on” state.
- the motion chip may have a new sensing range of ⁇ 2 G, where G is defined as 9.81 m/s 2 for gravitational acceleration.
- the motion sensor 1014 has a plurality of sensing ranges. As shown in FIG. 3D-3E , the TX side 12 detects whether a switch signal is invoked by an application program 1220 (e.g. game software/program), and an adjustment signal is generated in response to the invoked switch signal. Then, the motion sensor 1014 selects one of the plurality of sensing ranges of the motion chip of the motion sensor 1014 based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between the RX side 10 and the TX side 12 .
- an application program 1220 e.g. game software/program
- the game software may provide an input user interface to allow the user to choose one of three options for selecting particular sensing range among the plurality of sensing ranges where Option 1 stands for ⁇ 1 G; Option 2 for ⁇ 2 G; and Option 3 for ⁇ 3 G, respectively.
- a particular switch signal will be sent out to the RX side after the selected Option 2 is chosen by the user, for instance.
- the motion sensor 1014 (RX) is requested to change its sensing range to ⁇ 2 G directly by the switch signal, or by an adjusted signal in response to the switch signal.
- the motion sensor 1014 has a particular sensing range for generating the sensor data.
- the TX side 12 detects whether a switch signal is invoked by an application program 1222 of the operating system 1204 , and an adjustment signal is sent in response to the invoked switch signal. Then, the TX side 12 may provide a user interface to allow the user to adjust a threshold value of the application program 1222 (e.g. game software/program) based on either the adjustment signal or the invoked switch signal such that the threshold value of the application program 1222 can be increased at a ratio corresponding to the adjustment signal or the invoked switch signal. Subsequently, the motion data may be generated based on sensor data that is sent from the motion sensor 1014 (on the TX side) triggered by the switch signal, so as to enable an interaction between the RX side 10 and the TX side 12 .
- a threshold value of the application program 1222 e.g. game software/program
- the motion sensor 1014 has a particular sensing range for generating the sensor data.
- the TX side 12 detects whether a switch signal is invoked by the application program 1222 of the operating system 1204 , and issues an adjustment signal in response to the invoked switch signal. Then, the TX side 12 may provide a user interface to allow the user to adjust a magnitude of the motion data by means of increasing the magnitude's value at a ratio based on either the adjustment signal or the switch signal, so as to enable an interaction between the RX side 10 and the TX side 12 .
- the TX side 12 can be configured to adjust the threshold value of the application program 1222 and the magnitude of the motion data, respectively, but the TX side 12 can also simultaneously adjust the threshold value of the application program 1222 and the matmitude of the motion data, while retaining both merits.
- the spirit of the present invention is to utilize a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side.
- the RX side 10 can be foldable, portable, or any other types to meet each user's need(s).
- the network communication NC may preferably include a Wi-Fi communication and a Bluetooth (BT) communication, and since the touch data and sensor data have less data volume while the multimedia contents have greater data volume, the touch data and the motion date are preferably transmitted to the TX side 12 via the BT communication, while the multimedia contents are preferably transmitted to the RX side 10 via the Wi-Fi communication.
- the transceiver 1006 and 1206 of the present invention may be a wireless module adapting other communication standards or profiles including such as IEEE 802.16 (WiMAX), CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth-profile and combination thereof or a wired module.
- the video data of the multimedia contents decompressed and outputted by the decompressing unit 1010 may preferably from a MJPEG format to a RGB format and transmitted or outputted to the screen 1002 in a LVDS format via a display interface converter such as a RGB-to-LVDS converter. It can be understood that other video stream formats such as ITDMI and related interfaces may also be possible.
- the audio data of the multimedia contents processed and outputted may preferably be transmitted via an audio interface such as an I 2 S interface or standard and further to the speaker 1008 .
- the multimedia contents transmitted and processed in video streaming may be made “real-time” or without significant delays with respect to time for the sake of low latency.
- the multimedia contents may be communicated between the RX side 10 and the TX side 12 and processed to output “real-time” video and audio streams to the screen 1002 and/or the speaker 1008 of the RX side 10 of the present invention.
- the multimedia contents may be processed in video streaming (i.e.
- encoding and decoding and communicated there between as data packets and the duration or delay of time involved in a “real-time” processing and transmission may be less than 100 ms for the sake of low latency; in one example, it may preferably be less than 10 ms, and in another example, it may approximately be 20 ms.
- the controlling process 40 includes the following steps:
- the receiving process 50 includes the following steps:
- the receiving and transmitting process 60 includes the following steps:
- FIG. 7A-7B is a schematic diagram of the RX side 10 according to another alteration of the present invention.
- the TX side 12 further comprises a charging application 1224 , charging detector 1226 and battery cells 1228
- the RX side 10 further comprises a charging controller 1018 and battery cells 1020 .
- system default charging operations of the TX side 12 and the RX side 10 shown in FIG. 7A-7B can be summarized into charging process 80 as shown in FIG. 8 .
- the charging process 80 includes the following steps:
- the charging process 90 includes the following steps:
- the event of a charging connection can be detected either by a charging detector 1226 on the TX side 12 or by a charging controller 1018 on the RX side 10 .
- the charging controller 1018 can be implemented by a power management chip.
- the charging detector 1226 can receive the remaining battery capacity of the battery cells 1228 on the TX side 12 ; so as to send the battery capacity information to the user interface by means of the charging application 1224 managed by OS 1204 , and thus users can view the icon that indicates the actual measurement of the TX's battery capacity.
- the charging controller 1218 on the RX side 10 can receive the actual measurement of the RX's battery capacity, and sends the measurement to the charging detector 1226 on the TX side 12 , so as to facilitate the charging detector 1226 to determine what condition can be meet to implement the appropriate charging operation.
- a power adaptor is used to charge both the RX side 10 and the TX side 12 ; when the remaining battery amount of the RX side 10 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the RX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), the RX side 10 is used to charge the TX side 12 ; when the remaining battery amount of the TX side 12 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the TX side 12 is higher than that of the RX side 10 (i.e.
- the TX side 12 is used to charge the RX side 10 . It is noted that the RX side may be compulsorily switched to the power saving mode by the override mechanism when the user receives an incoming call from the TX side, that is, the user currently uses the TX side rather the RX side to either answer or dial the call. Therefore, it makes sense for the user to temporarily keep the RX side idle while he or she is busy for phone call.
- the present invention utilizes a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side. Furthermore, the present invention can compress multimedia contents a bandwidth of network communication, such that the multimedia contents can be smoothly displayed in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 .
- the TX side 12 can also be configured as a server having encoding hardware and the RX side 10 as a client having decoding hardware adapted for cloud computing environment running video streaming for the purpose of low latency requirement.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Telephone Function (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Mobile Radio Communication Systems (AREA)
Abstract
Method for performing actual operation on a second device having a second screen or a touch panel from a first device having a first screen or a touch panel in real time are provided. First screen is configured to display mirrored content of second screen. Method includes detecting a first touch signal on first screen of first device, converting first touch signal into a first touch data associated with a first position information, and transmitting first touch data to second device so actual operation is performed on second device. For first and second devices having touch panels, method includes receiving a first touch data associated with a first position information of first touch panel, calculating a second position information of second touch panel based on first position information, and performing actual operation on second device based on second position information. Second device is coupled to first device by network communication.
Description
- This application is a continuation application of U.S. non-provisional application Ser. No. 13/184584, filed on Jul. 18, 2011 and now pending, which in turn claims the benefit of U.S. provisional application Ser. No. 61/410,371, filed on Nov. 05, 2010, now abandoned, the contents of which are incorporated herein by reference and made a part of this disclosure.
- 1. Field of the Invention
- The present invention relates to a method and an apparatus for controlling contents in realtime fashion, and more particularly, to a method and an apparatus for controlling multimedia contents in realtime fashion by means of using wireless touch, motion sensing and battery charging.
- 2. Description of the Prior Art
- As the technology related to portable electronic device advances and the need for portability increases in relation to the current market trend, traditional personal computers are transforming to lighter and more dedicated devices provided for or targeted at different purposes and user groups including such as gaming, telecommunication, web browsing and media streaming. Such shift of market and technology trends related to advanced needs of portability and interactions demanded by users of such electronic devices are in fact facilitating not only the improvement of traditional personal computers, laptops or smart-phones to be more powerful and/or integrated with more features and functions than ever but also a new category or breed of portable electronic devices with more dedicated functions for certain purposes and greater portability at the same time.
- In view of the merits of greater portability as well as interaction provided by such newly evolved breed of electronic devices previously mentioned, the inventor realizes that electronic designers and manufacturers are introducing a “lighter” version of portable personal computers or laptop to the market while telecommunication device designers and manufacturers are, in general, offering a “heavier” version of mobile phones such as smart-phones such that the new growing market and user demand of high portability as well as interactive functions may be fulfilled. Some known examples of such new breed of devices may include: iPad by Apple®, Playbook by RIM®, Cius by CISCO®, Galaxy Tab by Samsung®, Libretto by Toshiba®; and examples of known smart phones may be: iPhone by Apple®, Blackberry Bold by RIM®, Galaxy S by Samsung®, HD7 by HTC™. In general, these portable devices share some common building blocks of traditional personal computers but with greater portability in terms of their size, weight as well as with greater interactive and integrated functions with more dedicated applications and telecommunication capabilities including via such as known 3G, 4G or Wi-Fi™.
- As such new breed of portable electronic devices are being introduced to the market, the inventor also realizes that each one of such devices may be operated under a specific platform or OS that may be unique from each other, which may be good in providing varieties to the market in general but may too raise a preference issue to the user depending on his/her own demands in terms of such as function, design or performance of different operating systems and devices. Instead of choosing one particular type of portable electronic device miming a certain type of operating systems including such as Windows® by Microsoft®, iOS® by Apple®, Android® or Linux® by Google®, Blackberry® OS by RIM®, the inventor realizes an user may advantageously be provided with an universal apparatus capable of communicating with the any example of new breed of electronic devices running any certain type of operating systems and may be advantageously free from the preference selection issue while still being able to utilize or interact with the any one such new breed of electronic devices running different type of operating systems.
- However, there are at least two major concerns to be considered in providing an electronic apparatus capable of advantageously communicating and interacting with the newly introduced breed of portable electronic devices, one is related to the various different platforms or operating systems being adapted by different electronic devices and the other is to preserve the merits of great portability and being light in terms of operation and hardware structures or components of portable electronic devices demanded by the market trend. Therefore, there is a need to provide an electronic apparatus capable of communicating with external device(s) of various types of operating systems while preserving a relatively “light” hardware structures to provide a highly portable and interactive devices to users.
- Furthermore, as previously mentioned that portable devices are integrated with more functions to be easily accessed by and interacted with users for more dedicated purposes or services including such as gaming, web browsing, teleconferencing and navigation. The inventor realizes that in order to provide an electronic apparatus capable of communicating with external devices of various of different operating systems offering more dedicated services or applications, the most accessible functions or main services of the communicated external devices shall be preferably preserved, controlled and/or interacted by users while preserving a relatively less hardware structures to provide a highly portable and interactive devices to users.
- In view of the foregoing, it is desirable to provide an electronic apparatus capable of communicating with external device(s) of various of different types of operating systems such that applications and services of the connected/communicated external device of different types of operating systems in relation to such as interactive gaming, teleconferencing, web browsing, media playing, navigation, image displaying or taking may all be preserved or advantageously made available, and preferably with enhanced data transmission, while realizing or having a relatively “light” hardware structures with great portability and interactive feasibility to users.
- In one aspect, the present invention provides a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time, in which the first device is coupled to the second device via a network communication, the first screen is configured to display a mirrored content of the second screen, and the first screen includes touch panel capability.
- In an embodiment, the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time includes the following steps: detecting a first touch signal on the first screen of the first device; converting the first touch signal into a first touch data associated with a first position information with respect to a first coordinate system of the first screen; transmitting the first touch data to the second device via the network communication, so that the actual operation is performed on the second device. In this embodiment, the actual operation performed on the second device can be overridden if a detecting event is detected on the second screen of the second device. In this embodiment, the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device.
- In addition, optional steps of the embodiment include of the following: detecting a motion event on the first device and generating a sensor data, and sending the sensor data to the second device via the network communication for enabling a corresponding operation on the second device, in which the corresponding operation includes changing orientation of the content displayed on the second screen. In the method of the present embodiment, one of a plurality of sensing ranges can be selected for detecting the motion event, and a threshold value can be provided for detecting the motion event on the first device. The method further includes charging capability to charge the first device with the second device or to charge the second device with the first device by detecting a charging connection between the first device and the second device, retrieving remaining battery amounts from the first device and the second device respectively, and providing an option to charge the first device with the second device, or to charge the second device with the first device.
- In an another embodiment, a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time further includes calculating a second position information with respect to a second coordinate system of the second screen based on the first position information.
- In yet another embodiment, the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time of the embodiment further includes additional steps which include: receiving compressed multimedia content compressed with a compression ratio from the second device via the network communication, and decompressing and displaying the multimedia content on the first screen of the first device. The compression ratio is determined based on a bandwidth of the network communication. In this embodiment, the decompressed multimedia content can be scaled based on a resolution of the first screen.
- In another aspect, the present invention provides a method for performing an actual operation on a second device having a second touch panel from a first device having a first touch panel in real time, and the first device is coupled to the second device via a network communication.
- In one embodiment, the method for performing the actual operation on the second device having the second touch panel from the first device having the first touch panel in real time includes steps as follow: receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel; calculating a second position information with respect to a second coordinate system of the second touch panel based on the first position information; and performing the actual operation on the second device based on the second position information. In this embodiment, the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device. In this embodiment, the performing of the actual operation is overridden if a detecting event is detected on the second touch panel of the second device. The method of this embodiment may further include steps for compressing and decompressing of multimedia content from the second device to the first device as follow: compressing a multimedia content displayed on the second touch panel, and sending the compressed multimedia content from the second device to the first device for decompressing and displaying on the first touch panel via the network communication. This embodiment may further include steps for performing operations on the second device in response to sensor data indicating a motion event on the first device as follow: receiving a sensor data from the first device, the sensor data indicating a motion event on the first device, and performing a corresponding operation on the second device in response to the sensor data. The corresponding operation may include changing orientation of the multimedia content displayed on the second display.
- In another aspect, the present invention provides a second device configured to be controlled by a first device with a first touch panel, and the first device is coupled to the second device by a network communication.
- In one embodiment, the second device includes a second touch panel, a compression unit for compressing multimedia content with a compression ratio, a transceiver for transmitting the compressed multimedia content to the first device for decompressing and displaying on the first touch panel and for receiving a touch data associated with a first position information with respect to a first coordinate system of the first touch panel, a coordinate calculating unit for calculating a second position information with respect to a second coordinate system of the second touch panel, and a processor for performing an actual operation on the second device according to the touch data and the second position information. In the second device, the actual operation can be overridden if the second touch panel detects a detecting event or if the second device receives an incoming call.
- The compression ratio is determined based on a bandwidth of the network communication.
- In the above embodiments, the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device; in addition, the first touch data is referred to the touch data being formed after converting of the at least one touch signal into touch data.
- This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
- Embodiments of the present invention will now be described, by way of example, with reference to the accompanying drawings, in which:
-
FIG. 1A is a schematic diagram showing wireless control from a RX side over a TX side according to an embodiment of the present invention. -
FIG. 1B andFIG. 1C are schematic diagrams showing touch control operations of from the RX side over the TX side shown inFIG. 1A according to an embodiment of the present invention. -
FIG. 2 is a schematic diagram showing a compression ratio in relation to an available bandwidth between the RX side and the TX side according to an alteration of the present invention. -
FIG. 3A andFIG. 3B are schematic diagrams showing an image orientation on the TX side in response to a motion detection on the RX side whereFIG. 3B is further provided with a switch element for adjusting motion sensitivity according to another alteration of the present invention. -
FIG. 3C is a schematic diagram showing motion control operations of the RX side over the TX side corresponding toFIGS. 3A and 3B according to an embodiment of the present invention. -
FIG. 3D andFIG. 3E are schematic diagrams showing adjustment of motion sensitivity on the RX side implemented by an application program on the TX side according to a further alteration of the present invention. -
FIG. 3F andFIG. 3G are schematic diagrams showing adjustment of magnitude and/or threshold values of application programs on the TX side after motion data is received according to a further alteration of the present invention. -
FIG. 4 is a flow diagram showing a process for controlling multimedia contents in realtime fashion according to an embodiment of the present invention. -
FIG. 5 is a flow diagram showing a process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention. -
FIG. 6 is a flow diagram showing another process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention. -
FIG. 7A andFIG. 7B are schematic diagrams showing battery charging between the TX side and the RX side according to a further alteration of the present invention. -
FIG. 8 is a flow diagram of a battery charging process according to an embodiment of the present invention. -
FIG. 9 is a flow diagram of a battery charging process according to another embodiment of the present invention. - Refer to
FIG. 1A , which is a schematic diagram of a RX side 10 (e.g. pad-like device having a touch-control display) and a TX side 12 (e.g. phone-like device having a computing power & operating system therein) according to an embodiment of the present invention. Advantageously, theRX side 10 is provided for controlling in realtime fashion multimedia contents that are originated from theTX side 12. The multimedia contents can include audio and video streams and be processed to present at the TX side 12 (such as cellphone, smartphone, PDA, and GPS navigation, or any other portable wireless device) having a phone-like screen 1202 and theTX side 12 is coupled to theRX side 10 via a network communication NC, which is preferably a wireless communication and a two-way communication. Noticeably, the multimedia contents may include video and audio data, i.e. the video data can be displayed on thescreen 1202 while the audio data can be played via a speaker (not shown) in theTX side 12. Besides, theRX side 10 may be advantageously provided in a CPU-less (without central processing unit) and/or GPU-less (without graphical processing unit) environment for the sake of less power consumption and/or simplified circuit complexity. - As shown in
FIG. 1A , theTX side 12 may be a computing system or mobile device equipped with CPU and/or GPU computing power, such as a PC, laptop, tablet PC, mobile phone, smart-phone, and may include a processor (not shown) for executing an application program (not shown) of an operation system (OS) 1204 installed in a storage medium (not shown) and atransceiver 1206 coupled to the processor and capable of transmitting and receiving the multimedia contents generated by the application program ofOS 1204 executed by the processor. TheOS 1204 may be any one of existing or known OS, such as the abovementioned Windows®, iOS®, Android®, Linux®, Blackberry®, Symbian®, Palm®, installed on theTX side 12 such that an application program may be performed or run thereon. The network communication NC between theRX side 10 and theTX side 12 may preferably be any one of the network communication standards or profiles of: IEEE 802.11, IEEE 802.16, CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth, Zigbee and combination thereof, or wired communication such as ADSL network, Optical/Fiber network or Cable network. - The RX side 10 (refer back to
FIG. 1A ) includes a pad-like screen 1002 (e.g. touch panel), a coordinatemapping unit 1004 and atransceiver 1006. Thescreen 1002 is provided for detecting at least one touch signal, and is preferably larger than thescreen 1202. The coordinatemapping unit 1004 can be controlled and/or controllable by an operating system (not shown). The coordinatemapping unit 1004 is provided for converting the at least one touch signal (from the screen 1002) into touch data associated with a first position information that indicates a virtual operation on thefirst screen 1002 corresponding to an actual operation on thesecond screen 1202, wherein the first position information is with respect to a first coordinate system of the first screen. For example, referred toFIG. 1B , the first position information represents a current cursor position (right side ofFIG. 1B ) on the pad-like screen 1202 moved by the user's finger (not shown), and thus the mirrored cursor on the phone-like screen 1202 will be moved to the corresponding position (left side ofFIG. 1B ) such that the user can do any operation on the phone-like screen by touching the pad-like screen 1202 while without touching the phone-like screen 1202. Thetransceiver 1006 is provided for transmitting the touch data (e.g. the first position information indicative of the current position of cursor on screen 1002) sent from the coordinatemapping unit 1004 to theTX side 12 via the network communication NC. On the other hand, theTX side 12 further includes a coordinate calculatingunit 1208 for calculating a second position information (e.g. new cursor position on phone-like screen) with respect to a second coordinate system of thescreen 1202 based on the first position information. For example, when the user “points and clicks” the new object on the touch screen 1002 (RX), and thus the same new object on the touch screen 1202 (TX) will be “pointed and clicked” simultaneously because an absolute position (x′, y′) of the same new object on thetouch screen 1002 can be obtained as following expression: -
- the mapping absolute position (x′, y′) on the phone-like screen=the current position (x, y)×mapping_ratio,
where - mapping_ratio along x-axis=x-axis_resolution of phone-like screen/x-axis_resolution of pad-like screen;
- mapping_ratio along y-axis=y-axis_resolution of phone-like screen/y-axis_resolution of pad-like screen;
- given the same origin for both
screens
- the mapping absolute position (x′, y′) on the phone-like screen=the current position (x, y)×mapping_ratio,
- Alternatively, the mapping movement (delta_x′, delta_y′) on the phone-like screen can be obtained as following expression:
-
- mapping movement (delta_x′, delta_y′) on the phone-like screen=current movement (delta_x, delta_v) on the pad-like screen×mapping ratio,
where - delta_x=x1−x0; delta_y=y1−y0;
- mapping_ratio along x-axis=x-axis_resolution of phone-like screen/x-axis resolution of pad-like screen;
- mapping_ratio along y-axis=y-axis_resolution of phone-like screen/y-axis resolution of pad-like screen;
given the same origin for bothscreens - where xo and y0 represents the old position on the pad-like screen, and x1 and y1 represents the new position on the pad-like screen while each of the mapping ratios has a value less than one when the display size of the pad-like screen is larger than that of the phone-like screen. Thus, the
operating system 1204 can be further provided for performing the actual operation at the second screen based on the second position information, such that thescreen 1002 can display results from the actual operation. As a result, theRX side 10 can control multimedia contents played atTX side 12 in realtime fashion due to up-scale and down-scale operations by the coordinate mapping unit 1004 (RX side) and the coordinate calculating unit 1208 (TX side).
- mapping movement (delta_x′, delta_y′) on the phone-like screen=current movement (delta_x, delta_v) on the pad-like screen×mapping ratio,
- Noticeably, the first position information may further include the current position of the cursor on the pad-
like screen 1002 of theRX side 10, and the resolution data of theRX side 10, e.g. 1920×1080, and the second position information can be calculated so as to obtain the mapping position of the cursor on the phone-like screen 1202, e.g. scaling down to a resolution 480×320. For example, the actual action can be “play a video movie” when the mapping position of the cursor on thescreen 1202 is pointed on a “play” icon. As a result, the user can “touch” on the pad-like screen 1002 of theRX side 10 to implement operations without access to the phone-like screen 1202 as if what he/she usually does on the phone-like screen 1202 of theTX side 12. On the other hand, all the tasks requiring extensive computing power/resources are handled by theTX side 12, and only the display task is duplicated (mirrored) to theRX side 10 from theTX side 12 such that the multimedia contents can be presented in realtime fashion between the pad-like screen 1002 of the RX side and the phone-like screen 1202 of theTX side 12. - For example, please refer to
FIG. 1B andFIG. 1C , which are schematic diagrams of touch control operations of theRX side 10 and theTX side 12 shown inFIG. 1A according to an embodiment of the present invention. As shown inFIG. 1B , a user can do the “open” operation to open a file or folder on thelarger screen 1002 of the RX side 10 (right side) and quasi-simultaneously thesmaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in theRX side 10, and thus the user can do the “double-click” operation to open some object like file/folder on thelarger screen 1002 while utilize computing power from theTX side 12, i.e. theoperating system 1204, to run the archive application. As shown inFIG. 1C , a user can do the “key-in” operation on the larger RX side 10 (right side) to type words and simultaneously thesmaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in theRX side 10, and thus the user can do the “key-in” operation on thelarger screen 1002 while utilize computing power from theTX side 12, i.e. theoperating system 1204, to run the text-editing application. - In detail, please continue to refer to
FIG. 1A , theTX side 12 may further include acompressing unit 1210, for compressing the multimedia contents with a compression ratio, e.g. compressing RGB data into MJPEG data, which requires less time for compression and thus facilitates operating in realtime fashion, and thetransceiver 1206 can be further utilized for transmitting the compressed multimedia contents to thescreen 1002 for displaying. Specifically, thetransceiver 1006 receives the compressed multimedia contents from thetransceiver 1206 via the network communication NC, and then compressed multimedia contents are decompressed by adecompressing unit 1010 with the compression ratio, such that the video data of the multimedia contents can be displayed on thescreen 1002 while the audio data of the multimedia contents can be played by aspeaker 1008. As a result, the multimedia contents can be displayed and controlled in realtime fashion between thescreen 1002 of theRX side 10 and thescreen 1202 of theTX side 12. - Moreover, please refer to
FIG. 2 , which is a schematic diagram of theRX side 10 and theTX side 12 according to an alteration of the present invention. In the embodiment shown inFIG. 2 , thetransceiver 1206 can further detect a bandwidth of the network communication NC, and theTX side 12 further includes adynamic compression unit 1212, for determining the compression ratio of the multimedia content based on the bandwidth of the network communication NC, such that thecompressing unit 1210 can compress the multimedia contents at the compression ratio, which is determined by the bandwidth of the network communication NC. Correspondingly, thetransceiver 1006 also can detect the bandwidth of the network communication NC, and then thedecompressing unit 1010 decompresses the compressed multimedia contents by the compression ratio provided by abandwidth calculating unit 1012. As a result, the multimedia contents can be compressed with a compression ratio according to a bandwidth of the network communication NC, such that the multimedia contents can be smoothly displayed in realtime fashion between thescreen 1002 of theRX side 10 and thescreen 1202 of theTX side 12 by facilitating the realtime codec function (e.g. the video compression ratio is determined by the actual available bandwidth). Besides, the transceiver 1006 (RX side) also can directly receive the compression ratio from the dynamic compression unit 1212 (TX side) rather than thebandwidth calculating unit 1012, and then to pass the compression ratio into thedecompressing unit 1010 so that thebandwidth calculating unit 1012 can be removed as well. - Besides, the present invention discloses another method for controlling in realtime fashion multimedia contents between a RX side having a first touch panel and a TX side having a second touch panel via a network communication. The multimedia contents comprise at least a screen frame data (not shown) where the screen frame data includes the screen resolution data, image data displayed on the screen and audio data. The method comprising the steps of:
-
- Step 1: detecting at least one touch signal from the
second touch panel 1202, for example, the user may touch the touch-sensitive phone-like screen 1202 to do something, and the “touch” can be detected and a touch signal will be sent to OS for particular operation; - Step 2: in response to detecting the touch signal from the second touch panel, compressing the screen frame data on the TX side and transmitting to the RX side so that the RX side decompresses the compressed screen frame data so as to obtain a TX's resolution data, where the TX's resolution data is provided together with a RX's resolution data to scale the screen frame data so as to display the screen frame data to fit with the RX's touch panel;
- Step 3: detecting at least one touch signal from the first touch panel;
- Step 4: in response to detecting the touch signal from the first touch panel, the further steps including:
- Step 4-1: converting the touch signal on the RX side into touch data associated with a first position information defining a virtual operation on the first screen corresponding to an actual operation on the second screen, the first position information being with respect to a first coordinate system of the first screen;
- Step 4-2: transmitting the touch data from the RX side to the TX side via the network communication, and calculating on the TX side a second position information with respect to a second coordinate system of the second screen based on the first position information, and
- Step 4-3: performing the actual operation at the second screen based on the second position information; and
- Step 5: implementing an override mechanism to assign the TX side a higher priority overriding the RX side when the two detecting events (respectively from the TX and RX sides) occur simultaneously, for example, the phone-
like screen 1202 will override the pad-like screen 1002 to take control when the user receives an incoming call from the phone-like screen 1202, and the override mechanism is implemented by a software program installed at the TX side. That means the TX side can dominant the RX side to facilitate the communication since the user has paid attention on doing the task at the TX side rather the RX side (e.g. pick-up the incoming call).
- Step 1: detecting at least one touch signal from the
- Furthermore, please refer to
FIGS. 3A-3B , which are schematic diagrams of theTX side 12 and theRX side 10 according to another alteration of the present invention. In one embodiment shown inFIG. 3B , theRX side 10 may further include a motion sensor 1014 (including at least one of an accelerometer, a gyroscope and a magnetic compass) for detecting at least one motion event. Thereceiver 1006 transmits the sensor data (RX) from themotion sensor 1014 after themotion sensor 1014 has detected the motion event on the RX side, and then the sensor data (RX) is sent to theTX side 12 via the network communication NC. A sensor decision unit (TX) may disable sensor data (TX) generated from amotion sensor 1216 on theTX side 12 only if the detecting exists (i.e. the sensor data (RX) from the RX side has being received on the TX side), and then send the sensor data (RX) to amotion process unit 1214. Themotion process unit 1214 can convert the sensor data (RX) into motion data (indicating a translation representative of linear acceleration in three-axis directions, and/or an angular rotation representative of Yaw, Roll and Pitch in three-axis directions) on theTX side 12, so as to present the multimedia contents on thescreen 1202 and subsequently on thescreen 1002 in a same particular orientation (e.g. portrait orientation or landscape orientation). The particular orientation is determined by theoperating system 1204 according to the motion data (RX) such that the image at thescreen 1202 will be presented at the landscape or portrait orientation. As a result, the multimedia contents can be displayed and controlled in realtime fashion between thescreen 1002 of theRX side 10 and thescreen 1202 of theTX side 12 through motion operations. - For example, please refer to
FIG. 3C , which is a schematic diagrams of motion control operations between theRX side 10 and theTX side 12 shown inFIGS. 3A-3B according to one embodiment of the present invention. As shown in the right side ofFIG. 3C , a user can do the “game play” driving motion, e.g. moving the image (RX) of the larger pad-like screen 1002 from Landscape to Portrait, and thus the mirrored image (TX) on the smaller phone-like screen 1002 (shown on left side ofFIG. 3C ) is quasi-simultaneously adjusted from Landscape to Portrait on theTX side 12 performed by theoperating system 1204 based on the motion data converted from the sensor data (RX) generated by the motion sensor 1014 (RX). Thus, the user can play the game on the pad-like screen 1002 while utilize computing power from theTX side 12, i.e. theoperating system 1204, to run the game application actually on the phone-like screen 1202. - Noticeably, the
motion sensor 1014 may have a plurality of sensing ranges, i.e. different sensing sensitivities, or a particular sensing range for generating the sensor data. When themotion sensor 1014 has a plurality of sensing ranges, one of the multiple sensing ranges can be selected by a control code determined by a switch signal so as to enable a motion chip to generate one motion data based on the selected sensing range determined by the control code. - In a first embodiment, the
motion sensor 1014 has a plurality of sensing ranges. As shown inFIG. 3A-3B , theRX side 10 detects whether a switch signal is invoked by aswitch element 1016 operative to a user, and theswitch element 1016 may issue an adjustment signal to a motion chip (i.e. motion sensor 1014) after the switch signal was invoked (e.g. user turns theswitch element 1016 on). Then, themotion sensor 1014 selects one of the plurality of sensing ranges of the motion chip based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between theRX side 10 and theTX side 12. For example, the motion chip has a sensing range of ±1 G at its “off” (original) state and ±2 G at “on” state. When theswitch element 1016 is turned to “on” from “off” by the user, the motion chip may have a new sensing range of ±2 G, where G is defined as 9.81 m/s2 for gravitational acceleration. - In a second embodiment, the
motion sensor 1014 has a plurality of sensing ranges. As shown inFIG. 3D-3E , theTX side 12 detects whether a switch signal is invoked by an application program 1220 (e.g. game software/program), and an adjustment signal is generated in response to the invoked switch signal. Then, themotion sensor 1014 selects one of the plurality of sensing ranges of the motion chip of themotion sensor 1014 based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between theRX side 10 and theTX side 12. For example, the game software may provide an input user interface to allow the user to choose one of three options for selecting particular sensing range among the plurality of sensing ranges where Option1 stands for ±1 G; Option2 for ±2 G; and Option3 for ±3 G, respectively. A particular switch signal will be sent out to the RX side after the selected Option2 is chosen by the user, for instance. The motion sensor 1014 (RX) is requested to change its sensing range to ±2 G directly by the switch signal, or by an adjusted signal in response to the switch signal. - In a third embodiment, the
motion sensor 1014 has a particular sensing range for generating the sensor data. As shown inFIG. 3F-3G , theTX side 12 detects whether a switch signal is invoked by anapplication program 1222 of theoperating system 1204, and an adjustment signal is sent in response to the invoked switch signal. Then, theTX side 12 may provide a user interface to allow the user to adjust a threshold value of the application program 1222 (e.g. game software/program) based on either the adjustment signal or the invoked switch signal such that the threshold value of theapplication program 1222 can be increased at a ratio corresponding to the adjustment signal or the invoked switch signal. Subsequently, the motion data may be generated based on sensor data that is sent from the motion sensor 1014 (on the TX side) triggered by the switch signal, so as to enable an interaction between theRX side 10 and theTX side 12. - In a fourth embodiment, the
motion sensor 1014 has a particular sensing range for generating the sensor data. As shown inFIG. 3D , theTX side 12 detects whether a switch signal is invoked by theapplication program 1222 of theoperating system 1204, and issues an adjustment signal in response to the invoked switch signal. Then, theTX side 12 may provide a user interface to allow the user to adjust a magnitude of the motion data by means of increasing the magnitude's value at a ratio based on either the adjustment signal or the switch signal, so as to enable an interaction between theRX side 10 and theTX side 12. Noticeably, in the above third and fourth embodiment, theTX side 12 can be configured to adjust the threshold value of theapplication program 1222 and the magnitude of the motion data, respectively, but theTX side 12 can also simultaneously adjust the threshold value of theapplication program 1222 and the matmitude of the motion data, while retaining both merits. - Noticeably the spirit of the present invention is to utilize a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side. Those skilled in the art should make modifications or alterations accordingly. For example, the
RX side 10 can be foldable, portable, or any other types to meet each user's need(s). - Besides, the network communication NC may preferably include a Wi-Fi communication and a Bluetooth (BT) communication, and since the touch data and sensor data have less data volume while the multimedia contents have greater data volume, the touch data and the motion date are preferably transmitted to the
TX side 12 via the BT communication, while the multimedia contents are preferably transmitted to theRX side 10 via the Wi-Fi communication. It can be understood that thetransceiver - Furthermore, the video data of the multimedia contents decompressed and outputted by the
decompressing unit 1010 may preferably from a MJPEG format to a RGB format and transmitted or outputted to thescreen 1002 in a LVDS format via a display interface converter such as a RGB-to-LVDS converter. It can be understood that other video stream formats such as ITDMI and related interfaces may also be possible. On the other hand, the audio data of the multimedia contents processed and outputted may preferably be transmitted via an audio interface such as an I2S interface or standard and further to thespeaker 1008. - Moreover, it is preferable that the multimedia contents transmitted and processed in video streaming may be made “real-time” or without significant delays with respect to time for the sake of low latency. In other words, in a preferred explanatory example of the present invention recited herein, the multimedia contents may be communicated between the
RX side 10 and theTX side 12 and processed to output “real-time” video and audio streams to thescreen 1002 and/or thespeaker 1008 of theRX side 10 of the present invention. Preferably, the multimedia contents may be processed in video streaming (i.e. encoding and decoding) and communicated there between as data packets and the duration or delay of time involved in a “real-time” processing and transmission may be less than 100 ms for the sake of low latency; in one example, it may preferably be less than 10 ms, and in another example, it may approximately be 20 ms. - Operations of the
RX side 10 and theTX side 12 shown in FIG IA can be summarized into a controlling process 40 shown inFIG. 4 . The controlling process 40 includes the following steps: -
- Step 400: Detect at least one touch signal via the
screen 1002, and convert the at least one touch signal into touch data associated with a first position information defining a virtual operation on thescreen 1002 corresponding to an actual operation on thescreen 1202, the first position information being with respect to a first coordinate system of thescreen 1002. - Step 402: Transmit the touch data to the
TX side 12 via the network communication NC, and calculate a second position information with respect to a second coordinate system of thescreen 1202 based on the first position information. - Step 404: Perform the actual operation at the
screen 1202 based on the second position information, and display results from the actual operation onto thescreen 1002 so as to control the multimedia contents in realtime fashion between thescreen 1002 of theRX side 10 and thescreen 1202 of theTX side 12. - Step 406: End.
- Step 400: Detect at least one touch signal via the
- Operations of the
RX side 10 shown inFIG. 1A can be summarized into a receiving process 50 shown inFIG. 5 . The receiving process 50 includes the following steps: -
- Step 500: Receive compressed multimedia contents from the
TX side 12. - Step 502: Convert the multimedia contents for decompression.
- Step 504: Scale the decompressed data based on a resolution of the
RX side 10. - Step 506: Output displayable multimedia contents to the
RX side 10. - Step 508: Detect whether there is touch signal from the
RX side 10. If yes, go to step 510; otherwise go to step 500. - Step 510: Output touch data to the
TX side 12. Go to step 500.
- Step 500: Receive compressed multimedia contents from the
- Operations of the
RX side 10 and theTX side 12 shown inFIG. 2A can be summarized into a receiving and transmitting process 60 shown inFIG. 6 . The receiving and transmitting process 60 includes the following steps: -
- Step 600: Receive compressed multimedia contents from the
TX side 12. - Step 602: Determine a compression ratio based on a bandwidth of the network communication NC.
- Step 604: Convert the multimedia contents based on the compression ratio.
- Step 606: Decompress the compressed data, and scale thereof if needed.
- Step 608: Output displayable multimedia contents to the
RX side 10. - Step 610: Detect whether there is touch signal from the
RX side 10. If yes, go to step 510; otherwise go to step 600. - Step 612: Output touch data to obtain position data.
- Step 614: Perform actions based on the position data at the
TX side 12. Go to step 600.
- Step 600: Receive compressed multimedia contents from the
- Detail descriptions of the controlling process 40, the receiving processes 50 and the receiving and transmitting process 60 can be derived by referring to the above description, and are not narrated hereinafter.
- On the other hand, please refer to
FIG. 7A-7B , which is a schematic diagram of theRX side 10 according to another alteration of the present invention. As shown inFIG. 7A-7B , theTX side 12 further comprises acharging application 1224, chargingdetector 1226 andbattery cells 1228, and theRX side 10 further comprises a chargingcontroller 1018 andbattery cells 1020. - In one embodiment, system default charging operations of the
TX side 12 and theRX side 10 shown inFIG. 7A-7B can be summarized into chargingprocess 80 as shown inFIG. 8 . The chargingprocess 80 includes the following steps: -
- Step 800: Detect an event of a charging connection between the
TX side 12 and theRX side 10. The charging connection is advantageously provided for conforming to the USB standard. It is advantageously applicable to the two-way charging process. - Step 802: Retrieve remaining battery amounts (Rt and Rr) respectively from
battery cells 1228 on theTX side 12 andbattery cells 1020 on theRX side 10. - Step 804: Compare the two remaining battery amounts (Rt and Rr) respectively with two predetermined battery capacity percentages (Tt and Tr) by a power management mechanism that is executed by either the
charging application 1224 in the operating system or the chargingcontroller 1018. If each of the remaining battery amounts is below the predetermined battery capacity percentage (i.e. Rt<Tt and Rr<Tr), go to step 806; if the remaining battery amount of theRX side 10 is higher than its predetermined battery capacity percentage while the remaining battery amount of theRX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), go to step 810; if the remaining battery amount of theTX side 12 is higher than its predetermined battery capacity percentage while the remaining battery amount of theTX side 12 is higher than that of the RX side 10 (i.e. Rt>Tt and Rt>Rr), go to step 814. - Step 806: Send a first message of recharging information shown as “Please Use Power Adaptor to charge Pad and Cellphone”.
- Step 808: Charging operation 1: Use Power Adaptor to charge Pad and Cellphone.
- Step 810: Send a second message of recharging information shown as “Please Use Pad to charge Cellphone”.
- Step 812: Charging operation 2: Use Pad to charge Cellphone.
- Step 814: Send a third message of recharging information “Please Use Cellphone to charge Pad”.
- Step 816: Charging operation 3: Use Cellphone to charge Pad.
- Step 800: Detect an event of a charging connection between the
- In another embodiment, user selection charging operations of the
TX side 12 and theRX side 10 shown inFIG. 7A-7B can be summarized into chargingprocess 90 as shown inFIG. 9 . The chargingprocess 90 includes the following steps: -
- Step 900: Detect an event of a charging connection between the
TX side 12 and theRX side 10. The charging connection is advantageously provided for conforming to the USB standard. It is also applicable to the two-way charging process. - Step 902: Retrieve remaining battery amounts (Rt and Rr) respectively from
battery cells 1228 on theTX side 12 andbattery cells 1020 on theRX side 10. - Step 904: Present a user interface to allow a user's selection of battery charging.
- Step 905: If a user chooses
option 1 for charging Cellphone, go to step 906; if the user choosesoption 2 for charging Pad, go to step 908. - Step 906: Use Pad to charge Cellphone.
- Step 908: Use Cellphone to charge Pad.
- Step 910: Compare the two remaining battery amounts with two predetermined battery capacity percentages (Tt and Tr) by a power management mechanism that is executed by either the
charging application 1224 in the operating system or the chargingcontroller 1018. If each of the remaining battery amounts is below the predetermined battery capacity percentage (i.e. Rt<Tt and Rr<Tr), go to step 912; if the remaining battery amount of theRX side 10 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of theRX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), go to step 916; if the remaining battery amount of theTX side 12 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of theTX side 12 is higher than that of the RX side 10 (i.e. Rt>Tt and Rt>Rr), go to step 920. - Step 912: Send a first message of recharging information shown as “Please Use Power Adaptor to charge Pad and Cellphone”.
- Step 914: Charging operation 1: Use Power Adaptor to charge Pad and Cellphone. Come to END.
- Step 916: Send a second message of recharging information shown as “Please Use Pad to charge Cellphone”.
- Step 918: Charging operation 2: Use Pad to charge Cellphone. Come to END.
- Step 920: Send a third message of recharging information “Please Use Cellphone to charge Pad”.
- Step 922: Charging operation 3: Use Cellphone to charge Pad. Come to END.
- Step 900: Detect an event of a charging connection between the
- The event of a charging connection can be detected either by a charging
detector 1226 on theTX side 12 or by a chargingcontroller 1018 on theRX side 10. The chargingcontroller 1018 can be implemented by a power management chip. The chargingdetector 1226 can receive the remaining battery capacity of thebattery cells 1228 on theTX side 12; so as to send the battery capacity information to the user interface by means of thecharging application 1224 managed byOS 1204, and thus users can view the icon that indicates the actual measurement of the TX's battery capacity. The chargingcontroller 1218 on theRX side 10 can receive the actual measurement of the RX's battery capacity, and sends the measurement to thecharging detector 1226 on theTX side 12, so as to facilitate thecharging detector 1226 to determine what condition can be meet to implement the appropriate charging operation. - For example, when each of the remaining battery amounts is below the predetermined battery capacity percentage (i.e. Rt<Tt and Rr<Tr), a power adaptor is used to charge both the
RX side 10 and theTX side 12; when the remaining battery amount of theRX side 10 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of theRX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), theRX side 10 is used to charge theTX side 12; when the remaining battery amount of theTX side 12 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of theTX side 12 is higher than that of the RX side 10 (i.e. Rt>Tt and Rt>Rr), theTX side 12 is used to charge theRX side 10. It is noted that the RX side may be compulsorily switched to the power saving mode by the override mechanism when the user receives an incoming call from the TX side, that is, the user currently uses the TX side rather the RX side to either answer or dial the call. Therefore, it makes sense for the user to temporarily keep the RX side idle while he or she is busy for phone call. - In the prior art, a universal apparatus capable of communicating with the any example of new breed of electronic devices running any certain type of operating systems is lack. In comparison, the present invention utilizes a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side. Furthermore, the present invention can compress multimedia contents a bandwidth of network communication, such that the multimedia contents can be smoothly displayed in realtime fashion between the
screen 1002 of theRX side 10 and thescreen 1202 of theTX side 12. Besides, theTX side 12 can also be configured as a server having encoding hardware and theRX side 10 as a client having decoding hardware adapted for cloud computing environment running video streaming for the purpose of low latency requirement. Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.
Claims (20)
1. A method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time, the first device is coupled to the second device via a network communication, the first screen is configured to display a mirrored content of the second screen, the first screen includes touch panel capability, the method comprises:
detecting a first touch signal on the first screen of the first device;
converting the first touch signal into a first touch data associated with a first position information with respect to a first coordinate system of the first screen; and
transmitting the first touch data to the second device via the network communication, so that the actual operation is performed on the second device.
2. The method of claim 1 , further comprising:
calculating a second position information with respect to a second coordinate system of the second screen based on the first position information.
3. The method of claim 1 , further comprising:
receiving compressed multimedia content compressed with a compression ratio from the second device via the network communication; and
decompressing and displaying the multimedia content on the first screen of the first device.
4. The method of claim 3 , wherein the compression ratio is determined based on a bandwidth of the network communication.
5. The method of claim 3 , further comprising:
scaling the decompressed multimedia content based on a resolution of the first screen.
6. The method of claim 1 , wherein the actual operation performed on the second device is overridden if a detecting event is detected on the second screen of the second device.
7. The method of claim 1 , further comprising:
detecting a motion event on the first device and generating a sensor data; and
sending the sensor data to the second device via the network communication for enabling a corresponding operation on the second device.
8. The method of claim 7 , wherein the corresponding operation includes changing orientation of the content displayed on the second screen.
9. The method of claim 7 , further comprising:
selecting one of a plurality of sensing ranges for detecting the motion event.
10. The method of claim 7 , further comprising:
providing a threshold value for detecting the motion event on the first device.
11. The method of claim 1 , further comprising:
detecting a charging connection between the first device and the second device;
retrieving remaining battery amounts from the first device and the second device respectively;
providing an option to charge the first device with the second device, or to charge the second device with the first device.
12. A method for performing an actual operation on a second device having a second touch panel from a first device having a first touch panel in real time, the first device is coupled to the second device via a network communication, the method comprises:
receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel;
calculating a second position information with respect to a second coordinate system of the second touch panel based on the first position information; and
performing the actual operation on the second device based on the second position information.
13. The method of claim 12 , further comprising:
compressing a multimedia content displayed on the second touch panel; and
sending the compressed multimedia content from the second device to the first device for decompressing and displaying on the first touch panel via the network communication.
14. The method of claim 12 , wherein the performing of the actual operation is overridden if a detecting event is detected on the second touch panel of the second device.
15. The method of claim 12 , further comprising:
receiving a sensor data from the first device, the sensor data indicating a motion event on the first device; and
performing a corresponding operation on the second device in response to the sensor data.
16. The method of claim 15 , wherein the corresponding operation includes changing orientation of the multimedia content displayed on the second display.
17. A second device configured to be controlled by a first device with a first touch panel, the first device is coupled to the second device by a network communication, the second device comprising:
a second touch panel;
a compression unit for compressing multimedia content with a compression ratio;
a transceiver for transmitting the compressed multimedia content to the first device for decompressing and displaying on the first touch panel, and for receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel;
a coordinate calculating unit for calculating a second position information with respect to a second coordinate system of the second touch panel; and
a processor for performing an actual operation on the second device according to the first touch data and the second position information.
18. The second device of claim 17 , wherein the actual operation is overridden if the second touch panel detects a detecting event.
19. The second device of claim 17 , wherein the actual operation is overridden if the second device receives an incoming call.
20. The second device of claim 17 , wherein the compression ratio is determined based on a bandwidth of the network communication.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/069,384 US20140055401A1 (en) | 2010-11-05 | 2013-11-01 | Method and apparatus for controlling contents in realtime fashion |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US41037110P | 2010-11-05 | 2010-11-05 | |
US13/184,584 US8605048B2 (en) | 2010-11-05 | 2011-07-18 | Method and apparatus for controlling multimedia contents in realtime fashion |
US14/069,384 US20140055401A1 (en) | 2010-11-05 | 2013-11-01 | Method and apparatus for controlling contents in realtime fashion |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/184,584 Continuation US8605048B2 (en) | 2010-11-05 | 2011-07-18 | Method and apparatus for controlling multimedia contents in realtime fashion |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140055401A1 true US20140055401A1 (en) | 2014-02-27 |
Family
ID=46019153
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/184,584 Expired - Fee Related US8605048B2 (en) | 2010-11-05 | 2011-07-18 | Method and apparatus for controlling multimedia contents in realtime fashion |
US13/253,993 Abandoned US20120114047A1 (en) | 2010-11-05 | 2011-10-06 | Wireless communication system for transmitting high resolution multimedia data and method thereof |
US14/069,384 Abandoned US20140055401A1 (en) | 2010-11-05 | 2013-11-01 | Method and apparatus for controlling contents in realtime fashion |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/184,584 Expired - Fee Related US8605048B2 (en) | 2010-11-05 | 2011-07-18 | Method and apparatus for controlling multimedia contents in realtime fashion |
US13/253,993 Abandoned US20120114047A1 (en) | 2010-11-05 | 2011-10-06 | Wireless communication system for transmitting high resolution multimedia data and method thereof |
Country Status (2)
Country | Link |
---|---|
US (3) | US8605048B2 (en) |
CN (1) | CN202602826U (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140247196A1 (en) * | 2010-12-31 | 2014-09-04 | Ebay Inc. | Methods and systems for displaying content on multiple network devices with a simple command |
US20140282068A1 (en) * | 2013-03-15 | 2014-09-18 | SingTel Idea Factory Pte. Ltd. | Systems and methods for transferring of objects among mobile devices based on pairing and matching using actions and/or gestures associated with the mobile device |
US20150042559A1 (en) * | 2013-08-09 | 2015-02-12 | Lenovo (Beijing) Co., Ltd. | Information Processing Method And Electronic Device Thereof, Image Calibration Method And Apparatus, And Electronic Device Thereof |
US20160313960A1 (en) * | 2015-04-24 | 2016-10-27 | Lenovo (Beijing) Co., Ltd. | Control Method, Control Apparatus, And Electronic Device |
US10171261B2 (en) * | 2014-04-24 | 2019-01-01 | Vivint, Inc. | Remote talk down to panel, camera and speaker |
US11991249B2 (en) * | 2012-06-13 | 2024-05-21 | Huawei Device Co., Ltd. | Method, user equipment, server, and apparatus for implementing information sharing |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9198084B2 (en) | 2006-05-26 | 2015-11-24 | Qualcomm Incorporated | Wireless architecture for a traditional wire-based protocol |
US9398089B2 (en) | 2008-12-11 | 2016-07-19 | Qualcomm Incorporated | Dynamic resource sharing among multiple wireless devices |
US9264248B2 (en) | 2009-07-02 | 2016-02-16 | Qualcomm Incorporated | System and method for avoiding and resolving conflicts in a wireless mobile display digital interface multicast environment |
US9582238B2 (en) | 2009-12-14 | 2017-02-28 | Qualcomm Incorporated | Decomposed multi-stream (DMS) techniques for video display systems |
US8781000B2 (en) * | 2010-12-30 | 2014-07-15 | Vixs Systems, Inc. | Dynamic video data compression |
US10135900B2 (en) | 2011-01-21 | 2018-11-20 | Qualcomm Incorporated | User input back channel for wireless displays |
US9787725B2 (en) | 2011-01-21 | 2017-10-10 | Qualcomm Incorporated | User input back channel for wireless displays |
US9413803B2 (en) | 2011-01-21 | 2016-08-09 | Qualcomm Incorporated | User input back channel for wireless displays |
US9065876B2 (en) | 2011-01-21 | 2015-06-23 | Qualcomm Incorporated | User input back channel from a wireless sink device to a wireless source device for multi-touch gesture wireless displays |
US9582239B2 (en) * | 2011-01-21 | 2017-02-28 | Qualcomm Incorporated | User input back channel for wireless displays |
US9503771B2 (en) | 2011-02-04 | 2016-11-22 | Qualcomm Incorporated | Low latency wireless display for graphics |
US10108386B2 (en) | 2011-02-04 | 2018-10-23 | Qualcomm Incorporated | Content provisioning for wireless back channel |
KR20130054579A (en) * | 2011-11-17 | 2013-05-27 | 삼성전자주식회사 | Display apparatus and control method thereof |
JP6106983B2 (en) * | 2011-11-30 | 2017-04-05 | 株式会社リコー | Image display device, image display system, method and program |
US9339691B2 (en) | 2012-01-05 | 2016-05-17 | Icon Health & Fitness, Inc. | System and method for controlling an exercise device |
US9525998B2 (en) | 2012-01-06 | 2016-12-20 | Qualcomm Incorporated | Wireless display with multiscreen service |
KR102058947B1 (en) * | 2012-05-24 | 2019-12-24 | 엘지전자 주식회사 | Mobile terminal and gps engine control method therof |
US9632648B2 (en) | 2012-07-06 | 2017-04-25 | Lg Electronics Inc. | Mobile terminal, image display device and user interface provision method using the same |
TWI637312B (en) * | 2012-09-19 | 2018-10-01 | 三星電子股份有限公司 | Method for displaying information on transparent display device, display device therewith, and computer-readable recording medium therefor |
JP2014068145A (en) * | 2012-09-25 | 2014-04-17 | Kyocera Corp | Portable terminal, display control program and display control method |
EP2969058B1 (en) | 2013-03-14 | 2020-05-13 | Icon Health & Fitness, Inc. | Strength training apparatus with flywheel and related methods |
US9403047B2 (en) | 2013-12-26 | 2016-08-02 | Icon Health & Fitness, Inc. | Magnetic resistance mechanism in a cable machine |
WO2015138339A1 (en) | 2014-03-10 | 2015-09-17 | Icon Health & Fitness, Inc. | Pressure sensor to quantify work |
US10426989B2 (en) | 2014-06-09 | 2019-10-01 | Icon Health & Fitness, Inc. | Cable system incorporated into a treadmill |
WO2015195965A1 (en) | 2014-06-20 | 2015-12-23 | Icon Health & Fitness, Inc. | Post workout massage device |
US9883101B1 (en) * | 2014-07-23 | 2018-01-30 | Hoyos Integrity Corporation | Providing a real-time via a wireless communication channel associated with a panoramic video capture device |
US10391361B2 (en) | 2015-02-27 | 2019-08-27 | Icon Health & Fitness, Inc. | Simulating real-world terrain on an exercise device |
US10625137B2 (en) | 2016-03-18 | 2020-04-21 | Icon Health & Fitness, Inc. | Coordinated displays in an exercise device |
US10272317B2 (en) | 2016-03-18 | 2019-04-30 | Icon Health & Fitness, Inc. | Lighted pace feature in a treadmill |
US10493349B2 (en) | 2016-03-18 | 2019-12-03 | Icon Health & Fitness, Inc. | Display on exercise device |
US10671705B2 (en) | 2016-09-28 | 2020-06-02 | Icon Health & Fitness, Inc. | Customizing recipe recommendations |
KR102447016B1 (en) * | 2017-11-01 | 2022-09-27 | 삼성디스플레이 주식회사 | Display driver integrated circuit, display system, and method for driving display driver integrated circuit |
CN113918104A (en) * | 2020-07-07 | 2022-01-11 | 华为技术有限公司 | Method and equipment for opening file |
Citations (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3708785A (en) * | 1970-07-31 | 1973-01-02 | Searle Medidata Inc | Data scanner for real time interfacing of a computer and plural remote units |
US5062059A (en) * | 1986-12-08 | 1991-10-29 | Sunriver Corporation | Apparatus and method for communication between host CPU and remote terminal |
US5638426A (en) * | 1993-10-12 | 1997-06-10 | Multimedia Systems Corporation | Interactive system for a closed cable network |
US5721878A (en) * | 1995-06-07 | 1998-02-24 | International Business Machines Corporation | Multimedia control system and method for controlling multimedia program presentation |
US5721815A (en) * | 1995-06-07 | 1998-02-24 | International Business Machines Corporation | Media-on-demand communication system and method employing direct access storage device |
US20030137426A1 (en) * | 2001-12-07 | 2003-07-24 | Eric Anthony | Early warning near-real-time security system |
US20040060063A1 (en) * | 2002-09-24 | 2004-03-25 | Russ Samuel H. | PVR channel and PVR IPG information |
US20040240650A1 (en) * | 2003-05-05 | 2004-12-02 | Microsoft Corporation | Real-time communications architecture and methods for use with a personal computer system |
US20050015803A1 (en) * | 2002-11-18 | 2005-01-20 | Macrae Douglas B. | Systems and methods for providing real-time services in an interactive television program guide application |
US20050021642A1 (en) * | 2003-05-27 | 2005-01-27 | Shunichiro Nonaka | Method and apparatus for moving image conversion, method and apparatus for moving image transmission, and programs therefor |
US20050093868A1 (en) * | 2003-10-30 | 2005-05-05 | Microsoft Corporation | Distributed sensing techniques for mobile devices |
US20050155043A1 (en) * | 2004-01-08 | 2005-07-14 | Schulz Kurt S. | Human-machine interface system and method for remotely monitoring and controlling a machine |
US20060031787A1 (en) * | 2003-08-02 | 2006-02-09 | Viswanath Ananth | System and method for real-time configurable monitoring and management of task performance systems |
US20060041655A1 (en) * | 2004-05-06 | 2006-02-23 | Marty Holloway | Bi-directional remote control for remotely controllable apparatus |
US20060242589A1 (en) * | 2005-04-26 | 2006-10-26 | Rod Cooper | System and method for remote examination services |
US20070035528A1 (en) * | 2004-02-10 | 2007-02-15 | Bruce Hodge | Method and apparatus for determining and retrieving positional information |
US7224382B2 (en) * | 2002-04-12 | 2007-05-29 | Image Masters, Inc. | Immersive imaging system |
US20090019492A1 (en) * | 2007-07-11 | 2009-01-15 | United Video Properties, Inc. | Systems and methods for mirroring and transcoding media content |
US20100064230A1 (en) * | 2008-09-09 | 2010-03-11 | Applied Systems, Inc. | Method and apparatus for remotely displaying screen files and efficiently handling remote operator input |
US20100076993A1 (en) * | 2008-09-09 | 2010-03-25 | Applied Systems, Inc. | Method and apparatus for remotely displaying a list by determining a quantity of data to send based on the list size and the display control size |
US20100073454A1 (en) * | 2008-09-17 | 2010-03-25 | Tandberg Telecom As | Computer-processor based interface for telepresence system, method and computer program product |
US20100179005A1 (en) * | 2000-06-16 | 2010-07-15 | Skyhawke Technologies, Llc. | Personal golfing assistant and method and system for graphically displaying golf related information and for collection, processing and distribution of golf related data |
US20110188760A1 (en) * | 2010-02-03 | 2011-08-04 | Oculus Info Inc. | System and Method for Creating and Displaying Map Projections related to Real-Time Images |
US8208007B2 (en) * | 2004-04-21 | 2012-06-26 | Telepresence Technologies, Llc | 3-D displays and telepresence systems and methods therefore |
US20140002361A1 (en) * | 2012-06-28 | 2014-01-02 | Lee E. Ballard | Systems and methods for remote mouse pointer management |
US20140053189A1 (en) * | 2012-08-17 | 2014-02-20 | Electronics And Telecommunications Research Institute | Apparatus and method for providing personalized home screen |
US20140223490A1 (en) * | 2013-02-07 | 2014-08-07 | Shanghai Powermo Information Tech. Co. Ltd. | Apparatus and method for intuitive user interaction between multiple devices |
US20140380193A1 (en) * | 2013-06-24 | 2014-12-25 | Microsoft Corporation | Showing interactions as they occur on a whiteboard |
US20150163281A1 (en) * | 2013-12-06 | 2015-06-11 | Vmware, Inc. | Moving objects of a remote desktop in unstable network environments |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6018765A (en) * | 1996-01-23 | 2000-01-25 | Storage Concepts, Inc. | Multi-channel multimedia data server |
US6970183B1 (en) * | 2000-06-14 | 2005-11-29 | E-Watch, Inc. | Multimedia surveillance and monitoring system including network configuration |
US6203432B1 (en) | 1999-05-11 | 2001-03-20 | Madcatz, Inc. | System for adjusting the response characteristic of an electronic game input device |
US7628730B1 (en) * | 1999-07-08 | 2009-12-08 | Icon Ip, Inc. | Methods and systems for controlling an exercise apparatus using a USB compatible portable remote device |
US6351991B1 (en) | 2000-06-05 | 2002-03-05 | Schlumberger Technology Corporation | Determining stress parameters of formations from multi-mode velocity data |
KR100396892B1 (en) * | 2000-10-05 | 2003-09-03 | 삼성전자주식회사 | Video MP3 system, apparatus and method for generating reduced video data, and apparatus and method for restoring the reduced video data |
JP2003046938A (en) * | 2001-07-30 | 2003-02-14 | Fujitsu Ltd | Semiconductor device |
US20030059014A1 (en) * | 2001-09-24 | 2003-03-27 | Teleware, Inc. | Multi-media communication management system with caller managed hold system |
JP2004220133A (en) | 2003-01-10 | 2004-08-05 | Toshiba Corp | Information processor, and operation control method for information processor |
JP2005267257A (en) | 2004-03-18 | 2005-09-29 | Nara Institute Of Science & Technology | Handwritten information input system |
US20050229228A1 (en) * | 2004-04-07 | 2005-10-13 | Sandeep Relan | Unicast cable content delivery |
US8370888B2 (en) * | 2004-06-22 | 2013-02-05 | University Of Southern California | Hydra: high-performance data recording architecture for streaming media |
JP4220943B2 (en) | 2004-07-15 | 2009-02-04 | ソフトバンクモバイル株式会社 | Electronics |
TWI332797B (en) * | 2005-06-30 | 2010-11-01 | Realtek Semiconductor Corp | Cordless multimedia system and image display system |
US7927216B2 (en) | 2005-09-15 | 2011-04-19 | Nintendo Co., Ltd. | Video game system with wireless modular handheld controller |
JP4262726B2 (en) | 2005-08-24 | 2009-05-13 | 任天堂株式会社 | Game controller and game system |
US8065710B2 (en) * | 2006-03-02 | 2011-11-22 | At& T Intellectual Property I, L.P. | Apparatuses and methods for interactive communication concerning multimedia content |
CN101110759A (en) * | 2006-07-20 | 2008-01-23 | 朗迅科技公司 | Peer-to-peer file download system for IPTV network |
JP3133600U (en) | 2007-04-10 | 2007-07-19 | 永之助 古閑 | Accelerometer type pen-type pointing device. |
US8194732B2 (en) * | 2007-07-30 | 2012-06-05 | Broadcom Corporation | Integrated circuit with conversion capability for portable media player |
TWI362605B (en) | 2007-08-30 | 2012-04-21 | Ind Tech Res Inst | Method for adjusting sensing range and sensitivity and inertia interactive apparatus and system using thereof |
US8416848B2 (en) * | 2007-12-21 | 2013-04-09 | Broadcom Corporation | Device adaptive video transmission system for use with layered video coding and methods for use therewith |
US20090238263A1 (en) * | 2008-03-20 | 2009-09-24 | Pawan Jaggi | Flexible field based energy efficient multimedia processor architecture and method |
TWM342547U (en) * | 2008-05-29 | 2008-10-11 | Grandtec Electronic Corp | Transmission switching device between USB and HDMI |
US9166633B2 (en) * | 2010-01-21 | 2015-10-20 | Qualcomm Incorporated | Systems and methods for interfacing a white space device with a host device |
US20110276624A1 (en) * | 2010-05-07 | 2011-11-10 | Haseeb Akhtar | Converged communication with web 2.0 based screen triggers |
-
2011
- 2011-07-18 US US13/184,584 patent/US8605048B2/en not_active Expired - Fee Related
- 2011-10-06 US US13/253,993 patent/US20120114047A1/en not_active Abandoned
- 2011-11-03 CN CN201120431207.6U patent/CN202602826U/en not_active Expired - Fee Related
-
2013
- 2013-11-01 US US14/069,384 patent/US20140055401A1/en not_active Abandoned
Patent Citations (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3708785A (en) * | 1970-07-31 | 1973-01-02 | Searle Medidata Inc | Data scanner for real time interfacing of a computer and plural remote units |
US5062059A (en) * | 1986-12-08 | 1991-10-29 | Sunriver Corporation | Apparatus and method for communication between host CPU and remote terminal |
US5638426A (en) * | 1993-10-12 | 1997-06-10 | Multimedia Systems Corporation | Interactive system for a closed cable network |
US5721878A (en) * | 1995-06-07 | 1998-02-24 | International Business Machines Corporation | Multimedia control system and method for controlling multimedia program presentation |
US5721815A (en) * | 1995-06-07 | 1998-02-24 | International Business Machines Corporation | Media-on-demand communication system and method employing direct access storage device |
US20100179005A1 (en) * | 2000-06-16 | 2010-07-15 | Skyhawke Technologies, Llc. | Personal golfing assistant and method and system for graphically displaying golf related information and for collection, processing and distribution of golf related data |
US20030137426A1 (en) * | 2001-12-07 | 2003-07-24 | Eric Anthony | Early warning near-real-time security system |
US7224382B2 (en) * | 2002-04-12 | 2007-05-29 | Image Masters, Inc. | Immersive imaging system |
US20040060063A1 (en) * | 2002-09-24 | 2004-03-25 | Russ Samuel H. | PVR channel and PVR IPG information |
US20050015803A1 (en) * | 2002-11-18 | 2005-01-20 | Macrae Douglas B. | Systems and methods for providing real-time services in an interactive television program guide application |
US20040240650A1 (en) * | 2003-05-05 | 2004-12-02 | Microsoft Corporation | Real-time communications architecture and methods for use with a personal computer system |
US20050021642A1 (en) * | 2003-05-27 | 2005-01-27 | Shunichiro Nonaka | Method and apparatus for moving image conversion, method and apparatus for moving image transmission, and programs therefor |
US20060031787A1 (en) * | 2003-08-02 | 2006-02-09 | Viswanath Ananth | System and method for real-time configurable monitoring and management of task performance systems |
US20050093868A1 (en) * | 2003-10-30 | 2005-05-05 | Microsoft Corporation | Distributed sensing techniques for mobile devices |
US20050155043A1 (en) * | 2004-01-08 | 2005-07-14 | Schulz Kurt S. | Human-machine interface system and method for remotely monitoring and controlling a machine |
US20070035528A1 (en) * | 2004-02-10 | 2007-02-15 | Bruce Hodge | Method and apparatus for determining and retrieving positional information |
US8208007B2 (en) * | 2004-04-21 | 2012-06-26 | Telepresence Technologies, Llc | 3-D displays and telepresence systems and methods therefore |
US20060041655A1 (en) * | 2004-05-06 | 2006-02-23 | Marty Holloway | Bi-directional remote control for remotely controllable apparatus |
US20060242589A1 (en) * | 2005-04-26 | 2006-10-26 | Rod Cooper | System and method for remote examination services |
US20090019492A1 (en) * | 2007-07-11 | 2009-01-15 | United Video Properties, Inc. | Systems and methods for mirroring and transcoding media content |
US20100064230A1 (en) * | 2008-09-09 | 2010-03-11 | Applied Systems, Inc. | Method and apparatus for remotely displaying screen files and efficiently handling remote operator input |
US20100076993A1 (en) * | 2008-09-09 | 2010-03-25 | Applied Systems, Inc. | Method and apparatus for remotely displaying a list by determining a quantity of data to send based on the list size and the display control size |
US20100073454A1 (en) * | 2008-09-17 | 2010-03-25 | Tandberg Telecom As | Computer-processor based interface for telepresence system, method and computer program product |
US20110188760A1 (en) * | 2010-02-03 | 2011-08-04 | Oculus Info Inc. | System and Method for Creating and Displaying Map Projections related to Real-Time Images |
US20140002361A1 (en) * | 2012-06-28 | 2014-01-02 | Lee E. Ballard | Systems and methods for remote mouse pointer management |
US20140053189A1 (en) * | 2012-08-17 | 2014-02-20 | Electronics And Telecommunications Research Institute | Apparatus and method for providing personalized home screen |
US20140223490A1 (en) * | 2013-02-07 | 2014-08-07 | Shanghai Powermo Information Tech. Co. Ltd. | Apparatus and method for intuitive user interaction between multiple devices |
US20140380193A1 (en) * | 2013-06-24 | 2014-12-25 | Microsoft Corporation | Showing interactions as they occur on a whiteboard |
US20150163281A1 (en) * | 2013-12-06 | 2015-06-11 | Vmware, Inc. | Moving objects of a remote desktop in unstable network environments |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10007477B2 (en) * | 2010-12-31 | 2018-06-26 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US10747491B2 (en) * | 2010-12-31 | 2020-08-18 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US9367281B2 (en) * | 2010-12-31 | 2016-06-14 | Ebay Inc. | Methods and systems for displaying content on multiple network devices with a simple command |
US20140247196A1 (en) * | 2010-12-31 | 2014-09-04 | Ebay Inc. | Methods and systems for displaying content on multiple network devices with a simple command |
US11650781B2 (en) * | 2010-12-31 | 2023-05-16 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US20220147306A1 (en) * | 2010-12-31 | 2022-05-12 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US20160259616A1 (en) * | 2010-12-31 | 2016-09-08 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US20180293044A1 (en) * | 2010-12-31 | 2018-10-11 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US11269583B2 (en) * | 2010-12-31 | 2022-03-08 | Ebay Inc. | Methods and systems for displaying content on multiple networked devices with a simple command |
US11991249B2 (en) * | 2012-06-13 | 2024-05-21 | Huawei Device Co., Ltd. | Method, user equipment, server, and apparatus for implementing information sharing |
US20140282068A1 (en) * | 2013-03-15 | 2014-09-18 | SingTel Idea Factory Pte. Ltd. | Systems and methods for transferring of objects among mobile devices based on pairing and matching using actions and/or gestures associated with the mobile device |
US9691357B2 (en) * | 2013-08-09 | 2017-06-27 | Lenovo (Beijing) Co., Ltd. | Information processing method and electronic device thereof, image calibration method and apparatus, and electronic device thereof |
US20150042559A1 (en) * | 2013-08-09 | 2015-02-12 | Lenovo (Beijing) Co., Ltd. | Information Processing Method And Electronic Device Thereof, Image Calibration Method And Apparatus, And Electronic Device Thereof |
US10171261B2 (en) * | 2014-04-24 | 2019-01-01 | Vivint, Inc. | Remote talk down to panel, camera and speaker |
US10001962B2 (en) * | 2015-04-24 | 2018-06-19 | Lenovo (Beijing) Co., Ltd. | Control method, control apparatus, and electronic device |
US20160313960A1 (en) * | 2015-04-24 | 2016-10-27 | Lenovo (Beijing) Co., Ltd. | Control Method, Control Apparatus, And Electronic Device |
Also Published As
Publication number | Publication date |
---|---|
US8605048B2 (en) | 2013-12-10 |
US20120113029A1 (en) | 2012-05-10 |
CN202602826U (en) | 2012-12-12 |
US20120114047A1 (en) | 2012-05-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8605048B2 (en) | Method and apparatus for controlling multimedia contents in realtime fashion | |
US11693496B2 (en) | Display method and device | |
US20230289131A1 (en) | Method and system for dynamically generating different user environments with secondary devices with displays of various form factors | |
US12061833B2 (en) | Multi-window display method, electronic device, and system | |
WO2021057830A1 (en) | Information processing method and electronic device | |
US9310613B2 (en) | Mobile wireless display for accessing data from a host and method for controlling | |
US20220201205A1 (en) | Video stream processing method, device, terminal device, and computer-readable storage medium | |
US20110001699A1 (en) | Remote control of host application using motion and voice commands | |
WO2015130793A1 (en) | Backward-compatible apparatus and method for providing video with both standard and high dynamic range | |
WO2018161534A1 (en) | Image display method, dual screen terminal and computer readable non-volatile storage medium | |
WO2016197697A1 (en) | Gesture control method, device and system | |
CN106406924B (en) | Control method and device for starting and quitting picture of application program and mobile terminal | |
US11886894B2 (en) | Display control method and terminal device for determining a display layout manner of an application | |
WO2020007116A1 (en) | Split-screen window adjustment method and apparatus, storage medium and electronic device | |
KR20140134088A (en) | Method and apparatus for using a electronic device | |
CN103488450A (en) | Method, device and terminal equipment for projecting picture | |
US20240103717A1 (en) | Multi-Interface Display Method and Electronic Device | |
US20220294844A1 (en) | Information Transmission Method And Electronic Device | |
EP3704861B1 (en) | Networked user interface back channel discovery via wired video connection | |
CN109445589B (en) | Multimedia file playing control method and terminal equipment | |
US9787746B2 (en) | Method and apparatus for processing multimedia content on a graphic cloud | |
WO2021248988A1 (en) | Cross-terminal screen recording method, terminal device, and storage medium | |
CN111447598B (en) | Interaction method and display device | |
CN112905132B (en) | Screen projection method and device | |
WO2021012128A1 (en) | Mobile terminal-based image display apparatus and method, medium, and electronic device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: BLUESPACE CORPORATION, SAMOA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YE, ZHOU;LIU, PEI-CHUAN;HUANG, SAN-YUAN;AND OTHERS;REEL/FRAME:031525/0502 Effective date: 20110715 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |