CN110096150B - Search interaction control method, equipment and computer readable storage medium - Google Patents
Search interaction control method, equipment and computer readable storage medium Download PDFInfo
- Publication number
- CN110096150B CN110096150B CN201910351202.3A CN201910351202A CN110096150B CN 110096150 B CN110096150 B CN 110096150B CN 201910351202 A CN201910351202 A CN 201910351202A CN 110096150 B CN110096150 B CN 110096150B
- Authority
- CN
- China
- Prior art keywords
- search
- area
- display area
- display
- item
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000003993 interaction Effects 0.000 title claims abstract description 89
- 238000000034 method Methods 0.000 title claims abstract description 57
- 230000002452 interceptive effect Effects 0.000 claims abstract description 116
- 238000006073 displacement reaction Methods 0.000 claims description 25
- 210000005252 bulbus oculi Anatomy 0.000 claims description 6
- 238000004590 computer program Methods 0.000 claims description 6
- 230000006870 function Effects 0.000 description 18
- 238000004891 communication Methods 0.000 description 12
- 230000008569 process Effects 0.000 description 10
- 238000012545 processing Methods 0.000 description 5
- 238000001514 detection method Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000010295 mobile communication Methods 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000007774 longterm Effects 0.000 description 2
- 238000003672 processing method Methods 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000005452 bending Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 239000012769 display material Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 229910021389 graphene Inorganic materials 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 238000010897 surface acoustic wave method Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The application discloses a search interaction control method, equipment and a computer readable storage medium, wherein the method comprises the following steps: the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display. The humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Description
Technical Field
The present invention relates to the field of mobile communications, and in particular, to a search interaction control method, apparatus, and computer-readable storage medium.
Background
In the prior art, with the rapid development of intelligent terminal equipment, wearable equipment different from conventional smart phones, such as wearable equipment like a smart watch or a smart bracelet, appears. Because wearing formula equipment compares in traditional smart mobile phone, its particularities such as software, hardware environment, operation mode and operation environment, if the scheme of controlling of traditional smart mobile phone is transferred to wearing formula equipment, can bring inconvenience, user experience bad for user's operation.
Disclosure of Invention
In order to solve the technical defects in the prior art, the invention provides a search interaction control method, which comprises the following steps:
acquiring a wearing state of a wearing device, and determining a first display area and a second display area according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area;
acquiring a first search signal in the interactive content area, and determining a first search item corresponding to the first search signal according to the first search signal;
Dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area;
dividing a search item triggering area according to the interaction characteristics of the interaction operation area, and sending selected items in the first search items to the search item triggering area for display.
Optionally, the acquiring the wearing state of the wearing device determines a first display area and a second display area according to the wearing state, where the first display area includes an interactive operation area, and the second display area includes an interactive content area, and includes:
acquiring a wearing state of the wearing equipment, wherein the wearing state comprises a wearing position;
and determining a display area and an operation area of the wearing equipment according to the wearing position and the eyeball position of the wearing object.
Optionally, the acquiring the wearing state of the wearing device determines a first display area and a second display area according to the wearing state, where the first display area includes an interactive operation area, and the second display area includes an interactive content area, and further includes:
dividing the display area into a first display area and a second display area according to the current display requirement in the display area;
And according to the current operation requirement, the interactive operation area is arranged in the first display area, and according to the current content requirement, the interactive content area is arranged in the second display area.
Optionally, the acquiring a first search signal in the interactive content area, determining a first search item corresponding to the first search signal according to the first search signal, and including:
acquiring sliding information or space displacement information of a first gesture in the interactive content area;
and generating the first search signal according to the sliding information or the spatial displacement information.
Optionally, the acquiring a first search signal in the interactive content area, determining a first search item corresponding to the first search signal according to the first search signal, and further includes:
extracting sliding features or spatial displacement features contained in the first search signal;
determining a search word corresponding to the first search signal according to the sliding characteristic or the spatial displacement characteristic;
and obtaining the search information containing the search word to obtain the first search item.
Optionally, the dividing a search item display area in the interactive operation area and displaying the first search item in the search item display area includes:
Determining the display size of the cable entry display area according to the number of the first search entries;
dividing the interactive operation area according to the display size to obtain the search item display area.
Optionally, the dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area, further includes:
identifying an item association within the first search item;
and in the search item display area, displaying and sorting the items by combining the item relevance and the operation priority of the operation area.
Optionally, the dividing a search item triggering area according to the interaction feature of the interaction operation area, and sending the selected item in the first search item to the search item triggering area for display includes:
identifying interactive features of the interactive operation region, wherein the interactive features comprise an operation layout and a content layout;
determining a display requirement and an operation requirement of the selected item;
dividing the search item triggering area according to the display requirement, and or dividing the search item triggering area according to the operation requirement.
The invention also provides a search interaction control device, which comprises:
A memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
The invention also proposes a computer readable storage medium having stored thereon a search interaction control program which, when executed by a processor, implements the steps of the search interaction control method as described in any of the preceding claims.
The method has the beneficial effects that the first display area and the second display area are determined according to the wearing state of the wearing equipment by acquiring the wearing state of the wearing equipment, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display. The humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention.
In order to more clearly illustrate the embodiments of the invention or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, and it will be obvious to a person skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of an implementation manner of a wearable device according to an embodiment of the present invention;
fig. 2 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 3 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 4 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 5 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
FIG. 6 is a flow chart of a first embodiment of a search interaction control method of the present invention;
FIG. 7 is a flow chart of a second embodiment of the search interaction control method of the present invention;
FIG. 8 is a flow chart of a third embodiment of a search interaction control method of the present invention;
FIG. 9 is a flow chart of a fourth embodiment of a search interaction control method of the present invention;
FIG. 10 is a flowchart of a fifth embodiment of the search interaction control method of the present invention;
FIG. 11 is a flowchart of a sixth embodiment of a search interaction control method of the present invention;
FIG. 12 is a flowchart of a seventh embodiment of a search interaction control method of the present invention;
fig. 13 is a flowchart of an eighth embodiment of the search interaction control method of the present invention.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present invention. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. In this embodiment, the device may be a wristwatch, a bracelet, or a wearable device, and the device includes a screen and a connection portion. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to this embodiment.
Example 1
Fig. 6 is a flowchart of a first embodiment of the search interaction control method of the present invention. A search interaction control method, the method comprising:
s1, acquiring a wearing state of a wearing device, and determining a first display area and a second display area according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area;
s2, acquiring a first search signal in the interactive content area, and determining a first search item corresponding to the first search signal according to the first search signal;
s3, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area;
s4, dividing a search item triggering area according to the interaction characteristics of the interaction operation area, and sending selected items in the first search items to the search item triggering area for display.
In this embodiment, firstly, a wearing state of a wearing device is obtained, and a first display area and a second display area are determined according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display.
Optionally, in this embodiment, the wearing state of the wearable device is obtained, and the first display area and the second display area are determined according to the wearing state, where the first display area includes an interactive operation area, and the second display area includes an interactive content area, it may be understood that, when the user is in normal wearing or in normal use, only a part of the displayable area is generally enabled, and meanwhile, due to the gesture of wearing and the visual range of the user, an area easy to view by the user or an area easy to operate by the user is generally used as a current use area in the displayable area, and therefore, if the user needs to perform a corresponding search in the actual use process, a separate search interface needs to be started, and the separate search interface occupies all or most of the use area, so that other display content is blocked, or, a smaller search interface is provided, and is inconvenient for the user to view or operate. In this embodiment, the wearing state is determined to be a first display area and a second display area, where the first display area includes an interactive operation area, and the second display area includes an interactive content area, so as to make a reasonable plan for the display area of the wearing device, thereby facilitating subsequent search display;
Optionally, a first search signal is acquired in the interactive content area, and a first search item corresponding to the interactive content area is determined according to the first search signal, wherein the interactive content area refers to that the content in the area is biased to display properties rather than operation properties, for example, a weather display area in a desktop, a small clock area in the desktop and the like, and the acquisition of the corresponding operation signal in the area does not influence the normal function display of the area, so in the embodiment, in order to maximize the utilization rate of the display area, the display area with the property characteristics is taken as the interactive content area, and meanwhile, the first search signal is acquired in the area;
optionally, a search item display area is divided in the interactive operation area, and the first search item is displayed in the search item display area, which can be understood that the search item display area needs to receive a search operation of a user according to a search requirement of the user in addition to displaying a corresponding search item, so that the area is divided in the interactive operation area, thereby being convenient for the user to execute a corresponding search selection operation;
optionally, the search entry trigger area is divided according to the interaction characteristic of the interaction operation area, and the selected entry in the first search entry is sent to the search entry trigger area for display, where the interaction characteristic refers to an interaction type in the interaction operation area, and the characteristic is further subdivision of the interaction operation area, that is, in order to avoid the influence of the search entry trigger area on the operation of the current interaction operation area, and in order to ensure that the search entry trigger area is in the operable range of the user, in this embodiment, the interaction operation area is further divided to obtain the search entry trigger area.
The method has the advantages that the wearing state of the wearing equipment is obtained, and the first display area and the second display area are determined according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display. The humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example two
Fig. 7 is a flowchart of a second embodiment of a search interaction control method according to the present invention, based on the foregoing embodiment, the acquiring a wearing state of a wearable device, and determining a first display area and a second display area according to the wearing state, where the first display area includes an interaction operation area, and the second display area includes an interaction content area, and includes:
S11, acquiring a wearing state of the wearing equipment, wherein the wearing state comprises a wearing position;
s12, determining a display area and an operation area of the wearing equipment according to the wearing position and the eyeball position of the wearing object.
In this embodiment, first, a wearing state of the wearable device is obtained, where the wearing state includes a wearing position; then, a display area and an operation area of the wearing device are determined according to the wearing position and the eyeball position of the wearing object.
Optionally, acquiring a wearing state of the wearing device, wherein the wearing state comprises wearing positions, and determining a corresponding display area and an operation area according to different wearing positions;
optionally, the accurate areas of the display area and the operation area are further determined in combination with the current eyeball position, and it can be understood that, due to the specificity of the form of the display area of the wearable device, the display area and the operation area are not completely the same, so in this embodiment, the respective ranges of the two areas will be determined respectively.
The method and the device have the beneficial effects that the wearing state of the wearing equipment is obtained, wherein the wearing state comprises the wearing position; then, a display area and an operation area of the wearing device are determined according to the wearing position and the eyeball position of the wearing object. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example III
Fig. 8 is a flowchart of a third embodiment of a search interaction control method according to the present invention, based on the foregoing embodiment, the acquiring a wearing state of a wearable device, and determining a first display area and a second display area according to the wearing state, where the first display area includes an interaction operation area, and the second display area includes an interaction content area, and further includes:
s13, dividing the display area into a first display area and a second display area according to the current display requirement in the display area;
s14, according to the current operation requirement, the interactive operation area is arranged in the first display area, and according to the current content requirement, the interactive content area is arranged in the second display area.
In this embodiment, first, in the display area, the display area is divided into the first display area and the second display area according to a current display requirement; then, the interactive operation region is arranged in the first display region according to the current operation requirement, and the interactive content region is arranged in the second display region according to the current content requirement.
Optionally, in the display area, the display area is divided into the first display area and the second display area according to a current display requirement, where the display requirement refers to a ratio of display content to operation content in the first display area, and when the ratio is higher, the display requirement is higher, whereas the display requirement is lower, in this embodiment, the display area is divided into the first display area and the second display area, where the display requirement of the first display area is lower than the display requirement of the second display area, so that the interactive operation area is arranged in the first display area, and the interactive content area is arranged in the second display area according to the current content requirement.
The method has the advantages that the display area is divided into the first display area and the second display area according to the current display requirement in the display area; then, the interactive operation region is arranged in the first display region according to the current operation requirement, and the interactive content region is arranged in the second display region according to the current content requirement. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example IV
Fig. 9 is a flowchart of a fourth embodiment of a search interaction control method according to the present invention, based on the above embodiment, the obtaining a first search signal in the interaction content area, determining a first search entry corresponding to the first search signal according to the first search signal, including:
s21, acquiring sliding information or space displacement information of a first gesture in the interactive content area;
s22, generating the first search signal according to the sliding information or the spatial displacement information.
In this embodiment, first, sliding information or spatial displacement information of a first gesture is acquired in the interactive content area; then, the first search signal is generated according to the sliding information or the spatial displacement information.
Optionally, sliding information or spatial displacement information of a first gesture is acquired in the interactive content area, wherein the sliding information of the first gesture refers to signal information obtained by a sliding touch signal acquired in the interactive content area, and the spatial displacement information refers to aerial gesture signal information acquired in a space above the interactive content area;
optionally, generating the first search signal according to the sliding information or the spatial displacement information;
Optionally, a second search signal is generated according to the sliding information and the spatial displacement information in a continuous time period.
The method has the advantages that sliding information or space displacement information of the first gesture is obtained in the interactive content area; then, the first search signal is generated according to the sliding information or the spatial displacement information. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example five
Fig. 10 is a flowchart of a fifth embodiment of a search interaction control method according to the present invention, based on the above embodiment, the method obtains a first search signal in the interaction content area, determines a first search entry corresponding to the first search signal according to the first search signal, and further includes:
s23, extracting sliding features or spatial displacement features contained in the first search signal;
s24, determining search words corresponding to the first search signals according to the sliding characteristics or the spatial displacement characteristics;
and obtaining the search information containing the search word to obtain the first search item.
In this embodiment, first, a sliding feature or a spatial displacement feature included in the first search signal is extracted; then, determining a search word corresponding to the first search signal according to the sliding characteristic or the spatial displacement characteristic; and finally, obtaining the search information containing the search word to obtain the first search item.
Alternatively, by obtaining the pattern or letter corresponding to the sliding feature or the spatial displacement feature, the corresponding application may be retrieved according to the rule, for example, by defining an initial rule: if the 'T' can search the 'weather', 'gallery', the user inputs the 'TQ/TQ', and the 'weather' is directly searched; or a special definition: "@" refers directly to "WeChat".
The embodiment has the beneficial effects that the sliding characteristic or the spatial displacement characteristic contained by the first search signal is extracted; then, determining a search word corresponding to the first search signal according to the sliding characteristic or the spatial displacement characteristic; and finally, obtaining the search information containing the search word to obtain the first search item. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example six
Fig. 11 is a flowchart of a sixth embodiment of the search interactive control method according to the present invention, based on the above embodiment, the dividing a search item display area within the interactive operation area and displaying the first search item within the search item display area, including:
s31, determining the display size of the cable entry display area according to the number of the first search entries;
s32, dividing the interactive operation area according to the display size to obtain the search item display area.
In this embodiment, first, a display size of the cable entry display area is determined according to the number of the first search entries; and dividing the interactive operation area according to the display size to obtain the search item display area.
Optionally, determining a display size of the searched item display area according to the number of the first searched items, wherein if two searches are simultaneously performed and a first search signal and a second search signal are simultaneously input, determining the display size of the searched item display area according to the sum of the number of the first searched items and the number of the second searched items;
optionally, if two searches are performed simultaneously, and the first search signal and the second search signal are simultaneously input, determining the display size of the searched item display area according to the sum of the number of the first search item and the number of the second search item after duplication removal.
The method has the advantages that the display size of the cable entry display area is determined through the number of the first search entries; and dividing the interactive operation area according to the display size to obtain the search item display area. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example seven
Fig. 12 is a flowchart of a seventh embodiment of the search interactive control method according to the present invention, based on the above embodiment, the dividing a search item display area within the interactive operation area and displaying the first search item within the search item display area, further including:
s33, identifying item relevance in the first search item;
and S34, in the search item display area, displaying and sorting the items by combining the item relevance and the operation priority of the operation area.
In this embodiment, first, an item association within the first search item is identified; and then, in the search item display area, carrying out display sorting on the items by combining the item relevance and the operation priority of the operation area.
Optionally, determining the operation priority of the operation area according to the convenience of operation;
optionally, determining the operation priority of the operation area by combining the centering degree of the sight range and the convenience of operation;
optionally, in the search item display area, the items are displayed and ordered in combination with the item relevance and the operation priority of the operation area, wherein the display ordering scheme is dynamically adjusted according to the dynamic change amount of the current operation range of the wearable device.
The method has the advantages that the item relevance in the first search item is identified; and then, in the search item display area, carrying out display sorting on the items by combining the item relevance and the operation priority of the operation area. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example eight
FIG. 13 is a flowchart of an eighth embodiment of the search interactive control method according to the present invention, based on the above embodiment, the dividing a search entry trigger area according to the interactive feature of the interactive operation area, and sending a selected entry of the first search entries to the search entry trigger area for display, including:
S41, identifying interaction characteristics of the interaction operation area, wherein the interaction characteristics comprise operation layout and content layout;
s42, determining the display requirement and the operation requirement of the selected item;
s43, dividing the search item triggering area according to the display requirement, and or dividing the search item triggering area according to the operation requirement.
In this embodiment, first, identifying interactive features of the interactive operation region, wherein the interactive features include an operation layout and a content layout; then, determining display requirements and operational requirements of the selected item; and finally, dividing the search item triggering area according to the display requirement, and or dividing the search item triggering area according to the operation requirement.
Optionally, as described above for example, also, in order to improve the convenience of operation and maximize the utilization of the display area, in this embodiment, the display requirement and the operation requirement of the selected item are determined; then dividing the search item triggering area according to the display requirement and/or dividing the search item triggering area according to the operation requirement, wherein the search item triggering area is the operation area of the search item;
Alternatively, the display requirements and the operation requirements of the selected items may be different, for example, some search items only have an information display function, for example, images or videos obtained by searching, and some search items also have an operation function, for example, text information obtained by searching, and may also receive corresponding word processing operations such as copy and paste.
The method has the advantages that the interactive characteristics of the interactive operation area are identified, wherein the interactive characteristics comprise operation layout and content layout; then, determining display requirements and operational requirements of the selected item; and finally, dividing the search item triggering area according to the display requirement, and or dividing the search item triggering area according to the operation requirement. The more humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Example nine
Based on the above embodiment, the present invention further provides a search interaction control device, including:
A memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
In this embodiment, firstly, a wearing state of a wearing device is obtained, and a first display area and a second display area are determined according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display.
Optionally, in this embodiment, the wearing state of the wearable device is obtained, and the first display area and the second display area are determined according to the wearing state, where the first display area includes an interactive operation area, and the second display area includes an interactive content area, it may be understood that, when the user is in normal wearing or in normal use, only a part of the displayable area is generally enabled, and meanwhile, due to the gesture of wearing and the visual range of the user, an area easy to view by the user or an area easy to operate by the user is generally used as a current use area in the displayable area, and therefore, if the user needs to perform a corresponding search in the actual use process, a separate search interface needs to be started, and the separate search interface occupies all or most of the use area, so that other display content is blocked, or, a smaller search interface is provided, and is inconvenient for the user to view or operate. In this embodiment, the wearing state is determined to be a first display area and a second display area, where the first display area includes an interactive operation area, and the second display area includes an interactive content area, so as to make a reasonable plan for the display area of the wearing device, thereby facilitating subsequent search display;
Optionally, a first search signal is acquired in the interactive content area, and a first search item corresponding to the interactive content area is determined according to the first search signal, wherein the interactive content area refers to that the content in the area is biased to display properties rather than operation properties, for example, a weather display area in a desktop, a small clock area in the desktop and the like, and the acquisition of the corresponding operation signal in the area does not influence the normal function display of the area, so in the embodiment, in order to maximize the utilization rate of the display area, the display area with the property characteristics is taken as the interactive content area, and meanwhile, the first search signal is acquired in the area;
optionally, a search item display area is divided in the interactive operation area, and the first search item is displayed in the search item display area, which can be understood that the search item display area needs to receive a search operation of a user according to a search requirement of the user in addition to displaying a corresponding search item, so that the area is divided in the interactive operation area, thereby being convenient for the user to execute a corresponding search selection operation;
optionally, the search entry trigger area is divided according to the interaction characteristic of the interaction operation area, and the selected entry in the first search entry is sent to the search entry trigger area for display, where the interaction characteristic refers to an interaction type in the interaction operation area, and the characteristic is further subdivision of the interaction operation area, that is, in order to avoid the influence of the search entry trigger area on the operation of the current interaction operation area, and in order to ensure that the search entry trigger area is in the operable range of the user, in this embodiment, the interaction operation area is further divided to obtain the search entry trigger area.
The method has the advantages that the wearing state of the wearing equipment is obtained, and the first display area and the second display area are determined according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display. The humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
Examples ten
Based on the above embodiments, the present invention also proposes a computer readable storage medium having a bitmap processing program stored thereon, which when executed by a processor implements the steps of the bitmap processing method according to any one of the above.
By means of the bitmap processing method, the bitmap processing device and the computer readable storage medium, a first display area and a second display area are determined according to the wearing state of the wearing device, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area; then, a first search signal is acquired in the interactive content area, and a first search item corresponding to the first search signal is determined according to the first search signal; then, dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area; and finally, dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display. The humanized search interaction control scheme is realized, so that a user can still obtain better search experience in the display area of the wearable device, the display utilization rate of the display area is maximized, the user operation is facilitated, and the user experience is enhanced.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present invention.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.
Claims (3)
1. A search interaction control method, the method comprising:
acquiring a wearing state of a wearing device, and determining a first display area and a second display area according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area;
acquiring a first search signal in the interactive content area, and determining a first search item corresponding to the first search signal according to the first search signal;
dividing a search item display area in the interactive operation area, and displaying the first search item in the search item display area;
dividing a search item triggering area according to the interaction characteristics of the interaction operation area, and sending selected items in the first search items to the search item triggering area for display;
the method comprises the steps of obtaining a wearing state of a wearing device, determining a first display area and a second display area according to the wearing state, wherein the first display area comprises an interactive operation area, and the second display area comprises an interactive content area and comprises the following steps:
acquiring a wearing state of the wearing equipment, wherein the wearing state comprises a wearing position;
Determining a display area and an operation area of the wearing equipment according to the wearing position and the eyeball position of the wearing object;
dividing the display area into a first display area and a second display area according to the current display requirement in the display area;
according to the current operation requirement, the interactive operation area is arranged in the first display area, and according to the current content requirement, the interactive content area is arranged in the second display area;
the step of acquiring a first search signal in the interactive content area, and determining a first search item corresponding to the first search signal according to the first search signal comprises the following steps:
acquiring sliding information or space displacement information of a first gesture in the interactive content area;
generating the first search signal according to the sliding information or the spatial displacement information;
extracting sliding features or spatial displacement features contained in the first search signal;
determining a search word corresponding to the first search signal according to the sliding characteristic or the spatial displacement characteristic;
obtaining search information containing the search word to obtain the first search item;
the dividing a search item display area in the interactive operation area and displaying the first search item in the search item display area includes:
Determining the display size of the search item display area according to the number of the first search items;
dividing the interactive operation area according to the display size to obtain the search item display area;
identifying an item association within the first search item;
in the search item display area, displaying and sorting the items by combining the item relevance and the operation priority of the operation area;
the step of dividing a search item trigger area according to the interaction characteristics of the interaction operation area, and sending the selected item in the first search item to the search item trigger area for display comprises the following steps:
identifying interactive features of the interactive operation region, wherein the interactive features comprise an operation layout and a content layout;
determining a display requirement and an operation requirement of the selected item;
dividing the search item triggering area according to the display requirement, and or dividing the search item triggering area according to the operation requirement.
2. A search interaction control apparatus, the apparatus comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
The computer program implementing the steps of the method of claim 1 when executed by the processor.
3. A computer readable storage medium, wherein a search interaction control program is stored on the computer readable storage medium, which when executed by a processor, implements the steps of the search interaction control method of claim 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910351202.3A CN110096150B (en) | 2019-04-28 | 2019-04-28 | Search interaction control method, equipment and computer readable storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910351202.3A CN110096150B (en) | 2019-04-28 | 2019-04-28 | Search interaction control method, equipment and computer readable storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110096150A CN110096150A (en) | 2019-08-06 |
CN110096150B true CN110096150B (en) | 2024-01-12 |
Family
ID=67446285
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910351202.3A Active CN110096150B (en) | 2019-04-28 | 2019-04-28 | Search interaction control method, equipment and computer readable storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110096150B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113010766B (en) * | 2019-12-18 | 2023-08-18 | 腾讯科技(深圳)有限公司 | Information searching method, device and storage medium |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105468285A (en) * | 2015-12-04 | 2016-04-06 | 北京奇虎科技有限公司 | Wearable apparatus and touch event response method of wearable apparatus |
CN106484283A (en) * | 2016-09-19 | 2017-03-08 | 广东欧珀移动通信有限公司 | A kind of display control method and mobile terminal |
CN107145348A (en) * | 2017-04-28 | 2017-09-08 | 维沃移动通信有限公司 | A kind of application program split screen processing method and mobile terminal |
CN107209534A (en) * | 2015-02-12 | 2017-09-26 | 高通股份有限公司 | Content shows through control in wearable display |
CN107454946A (en) * | 2015-02-09 | 2017-12-08 | 卡西欧计算机株式会社 | Display device, method for information display and the computer-readable non-volatile memory medium for being stored with information display program |
CN108196783A (en) * | 2018-01-31 | 2018-06-22 | 努比亚技术有限公司 | Acquisition parameters adjusting method, mobile terminal and computer readable storage medium |
CN108459813A (en) * | 2018-01-23 | 2018-08-28 | 维沃移动通信有限公司 | A kind of searching method and mobile terminal |
CN108984082A (en) * | 2018-07-09 | 2018-12-11 | 维沃移动通信有限公司 | A kind of image display method and mobile terminal |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102190062B1 (en) * | 2014-06-02 | 2020-12-11 | 엘지전자 주식회사 | Wearable device and method for controlling the same |
KR20170011583A (en) * | 2015-07-23 | 2017-02-02 | 삼성전자주식회사 | Operating Method For Contents Searching Function and electronic device supporting the same |
CN106371505A (en) * | 2015-07-24 | 2017-02-01 | 中兴通讯股份有限公司 | Display realization method and apparatus, and wearable device |
-
2019
- 2019-04-28 CN CN201910351202.3A patent/CN110096150B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107454946A (en) * | 2015-02-09 | 2017-12-08 | 卡西欧计算机株式会社 | Display device, method for information display and the computer-readable non-volatile memory medium for being stored with information display program |
CN107209534A (en) * | 2015-02-12 | 2017-09-26 | 高通股份有限公司 | Content shows through control in wearable display |
CN105468285A (en) * | 2015-12-04 | 2016-04-06 | 北京奇虎科技有限公司 | Wearable apparatus and touch event response method of wearable apparatus |
CN106484283A (en) * | 2016-09-19 | 2017-03-08 | 广东欧珀移动通信有限公司 | A kind of display control method and mobile terminal |
CN107145348A (en) * | 2017-04-28 | 2017-09-08 | 维沃移动通信有限公司 | A kind of application program split screen processing method and mobile terminal |
CN108459813A (en) * | 2018-01-23 | 2018-08-28 | 维沃移动通信有限公司 | A kind of searching method and mobile terminal |
CN108196783A (en) * | 2018-01-31 | 2018-06-22 | 努比亚技术有限公司 | Acquisition parameters adjusting method, mobile terminal and computer readable storage medium |
CN108984082A (en) * | 2018-07-09 | 2018-12-11 | 维沃移动通信有限公司 | A kind of image display method and mobile terminal |
Non-Patent Citations (2)
Title |
---|
"Comprehensive deformed map generation for wristwatch-type wearable devices based on landmark-based partitioning";Keisuke Kono 等;2016 IEEE 5th Global Conference on Consumer Electronics;全文 * |
"一种可穿戴人体运动轨迹测评装置的液晶显示设计";梁文 等;《液晶与显示》;第31卷(第5期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN110096150A (en) | 2019-08-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107734170B (en) | Notification message processing method, mobile terminal and wearable device | |
CN110096195B (en) | Sports icon display method, wearable device and computer readable storage medium | |
CN110399195B (en) | Desktop icon dynamic replacement method, equipment and computer readable storage medium | |
CN110225282B (en) | Video recording control method, device and computer readable storage medium | |
CN110177209B (en) | Video parameter regulation and control method, device and computer readable storage medium | |
CN110139270B (en) | Wearable device pairing method, wearable device and computer readable storage medium | |
CN109933294B (en) | Data processing method and device, wearable device and storage medium | |
CN110177208B (en) | Video recording association control method, equipment and computer readable storage medium | |
CN110083205B (en) | Page switching method, wearable device and computer-readable storage medium | |
CN110072071B (en) | Video recording interaction control method, equipment and computer readable storage medium | |
CN110162369B (en) | Wearable device, icon arrangement method thereof and computer readable storage medium | |
CN110198411B (en) | Depth of field control method and device in video shooting process and computer readable storage medium | |
CN109933400B (en) | Display interface layout method, wearable device and computer readable storage medium | |
CN110086808B (en) | Account login control method and device and computer readable storage medium | |
CN110069200B (en) | Wearable device input control method, wearable device and storage medium | |
CN109918014B (en) | Page display method, wearable device and computer-readable storage medium | |
CN110113529B (en) | Shooting parameter regulation and control method and device and computer readable storage medium | |
CN110096150B (en) | Search interaction control method, equipment and computer readable storage medium | |
CN110083427B (en) | Application program interaction control method, device and computer readable storage medium | |
CN110109605B (en) | Image sharing method and device and computer readable storage medium | |
CN110650289B (en) | Shooting depth of field control method, equipment and computer readable storage medium | |
CN110049253B (en) | Focusing control method and device and computer readable storage medium | |
CN110174981B (en) | Icon interaction control method and device and computer readable storage medium | |
CN110187950B (en) | Method for adjusting picture display position, wearable device and storage medium | |
CN110620876B (en) | Image preview interaction method, device and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |