WO2010095255A1 - 情報処理装置、表示制御方法および表示制御プログラム - Google Patents
情報処理装置、表示制御方法および表示制御プログラム Download PDFInfo
- Publication number
- WO2010095255A1 WO2010095255A1 PCT/JP2009/053145 JP2009053145W WO2010095255A1 WO 2010095255 A1 WO2010095255 A1 WO 2010095255A1 JP 2009053145 W JP2009053145 W JP 2009053145W WO 2010095255 A1 WO2010095255 A1 WO 2010095255A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- window
- display
- unit
- display control
- touch operation
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
Definitions
- This case relates to an information processing apparatus, a display control method, and a display control program.
- a window system In an information processing apparatus equipped with GUI (Graphical User Interface), a window system is widely adopted.
- the window system one or more windows are displayed on the operation screen.
- the user can perform various operations such as character input and graphic drawing within the displayed window area.
- various window operations such as moving, resizing, maximizing, minimizing (iconizing), closing, and scrolling can be performed on the displayed window.
- the window operation can be performed using a pointing device such as a mouse.
- a pointing device such as a mouse.
- the user moves the window by dragging the title bar provided on the upper side of the window with the mouse.
- the window is maximized / minimized / closed by clicking the button provided at the end of the title bar with the mouse.
- an input device dedicated to window operation can be provided in the information processing apparatus (see, for example, Patent Document 1).
- a device for example, a touch panel
- the user can directly specify the position on the operation screen with a pen or a finger.
- a handwriting input method of opening a memo input area when a predetermined memo command is entered with a pen and then closing the memo input area when a predetermined box is touched (For example, refer to Patent Document 2).
- This case has been made in view of these points, and an object thereof is to provide an information processing apparatus, a display control method, and a display control program that improve the operability of window operations.
- an information processing apparatus having a display unit, a detection unit, and a control unit.
- the display unit displays one or more windows.
- the detection unit detects a touch operation on the display unit.
- the control unit specifies a graphic represented by the locus of the touch operation detected by the detection unit, and performs display control corresponding to the graphic for the window displayed on the display unit.
- a display control method for an information processing apparatus provided with a display unit is provided.
- this display control method one or more windows are displayed on the display unit.
- a touch operation on the display unit is detected.
- the figure represented by the locus of the touch operation is specified, and display control corresponding to the figure is performed for the window displayed on the display unit.
- a display control program to be executed by a computer connected to a display device is provided.
- a computer that executes the display control program displays one or more windows on the display device.
- a touch operation on the display device is detected.
- the figure represented by the locus of the touch operation is specified, and display control corresponding to the figure is performed for the window displayed on the display device.
- FIG. 1 shows the example of the information processing apparatus which concerns on this Embodiment. It is a figure which shows the external appearance of a portable information terminal device. It is a block diagram which shows the hardware of a portable information terminal device. It is a block diagram which shows the function of a portable information terminal device. It is a figure which shows the example of a data structure of a window management table. It is a figure which shows the example of a data structure of a figure definition table. It is a flowchart which shows window display control. It is a figure which shows the example of a 1st figure determination. It is a figure which shows the example of 1st window operation. It is a figure which shows the example of a 2nd figure judgment.
- FIG. 1 is a diagram illustrating an example of an information processing apparatus according to the present embodiment.
- the information processing apparatus 1 illustrated in FIG. 1 can perform window display.
- various apparatuses such as a laptop computer, a UMPC (Ultra Mobile Personal Computer), a PDA (Personal Data Assistance), and a mobile phone can be considered.
- the information processing apparatus 1 includes a display unit 1a, a detection unit 1b, and a control unit 1c.
- the display unit 1a displays an operation screen including one or more windows.
- a liquid crystal display or an organic EL (Electro-Luminescence) display can be used as the display unit 1a.
- the user can perform various operations by operating the window displayed on the display unit 1a.
- the detection unit 1b detects a touch operation on the operation screen on the display unit 1a.
- a touch panel can be used as the detection unit 1b.
- the detection unit 1b notifies the control unit 1c of the touched position on the operation screen.
- the position notification is performed regularly (for example, every 50 ms) or irregularly.
- the detection unit 1b may allow the touch operation to be performed with the user's finger or allow the touch operation to be performed using a stylus pen.
- the control part 1c specifies the figure corresponding to the locus
- the control unit 1c can also distinguish and specify a plurality of types of figures. In that case, it is conceivable to hold a table in which graphics and shape features are associated with each other.
- the figure to be detected may be fixed or may be set by the user.
- control part 1c performs display control according to the specified figure with respect to the window currently displayed on the display part 1a.
- Display control includes window maximization, minimization, closing, scrolling, and the like. For example, when “x” is specified as a graphic, the window is closed.
- the target of display control may be all the windows displayed on the display unit 1a or a part of the windows. In the latter case, for example, it is conceivable to target only the window (active window) currently being operated.
- one or more windows are displayed on the display unit 1a.
- the touch operation on the display unit 1a is detected by the detection unit 1b.
- the control unit 1c specifies a graphic represented by the detected locus of the touch operation, and performs display control corresponding to the graphic for the window displayed on the display unit 1a.
- window operations improves the operability of window operations by the user.
- the user can perform window operations such as maximization, minimization, close, and scroll without accurately touching a title bar or button that is displayed relatively small in the operation screen. This is particularly effective in a portable information terminal device in which the area of the operation screen cannot be made sufficiently large.
- the display control method described above can be applied to a portable information terminal device including a detection unit other than the touch panel, and can also be applied to other types of information processing devices.
- FIG. 2 is a diagram illustrating an appearance of the portable information terminal device.
- the portable information terminal device 100 according to the present embodiment is a small information terminal device called UMPC and is easy to carry.
- the portable information terminal device 100 includes housings 101 and 102.
- the housing 101 can be opened and closed with respect to the housing 102.
- FIG. 2 shows a state where the housing 101 is opened.
- the housing 101 has a display 103 (not shown in FIG. 2) and a touch panel 104.
- the housing 102 has a keyboard 105.
- the display 103 is arranged below the touch panel 104 and displays an operation screen.
- the operation screen displayed on the display 103 is visible from the front surface of the housing 101 through the touch panel 104.
- the user can perform a touch operation on the touch panel 104 while viewing the operation screen displayed on the display 103.
- the display 103 may be arranged on the upper layer side of the touch panel 104.
- the touch panel 104 detects a touch operation.
- the user can touch the touch panel 104 with the stylus pen 106 or a finger.
- the touch panel 104 detects the position where the tip of the stylus pen 106 and the finger are in contact.
- a method for detecting a touch operation for example, a method using a resistance film, a method of forming an electric field on the surface, a method using infrared rays, and the like are conceivable.
- the position may be detected without contact.
- the keyboard 105 is an input device having various keys such as character keys, symbol keys, numeric keys, and function keys. When any key is pressed by the user, the keyboard 105 outputs an electrical signal corresponding to the pressed key. The user can input a sentence and call a function by a key operation.
- another input / output device may be provided in the portable information terminal device 100.
- a speaker or a microphone may be provided in the portable information terminal device 100.
- another input / output device can be connected to the portable information terminal device 100.
- FIG. 3 is a block diagram showing hardware of the portable information terminal device.
- the portable information terminal device 100 includes a CPU (Central Processing Unit) 111, a RAM (Random Access Memory) 112, an HDD (Hard Disk Drive) 113, a communication interface 114, a display processing unit 115, and an input processing unit 116. These components are provided in the housing 101 or the housing 102.
- CPU Central Processing Unit
- RAM Random Access Memory
- HDD Hard Disk Drive
- CPU 111 controls the entire operation of portable information terminal device 100. At that time, the program stored in the HDD 113 is read out, expanded in the RAM 112 and executed. Specifically, the CPU 111 starts executing an OS (Operating System) program or a window display control program when the portable information terminal device 100 is activated. Thereafter, the execution of the application program is started and ended as appropriate. Note that a plurality of CPUs may be mounted on the portable information terminal device 100 to perform distributed processing.
- OS Operating System
- the RAM 112 temporarily stores at least a part of programs executed by the CPU 111 and various data used for processing.
- the CPU 111 appropriately writes and reads programs and data to and from the RAM 112. Note that another type of memory may be used instead of the RAM 112.
- the HDD 113 stores programs executed by the CPU 111 and various data used for processing.
- Programs stored in the HDD 113 include an OS program, a window display control program, and an application program.
- the stored data includes window display control setting data.
- a nonvolatile memory device such as an SSD (Solid State Drive) may be used.
- the communication interface 114 performs data communication with other information processing apparatuses via a network. That is, the communication interface 114 encodes and modulates transmission data under the control of the CPU 111 and outputs it to the network. It also demodulates and decodes data received from the network. Note that the network accessed by the communication interface 114 may be wired or wireless.
- the display processing unit 115 is connected to the display 103.
- the display processing unit 115 performs image processing on the operation screen displayed on the display 103 under the control of the CPU 111 and continuously outputs image signals to the display 103.
- the operation screen may include one or a plurality of windows.
- the input processing unit 116 is connected to the touch panel 104 and the keyboard 105.
- the input processing unit 116 detects an input signal indicating touch start (DOWN event), touch end (UP event), coordinates of the current touch position, and the like output from the touch panel 104. Further, an input signal indicating the pressed key output from the keyboard 105 is detected. The detected input signal is output to the CPU 111.
- the display 103 corresponds to the display unit 1a in FIG.
- the touch panel 104 corresponds to the detection unit 1b in FIG.
- the CPU 111 and the RAM 112 correspond to the control unit 1c in FIG.
- FIG. 4 is a block diagram showing functions of the portable information terminal device.
- the portable information terminal device 100 includes a window information storage unit 120, a graphic information storage unit 130, a touch detection unit 140, a stylus monitoring unit 150, and a window operation unit 160.
- the functions of these modules are realized by the CPU 111 developing and executing a program on the RAM 112. However, some or all of them may be realized as hardware components.
- the window information storage unit 120 stores window management information indicating the state of each opened window.
- the state indicated by the window management information includes a window display state and a use state.
- the display status includes, for example, a maximized state (displayed in the maximum display area that can be secured), a minimized state (displayed as an icon), and a normal display state (state other than maximized and minimized) There is.
- the usage state includes, for example, active (a state that is a user's operation target) and inactive (a state that is not a user's operation target).
- the graphic information storage unit 130 stores graphic information defining window operation commands.
- a command is defined as a figure drawn by a touch operation.
- the graphic information includes, for example, information indicating the type of window operation and feature information for specifying a graphic from the locus of the touch operation. Examples of the window operation include window maximization / minimization / close / scroll. Note that the user may be able to update the graphic information.
- the touch detection unit 140 acquires an input signal indicating a touch operation output from the touch panel 104 via the input processing unit 116.
- the touch detection unit 140 detects a drag operation based on the UP event and the DOWN event. While the drag is being performed, the stylus monitoring unit 150 is notified of coordinates indicating the current touch position continuously (for example, in a cycle of 50 ms). However, the change amount from the previously notified coordinate may be notified.
- the touch detection unit 140 may set the area for recognizing the touch operation for the window operation as the entire area on the touch panel 104 or may be limited to a part of the area. As the latter, for example, a method of limiting to the area where the foreground window or the active window is displayed can be considered. That is, a method of determining a predetermined touch operation on an area where a predetermined window is displayed as a touch operation for the window operation can be considered. In that case, a part of the trajectory may straddle the window.
- the stylus monitoring unit 150 continuously acquires information on the current touch position from the touch detection unit 140, and temporarily stores the locus of the touch operation. For example, the coordinates of the drag operation up to a predetermined time or the drag operation up to a predetermined number of times are held. Then, the stylus monitoring unit 150 refers to the graphic information stored in the graphic information storage unit 130 and identifies the command indicated by the touch operation locus. For example, the touch operation trajectory is compared with the feature information of each command, and the most suitable command is selected. Thereafter, the stylus monitoring unit 150 instructs the window operation unit 160 to execute the window operation indicated by the identified command.
- the window operation unit 160 controls the display state of the window. That is, when there is a window operation instruction from the stylus monitoring unit 150, the window operation unit 160 refers to the window information stored in the window information storage unit 120 and identifies the currently active window. Then, the window operation for the active window is executed, and the operation screen after the window operation is displayed on the display 103 via the display processing unit 115. Further, the window operation unit 160 updates the window information.
- the target of the window operation may be other windows such as the frontmost window and the rearmost window instead of the active window.
- the operation target may be selected according to the command, or the operation target may be selected according to the position where the touch operation is performed. In that case, the stylus monitoring unit 150 specifies an operation target window to the window operation unit 160.
- the functions of the above modules can also be realized by linking multiple types of programs.
- the functions of the touch detection unit 140 may be realized by a driver program
- the functions of the graphic information storage unit 130 and the stylus monitoring unit 150 may be realized by an application program
- the functions of the window information storage unit 120 and the window operation unit 160 may be realized by an OS program.
- FIG. 5 is a diagram showing an example of the data structure of the window management table.
- the window management table 121 is stored in the window information storage unit 120 and is referred to and updated by the window operation unit 160.
- the window management table 121 includes items indicating a window ID (IDentifier), a display state, and active. Information arranged in the horizontal direction of each item is associated with each other.
- a character string for identifying each window is set.
- the window ID is automatically given, for example, when the window is opened.
- a character string indicating the current display state of the window is set. For example, “maximize”, “minimize” or “normal” is set.
- a flag indicating whether or not the window is currently active is set. For example, “Yes” is set when active, and “No” is set when inactive.
- the number of active windows may be limited to one, or multiple windows may be allowed to exist at the same time.
- information that the window ID is “window # A”, the display state is “normal”, and the active flag is “Yes” is registered in the window management table 121. This indicates that a certain window is currently displayed in a normal display state on the operation screen and is active.
- information that the window ID is “window # C”, the display state is “minimized”, and the active flag is “No” is registered. This indicates that another window is currently displayed in an iconized state on the operation screen and is not active.
- the window operation unit 160 registers information on the opened window in the window management table 121 when the window is opened.
- a newly opened window is normally active with a display state of “normal”. Thereafter, the display state and active / inactive change according to the user's input operation. Further, the window operation unit 160 deletes information on the closed window from the window management table 121 when the window is closed. However, it is possible to set the display state to “closed” without deleting the window information.
- FIG. 6 is a diagram showing an example of the data structure of the figure definition table.
- the graphic definition table 131 is stored in the graphic information storage unit 130 and is referred to by the stylus monitoring unit 150.
- items indicating figures, operations, and features are provided in the figure definition table 131.
- Information arranged in the horizontal direction of each item is associated with each other.
- Symbols for identifying each command are set in the items indicating graphics.
- the identification symbol for example, a simple representation of the shape of a figure is used.
- a character string indicating the type of window operation is set. For example, “maximize”, “minimize”, “close” or “scroll” is set.
- Feature information for recognizing a graphic drawn by a touch operation is set in the item indicating the feature.
- a plurality of feature items may be defined for one figure. Note that the feature of the graphic may be expressed as a mathematical expression that defines a relationship between a plurality of coordinates so that the stylus monitoring unit 150 can easily interpret it.
- information that the figure is “x” and the operation is “close” is set. This indicates that when a figure that can be recognized as “ ⁇ ” is drawn by a touch operation, the active window is closed.
- information that the graphic is “ ⁇ ” and the operation is “maximization” is set. This indicates that the active window is maximized when a figure that can be recognized as “ ⁇ ” is drawn by a touch operation.
- the stylus monitoring unit 150 compares the trajectory of the drag operation from a predetermined time before or a predetermined number of times to the present with the characteristic information of each graphic, and determines whether there is a registered graphic that matches the characteristic. When a plurality of feature items are set for one figure, for example, it is determined that the figure matches the figure when all the feature items are matched. Note that the user can change the correspondence between the figure and the type of window operation by updating the figure definition table 131. It is also possible to register additional new figures.
- FIG. 7 is a flowchart showing window display control. The process shown in FIG. 7 is repeatedly executed in the portable information terminal device 100. Hereinafter, the processing content of the window display control will be described along with step numbers.
- Step S1 The touch detection unit 140 detects a drag operation on the touch panel 104 based on an input signal acquired from the touch panel 104. Then, the touch detection unit 140 notifies the stylus monitoring unit 150 of a change in the touch position during the drag operation.
- the stylus monitoring unit 150 refers to the graphic definition table 131 stored in the graphic information storage unit 130 and collates the locus of the drag operation with the graphic registered as a command. That is, it is determined whether the trajectory of the drag operation matches the feature of each figure.
- Step S3 The stylus monitoring unit 150 determines whether the locus of the drag operation matches any registered figure in step S2. If it matches any registered figure, the process proceeds to step S4. If it does not match any registered figure, the process is terminated.
- Step S4 The stylus monitoring unit 150 determines the contents of the window operation corresponding to the graphic specified in step S2. Then, the stylus monitoring unit 150 instructs the window operation unit 160 to perform the window operation on the active window.
- the window operation unit 160 refers to the window management table 121 stored in the window information storage unit 120 to identify an active window. Then, the window operation unit 160 executes the operation designated by the stylus monitoring unit 150 on the specified window, and updates the operation screen displayed on the display 103. At the same time, the display state information registered in the window management table 121 is also updated.
- the portable information terminal device 100 detects a touch operation on the touch panel 104 and specifies a figure indicated by the locus of the touch operation. Then, the operation associated with the registered figure is executed on the active window to change the display state, and the operation screen displayed on the display 103 is updated.
- FIG. 8 is a diagram illustrating an example of first graphic determination.
- FIG. 8 shows a method of detecting “x” from the locus of the touch operation.
- the following feature items A1 and A2 are registered in the figure definition table 131.
- [A1] Two independent lines intersect at only one point.
- [A2] The intersection of each of the two lines must be within the range of “25% of the total length of the line from the center of the line”.
- Example 1 The trajectory of the touch operation in Example 1 satisfies both feature items A1 and A2. Therefore, in Example 1, it is determined that the figure “x” is drawn.
- the trajectory of the touch operation in Example 2 is such that two independent lines intersect at two points and do not satisfy the feature item A1. Further, the trajectory of the touch operation in Example 3 does not satisfy the feature item A2 because the intersection does not exist within the range of “25% of the total length of the line from the center of the line”. Therefore, in Examples 2 and 3, it is not determined that the figure “x” is drawn.
- FIG. 9 is a diagram showing an example of the first window operation.
- Operation screens 211 and 212 shown in FIG. 9 are examples of operation screens displayed on the display 103.
- windows A, B, and C are opened. Of these, windows A and B are displayed in a normal state, and window C is displayed in an iconized state on the task bar.
- Window A is displayed in the foreground and is active.
- FIG. 10 is a diagram illustrating an example of second graphic determination.
- FIG. 10 shows a method of detecting “ ⁇ ” from the trajectory of the touch operation.
- the following feature items B 1 and B 2 are registered in the figure definition table 131.
- the length between the start point and the intersection point is 10% or less of the total length of the line.
- Example 1 The trajectory of the touch operation in Example 1 satisfies both feature items B1 and B2. Therefore, in Example 1, it is determined that the figure “ ⁇ ” is drawn.
- the trajectory of the touch operation in Example 2 does not satisfy the feature item B1 because there is no intersection between the start point and the end point. Further, in the touch operation locus of Example 3, the length between the start point and the intersection is greater than 10% of the total length of the line, and does not satisfy the feature item B2. Therefore, in Examples 2 and 3, it is not determined that the figure “ ⁇ ” is drawn.
- FIG. 11 is a diagram illustrating an example of the second window operation.
- Operation screens 221 and 222 shown in FIG. 11 are examples of operation screens displayed on the display 103.
- windows A, B, and C are opened as in the operation screen 211 shown in FIG. Of these, windows A and B are displayed in a normal state, and window C is displayed in an iconized state on the task bar.
- Window A is displayed in the foreground and is active.
- window A is displayed using the maximum display area of display 103. After window A is maximized, for example, window A continues to be active.
- FIG. 12 is a diagram illustrating an example of third graphic determination.
- FIG. 12 shows a method for detecting “ ⁇ ” from the trajectory of the touch operation.
- the following feature items C1 and C2 are registered in the figure definition table 131.
- the length between the center point M of one line and the highest vertex P (the point with the smallest y coordinate) is 10% or less of the total length of the line.
- the direction from the top to the bottom of the display area is the y-axis.
- the difference between the y coordinate of the start point and the y coordinate of the end point is 10% or less of the difference between the y coordinate of the start point and the y coordinate of the top vertex P. Note that either the start point y coordinate or the end point y coordinate may be small.
- Example 2 The trajectory of the touch operation in Example 1 satisfies both the feature items C1 and C2. Therefore, in Example 1, it is determined that the figure “ ⁇ ” is drawn.
- the length between the center point M and the highest vertex P is larger than 10% of the total length of the line, and does not satisfy the feature item C1.
- the difference between the y-coordinate of the start point and the y-coordinate of the end point is larger than 10% of the difference between the y-coordinate of the start point and the y-coordinate of the highest vertex P, and satisfies the feature item C2. Absent. Therefore, in Examples 2 and 3, it is not determined that the figure “ ⁇ ” is drawn.
- FIG. 13 is a diagram illustrating an example of the third window operation.
- Operation screens 231 and 232 shown in FIG. 13 are examples of operation screens displayed on the display 103.
- windows A, B, and C are opened as in the operation screen 211 shown in FIG. Of these, windows A and B are displayed in a normal state, and window C is displayed in an iconized state on the task bar.
- Window A is displayed in the foreground and is active.
- the operation screen of the display 103 transitions from the operation screen 231 to the operation screen 232. That is, the window A is displayed only as an icon on the task bar, and the window B displayed on the lower layer of the window A is displayed in the foreground in an area other than the task bar. If the icon is subsequently touched, the window A returns to the normal display.
- FIG. 14 is a diagram illustrating an example of the fourth graphic determination.
- FIG. 14 shows a method of detecting “Le” from the trajectory of the touch operation.
- the following feature items D1 and D2 are registered in the figure definition table 131.
- D1 The difference between the x coordinate of the lowest point Q (the point having the largest y coordinate) of one line and the x coordinate of the starting point is 10% or less of the total length of the line. However, the direction from the left to the right of the display area is the x-axis.
- the y coordinate of the end point is between the y coordinate of the start point and the y coordinate of the lowest point Q, and is within the range of “25% of the y coordinate difference from the center of both points”.
- Example 1 The trajectory of the touch operation in Example 1 satisfies both the feature items D1 and D2. Therefore, in Example 1, it is determined that the figure “R” is drawn.
- the difference between the x coordinate of the lowest point Q and the x coordinate of the start point is larger than 10% of the total length of the line, and does not satisfy the feature item D1.
- the locus of the touch operation in Example 3 does not satisfy the feature item D2 because the y coordinate of the end point is too small (too much to extend upward). Therefore, in Examples 2 and 3, it is not determined that the figure “R” is drawn.
- FIG. 15 is a diagram showing an example of the fourth window operation.
- Operation screens 241 and 242 shown in FIG. 15 are examples of operation screens displayed on the display 103.
- windows A, B, and C are opened as in the operation screen 211 shown in FIG. 9.
- windows A and B are displayed in a normal state
- window C is displayed in an iconized state on the task bar.
- Window A is displayed in the foreground and is active.
- Examples of four types of window operations have been shown, but the operation content is not limited to these, and various window operations can be set.
- graphics that can be used as commands are not limited to these, and various graphics can be registered.
- the correspondence between the figure and the type of window operation does not have to be 1: 1, and a plurality of figures may be associated with one type of window operation.
- the locus drawn by the stylus pen 106 or the finger may be invisible or may be displayed on the display 103 for a predetermined time.
- window A may be closed, maximized, or minimized. Further, when the scroll bar of window A is dragged, screen scrolling may be performed.
- the method using graphic commands and the method using buttons and scroll bars may be used in combination, or only one of them may be used. The user may be allowed to set which method to turn on. In addition, it is possible to suppress erroneous operations by using only figures of a predetermined size or more as detection targets.
- a figure drawn by a touch operation is specified, and an operation corresponding to the figure is executed on a currently opened window (for example, an active window).
- a currently opened window for example, an active window.
- the display control can be realized by causing a computer to execute a predetermined program.
- a program describing the processing contents to be realized is provided.
- the program describing the processing contents can be recorded on a computer-readable recording medium.
- the computer-readable recording medium include a magnetic recording device, an optical disk, a magneto-optical recording medium, and a semiconductor memory.
- the magnetic recording device include a hard disk device, a flexible disk (FD), and a magnetic tape (MT).
- Optical discs include DVD (Digital Versatile Disc), DVD-RAM, CD-ROM (Compact Disc-Read Only Memory), CD-R (Recordable), CD-RW (ReWritable), and the like.
- Magneto-optical recording media include MO (Magneto-Optical disk).
- a portable recording medium such as a DVD or CD-ROM in which the program is recorded is sold. It is also possible to store the program in a storage device of a server computer and transfer the program from the server computer to another computer via a network.
- the computer that executes the program stores, for example, the program recorded on the portable recording medium or the program transferred from the server computer in its own storage device. Then, the computer reads the program from its own storage device and executes processing according to the program. The computer can also read the program directly from the portable recording medium and execute processing according to the program. In addition, each time the program is transferred from the server computer, the computer can sequentially execute processing according to the received program.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
- Telephone Function (AREA)
Abstract
Description
本発明の上記および他の目的、特徴および利点は本発明の例として好ましい実施の形態を表す添付の図面と関連した以下の説明により明らかになるであろう。
図1は、本実施の形態に係る情報処理装置の例を示す図である。図1に示す情報処理装置1は、ウィンドウ表示を行うことができる。情報処理装置1としては、ラップトップ型コンピュータ、UMPC(Ultra Mobile Personal Computer)、PDA(Personal Data Assistance)、携帯電話機などの種々の装置が考えられる。情報処理装置1は、表示部1a、検出部1bおよび制御部1cを有する。
携帯情報端末装置100は、筐体101,102を有する。筐体101は、筐体102に対して開閉することができる。図2は、筐体101が開いた状態を示している。筐体101は、ディスプレイ103(図2では図示せず)およびタッチパネル104を有する。筐体102は、キーボード105を有する。
[A1]2本の独立した線が1点のみで交差すること。
[A2]2本の線それぞれについて、交点が「線の中心から線の全長の25%」の範囲内に存在すること。
[B1]1本の線が始点(DOWNイベントを検出した点)から終点(UPイベントを検出した点)までの間に1回だけ交差すること。
[B2]始点と交点との間の長さが線の全長の10%以下であること。
[C1]1本の線の中心点Mと最頂点P(y座標が最も小さい点)との間の長さが、線の全長の10%以下であること。ただし、表示領域の上から下の方向をy軸とする。
[C2]始点のy座標と終点のy座標との差が、始点のy座標と最頂点Pのy座標との差の10%以下であること。なお、始点のy座標と終点のy座標の何れが小さくてもよい。
[D1]1本の線の最下点Q(y座標が最も大きい点)のx座標と始点のx座標との差が線の全長の10%以下であること。ただし、表示領域の左から右の方向をx軸とする。
[D2]終点のy座標が、始点のy座標と最下点Qのy座標との間であって、「両点の中心からy座標差の25%」の範囲内に存在すること。
1a 表示部
1b 検出部
1c 制御部
Claims (7)
- 1またはそれ以上のウィンドウを表示する表示部と、
前記表示部に対するタッチ操作を検出する検出部と、
前記検出部で検出した前記タッチ操作の軌跡が表す図形を特定し、前記表示部に表示されているウィンドウについて前記図形に応じた表示制御を行う制御部と、
を有することを特徴とする情報処理装置。 - 前記制御部は、前記表示部に表示されているウィンドウのうちアクティブなウィンドウを制御対象とすることを特徴とする請求の範囲第1項記載の情報処理装置。
- 前記図形に応じた表示制御には、ウィンドウの最大化、最小化、クローズおよびスクロールの少なくとも1つが含まれることを特徴とする請求の範囲第1項記載の情報処理装置。
- 前記タッチ操作の軌跡の特徴と前記図形とを対応付けた情報を記憶する記憶部を有し、
前記制御部は、前記記憶部に記憶された情報を参照して前記図形を特定する、
ことを特徴とする請求の範囲第1項記載の情報処理装置。 - 前記図形と表示制御の内容とを対応付けた情報を記憶する記憶部を有し、
前記制御部は、前記記憶部に記憶された情報を参照して実行する表示制御を特定する、
ことを特徴とする請求の範囲第1項記載の情報処理装置。 - 表示部を備えた情報処理装置の表示制御方法であって、
前記表示部に1またはそれ以上のウィンドウを表示し、
前記表示部に対するタッチ操作を検出し、
前記タッチ操作の軌跡が表す図形を特定し、前記表示部に表示されているウィンドウについて前記図形に応じた表示制御を行う、
ことを特徴とする表示制御方法。 - 表示装置と接続されたコンピュータに、
前記表示装置に1またはそれ以上のウィンドウを表示し、
前記表示装置に対するタッチ操作を検出し、
前記タッチ操作の軌跡が表す図形を特定し、前記表示装置に表示されているウィンドウについて前記図形に応じた表示制御を行う、
処理を実行させることを特徴とする表示制御プログラム。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020117018226A KR20110104088A (ko) | 2009-02-23 | 2009-02-23 | 정보 처리 장치, 표시 제어 방법 및 표시 제어 프로그램을 기록한 컴퓨터 판독 가능한 매체 |
JP2011500424A JPWO2010095255A1 (ja) | 2009-02-23 | 2009-02-23 | 情報処理装置、表示制御方法および表示制御プログラム |
CN2009801566606A CN102317895A (zh) | 2009-02-23 | 2009-02-23 | 信息处理装置、显示控制方法以及显示控制程序 |
EP09840361A EP2400378A1 (en) | 2009-02-23 | 2009-02-23 | Information processing device, display control method and display control program |
PCT/JP2009/053145 WO2010095255A1 (ja) | 2009-02-23 | 2009-02-23 | 情報処理装置、表示制御方法および表示制御プログラム |
US13/204,447 US20110289449A1 (en) | 2009-02-23 | 2011-08-05 | Information processing apparatus, display control method, and display control program |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2009/053145 WO2010095255A1 (ja) | 2009-02-23 | 2009-02-23 | 情報処理装置、表示制御方法および表示制御プログラム |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/204,447 Continuation US20110289449A1 (en) | 2009-02-23 | 2011-08-05 | Information processing apparatus, display control method, and display control program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010095255A1 true WO2010095255A1 (ja) | 2010-08-26 |
Family
ID=42633555
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/053145 WO2010095255A1 (ja) | 2009-02-23 | 2009-02-23 | 情報処理装置、表示制御方法および表示制御プログラム |
Country Status (6)
Country | Link |
---|---|
US (1) | US20110289449A1 (ja) |
EP (1) | EP2400378A1 (ja) |
JP (1) | JPWO2010095255A1 (ja) |
KR (1) | KR20110104088A (ja) |
CN (1) | CN102317895A (ja) |
WO (1) | WO2010095255A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013171195A (ja) * | 2012-02-21 | 2013-09-02 | Canon Inc | 表示装置、その制御方法及びプログラム |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120102400A1 (en) * | 2010-10-22 | 2012-04-26 | Microsoft Corporation | Touch Gesture Notification Dismissal Techniques |
WO2014042481A1 (ko) * | 2012-09-14 | 2014-03-20 | Go Jae Yong | 자기장 센서를 이용하여 사용자 입력을 판단하는 전기 장치 |
TWI496069B (zh) * | 2013-06-28 | 2015-08-11 | Insyde Software Corp | Method of Judging Electronic Device and Multi - window Touch Command |
CN104571798A (zh) * | 2013-10-28 | 2015-04-29 | 联想(北京)有限公司 | 一种信息处理方法及电子设备 |
CN104049900A (zh) * | 2014-06-30 | 2014-09-17 | 北京安兔兔科技有限公司 | 悬浮窗口关闭方法及装置 |
CN105739848B (zh) * | 2014-12-12 | 2019-06-25 | 联想(北京)有限公司 | 一种信息处理方法及电子设备 |
JP6292213B2 (ja) * | 2015-11-25 | 2018-03-14 | コニカミノルタ株式会社 | 画像形成装置および制御プログラム |
US10592228B2 (en) * | 2016-02-12 | 2020-03-17 | Microsoft Technology Licensing, Llc | Stylus firmware updates |
KR102391965B1 (ko) * | 2017-02-23 | 2022-04-28 | 삼성전자주식회사 | 가상현실 서비스를 위한 화면 제어 방법 및 장치 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07141092A (ja) | 1993-11-16 | 1995-06-02 | Casio Comput Co Ltd | 手書き入力装置 |
JP2002259001A (ja) | 2001-03-05 | 2002-09-13 | Shingo Tsuda | ウィンドウ操作方法及び装置 |
JP2003140791A (ja) * | 2001-10-31 | 2003-05-16 | Toshiba Corp | 情報処理装置及び情報処理装置の制御方法 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0481056A (ja) * | 1990-07-20 | 1992-03-13 | Nec Corp | ネットワーク環境定義誤り検出方式 |
US7411575B2 (en) * | 2003-09-16 | 2008-08-12 | Smart Technologies Ulc | Gesture recognition method and touch system incorporating the same |
JP5091180B2 (ja) * | 2009-03-27 | 2012-12-05 | ソニーモバイルコミュニケーションズ, エービー | 携帯端末装置 |
TWI391850B (zh) * | 2009-03-31 | 2013-04-01 | Au Optronics Corp | 輸入圖形判定方法及電腦可讀儲存媒體 |
JP5208267B2 (ja) * | 2009-04-28 | 2013-06-12 | 三菱電機株式会社 | 入力装置 |
KR101648747B1 (ko) * | 2009-10-07 | 2016-08-17 | 삼성전자 주식회사 | 복수의 터치 센서를 이용한 ui 제공방법 및 이를 이용한 휴대 단말기 |
EP2544078B1 (en) * | 2010-03-01 | 2021-02-17 | Panasonic Intellectual Property Management Co., Ltd. | Display device with adaptive capacitive touch panel |
-
2009
- 2009-02-23 JP JP2011500424A patent/JPWO2010095255A1/ja active Pending
- 2009-02-23 EP EP09840361A patent/EP2400378A1/en not_active Withdrawn
- 2009-02-23 KR KR1020117018226A patent/KR20110104088A/ko not_active Application Discontinuation
- 2009-02-23 WO PCT/JP2009/053145 patent/WO2010095255A1/ja active Application Filing
- 2009-02-23 CN CN2009801566606A patent/CN102317895A/zh active Pending
-
2011
- 2011-08-05 US US13/204,447 patent/US20110289449A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07141092A (ja) | 1993-11-16 | 1995-06-02 | Casio Comput Co Ltd | 手書き入力装置 |
JP2002259001A (ja) | 2001-03-05 | 2002-09-13 | Shingo Tsuda | ウィンドウ操作方法及び装置 |
JP2003140791A (ja) * | 2001-10-31 | 2003-05-16 | Toshiba Corp | 情報処理装置及び情報処理装置の制御方法 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013171195A (ja) * | 2012-02-21 | 2013-09-02 | Canon Inc | 表示装置、その制御方法及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
CN102317895A (zh) | 2012-01-11 |
JPWO2010095255A1 (ja) | 2012-08-16 |
US20110289449A1 (en) | 2011-11-24 |
EP2400378A1 (en) | 2011-12-28 |
KR20110104088A (ko) | 2011-09-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20220083214A1 (en) | Systems and Methods for Interacting with Multiple Applications that are Simultaneously Displayed on an Electronic Device with a Touch-Sensitive Display | |
US11989409B2 (en) | Device, method, and graphical user interface for displaying a plurality of settings controls | |
EP3750045B1 (en) | Systems, methods, and user interfaces for interacting with multiple application windows | |
WO2010095255A1 (ja) | 情報処理装置、表示制御方法および表示制御プログラム | |
US10365787B2 (en) | Apparatus and method for grouping and displaying icons on a screen | |
US10156967B2 (en) | Device, method, and graphical user interface for tabbed and private browsing | |
US8773370B2 (en) | Table editing systems with gesture-based insertion and deletion of columns and rows | |
EP2908232B1 (en) | Display control device, display control method and program for managing the layout of software applications | |
US20120212438A1 (en) | Methods and apparatuses for facilitating interaction with touch screen apparatuses | |
US20110175826A1 (en) | Automatically Displaying and Hiding an On-screen Keyboard | |
KR102168648B1 (ko) | 사용자 단말 장치 및 그 제어 방법 | |
EP3590034B1 (en) | Systems and methods for interacting with multiple applications that are simultaneously displayed on an electronic device with a touch-sensitive display | |
CN105229590A (zh) | 具有笔的用户终端装置和用户终端装置的控制方法 | |
JP2013529339A (ja) | 携帯用電子デバイスおよびそれを制御する方法 | |
US10895955B2 (en) | Apparatus and method for grouping and displaying icons on a screen | |
WO2022059386A1 (ja) | 操作領域を移動する情報処理装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980156660.6 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09840361 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2011500424 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20117018226 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009840361 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |