US20110022033A1 - System and Method for Wearable User Interface in Computer Assisted Surgery - Google Patents
System and Method for Wearable User Interface in Computer Assisted Surgery Download PDFInfo
- Publication number
- US20110022033A1 US20110022033A1 US12/896,309 US89630910A US2011022033A1 US 20110022033 A1 US20110022033 A1 US 20110022033A1 US 89630910 A US89630910 A US 89630910A US 2011022033 A1 US2011022033 A1 US 2011022033A1
- Authority
- US
- United States
- Prior art keywords
- input device
- surgical
- cursor
- user
- control signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000001356 surgical procedure Methods 0.000 title claims abstract description 19
- 238000000034 method Methods 0.000 title claims description 20
- 238000004891 communication Methods 0.000 claims abstract description 13
- 238000009877 rendering Methods 0.000 claims abstract description 11
- 230000000007 visual effect Effects 0.000 claims abstract description 6
- 210000000245 forearm Anatomy 0.000 claims description 21
- 239000000463 material Substances 0.000 claims description 11
- 230000000694 effects Effects 0.000 claims description 7
- 230000003287 optical effect Effects 0.000 description 28
- 210000000988 bone and bone Anatomy 0.000 description 12
- 239000012634 fragment Substances 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 230000002093 peripheral effect Effects 0.000 description 5
- 238000013459 approach Methods 0.000 description 3
- 210000000707 wrist Anatomy 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 230000000881 depressing effect Effects 0.000 description 2
- 238000002594 fluoroscopy Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 240000005020 Acaciella glauca Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 1
- 244000052616 bacterial pathogen Species 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000036772 blood pressure Effects 0.000 description 1
- 239000012530 fluid Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000002675 image-guided surgery Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012829 orthopaedic surgery Methods 0.000 description 1
- 229910052760 oxygen Inorganic materials 0.000 description 1
- 239000001301 oxygen Substances 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 235000003499 redwood Nutrition 0.000 description 1
- 230000001954 sterilising effect Effects 0.000 description 1
- 238000004659 sterilization and disinfection Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00199—Electrical control of surgical instruments with a console, e.g. a control panel with a display
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00207—Electrical control of surgical instruments with hand gesture control or hand gesture recognition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
Definitions
- This invention relates to computer assisted surgery and, more particularly, to user interface devices used in computer assisted surgery.
- CAS systems computer assisted surgery systems
- image guided surgery systems are becoming widespread.
- a computer and a position measurement device are used in order to measure the position of surgical instruments, devices and a body portion of the patient.
- CAS systems also incorporate a memory means to store medical data such as e.g. X-rays, computertomographs or magnetic resonance images (MRIs).
- the medical images may be gathered pre-operatively or intraoperatively.
- Computer assisted orthopaedic surgery systems include a) computertomogram (CT) based systems, may be used preoperatively to establish a three-dimensional anatomical model of a bone or bone fragment that is referenced during the surgical procedure to identify the respective bone or bone fragment through a landmark based or surface based registration or matching procedure; b) CT based and fluoroscopy systems, which use the same method as CT based systems to establish a three-dimensional anatomical model, whereby the preoperative CT of a bone or bone fragment is registered or matched to the intraoperative respective bone or bone fragment through using a surface model of the bone or bone fragment and its projections in the planes of the fluoroscopic images; and c) fluoroscopy based systems, which use calibrated fluoroscopes to generate undistorted images of a bone or bone fragment and virtual geometric representations of the projection of surgical tools.
- CT computertomogram
- the surgical data may include data stored preoperatively as well as data obtained during the surgery, such as blood pressure, heart rate, oxygen levels, etc.
- Many systems rely upon a non-sterile assistant to input instructions at a keyboard or with a mouse, but these systems are inefficient and risk miscommunication.
- an input device is physically located within the operating room.
- the input device is generally covered with a plastic material to provide a physical barrier against the spread of germs.
- control of the CAS is required, the surgeon or other personnel within the operating room go to the location of the input device and operate the device. This approach is cumbersome since it requires a dedicated location for the input device within an already crowded operating room. Moreover, the movement of personnel to the location of the input device creates undesired activity within the operating room.
- Some systems have attempted to address the shortcomings of the above described approach in various ways.
- One such approach is to provide an input device in the form of a foot operated device.
- this type of a device becomes a hazard when personnel are moving about the location of the foot operated device.
- the foot operated devices do not provide for cursor control.
- a method of performing a medical procedure in accordance with the invention includes mounting an input device on a user, establishing a communications link between the input device and a computer, placing a light permeable garment over the input device, sensing with the input device a desired cursor control input through the light permeable material, generating a cursor control signal based upon the control input, and controlling the visual rendering on a display based upon the cursor control signal.
- a computer assisted surgery system includes a display for visual rendering of surgical data, a processor for receiving a control signal and performing operations associated with the control signal and for controlling the surgical data rendered by the display, and a wearable input device in communication with the processor for receiving input from a user wearing the interface device, converting the input into the control signal and transmitting the control signal to the processor.
- a computer assisted surgery system includes a display for visual rendering of surgical data; a processor for receiving a control signal and performing operations associated with the control signal and for controlling the surgical data rendered by the display, and a wearable input device in communication with the processor.
- the wearable input device includes a mounting surface on a first side of the input device for mounting to a body portion of a user, and a sensor assembly located on the side of the input device opposite the mounting surface.
- FIG. 1 depicts a block diagram of a computer assisted surgery system including a wearable input device with an optical sensor in accordance with principles of the present invention
- FIG. 2 depicts a perspective view of the input device of FIG. 1 ;
- FIG. 3 depicts a perspective view of the input device of FIG. 1 mounted on the left forearm of a user in accordance with principles of the present invention
- FIG. 4 depicts a block diagram of the input device of FIG. 1 ;
- FIG. 5 depicts a perspective view of the input device of FIG. 1 mounted on the right forearm of a user in accordance with principles of the present invention
- FIG. 6 depicts an alternative embodiment of a wearable input device suitable for mounting to the chest, forearm, or other appendage of a user in accordance with principles of the present invention
- FIG. 7 depicts an alternative embodiment of a wearable input device incorporating mechanical switches suitable for mounting to the chest, forearm, or other appendage of a user in accordance with principles of the present invention.
- FIG. 8 shows a flow diagram of a method of performing a surgical procedure using the input device of FIG. 1 in accordance with principles of the present invention.
- FIG. 1 illustrates a block diagram of a computer system 10 for use in computer assisted surgery.
- the computer system 10 includes a surgical computer 12 which is in communication with a network 14 , a display 16 and an input device 18 .
- the surgical computer 12 may further be in communication with other peripherals such as peripheral 20 .
- Peripheral 20 may be an external memory, a surgical device, a printer, or other desired device.
- the surgical computer 12 includes an internal processor 22 and an internal memory 24 in communication with the processor 22 .
- the internal memory 24 includes instructions which allow the surgical computer 12 to interface with the network 14 and various peripherals such as input device 18 , display 16 and peripheral 20 as is known in the art.
- the internal memory 24 further includes instructions which, when executed by the processor 22 , allow the surgical computer 12 to be used to assist in performance of a surgical operation.
- the network 14 in this embodiment is a local area network for a medical facility which provides a surgeon or other medical personnel with access to resources useful in the conduct of a particular operation.
- the resources may include medical records, the internet, and video-conferencing capability.
- some or all of the resources may be provided in the internal memory 22 .
- Communication between the network 14 and the surgical computer 12 may be accomplished using cables, optical communications, radio frequency (RF) communications or any other desired mode of transferring data.
- RF radio frequency
- the display 16 in this embodiment is a cathode ray tube (CRT) used for rendering graphical representations of data under control of the surgical computer 12 .
- CTR cathode ray tube
- Any appropriate type of display may be used in accordance with the present invention including liquid crystal diode (LCD) displays, light emitting diode (LED) displays and heads-up displays (HUD).
- LCD liquid crystal diode
- LED light emitting diode
- HUD heads-up displays
- FIG. 2 depicts a perspective view of the input device 18 .
- the input device 18 includes a body 26 with an outer side 28 and an inner side 30 .
- a plurality of slots 32 , 33 (see FIG. 5 ), 34 and 36 extend through the body 26 from the inner side 30 to the outer side 28 .
- the slots 32 , 33 , 34 and 36 may be used with one or more straps to mount the input device 18 onto a user.
- the input device 18 is mounted onto the forearm 38 of a user by a strap 40 extending between the slot 34 and the slot 36 and a strap 42 extending between the slot 32 and slot 33 (see FIG. 5 ).
- the straps 40 and 42 may be made of a material such as VELCRO®.
- VELCRO® traps allows a single input device 18 to be used with forearms of various sizes. Nonetheless, it may be desired to provide a plurality of input devices configured with different curvatures. More specifically, the inner side of the input devices may have different radii of curvature. In such embodiments, small, medium and large curvature input devices may be provided so as to comfortably fit a number of different sized users.
- the input device may include a body made of expandable material. In this alternative embodiment, the body of the input device stretches to conform to the size of the user's forearm so as to maintain the input device snuggly in position on the user's forearm.
- the input device 18 further includes a USB connection 44 and an optical sensor well 46 for placement of a sensor assembly.
- the USB connection 44 provides a path for communications with the surgical computer 12 and further provides power to the input device 18 .
- the USB connection 44 may be connected to the surgical computer 12 by a USB cable 48 .
- the USB connection 44 may be connected to an RF transmitter so as to provide a wireless connection.
- the input device may be a self contained wireless unit with an internal power supply.
- a separate power supply may be used so that the battery or other power source may be placed at a different location on the body of a user, such as on a belt. This minimizes the weight of the input device 18 and facilitates replacement of the power source as well as facilitating the provision of redundant power sources.
- the input device 18 includes a micro controller 50 , the USB connector 44 , a USB connector 52 , two mechanical switches 54 and 56 , an optical sensor 58 which controls an LED 60 .
- the optical sensor 58 and the LED 60 are located in the optical sensor well 46 shown in FIG. 2 .
- the optical sensor 58 includes a charged-coupled device (CCD) array and a lens for focusing reflected light onto the array.
- the optical sensor 58 can have a photo-sensitive element other than a CCD array, such as a number of photo-diodes or photo-transistors.
- the input device 18 allows a user to interact (e.g., effect cursor movement, scrolling, or button action) with the surgical computer 12 . Movement relevant to the input device 18 is detected by a sensor assembly and translated into position data, and is communicated to the surgical computer 12 via the USB connection 44 .
- light from the LED 60 reflects off of a surface moving past the optical sensor well 46 , and causes an image of the surface or object to be generated. This image is detected by optical sensor 58 .
- the direction and distance of movement can be determined by a series of such detected images.
- the reflected images are focused by a lens onto the CCD array.
- Each image can be represented by a number of pixels on the CCD array.
- a difference between consecutive images indicates movement, while no difference between consecutive images indicates lack of movement.
- the image difference data is determined by the microcontroller 50 and communicated to the surgical computer 12 which in turn controls the position of a cursor rendered on the display 16 .
- the optical sensor 58 is calibrated such that the movement of material used in a surgical gown or other garment is not detected.
- surgical garments are typically impermeable to fluids, they will allow, to some extent, energy in the form of light to pass through.
- an optical sensor such as one available in a PocketMouseTM Wireless Mini or a PocketMouse Optical Mini computer mouse commercially available from Kensington Computer Products Group, of Redwood Shores, Calif., does not detect the movement of the material.
- the input device 18 may be worn and operated beneath a surgical gown. Additionally, because the use of a material that is somewhat light permeable allows a user to easily determine the location of the optical sensor well 46 even when the optical sensor well 46 is covered by the material.
- the USB connection 44 is used when the input device 18 is mounted on the left forearm of a user as shown in FIG. 3 . This allows the USB cable 48 to be directed toward the back of the user and away from the hand of the user.
- movement of an appendage such as a hand or finger in the direction generally defined by the direction from the elbow toward the wrist over the optical sensor well 46 is detected by the optical sensor 58 and translated by the microcontroller 50 into a signal corresponding to a desired movement of a cursor rendered on the display 16 from a first location on the display 16 to a position to the right of the first location.
- movement of a hand or finger in the direction generally defined by the direction from the slot 34 to the slot 36 over the optical sensor well 46 is detected by the optical sensor 58 and translated by the microcontroller 50 into a signal corresponding to a desired movement of a cursor rendered on the display 16 from a first location on the display 16 to a higher location on the display 16 .
- the USB connector 52 is similar to the USB connector 44 , however, the USB connector 52 is physically located at the end of the body 26 opposite to the USB connector 44 . Accordingly, when the input device 18 is mounted on the right forearm of a user, the USB connector 52 may be used to avoid having any cables extending out from the input device 18 toward the hand of the user. When the input device 18 is mounted on the right arm of a user, control of the movement of a cursor on the display 16 is effected in the same manner as described above with respect to USB connector 44 .
- the mechanical switches 54 and 56 function in the same manner as the “left-click” and “right-click” buttons on a standard mouse.
- the mechanical switch 54 is physically located to the left of the optical sensor well 46 and the mechanical switch 56 is physically located to the right of the optical sensor well 46 .
- the microcontroller 50 translates activation of the mechanical switch 54 into a “left-click” control signal and activation of the mechanical switch 56 into a “right-click” control signal.
- operation of the input device is the same as described above.
- the single USB connector device is worn on the right forearm, however, it may be desired to modify the correlation between movement sensed by the optical sensor 58 and the position of a cursor on the display 16 .
- a USB cable connected to the connector 44 will extend, to some extent, toward the wrist of the user. Typically, this is not desired.
- the surgical computer 12 may easily be programmed to account for this situation and be programmed to modify the correlation between movement sensed by the optical sensor 58 and the position of a cursor on the display 16 .
- the alternative correlation may be activated by, for example, designating desired correlation from a configuration menu or activating a switch.
- the input device 18 is configured to conform to the general shape of the forearm of a user. In alternative embodiments, the input device 18 may be configured to be mounted on other parts of a user's body such as the chest of the user.
- the input device 64 includes an optical sensor well 66 , a USB output connector 68 and two loops 70 and 72 .
- the input device 64 is relatively flat so as to be mounted comfortably on, for example, the chest of the user using the two loops 70 and 72 .
- the input device 64 may also be mounted on the arm of the user if desired.
- FIG. 7 depicts an input device 74 that may be mounted on a user so as to provide input to the surgical computer 12 during a medical operation through a mechanical sensor assembly.
- the mechanical input device 74 includes a control pad 76 , two attachment posts 78 and 80 , and a USB connector 82 . Desired movement of a cursor is effected by pressing the control pad 76 at a location near the head of one of the arrows 84 . Pressing the control pad 76 near the outer edges of the control pad 76 in this manner causes one or more of a plurality of mechanical switches to be closed, indicating the desired direction of movement.
- Depressing the circle 86 causes a centrally located mechanical switch to be closed, thereby “selecting” the location on the display 16 associated with the cursor.
- the control pad 76 and/or the arrows 84 and circle 86 may be lighted.
- FIG. 8 depicts a method 90 for performing a surgical procedure using the computer assisted surgery system of FIG. 1 .
- surgical data is stored in memory 24 of the host computer 12 .
- the data may be stored in a memory accessible through the network 14 .
- the data may include x-rays, steps for the planned surgical procedure, and special precautions or constraints.
- the materials or devices to be used in the surgical procedure, such as prosthetics, may also be identified.
- the input device 18 is mounted on the forearm of the surgeon at the step 96 .
- the input device 18 may be mounted on either the left forearm or the right forearm of the surgeon. In this example, the input device 18 is mounted on the right forearm of the surgeon.
- those devices are also mounted on the surgeon during this step.
- the input device 18 is placed upon the forearm of the user and the straps 40 and 42 are adjusted to provide a snug fit.
- a garment or padding may first be placed on the forearm prior to mounting the input device 18 on the forearm.
- the surgeon dons a surgical gown which is light permeable. Communication between the input device 18 and the computer 12 is then established by connecting the USB cable 48 between the USB connector 44 and a USB connector (not shown) on the computer 12 at the step 100 .
- the surgeon can visually verify that the input device is powered by looking for a reddish illumination on the sleeve of the surgical gown. If desired, the surgeon may modify the operational mode of the input device 18 at this point. This may be desired in single USB connector devices as discussed above.
- the surgeon controls the position of the cursor on the display 16 by moving a hand over the illuminated area of the surgical gown.
- the optical sensor 58 detects the movement of the hand and generates a cursor control signal based upon the detected movement.
- the cursor control signal is transmitted to the computer 12 which uses the cursor control signal to control the position on the display 16 at which the cursor is rendered.
- the surgeon commands the operation associated with the cursor position by depressing the mechanical switch 56 at step 104 .
- Depression of the switch 56 is detected by the input device 18 which generates a mechanical device control signal based upon the detected depression which is transmitted to the computer 12 .
- the computer 12 executes a program.
- the executed program causes a stored surgical data file to be accessed.
- a first surgical data is displayed.
- the surgeon performs the activity at the step 108 .
- Such an activity may be, for example, the selection of a desired prosthetic device, selecting a desired incision location, etc.
- the surgeon determines if another surgical data is needed at the step 110 . If so, the method returns to step 102 so as to display a second surgical data at the step 106 . If no additional surgical data is needed, the method ends.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Surgery (AREA)
- Heart & Thoracic Surgery (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Robotics (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Pathology (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
A computer assisted surgery system includes a display for visual rendering of surgical data, a processor for receiving a control signal and performing operations associated with the control signal and for controlling the surgical data rendered by the display, and a wearable input device in communication with the processor. The wearable input device in one embodiment includes a mounting surface on a first side of the input device for mounting to a body portion of a user and a sensor assembly located on the side of the input device opposite the mounting surface.
Description
- This application is a divisional of co-pending application Ser. No. 11/319,960, filed Dec. 28, 2005, the disclosure of which is herein totally incorporated by reference in its entirety.
- This invention relates to computer assisted surgery and, more particularly, to user interface devices used in computer assisted surgery.
- The use of computer assisted surgery systems (CAS systems) or image guided surgery systems is becoming widespread. In a typical CAS system, a computer and a position measurement device are used in order to measure the position of surgical instruments, devices and a body portion of the patient. CAS systems also incorporate a memory means to store medical data such as e.g. X-rays, computertomographs or magnetic resonance images (MRIs). The medical images may be gathered pre-operatively or intraoperatively.
- Computer assisted orthopaedic surgery systems include a) computertomogram (CT) based systems, may be used preoperatively to establish a three-dimensional anatomical model of a bone or bone fragment that is referenced during the surgical procedure to identify the respective bone or bone fragment through a landmark based or surface based registration or matching procedure; b) CT based and fluoroscopy systems, which use the same method as CT based systems to establish a three-dimensional anatomical model, whereby the preoperative CT of a bone or bone fragment is registered or matched to the intraoperative respective bone or bone fragment through using a surface model of the bone or bone fragment and its projections in the planes of the fluoroscopic images; and c) fluoroscopy based systems, which use calibrated fluoroscopes to generate undistorted images of a bone or bone fragment and virtual geometric representations of the projection of surgical tools.
- When used during a surgical procedure, most CAS systems require input from the surgeon in order to specify the data to be shown or to alter the program flow. The surgical data may include data stored preoperatively as well as data obtained during the surgery, such as blood pressure, heart rate, oxygen levels, etc. Many systems rely upon a non-sterile assistant to input instructions at a keyboard or with a mouse, but these systems are inefficient and risk miscommunication.
- In other systems, an input device is physically located within the operating room. The input device is generally covered with a plastic material to provide a physical barrier against the spread of germs. When control of the CAS is required, the surgeon or other personnel within the operating room go to the location of the input device and operate the device. This approach is cumbersome since it requires a dedicated location for the input device within an already crowded operating room. Moreover, the movement of personnel to the location of the input device creates undesired activity within the operating room.
- Some systems have attempted to address the shortcomings of the above described approach in various ways. One such approach is to provide an input device in the form of a foot operated device. Of course, this type of a device becomes a hazard when personnel are moving about the location of the foot operated device. Moreover, the foot operated devices do not provide for cursor control.
- What is needed, therefore, is a CAS input device that allows personnel to control the CAS during an operation.
- What is further needed is a CAS input device that does not require personnel to move to a specific location within the operating room to operate the device.
- What is also needed is a CAS input device that is conveniently located for ease of access and which allows for the control of the cursor on a display.
- What is needed is a CAS input device that does not present a hazard to individuals moving about an operating room.
- A method of performing a medical procedure in accordance with the invention includes mounting an input device on a user, establishing a communications link between the input device and a computer, placing a light permeable garment over the input device, sensing with the input device a desired cursor control input through the light permeable material, generating a cursor control signal based upon the control input, and controlling the visual rendering on a display based upon the cursor control signal.
- In one embodiment, a computer assisted surgery system includes a display for visual rendering of surgical data, a processor for receiving a control signal and performing operations associated with the control signal and for controlling the surgical data rendered by the display, and a wearable input device in communication with the processor for receiving input from a user wearing the interface device, converting the input into the control signal and transmitting the control signal to the processor.
- In a further embodiment, a computer assisted surgery system includes a display for visual rendering of surgical data; a processor for receiving a control signal and performing operations associated with the control signal and for controlling the surgical data rendered by the display, and a wearable input device in communication with the processor. The wearable input device includes a mounting surface on a first side of the input device for mounting to a body portion of a user, and a sensor assembly located on the side of the input device opposite the mounting surface.
- The above-described features and advantages, as well as others, will become more readily apparent to those of ordinary skill in the art by reference to the following detailed description and accompanying drawings.
-
FIG. 1 depicts a block diagram of a computer assisted surgery system including a wearable input device with an optical sensor in accordance with principles of the present invention; -
FIG. 2 depicts a perspective view of the input device ofFIG. 1 ; -
FIG. 3 depicts a perspective view of the input device ofFIG. 1 mounted on the left forearm of a user in accordance with principles of the present invention; -
FIG. 4 depicts a block diagram of the input device ofFIG. 1 ; -
FIG. 5 depicts a perspective view of the input device ofFIG. 1 mounted on the right forearm of a user in accordance with principles of the present invention; -
FIG. 6 depicts an alternative embodiment of a wearable input device suitable for mounting to the chest, forearm, or other appendage of a user in accordance with principles of the present invention; -
FIG. 7 depicts an alternative embodiment of a wearable input device incorporating mechanical switches suitable for mounting to the chest, forearm, or other appendage of a user in accordance with principles of the present invention; and -
FIG. 8 shows a flow diagram of a method of performing a surgical procedure using the input device ofFIG. 1 in accordance with principles of the present invention. -
FIG. 1 illustrates a block diagram of acomputer system 10 for use in computer assisted surgery. Thecomputer system 10 includes asurgical computer 12 which is in communication with anetwork 14, adisplay 16 and aninput device 18. Thesurgical computer 12 may further be in communication with other peripherals such as peripheral 20. Peripheral 20 may be an external memory, a surgical device, a printer, or other desired device. - The
surgical computer 12 includes aninternal processor 22 and aninternal memory 24 in communication with theprocessor 22. Theinternal memory 24 includes instructions which allow thesurgical computer 12 to interface with thenetwork 14 and various peripherals such asinput device 18,display 16 and peripheral 20 as is known in the art. Theinternal memory 24 further includes instructions which, when executed by theprocessor 22, allow thesurgical computer 12 to be used to assist in performance of a surgical operation. - The
network 14 in this embodiment is a local area network for a medical facility which provides a surgeon or other medical personnel with access to resources useful in the conduct of a particular operation. By way of example, the resources may include medical records, the internet, and video-conferencing capability. Alternatively, some or all of the resources may be provided in theinternal memory 22. Communication between thenetwork 14 and thesurgical computer 12 may be accomplished using cables, optical communications, radio frequency (RF) communications or any other desired mode of transferring data. - The
display 16 in this embodiment is a cathode ray tube (CRT) used for rendering graphical representations of data under control of thesurgical computer 12. Any appropriate type of display may be used in accordance with the present invention including liquid crystal diode (LCD) displays, light emitting diode (LED) displays and heads-up displays (HUD). -
FIG. 2 depicts a perspective view of theinput device 18. Theinput device 18 includes abody 26 with anouter side 28 and aninner side 30. A plurality ofslots 32, 33 (seeFIG. 5 ), 34 and 36 extend through thebody 26 from theinner side 30 to theouter side 28. Theslots input device 18 onto a user. By way of example, inFIG. 3 theinput device 18 is mounted onto theforearm 38 of a user by astrap 40 extending between theslot 34 and theslot 36 and astrap 42 extending between theslot 32 and slot 33 (seeFIG. 5 ). - The
straps single input device 18 to be used with forearms of various sizes. Nonetheless, it may be desired to provide a plurality of input devices configured with different curvatures. More specifically, the inner side of the input devices may have different radii of curvature. In such embodiments, small, medium and large curvature input devices may be provided so as to comfortably fit a number of different sized users. Alternatively, the input device may include a body made of expandable material. In this alternative embodiment, the body of the input device stretches to conform to the size of the user's forearm so as to maintain the input device snuggly in position on the user's forearm. - The
input device 18 further includes aUSB connection 44 and an optical sensor well 46 for placement of a sensor assembly. TheUSB connection 44 provides a path for communications with thesurgical computer 12 and further provides power to theinput device 18. TheUSB connection 44 may be connected to thesurgical computer 12 by aUSB cable 48. Alternatively, theUSB connection 44 may be connected to an RF transmitter so as to provide a wireless connection. Of course, the input device may be a self contained wireless unit with an internal power supply. Alternatively, a separate power supply may be used so that the battery or other power source may be placed at a different location on the body of a user, such as on a belt. This minimizes the weight of theinput device 18 and facilitates replacement of the power source as well as facilitating the provision of redundant power sources. - A block diagram of the
input device 18 is depicted inFIG. 4 . Theinput device 18 includes amicro controller 50, theUSB connector 44, aUSB connector 52, twomechanical switches optical sensor 58 which controls anLED 60. Theoptical sensor 58 and theLED 60 are located in the optical sensor well 46 shown inFIG. 2 . In this embodiment, theoptical sensor 58 includes a charged-coupled device (CCD) array and a lens for focusing reflected light onto the array. Alternatively, theoptical sensor 58 can have a photo-sensitive element other than a CCD array, such as a number of photo-diodes or photo-transistors. - The
input device 18 allows a user to interact (e.g., effect cursor movement, scrolling, or button action) with thesurgical computer 12. Movement relevant to theinput device 18 is detected by a sensor assembly and translated into position data, and is communicated to thesurgical computer 12 via theUSB connection 44. In this embodiment, light from theLED 60 reflects off of a surface moving past the optical sensor well 46, and causes an image of the surface or object to be generated. This image is detected byoptical sensor 58. - The direction and distance of movement can be determined by a series of such detected images. In one embodiment, the reflected images are focused by a lens onto the CCD array. Each image can be represented by a number of pixels on the CCD array. A difference between consecutive images indicates movement, while no difference between consecutive images indicates lack of movement. The image difference data is determined by the
microcontroller 50 and communicated to thesurgical computer 12 which in turn controls the position of a cursor rendered on thedisplay 16. - The
optical sensor 58 is calibrated such that the movement of material used in a surgical gown or other garment is not detected. By way of example, while surgical garments are typically impermeable to fluids, they will allow, to some extent, energy in the form of light to pass through. Accordingly, an optical sensor, such as one available in a PocketMouse™ Wireless Mini or a PocketMouse Optical Mini computer mouse commercially available from Kensington Computer Products Group, of Redwood Shores, Calif., does not detect the movement of the material. When the user's hand is moved over the top of the optical sensor well 46 and surgical gown material, however, the movement of the hand is detected. Accordingly, theinput device 18 may be worn and operated beneath a surgical gown. Additionally, because the use of a material that is somewhat light permeable allows a user to easily determine the location of the optical sensor well 46 even when the optical sensor well 46 is covered by the material. - The
USB connection 44 is used when theinput device 18 is mounted on the left forearm of a user as shown inFIG. 3 . This allows theUSB cable 48 to be directed toward the back of the user and away from the hand of the user. When theUSB connection 44 is used, movement of an appendage such as a hand or finger in the direction generally defined by the direction from the elbow toward the wrist over the optical sensor well 46 is detected by theoptical sensor 58 and translated by themicrocontroller 50 into a signal corresponding to a desired movement of a cursor rendered on thedisplay 16 from a first location on thedisplay 16 to a position to the right of the first location. - Similarly, movement of a hand or finger in the direction generally defined by the direction from the
slot 34 to theslot 36 over the optical sensor well 46 is detected by theoptical sensor 58 and translated by themicrocontroller 50 into a signal corresponding to a desired movement of a cursor rendered on thedisplay 16 from a first location on thedisplay 16 to a higher location on thedisplay 16. - The
USB connector 52 is similar to theUSB connector 44, however, theUSB connector 52 is physically located at the end of thebody 26 opposite to theUSB connector 44. Accordingly, when theinput device 18 is mounted on the right forearm of a user, theUSB connector 52 may be used to avoid having any cables extending out from theinput device 18 toward the hand of the user. When theinput device 18 is mounted on the right arm of a user, control of the movement of a cursor on thedisplay 16 is effected in the same manner as described above with respect toUSB connector 44. - The mechanical switches 54 and 56 function in the same manner as the “left-click” and “right-click” buttons on a standard mouse. In this embodiment, the
mechanical switch 54 is physically located to the left of the optical sensor well 46 and themechanical switch 56 is physically located to the right of the optical sensor well 46. Accordingly, themicrocontroller 50 translates activation of themechanical switch 54 into a “left-click” control signal and activation of themechanical switch 56 into a “right-click” control signal. - In embodiments having a single USB connector, operation of the input device is the same as described above. When the single USB connector device is worn on the right forearm, however, it may be desired to modify the correlation between movement sensed by the
optical sensor 58 and the position of a cursor on thedisplay 16. Specifically, using theinput device 18 as an example, when theinput device 18 is located on aright forearm 62 with theUSB connector 44 positioned to be adjacent the wrist of the user, a USB cable connected to theconnector 44 will extend, to some extent, toward the wrist of the user. Typically, this is not desired. - Turning the
input device 18 to position theUSB connector 44 adjacent to the elbow of the user as shown inFIG. 5 causes the optical sensor well 46, when viewed by the user, to be inverted and reversed and the position of themechanical switches optical sensor 46 to be reversed. Thesurgical computer 12, however, may easily be programmed to account for this situation and be programmed to modify the correlation between movement sensed by theoptical sensor 58 and the position of a cursor on thedisplay 16. The alternative correlation may be activated by, for example, designating desired correlation from a configuration menu or activating a switch. - In the embodiment of
FIG. 2 , theinput device 18 is configured to conform to the general shape of the forearm of a user. In alternative embodiments, theinput device 18 may be configured to be mounted on other parts of a user's body such as the chest of the user. One such embodiment is shown inFIG. 6 . Theinput device 64 includes an optical sensor well 66, aUSB output connector 68 and twoloops input device 64 is relatively flat so as to be mounted comfortably on, for example, the chest of the user using the twoloops input device 64 may also be mounted on the arm of the user if desired. -
FIG. 7 depicts aninput device 74 that may be mounted on a user so as to provide input to thesurgical computer 12 during a medical operation through a mechanical sensor assembly. Themechanical input device 74 includes acontrol pad 76, twoattachment posts USB connector 82. Desired movement of a cursor is effected by pressing thecontrol pad 76 at a location near the head of one of thearrows 84. Pressing thecontrol pad 76 near the outer edges of thecontrol pad 76 in this manner causes one or more of a plurality of mechanical switches to be closed, indicating the desired direction of movement. Depressing thecircle 86 causes a centrally located mechanical switch to be closed, thereby “selecting” the location on thedisplay 16 associated with the cursor. To assist in locating thecontrol pad 76 when it is mounted underneath a garment, thecontrol pad 76 and/or thearrows 84 andcircle 86 may be lighted. -
FIG. 8 depicts amethod 90 for performing a surgical procedure using the computer assisted surgery system ofFIG. 1 . At thestep 90, surgical data is stored inmemory 24 of thehost computer 12. Alternatively, the data may be stored in a memory accessible through thenetwork 14. The data may include x-rays, steps for the planned surgical procedure, and special precautions or constraints. The materials or devices to be used in the surgical procedure, such as prosthetics, may also be identified. - The surgical team including the individual identified to operate the
input device 18 who, in this example, is the surgeon, then undertake typical preparations for surgery such as sterilization, at thestep 92. Next, theinput device 18 is mounted on the forearm of the surgeon at thestep 96. Theinput device 18 may be mounted on either the left forearm or the right forearm of the surgeon. In this example, theinput device 18 is mounted on the right forearm of the surgeon. - In embodiments including a battery pack and or wearable transceiver, those devices are also mounted on the surgeon during this step. In this example, the
input device 18 is placed upon the forearm of the user and thestraps input device 18 on the forearm. - At the
step 98, the surgeon dons a surgical gown which is light permeable. Communication between theinput device 18 and thecomputer 12 is then established by connecting theUSB cable 48 between theUSB connector 44 and a USB connector (not shown) on thecomputer 12 at thestep 100. The surgeon can visually verify that the input device is powered by looking for a reddish illumination on the sleeve of the surgical gown. If desired, the surgeon may modify the operational mode of theinput device 18 at this point. This may be desired in single USB connector devices as discussed above. - At the
step 102, the surgeon controls the position of the cursor on thedisplay 16 by moving a hand over the illuminated area of the surgical gown. Theoptical sensor 58 detects the movement of the hand and generates a cursor control signal based upon the detected movement. The cursor control signal is transmitted to thecomputer 12 which uses the cursor control signal to control the position on thedisplay 16 at which the cursor is rendered. - Once the cursor on the display is at the desired location, the surgeon commands the operation associated with the cursor position by depressing the
mechanical switch 56 atstep 104. Depression of theswitch 56 is detected by theinput device 18 which generates a mechanical device control signal based upon the detected depression which is transmitted to thecomputer 12. Based upon the location of the cursor on thedisplay 16 and the receipt of the mechanical device control signal, thecomputer 12 executes a program. In this example, the executed program causes a stored surgical data file to be accessed. Thus, at thestep 106, a first surgical data is displayed. - In the event the surgical data is associated with a particular activity, the surgeon performs the activity at the
step 108. Such an activity may be, for example, the selection of a desired prosthetic device, selecting a desired incision location, etc. The surgeon then determines if another surgical data is needed at thestep 110. If so, the method returns to step 102 so as to display a second surgical data at thestep 106. If no additional surgical data is needed, the method ends. - While the present invention has been illustrated by the description of exemplary processes and system components, and while the various processes and components have been described in considerable detail, applicant does not intend to restrict or in any limit the scope of the appended claims to such detail. Additional advantages and modifications will also readily appear to those ordinarily skilled in the art. The invention in its broadest aspects is therefore not limited to the specific details, implementations, or illustrative examples shown and described. Accordingly, departures may be made from such details without departing from the spirit or scope of applicant's general inventive concept.
Claims (8)
1. A method of performing a medical procedure comprising:
mounting an input device on a user;
establishing a communications link between the input device and a computer;
placing a light permeable garment over the input device;
sensing with the input device a desired cursor control input through the light permeable material;
generating a cursor control signal based upon the control input; and
controlling the visual rendering on a display based upon the cursor control signal.
2. The method of claim 1 , wherein controlling the visual rendering further comprises:
rendering a cursor on the display at a first location based upon the cursor control signal;
rendering a first surgical data in support of a first surgical activity based upon the first location of the cursor;
moving the rendered cursor to a second location in response to a change in the cursor control signal; and
rendering a second surgical data in support of a second surgical activity based upon the second location of the cursor.
3. The method of claim 2 , wherein the first surgical data comprises a first x-ray of a surgical patient and the second surgical data comprises a second x-ray of the surgical patient.
4. The method of claim 1 , wherein sensing comprises:
detecting, through the garment, movement of an object.
5. The method of claim 4 , wherein detecting further comprises:
comparing a first image of the object with a second image of the object.
6. The method of claim 5 , wherein detecting further comprises:
detecting the location of an appendage of the user.
7. The method of claim 1 , wherein mounting comprises:
Mounting the input device to the forearm of the user.
8. The method of claim 7 , further comprising:
rendering medical data of a patient undergoing surgery based upon the cursor control signal.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/896,309 US20110022033A1 (en) | 2005-12-28 | 2010-10-01 | System and Method for Wearable User Interface in Computer Assisted Surgery |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/319,960 US7810504B2 (en) | 2005-12-28 | 2005-12-28 | System and method for wearable user interface in computer assisted surgery |
US12/896,309 US20110022033A1 (en) | 2005-12-28 | 2010-10-01 | System and Method for Wearable User Interface in Computer Assisted Surgery |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/319,960 Division US7810504B2 (en) | 2005-12-28 | 2005-12-28 | System and method for wearable user interface in computer assisted surgery |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110022033A1 true US20110022033A1 (en) | 2011-01-27 |
Family
ID=37890426
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/319,960 Expired - Fee Related US7810504B2 (en) | 2005-12-28 | 2005-12-28 | System and method for wearable user interface in computer assisted surgery |
US12/896,309 Abandoned US20110022033A1 (en) | 2005-12-28 | 2010-10-01 | System and Method for Wearable User Interface in Computer Assisted Surgery |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/319,960 Expired - Fee Related US7810504B2 (en) | 2005-12-28 | 2005-12-28 | System and method for wearable user interface in computer assisted surgery |
Country Status (4)
Country | Link |
---|---|
US (2) | US7810504B2 (en) |
EP (1) | EP1804151A1 (en) |
JP (1) | JP2007181683A (en) |
AU (1) | AU2006252285A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110278956A1 (en) * | 2010-05-11 | 2011-11-17 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | System including wearable power receiver and wearable power-output device |
US8941560B2 (en) | 2011-09-21 | 2015-01-27 | Google Inc. | Wearable computer with superimposed controls and instructions for external device |
WO2019079790A1 (en) * | 2017-10-21 | 2019-04-25 | Eyecam, Inc | Adaptive graphic user interfacing system |
US11262841B2 (en) | 2012-11-01 | 2022-03-01 | Eyecam Llc | Wireless wrist computing and control device and method for 3D imaging, mapping, networking and interfacing |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7887330B2 (en) | 2004-12-02 | 2011-02-15 | The United States Of America As Represented By The Secretary Of The Army | Trauma training system |
EP2026111B1 (en) * | 2007-08-10 | 2010-05-12 | Möller-Wedel GmbH | Controller for operating an operation microscope |
US8094009B2 (en) * | 2008-08-27 | 2012-01-10 | The Invention Science Fund I, Llc | Health-related signaling via wearable items |
US8125331B2 (en) * | 2008-08-27 | 2012-02-28 | The Invention Science Fund I, Llc | Health-related signaling via wearable items |
US8130095B2 (en) * | 2008-08-27 | 2012-03-06 | The Invention Science Fund I, Llc | Health-related signaling via wearable items |
US20100053175A1 (en) * | 2008-08-27 | 2010-03-04 | Bedros Tashjian | Wireless Display for Compensation or Cause |
US8284046B2 (en) | 2008-08-27 | 2012-10-09 | The Invention Science Fund I, Llc | Health-related signaling via wearable items |
GB201009952D0 (en) * | 2010-05-11 | 2010-07-21 | Hu Do Ltd | Hinge development |
US11688303B2 (en) | 2010-06-30 | 2023-06-27 | Strategic Operations, Inc. | Simulated torso for an open surgery simulator |
US8840403B2 (en) * | 2010-06-30 | 2014-09-23 | Stuart C. Segall | Wearable partial task surgical simulator |
US11495143B2 (en) | 2010-06-30 | 2022-11-08 | Strategic Operations, Inc. | Emergency casualty care trainer |
US10360817B2 (en) | 2010-06-30 | 2019-07-23 | Stuart Charles Segall | Wearable partial task surgical simulator |
US11854427B2 (en) | 2010-06-30 | 2023-12-26 | Strategic Operations, Inc. | Wearable medical trainer |
JP5266348B2 (en) * | 2011-02-01 | 2013-08-21 | 富士フイルム株式会社 | Ultrasonic diagnostic equipment |
JP5266351B2 (en) * | 2011-02-09 | 2013-08-21 | 富士フイルム株式会社 | Ultrasonic diagnostic equipment |
WO2012142254A2 (en) * | 2011-04-15 | 2012-10-18 | Saint Louis University | Input device |
WO2013166261A1 (en) * | 2012-05-03 | 2013-11-07 | Georgia Tech Research Corporation | Methods, controllers and computer program products for accessibility to computing devices |
WO2013168056A1 (en) * | 2012-05-10 | 2013-11-14 | Koninklijke Philips N.V. | Gesture control |
US10076254B2 (en) | 2014-12-16 | 2018-09-18 | Microsoft Technology Licensing, Llc | Optical communication with optical sensors |
US10665135B2 (en) | 2015-11-07 | 2020-05-26 | Strategic Operations, Inc. | Lateral cathotomy and cantholysis simulation device |
US10325524B2 (en) | 2015-11-07 | 2019-06-18 | Stuart Charles Segall | Lateral canthotomy and cantholysis simulation device |
CN107196161A (en) * | 2017-06-26 | 2017-09-22 | 恪道技术(深圳)有限公司 | A kind of portable data line |
Citations (78)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1924617A (en) * | 1931-03-21 | 1933-08-29 | Goodrich Co B F | Latex glove |
US4414537A (en) * | 1981-09-15 | 1983-11-08 | Bell Telephone Laboratories, Incorporated | Digital data entry glove interface device |
US4722625A (en) * | 1985-09-26 | 1988-02-02 | Triune Automated Painting Systems | Remote control device for powered painting system |
US4905001A (en) * | 1987-10-08 | 1990-02-27 | Penner Henry C | Hand-held finger movement actuated communication devices and systems employing such devices |
US4954817A (en) * | 1988-05-02 | 1990-09-04 | Levine Neil A | Finger worn graphic interface device |
US5332135A (en) * | 1992-11-24 | 1994-07-26 | Fletcher Stephen R | Apparatus for protecting watches and the like |
US5444462A (en) * | 1991-12-16 | 1995-08-22 | Wambach; Mark L. | Computer mouse glove with remote communication |
US5481265A (en) * | 1989-11-22 | 1996-01-02 | Russell; David C. | Ergonomic customizeable user/computer interface devices |
US5488362A (en) * | 1993-10-01 | 1996-01-30 | Anaphase Unlimited, Inc. | Apparatus for controlling a video game |
US5489922A (en) * | 1993-12-08 | 1996-02-06 | Hewlett-Packard Company | Hand worn remote computer mouse |
US5638092A (en) * | 1994-12-20 | 1997-06-10 | Eng; Tommy K. | Cursor control system |
US5638832A (en) * | 1995-06-07 | 1997-06-17 | Interval Research Corporation | Programmable subcutaneous visible implant |
US5726685A (en) * | 1994-06-30 | 1998-03-10 | Siemens Aktiengesellschaft | Input unit for a computer |
US5764164A (en) * | 1997-02-07 | 1998-06-09 | Reality Quest Corp. | Ergonomic hand-attachable controller |
US5796354A (en) * | 1997-02-07 | 1998-08-18 | Reality Quest Corp. | Hand-attachable controller with direction sensing |
US5911036A (en) * | 1995-09-15 | 1999-06-08 | Computer Motion, Inc. | Head cursor control interface for an automated endoscope system for optimal positioning |
US6141643A (en) * | 1998-11-25 | 2000-10-31 | Harmon; Steve | Data input glove having conductive finger pads and thumb pad, and uses therefor |
US6154199A (en) * | 1998-04-15 | 2000-11-28 | Butler; Craig L. | Hand positioned mouse |
US6184804B1 (en) * | 1996-01-26 | 2001-02-06 | Orang-Otang Computers, Inc. | Key palette |
US6304840B1 (en) * | 1998-06-30 | 2001-10-16 | U.S. Philips Corporation | Fingerless glove for interacting with data processing system |
US20020015022A1 (en) * | 2000-05-29 | 2002-02-07 | Moshe Ein-Gal | Wireless cursor control |
US20020024500A1 (en) * | 1997-03-06 | 2002-02-28 | Robert Bruce Howard | Wireless control device |
US6380923B1 (en) * | 1993-08-31 | 2002-04-30 | Nippon Telegraph And Telephone Corporation | Full-time wearable information managing device and method for the same |
US20020101401A1 (en) * | 2001-01-29 | 2002-08-01 | Mehran Movahed | Thumb mounted function and cursor control device for a computer |
US6450978B1 (en) * | 1998-05-28 | 2002-09-17 | Orthosoft, Inc. | Interactive computer-assisted surgical system and method thereof |
US20020130839A1 (en) * | 2001-03-16 | 2002-09-19 | Hugh Wallace | Optical screen pointing device with inertial properties |
US20020140674A1 (en) * | 2001-03-13 | 2002-10-03 | Canon Kabushiki Kaisha | Position/posture sensor or marker attachment apparatus |
US20020147455A1 (en) * | 2001-02-27 | 2002-10-10 | Carson Christopher P. | Total knee arthroplasty systems and processes |
US20020167422A1 (en) * | 2001-04-27 | 2002-11-14 | Jacques Andre | Device for controlling a three-dimensional movement |
US20020175894A1 (en) * | 2001-03-06 | 2002-11-28 | Vince Grillo | Hand-supported mouse for computer input |
US20020186200A1 (en) * | 2001-06-08 | 2002-12-12 | David Green | Method and apparatus for human interface with a computer |
US20030006962A1 (en) * | 2001-07-06 | 2003-01-09 | Bajramovic Mark B. | Computer mouse on a glove |
US20030011568A1 (en) * | 2001-06-15 | 2003-01-16 | Samsung Electronics Co., Ltd. | Glove-type data input device and sensing method thereof |
US20030011573A1 (en) * | 2001-07-16 | 2003-01-16 | Samsung Electronics Co., Ltd. | Information input method using wearable information input device |
US20030069897A1 (en) * | 2000-10-10 | 2003-04-10 | Roy Stephen C. | Systems and methods for enhancing the viewing of medical images |
US20030069591A1 (en) * | 2001-02-27 | 2003-04-10 | Carson Christopher Patrick | Computer assisted knee arthroplasty instrumentation, systems, and processes |
US6573883B1 (en) * | 1998-06-24 | 2003-06-03 | Hewlett Packard Development Company, L.P. | Method and apparatus for controlling a computing device with gestures |
US20030209604A1 (en) * | 1996-01-26 | 2003-11-13 | Harrison Shelton E. | Wearable computing system, method and device |
US20030214481A1 (en) * | 2002-05-14 | 2003-11-20 | Yongming Xiong | Finger worn and operated input device and method of use |
US20030227437A1 (en) * | 2002-06-05 | 2003-12-11 | Ramirez Nohl W. | Computer pointing device and utilization system |
US20040012564A1 (en) * | 2002-07-11 | 2004-01-22 | Frank Zngf | Glove mouse with virtual tracking ball |
US20040012559A1 (en) * | 2002-07-17 | 2004-01-22 | Kanazawa University | Input device |
US20040012557A1 (en) * | 2002-07-18 | 2004-01-22 | Sony Computer Entertainment Inc. | Hand-held computer interactive device |
US20040032346A1 (en) * | 2002-08-19 | 2004-02-19 | Samsung Electro-Mechanics Co., Ltd. | Information input device, information processing device and information input method |
US6696986B1 (en) * | 1998-03-16 | 2004-02-24 | Shelton E. Harrison, Jr. | Key palette improvements |
US20040046734A1 (en) * | 2002-09-25 | 2004-03-11 | Hart Timothy O. | Thumb-retained stylus |
US20040046736A1 (en) * | 1997-08-22 | 2004-03-11 | Pryor Timothy R. | Novel man machine interfaces and applications |
US6707447B1 (en) * | 1997-12-04 | 2004-03-16 | Richard Goranowski | Therapeutic and computer input gauntlet |
US20040051694A1 (en) * | 2001-01-03 | 2004-03-18 | Jonas Backman | Portable device for inputting control signals to a peripheral unit, and use of such a device |
US6744420B2 (en) * | 2000-06-01 | 2004-06-01 | Olympus Optical Co., Ltd. | Operation input apparatus using sensor attachable to operator's hand |
US20040164880A1 (en) * | 2003-02-24 | 2004-08-26 | Nguyen Quang Q. | Wearable data input device employing wrist and finger movements |
US20040207597A1 (en) * | 2002-07-27 | 2004-10-21 | Sony Computer Entertainment Inc. | Method and apparatus for light input device |
US20040263358A1 (en) * | 2001-09-14 | 2004-12-30 | Fredrik Madsen | Portable unit for inputting signals to a peripheral unit, and use of such a unit |
US6850224B2 (en) * | 2001-08-27 | 2005-02-01 | Carba Fire Technologies, Inc. | Wearable ergonomic computer mouse |
US20050035942A1 (en) * | 2003-03-21 | 2005-02-17 | Ruiz David M. | Hand-held on-screen control device |
US20050052291A1 (en) * | 2001-10-30 | 2005-03-10 | Perific Ab | Portable data input device and use of such a device |
US6885361B1 (en) * | 1987-03-24 | 2005-04-26 | Sun Microsystems, Inc. | Tactile feedback mechanism for a data processing system |
US20050151722A1 (en) * | 2004-01-14 | 2005-07-14 | Xerox Corporation | Methods and systems for collecting and generating ergonomic data utilizing an electronic portal |
US20050174326A1 (en) * | 2004-01-27 | 2005-08-11 | Samsung Electronics Co., Ltd. | Method of adjusting pointing position during click operation and 3D input device using the same |
US20050184884A1 (en) * | 2004-02-25 | 2005-08-25 | Samsung Electronics Co., Ltd. | Spatial information input apparatus and method for recognizing information-completion signal from a plurality of concurrent spatial motions |
US20050200602A1 (en) * | 2002-05-14 | 2005-09-15 | Christer Laurell | Control arrangement for a cursor |
US20050199250A1 (en) * | 2004-03-11 | 2005-09-15 | Green John M.Ii | System for determining a position of a point on an object |
US20050234332A1 (en) * | 2004-01-16 | 2005-10-20 | Murphy Stephen B | Method of computer-assisted ligament balancing and component placement in total knee arthroplasty |
US20050267353A1 (en) * | 2004-02-04 | 2005-12-01 | Joel Marquart | Computer-assisted knee replacement apparatus and method |
US20060001646A1 (en) * | 2004-07-02 | 2006-01-05 | Wei Hai | Finger worn and operated input device |
US20060012567A1 (en) * | 2004-07-13 | 2006-01-19 | Todd Sicklinger | Minature optical mouse and stylus |
US7042438B2 (en) * | 2003-09-06 | 2006-05-09 | Mcrae Michael William | Hand manipulated data apparatus for computers and video games |
US20060119578A1 (en) * | 2004-11-11 | 2006-06-08 | Thenkurussi Kesavadas | System for interfacing between an operator and a virtual object for computer aided design applications |
US20060214912A1 (en) * | 2000-07-01 | 2006-09-28 | Miller Stephen S | Apparatus for remotely controlling computers and other electronic appliances/devices using a combination of voice commands and finger movements |
US20060214000A1 (en) * | 2003-04-07 | 2006-09-28 | Silverbrook Research Pty Ltd | Wearable product reading device |
US7193606B2 (en) * | 2001-08-03 | 2007-03-20 | Nokia Corporation | Wearable electronic device |
US7205979B2 (en) * | 1987-03-17 | 2007-04-17 | Sun Microsystems, Inc. | Computer data entry and manipulation apparatus and method |
US7355584B2 (en) * | 2000-08-18 | 2008-04-08 | International Business Machines Corporation | Projector and camera arrangement with shared optics and optical marker for use with whiteboard systems |
US20080174549A1 (en) * | 2007-01-23 | 2008-07-24 | Chi Wen Chen | Wireless cursor control device with image-tracking |
US7405725B2 (en) * | 2003-01-31 | 2008-07-29 | Olympus Corporation | Movement detection device and communication apparatus |
US7427979B2 (en) * | 2003-09-08 | 2008-09-23 | Samsung Electronics Co., Ltd. | Pointing apparatus and method |
US7747311B2 (en) * | 2002-03-06 | 2010-06-29 | Mako Surgical Corp. | System and method for interactive haptic positioning of a medical device |
US7839383B2 (en) * | 2004-08-27 | 2010-11-23 | Lenovo (Beijing) Limited | Wearable signal input apparatus for data processing system |
Family Cites Families (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4221975A (en) * | 1978-04-19 | 1980-09-09 | Touch Activated Switch Arrays, Inc. | Touch activated controller and method |
US5191197A (en) * | 1988-05-11 | 1993-03-02 | Symbol Technologies, Inc. | Arm mounted scanner actuatable by hand movement |
US5305181A (en) * | 1989-05-15 | 1994-04-19 | Norand Corporation | Arm or wrist mounted terminal with a flexible housing |
US5230623A (en) * | 1991-12-10 | 1993-07-27 | Radionics, Inc. | Operating pointer with interactive computergraphics |
US5416310A (en) * | 1993-05-28 | 1995-05-16 | Symbol Technologies, Inc. | Computer and/or scanner system incorporated into a garment |
US6853293B2 (en) * | 1993-05-28 | 2005-02-08 | Symbol Technologies, Inc. | Wearable communication system |
WO1995002801A1 (en) * | 1993-07-16 | 1995-01-26 | Immersion Human Interface | Three-dimensional mechanical mouse |
US5995077A (en) * | 1994-07-20 | 1999-11-30 | The United States Of America As Represented By The Secretary Of The Navy | Portable, wearable read/write data device |
US6646541B1 (en) * | 1996-06-24 | 2003-11-11 | Computer Motion, Inc. | General purpose distributed operating room control system |
US5786804A (en) * | 1995-10-06 | 1998-07-28 | Hewlett-Packard Company | Method and system for tracking attitude |
GB9617288D0 (en) | 1996-08-16 | 1996-09-25 | Univ Dundee | Manually operated signalling apparatus |
US6847336B1 (en) * | 1996-10-02 | 2005-01-25 | Jerome H. Lemelson | Selectively controllable heads-up display system |
US6137675A (en) * | 1997-04-15 | 2000-10-24 | Perkins; Michael T. | Supportive belt system integrating computers, interfaces, and other devices |
US6057540A (en) * | 1998-04-30 | 2000-05-02 | Hewlett-Packard Co | Mouseless optical and position translation type screen pointer control for a computer system |
US6424335B1 (en) * | 1998-09-02 | 2002-07-23 | Fujitsu Limited | Notebook computer with detachable infrared multi-mode input device |
US7640007B2 (en) * | 1999-02-12 | 2009-12-29 | Fisher-Rosemount Systems, Inc. | Wireless handheld communicator in a process control environment |
DE10033680A1 (en) | 2000-07-11 | 2002-01-24 | Daniela Mayer | Command input for minimal invasive surgery using robotic device with surgeon input from arm based unit and joystick |
US6781570B1 (en) * | 2000-11-09 | 2004-08-24 | Logitech Europe S.A. | Wireless optical input device |
US7737942B2 (en) * | 2001-07-06 | 2010-06-15 | Bajramovic Mark B | Computer mouse on a glove |
US7400752B2 (en) * | 2002-02-21 | 2008-07-15 | Alcon Manufacturing, Ltd. | Video overlay system for surgical apparatus |
US6757156B2 (en) * | 2002-03-06 | 2004-06-29 | Xybernaut Corporation | Ergonomic hand held display |
AU2003245758A1 (en) | 2002-06-21 | 2004-01-06 | Cedara Software Corp. | Computer assisted system and method for minimal invasive hip, uni knee and total knee replacement |
US20040044295A1 (en) * | 2002-08-19 | 2004-03-04 | Orthosoft Inc. | Graphical user interface for computer-assisted surgery |
WO2005092230A2 (en) | 2004-03-22 | 2005-10-06 | Koninklijke Philips Electronics N.V. | Medical interventional system and method |
JP2005293505A (en) * | 2004-04-05 | 2005-10-20 | Sony Corp | Electronic equipment, input device and input method |
US7896869B2 (en) * | 2004-12-29 | 2011-03-01 | Depuy Products, Inc. | System and method for ensuring proper medical instrument use in an operating room |
KR100806029B1 (en) | 2005-04-27 | 2008-02-26 | 이문기 | computer input device using touch switch |
-
2005
- 2005-12-28 US US11/319,960 patent/US7810504B2/en not_active Expired - Fee Related
-
2006
- 2006-12-16 EP EP06256405A patent/EP1804151A1/en not_active Withdrawn
- 2006-12-27 JP JP2006352599A patent/JP2007181683A/en active Pending
- 2006-12-28 AU AU2006252285A patent/AU2006252285A1/en not_active Abandoned
-
2010
- 2010-10-01 US US12/896,309 patent/US20110022033A1/en not_active Abandoned
Patent Citations (90)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1924617A (en) * | 1931-03-21 | 1933-08-29 | Goodrich Co B F | Latex glove |
US4414537A (en) * | 1981-09-15 | 1983-11-08 | Bell Telephone Laboratories, Incorporated | Digital data entry glove interface device |
US4722625A (en) * | 1985-09-26 | 1988-02-02 | Triune Automated Painting Systems | Remote control device for powered painting system |
US7205979B2 (en) * | 1987-03-17 | 2007-04-17 | Sun Microsystems, Inc. | Computer data entry and manipulation apparatus and method |
US6885361B1 (en) * | 1987-03-24 | 2005-04-26 | Sun Microsystems, Inc. | Tactile feedback mechanism for a data processing system |
US4905001A (en) * | 1987-10-08 | 1990-02-27 | Penner Henry C | Hand-held finger movement actuated communication devices and systems employing such devices |
US4954817A (en) * | 1988-05-02 | 1990-09-04 | Levine Neil A | Finger worn graphic interface device |
US5481265A (en) * | 1989-11-22 | 1996-01-02 | Russell; David C. | Ergonomic customizeable user/computer interface devices |
US5444462A (en) * | 1991-12-16 | 1995-08-22 | Wambach; Mark L. | Computer mouse glove with remote communication |
US5332135A (en) * | 1992-11-24 | 1994-07-26 | Fletcher Stephen R | Apparatus for protecting watches and the like |
US6380923B1 (en) * | 1993-08-31 | 2002-04-30 | Nippon Telegraph And Telephone Corporation | Full-time wearable information managing device and method for the same |
US5488362A (en) * | 1993-10-01 | 1996-01-30 | Anaphase Unlimited, Inc. | Apparatus for controlling a video game |
US5489922A (en) * | 1993-12-08 | 1996-02-06 | Hewlett-Packard Company | Hand worn remote computer mouse |
US5726685A (en) * | 1994-06-30 | 1998-03-10 | Siemens Aktiengesellschaft | Input unit for a computer |
US5638092A (en) * | 1994-12-20 | 1997-06-10 | Eng; Tommy K. | Cursor control system |
US5638832A (en) * | 1995-06-07 | 1997-06-17 | Interval Research Corporation | Programmable subcutaneous visible implant |
US5911036A (en) * | 1995-09-15 | 1999-06-08 | Computer Motion, Inc. | Head cursor control interface for an automated endoscope system for optimal positioning |
US6184804B1 (en) * | 1996-01-26 | 2001-02-06 | Orang-Otang Computers, Inc. | Key palette |
US20030209604A1 (en) * | 1996-01-26 | 2003-11-13 | Harrison Shelton E. | Wearable computing system, method and device |
US5764164A (en) * | 1997-02-07 | 1998-06-09 | Reality Quest Corp. | Ergonomic hand-attachable controller |
US5796354A (en) * | 1997-02-07 | 1998-08-18 | Reality Quest Corp. | Hand-attachable controller with direction sensing |
US6747632B2 (en) * | 1997-03-06 | 2004-06-08 | Harmonic Research, Inc. | Wireless control device |
US20020024500A1 (en) * | 1997-03-06 | 2002-02-28 | Robert Bruce Howard | Wireless control device |
US20040046736A1 (en) * | 1997-08-22 | 2004-03-11 | Pryor Timothy R. | Novel man machine interfaces and applications |
US6707447B1 (en) * | 1997-12-04 | 2004-03-16 | Richard Goranowski | Therapeutic and computer input gauntlet |
US6696986B1 (en) * | 1998-03-16 | 2004-02-24 | Shelton E. Harrison, Jr. | Key palette improvements |
US20060113338A1 (en) * | 1998-03-16 | 2006-06-01 | Harrison Shelton E Jr | Key palette improvements |
US6154199A (en) * | 1998-04-15 | 2000-11-28 | Butler; Craig L. | Hand positioned mouse |
US6450978B1 (en) * | 1998-05-28 | 2002-09-17 | Orthosoft, Inc. | Interactive computer-assisted surgical system and method thereof |
US6573883B1 (en) * | 1998-06-24 | 2003-06-03 | Hewlett Packard Development Company, L.P. | Method and apparatus for controlling a computing device with gestures |
US6304840B1 (en) * | 1998-06-30 | 2001-10-16 | U.S. Philips Corporation | Fingerless glove for interacting with data processing system |
US6141643A (en) * | 1998-11-25 | 2000-10-31 | Harmon; Steve | Data input glove having conductive finger pads and thumb pad, and uses therefor |
US20020015022A1 (en) * | 2000-05-29 | 2002-02-07 | Moshe Ein-Gal | Wireless cursor control |
US6744420B2 (en) * | 2000-06-01 | 2004-06-01 | Olympus Optical Co., Ltd. | Operation input apparatus using sensor attachable to operator's hand |
US20060214912A1 (en) * | 2000-07-01 | 2006-09-28 | Miller Stephen S | Apparatus for remotely controlling computers and other electronic appliances/devices using a combination of voice commands and finger movements |
US7355584B2 (en) * | 2000-08-18 | 2008-04-08 | International Business Machines Corporation | Projector and camera arrangement with shared optics and optical marker for use with whiteboard systems |
US20030069897A1 (en) * | 2000-10-10 | 2003-04-10 | Roy Stephen C. | Systems and methods for enhancing the viewing of medical images |
US20040051694A1 (en) * | 2001-01-03 | 2004-03-18 | Jonas Backman | Portable device for inputting control signals to a peripheral unit, and use of such a device |
US20020101401A1 (en) * | 2001-01-29 | 2002-08-01 | Mehran Movahed | Thumb mounted function and cursor control device for a computer |
US7547307B2 (en) * | 2001-02-27 | 2009-06-16 | Smith & Nephew, Inc. | Computer assisted knee arthroplasty instrumentation, systems, and processes |
US20020147455A1 (en) * | 2001-02-27 | 2002-10-10 | Carson Christopher P. | Total knee arthroplasty systems and processes |
US20020198451A1 (en) * | 2001-02-27 | 2002-12-26 | Carson Christopher P. | Surgical navigation systems and processes for high tibial osteotomy |
US20030069591A1 (en) * | 2001-02-27 | 2003-04-10 | Carson Christopher Patrick | Computer assisted knee arthroplasty instrumentation, systems, and processes |
US20020175894A1 (en) * | 2001-03-06 | 2002-11-28 | Vince Grillo | Hand-supported mouse for computer input |
US20020140674A1 (en) * | 2001-03-13 | 2002-10-03 | Canon Kabushiki Kaisha | Position/posture sensor or marker attachment apparatus |
US6621483B2 (en) * | 2001-03-16 | 2003-09-16 | Agilent Technologies, Inc. | Optical screen pointing device with inertial properties |
US20020130839A1 (en) * | 2001-03-16 | 2002-09-19 | Hugh Wallace | Optical screen pointing device with inertial properties |
US20020167422A1 (en) * | 2001-04-27 | 2002-11-14 | Jacques Andre | Device for controlling a three-dimensional movement |
US20020186200A1 (en) * | 2001-06-08 | 2002-12-12 | David Green | Method and apparatus for human interface with a computer |
US7012593B2 (en) * | 2001-06-15 | 2006-03-14 | Samsung Electronics Co., Ltd. | Glove-type data input device and sensing method thereof |
US20030011568A1 (en) * | 2001-06-15 | 2003-01-16 | Samsung Electronics Co., Ltd. | Glove-type data input device and sensing method thereof |
US20030137489A1 (en) * | 2001-07-06 | 2003-07-24 | Bajramovic Mark B. | Computer mouse on a glove |
US20030006962A1 (en) * | 2001-07-06 | 2003-01-09 | Bajramovic Mark B. | Computer mouse on a glove |
US7057604B2 (en) * | 2001-07-06 | 2006-06-06 | Mikamed Health Technologies Inc. | Computer mouse on a glove |
US20030011573A1 (en) * | 2001-07-16 | 2003-01-16 | Samsung Electronics Co., Ltd. | Information input method using wearable information input device |
US7193606B2 (en) * | 2001-08-03 | 2007-03-20 | Nokia Corporation | Wearable electronic device |
US6850224B2 (en) * | 2001-08-27 | 2005-02-01 | Carba Fire Technologies, Inc. | Wearable ergonomic computer mouse |
US20040263358A1 (en) * | 2001-09-14 | 2004-12-30 | Fredrik Madsen | Portable unit for inputting signals to a peripheral unit, and use of such a unit |
US20050052291A1 (en) * | 2001-10-30 | 2005-03-10 | Perific Ab | Portable data input device and use of such a device |
US7747311B2 (en) * | 2002-03-06 | 2010-06-29 | Mako Surgical Corp. | System and method for interactive haptic positioning of a medical device |
US20050200602A1 (en) * | 2002-05-14 | 2005-09-15 | Christer Laurell | Control arrangement for a cursor |
US20030214481A1 (en) * | 2002-05-14 | 2003-11-20 | Yongming Xiong | Finger worn and operated input device and method of use |
US20030227437A1 (en) * | 2002-06-05 | 2003-12-11 | Ramirez Nohl W. | Computer pointing device and utilization system |
US20040012564A1 (en) * | 2002-07-11 | 2004-01-22 | Frank Zngf | Glove mouse with virtual tracking ball |
US20040012559A1 (en) * | 2002-07-17 | 2004-01-22 | Kanazawa University | Input device |
US7038658B2 (en) * | 2002-07-17 | 2006-05-02 | Kanazawa University | Input device |
US20040012557A1 (en) * | 2002-07-18 | 2004-01-22 | Sony Computer Entertainment Inc. | Hand-held computer interactive device |
US20070075966A1 (en) * | 2002-07-18 | 2007-04-05 | Sony Computer Entertainment Inc. | Hand-held computer interactive device |
US7161579B2 (en) * | 2002-07-18 | 2007-01-09 | Sony Computer Entertainment Inc. | Hand-held computer interactive device |
US7623115B2 (en) * | 2002-07-27 | 2009-11-24 | Sony Computer Entertainment Inc. | Method and apparatus for light input device |
US20040207597A1 (en) * | 2002-07-27 | 2004-10-21 | Sony Computer Entertainment Inc. | Method and apparatus for light input device |
US20040032346A1 (en) * | 2002-08-19 | 2004-02-19 | Samsung Electro-Mechanics Co., Ltd. | Information input device, information processing device and information input method |
US20040046734A1 (en) * | 2002-09-25 | 2004-03-11 | Hart Timothy O. | Thumb-retained stylus |
US7405725B2 (en) * | 2003-01-31 | 2008-07-29 | Olympus Corporation | Movement detection device and communication apparatus |
US20040164880A1 (en) * | 2003-02-24 | 2004-08-26 | Nguyen Quang Q. | Wearable data input device employing wrist and finger movements |
US20050035942A1 (en) * | 2003-03-21 | 2005-02-17 | Ruiz David M. | Hand-held on-screen control device |
US20060214000A1 (en) * | 2003-04-07 | 2006-09-28 | Silverbrook Research Pty Ltd | Wearable product reading device |
US7042438B2 (en) * | 2003-09-06 | 2006-05-09 | Mcrae Michael William | Hand manipulated data apparatus for computers and video games |
US7427979B2 (en) * | 2003-09-08 | 2008-09-23 | Samsung Electronics Co., Ltd. | Pointing apparatus and method |
US20050151722A1 (en) * | 2004-01-14 | 2005-07-14 | Xerox Corporation | Methods and systems for collecting and generating ergonomic data utilizing an electronic portal |
US20050234332A1 (en) * | 2004-01-16 | 2005-10-20 | Murphy Stephen B | Method of computer-assisted ligament balancing and component placement in total knee arthroplasty |
US20050174326A1 (en) * | 2004-01-27 | 2005-08-11 | Samsung Electronics Co., Ltd. | Method of adjusting pointing position during click operation and 3D input device using the same |
US20050267353A1 (en) * | 2004-02-04 | 2005-12-01 | Joel Marquart | Computer-assisted knee replacement apparatus and method |
US20050184884A1 (en) * | 2004-02-25 | 2005-08-25 | Samsung Electronics Co., Ltd. | Spatial information input apparatus and method for recognizing information-completion signal from a plurality of concurrent spatial motions |
US20050199250A1 (en) * | 2004-03-11 | 2005-09-15 | Green John M.Ii | System for determining a position of a point on an object |
US20060001646A1 (en) * | 2004-07-02 | 2006-01-05 | Wei Hai | Finger worn and operated input device |
US20060012567A1 (en) * | 2004-07-13 | 2006-01-19 | Todd Sicklinger | Minature optical mouse and stylus |
US7839383B2 (en) * | 2004-08-27 | 2010-11-23 | Lenovo (Beijing) Limited | Wearable signal input apparatus for data processing system |
US20060119578A1 (en) * | 2004-11-11 | 2006-06-08 | Thenkurussi Kesavadas | System for interfacing between an operator and a virtual object for computer aided design applications |
US20080174549A1 (en) * | 2007-01-23 | 2008-07-24 | Chi Wen Chen | Wireless cursor control device with image-tracking |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110278956A1 (en) * | 2010-05-11 | 2011-11-17 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | System including wearable power receiver and wearable power-output device |
US8427014B2 (en) * | 2010-05-11 | 2013-04-23 | The Invention Science Fund I, Llc | System including wearable power receiver and wearable power-output device |
US8941560B2 (en) | 2011-09-21 | 2015-01-27 | Google Inc. | Wearable computer with superimposed controls and instructions for external device |
US9678654B2 (en) | 2011-09-21 | 2017-06-13 | Google Inc. | Wearable computer with superimposed controls and instructions for external device |
US11262841B2 (en) | 2012-11-01 | 2022-03-01 | Eyecam Llc | Wireless wrist computing and control device and method for 3D imaging, mapping, networking and interfacing |
WO2019079790A1 (en) * | 2017-10-21 | 2019-04-25 | Eyecam, Inc | Adaptive graphic user interfacing system |
US11314399B2 (en) | 2017-10-21 | 2022-04-26 | Eyecam, Inc. | Adaptive graphic user interfacing system |
Also Published As
Publication number | Publication date |
---|---|
US20070200863A1 (en) | 2007-08-30 |
JP2007181683A (en) | 2007-07-19 |
US7810504B2 (en) | 2010-10-12 |
AU2006252285A1 (en) | 2007-07-12 |
EP1804151A1 (en) | 2007-07-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20110022033A1 (en) | System and Method for Wearable User Interface in Computer Assisted Surgery | |
US12133692B2 (en) | Systems and methods for implantation of spinal plate | |
US20090046146A1 (en) | Surgical communication and control system | |
JP7547241B2 (en) | Ensure correct selection of physical instrument attachments by displaying a virtual model of the proposed instrument attachment | |
KR102060390B1 (en) | System for establishing virtual constraint boundaries | |
EP2524289B1 (en) | Controlling and/or operating a medical device by means of a light pointer | |
JP2021090753A (en) | Augmented reality headset with varied opacity for navigated robotic surgery | |
US20100013812A1 (en) | Systems for Controlling Computers and Devices | |
JP2021176521A (en) | Extended reality headset camera system for computer-aided navigation in surgery | |
JP5642931B2 (en) | Apparatus in a dental environment and method for controlling devices belonging to the apparatus | |
JP2021171657A (en) | Alignment of surgical instruments with reference arrays tracked by cameras in augmented reality headsets for assisted navigation during surgery | |
US20050279368A1 (en) | Computer assisted surgery input/output systems and processes | |
JP2017534985A (en) | Control device based on foot gesture | |
JP2008529707A (en) | Automatic control of medical equipment | |
WO2007137093A2 (en) | Systems and methods for a hands free mouse | |
US10828119B2 (en) | RFID medical device control interface | |
JP2021122743A (en) | Extended Reality Instrument Interaction Zone for Navigated Robot Surgery | |
CN218899676U (en) | Operation simulation equipment based on virtual reality | |
Weber et al. | Application of different visualization concepts in the navigated image viewer |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: DEPUY PRODUCTS, INC., INDIANA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GUZMAN, JOSE F.;REEL/FRAME:025080/0135 Effective date: 20051227 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE |