[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN111831119A - Eyeball tracking method and device, storage medium and head-mounted display equipment - Google Patents

Eyeball tracking method and device, storage medium and head-mounted display equipment Download PDF

Info

Publication number
CN111831119A
CN111831119A CN202010664802.8A CN202010664802A CN111831119A CN 111831119 A CN111831119 A CN 111831119A CN 202010664802 A CN202010664802 A CN 202010664802A CN 111831119 A CN111831119 A CN 111831119A
Authority
CN
China
Prior art keywords
vision sensor
sensor module
dynamic vision
imaging
display screen
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010664802.8A
Other languages
Chinese (zh)
Inventor
杜鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority to CN202010664802.8A priority Critical patent/CN111831119A/en
Publication of CN111831119A publication Critical patent/CN111831119A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0187Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Position Input By Displaying (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the application discloses an eyeball tracking method, an eyeball tracking device, a storage medium and head-mounted display equipment, wherein event stream data which is shot and output by a dynamic vision sensor module is collected; acquiring imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data; calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information; according to the coordinate mapping relation and the imaging position coordinate, the fixation position coordinate of the eyeball on the display screen is obtained through calculation, and based on the coordinate mapping relation and the imaging position coordinate, when the user changes the fixation direction, the dynamic vision sensor can perform imaging according to the movement of the eyeball and calculate the fixation position coordinate of the eyeball on the display screen.

Description

Eyeball tracking method and device, storage medium and head-mounted display equipment
Technical Field
The application relates to the technical field of electronic equipment, in particular to an eyeball tracking method and device, a storage medium and head-mounted display equipment.
Background
With the development of human-computer interaction technology, the eyeball tracking technology is more and more widely applied in the field of human-computer interaction. The eyeball tracking technology is a technology for positioning the sight of human eyes, integrates knowledge in subject fields such as mechanics, electronics and optics, and has good application prospects in the fields such as artificial intelligence and machine vision, for example, fatigue monitoring, virtual reality, communication auxiliary tools and the like.
In a conventional eyeball tracking technology, sensing acquisition of an eyeball tracking image is generally realized by adopting an infrared supplementary lighting emitting and camera receiving mode, but the mode has the defect of poor real-time performance.
Disclosure of Invention
The embodiment of the application provides an eyeball tracking method, an eyeball tracking device, a storage medium and head-mounted display equipment, which can improve the instantaneity of eyeball tracking.
In a first aspect, an embodiment of the present application provides an eyeball tracking method, including:
acquiring event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user;
acquiring imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data;
calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information;
and calculating to obtain the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
In a second aspect, an embodiment of the present application further provides an eyeball tracking device, including:
the data acquisition module is used for acquiring event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user;
the imaging acquisition module is used for acquiring the imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data;
the first calculation module is used for calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information;
and the second calculation module is used for calculating and obtaining the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
In a third aspect, embodiments of the present application further provide a storage medium having a computer program stored thereon, where the computer program is executed on a computer, so as to enable the computer to execute the eyeball tracking method provided in any embodiment of the present application.
In a fourth aspect, embodiments of the present application further provide a head-mounted display device, including a processor and a memory, where the memory has a computer program, and the processor is configured to execute the eye tracking method according to any of the embodiments of the present application by calling the computer program.
According to the technical scheme provided by the embodiment of the application, the eyes of a user are shot through the dynamic vision sensor module, event stream data output by the dynamic vision sensor is collected, imaging information of eyeballs in the dynamic vision sensor module is obtained according to the event stream data, imaging position coordinates of the eyeballs are calculated according to the imaging information, the fixation position coordinates of the eyeballs on the display screen are calculated according to the coordinate mapping relation between the display screen and the dynamic vision sensor and the imaging position coordinates, when the user fixes the display screen and changes the fixation direction, the dynamic vision sensor can carry out imaging according to the movement of the eyeballs, the change of the fixation direction of the user is determined according to the change of the imaging position, and the fixation position coordinates of the eyeballs on the display screen are calculated, so that the scheme only captures the characteristic of target movement change, the method greatly reduces the redundancy of information, reduces the data volume, has the characteristic of low time delay, can efficiently acquire the fixation position coordinate, and improves the instantaneity of eyeball tracking.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a first flowchart of an eyeball tracking method according to an embodiment of the present disclosure.
Fig. 2 is an imaging schematic diagram of a dynamic vision sensor module according to an embodiment of the present application.
Fig. 3 is a schematic structural diagram of an eyeball tracking device according to an embodiment of the present application.
Fig. 4 is a schematic structural diagram of a head-mounted display device according to an embodiment of the present disclosure.
Fig. 5 is a second structural schematic diagram of a head-mounted display device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application. It is to be understood that the embodiments described are only a few embodiments of the present application and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without inventive step, are within the scope of the present application.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the application. The appearances of the phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. It is explicitly and implicitly understood by one skilled in the art that the embodiments described herein can be combined with other embodiments.
An implementation subject of the eyeball tracking method may be the eyeball tracking device provided in the embodiment of the application, or a head-mounted display device integrated with the eyeball tracking device, wherein the eyeball tracking device may be implemented in a hardware or software manner. The head-mounted display device may be an AR (augmented Reality) glasses, a VR (Virtual Reality) glasses, or the like.
Referring to fig. 1, fig. 1 is a first flowchart illustrating an eyeball tracking method according to an embodiment of the present disclosure. The specific process of the eyeball tracking method provided by the embodiment of the application can be as follows:
in 101, event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user is collected.
For convenience of explanation, the following describes embodiments of the present application with AR glasses integrated with an eye tracking device as an execution body. The augmented reality is a technology for increasing the perception of a user to the real world through information provided by a system, the augmented reality overlays contents such as virtual objects, scenes or prompt information generated by a computer to the real scene for display so as to enhance or modify the perception of the real world environment or data representing the real world environment, and the user can observe the effect of the augmented reality or mixed reality after the contents such as the virtual objects and the real world are overlaid by wearing head-mounted display equipment such as AR glasses in the real world environment. In the virtual reality technology, a virtual environment is simulated through a computer so as to give people an environmental immersion feeling, and a user can view a virtual object by wearing head-mounted display equipment such as VR glasses and the like.
AR glasses and VR glasses are all near-to-eye display device, and its control mode is not so convenient as the control to portable electronic equipment such as smart mobile phone, and if these equipment have possessed eyeball tracking function, the user can come the interactive operation of controlgear through the rotation of eyes, improves interactive efficiency. For example, studies have shown that the focus resolution area of human eyes can only be 10 ° of the pupil facing position, within which the user can clearly distinguish the displayed content, and outside which the user does not recognize the content in detail. Therefore, if the head-mounted display device tracks the eye gaze focusing position of the user, i.e., the gaze position, the gaze focusing position may be heavily rendered, while other positions may reduce the amount of rendering. Therefore, the image rendering workload of the GPU is reduced, the calculation amount is favorably reduced, and the load and the power consumption of a system are reduced.
Based on this, this scheme sets up a dynamic vision sensor module on head mounted display device, and this dynamic vision sensor module can be according to following position setting on head mounted display device: when the user wears the head-mounted display equipment, the eyeballs are located in the field of view of the dynamic vision sensor module.
The dynamic vision sensor module comprises a dynamic vision sensor and a lens module, the lens module shoots an external object, light is converted through the lens module to form an image on a pixel array of the dynamic vision sensor, when the position of the external object changes, the position of the image on the pixel array changes along with the change, and when the position of the image moves, the light sensitivity of a corresponding pixel also changes. For example, when the motion vision sensor module shoots an eye of a user, the position of an image reflected on the pixel array of the motion vision sensor moves when the eyeball of the user rotates, and the light sensitivity of the corresponding pixel changes.
The dynamic vision sensor module is an Event-driven (Event-drive) photoelectric sensor, the output of the sensor no longer has the concept of frame, but is an Event (Event), each pixel unit in the dynamic vision sensor module independently senses the change of light intensity in space, when the light sensitivity of the pixel changes and the pixel point with the light intensity change exceeding a threshold value is considered as an activated pixel point, the dynamic vision sensor module packs and encodes and outputs the information of the row and column positions, the time stamps and the like of the activated pixel point, the related information of the activated pixel points is the Event, and a plurality of activated pixel points form an Event stream.
When the position of an external object moves, the corresponding pixel position in the pixel array of the dynamic vision sensor module is changed in light intensity, and then the corresponding event is output, namely, the dynamic vision sensor module only captures the characteristic of target motion change, so that the redundancy of information is greatly reduced, the data volume is reduced, the computing resource and the bandwidth can be saved in practical application, and the dynamic vision sensor module is very suitable for being used in the field of machine vision. The scheme of this application embodiment utilizes dynamic vision sensor module to have high dynamic range, be good at catching high-speed object, can filter characteristics such as background and carry out eyeball tracking. And compared with the conventional mode of adopting infrared light supplement emission and camera receiving, the dynamic vision sensor module adopts a passive acquisition mode, does not need to actively emit infrared rays to irradiate the eyeballs of people, reduces safety risks, has the characteristics of low time delay and low power consumption, can save electric quantity, can efficiently acquire fixation position coordinates, and improves the real-time performance of eyeball tracking
In addition, it should be noted that, when the resolution of the dynamic vision sensor module is higher, the imaging of the moving object that can be obtained is clearer, the outline of the object that can be identified is clearer, and the details of the object can be embodied more, so that the scheme of the application can select the dynamic vision sensor module with higher resolution when applied, so as to improve the accuracy of eyeball tracking.
In this application embodiment, when the eyeball that is located the visual field scope of dynamic vision sensor module rotated, this kind of motion information was caught to the dynamic vision sensor module, gathered the event stream data of dynamic vision sensor module output.
At 102, imaging information of the eyeball in the dynamic vision sensor module is obtained according to the event flow data.
After the event stream data output by the dynamic vision sensor is acquired, the event stream data is extracted and calculated to obtain the imaging information of the eyeballs.
When the eyeball rotates, the light intensity of the corresponding pixel in the pixel array of the dynamic vision sensor module changes, and a corresponding event is output, wherein the output event comprises the following information { x, y, p, t } of four dimensions. Where t denotes an activation timestamp, x denotes a row position of an activated pixel, y denotes a column position of an activated pixel, and p denotes an activation polarity, where p-1 denotes an "on" polarity, and p-1 denotes an "off" polarity, and when an increase in light intensity sensed by a pixel exceeds a threshold, the pixel is activated, generating an "on" event, and when a decrease in light intensity sensed by a pixel exceeds a threshold, the pixel is activated, generating an "off" event.
And each activated pixel in the pixel array of the dynamic vision sensor module outputs a corresponding event according to the format to form event stream data. The event stream selects the activated pixels according to the principle of 'activation first and reading first' and reads information. And the activation polarities are not distinguished, and the detected activation pixel points are identified.
In addition, when the user wears the head-mounted display device, the parts except the eyeballs are static relative to the head-mounted display device, when the user changes the watching direction of the display screen, the eyeballs can rotate, and therefore the eyeball part of the eyes of the user is imaged on the dynamic vision sensor module according to the event stream data.
For example, in one embodiment, "acquiring imaging information of an eye on the dynamic data sensor from the event stream data" includes: determining a target tracking time point from the event stream data; and acquiring a plurality of activated pixel points corresponding to the target tracking time point from the event stream data, and forming imaging information of eyeballs on the dynamic data sensor by the plurality of activated pixel points.
Since the image of the eyeball part on the dynamic vision sensor module corresponds to a plurality of activated pixels, at the time t, the coordinates of the plurality of corresponding activated pixels, (x0, y0), (x1, y1) … … (xn, yn) can be obtained from the event stream data.
In addition, since the eyeball may be continuously rotated, event stream data composed of events may be continuously output, for example, an event may be output once at each interval t0, when new event data is received, a time point closest to the current time is determined from the event stream data as a target tracking time point, in such a manner that as time goes by, the target tracking time point gradually goes by, event data corresponding to each target tracking time point is obtained, an eyeball gaze position corresponding to each target tracking time point is calculated from the event data corresponding to each target tracking time point, and dynamic tracking of the eyeball gaze position is further achieved.
Referring to fig. 2, fig. 2 is an imaging schematic diagram of a dynamic vision sensor module according to an embodiment of the present application. Each square box corresponds to a pixel on the pixel array of the dynamic vision sensor module, and fig. 2 is a 7 × 7 pixel array, where a gray pixel is an activated pixel point and represents a detected moving object. Where t2 is t1+ t0, and t0 is the minimum time interval for detecting the dynamic vision sensor module, i.e. the minimum time interval of the event stream. From time t1 to time t2, the position of the activated pixel points on the pixel array as a result of the motion of the object changes as shown in FIG. 2. That is, the image obtained at the time t1 is composed of t1 gray pixels, and the image obtained at the time t2 is composed of t2 gray pixels.
It will be appreciated that the example of figure 2 simplifies the pixel array and the size of the image for the convenience of the reader, and in practice the number of pixels in the pixel array is much greater than that shown in figure 2.
In 103, calculating the coordinates of the imaging position of the eyeball on the dynamic vision sensor module according to the imaging information.
And after the imaging information of the eyeball is acquired, calculating the imaging position coordinate of the eyeball on the dynamic vision sensor module according to the imaging information. For example, in an embodiment, the calculating the coordinates of the imaging position of the eyeball on the dynamic vision sensor module according to the imaging information includes: carrying out feature point identification on the imaging information, and obtaining target activation pixel points forming an iris outline from the plurality of activation pixel points; and calculating the central position coordinates of the target activation pixel points, and taking the central position coordinates as the imaging position coordinates of the eyeballs.
In the embodiment, taking the time t2 as an example, all the activation pixels corresponding to the time t2 are obtained, feature extraction is performed on the activation pixels, target activation pixels capable of representing and forming an iris outline are obtained, and the center position coordinates of the target activation pixels are calculated.
In an embodiment, the target activation pixel points constituting the iris outline in the imaging information may be identified according to a pre-trained preset convolutional neural network, wherein the preset convolutional neural network may be obtained by training a sample imaging image, and the sample imaging image has labeling information of the activation pixel points constituting the iris outline.
And 104, calculating to obtain the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
After acquiring the coordinates of the imaging position of the eyeball on the dynamic vision sensor module, performing coordinate conversion processing according to a coordinate mapping relation between the display screen and the dynamic vision sensor module, wherein the coordinate mapping relation can be stored in the head-mounted display device after calibration when the head-mounted display device leaves a factory, or can be calibrated by a user according to needs.
For example, in an embodiment, before acquiring event stream data output by the dynamic vision sensor module for shooting eyes of the user, the method further includes: displaying calibration information on the display screen, and outputting prompt information, wherein the prompt information is used for prompting a user to watch the calibration information; acquiring a first position coordinate of an eyeball when a user watches the calibration information according to the dynamic vision sensor module; acquiring a plurality of first position coordinates according to the calibration information of a plurality of different positions; and calculating to obtain a coordinate mapping relation between the display screen and the dynamic vision sensor module according to the respective second position coordinates of the plurality of calibration information and the plurality of first position coordinates.
In this embodiment, the user, when using the device for the first time, controls the device to enter a calibration mode, which in turn displays a plurality of calibration information, e.g., calibration icons, at different locations on the display screen. Wherein the user is prompted to gaze at each calibration icon as it is displayed. After a user watches the calibration icon, a confirmation instruction can be triggered, after the equipment receives the confirmation instruction, the first position coordinates of eyeballs when the user watches the calibration information are obtained according to the dynamic vision sensor module, each calibration information is provided with a corresponding second position coordinate, when the calibration is completed according to the plurality of calibration information displayed at different positions of the display screen according to the mode, a plurality of groups of first position coordinates and second position coordinates can be obtained, and the coordinate mapping relation between the display screen and the dynamic vision sensor module can be obtained through calculation according to the plurality of groups of first position coordinates and second position coordinates.
Wherein, in some embodiments, after calculating the gaze location coordinates of the eye on the display screen, the method further comprises: and controlling the display of the display screen according to the target display content corresponding to the fixation position coordinates on the display screen.
After the gaze position coordinates of the user gazing at the display screen are obtained, the display of the display screen can be controlled in various ways based on the gaze position coordinates, for example, the target display content corresponding to the gaze position coordinates of the user is subjected to key rendering and the like.
In particular implementation, the present application is not limited by the execution sequence of the described steps, and some steps may be performed in other sequences or simultaneously without conflict.
As can be seen from the above, the eyeball tracking method provided in the embodiment of the present application captures an eye of a user through the dynamic vision sensor module, collects event stream data output by the dynamic vision sensor, obtains imaging information of the eyeball in the dynamic vision sensor module according to the event stream data, calculates an imaging position coordinate of the eyeball according to the imaging information, calculates a gaze position coordinate of the eyeball on the display screen according to a coordinate mapping relationship between the display screen and the dynamic vision sensor and the imaging position coordinate, and according to the scheme, when the user gazes at the display screen and changes a gaze direction, the dynamic vision sensor can image according to a movement of the eyeball, determines a change of the gaze direction of the user according to a change of the imaging position, and further calculates the gaze position coordinate of the eyeball on the display screen, so that the scheme captures only a characteristic of a change of a target movement, the method greatly reduces the redundancy of information, reduces the data volume, has the characteristic of low time delay, can efficiently acquire the fixation position coordinate, and improves the instantaneity of eyeball tracking.
An eye tracking device is also provided in one embodiment. Referring to fig. 3, fig. 3 is a schematic structural diagram of an eyeball tracking device 300 according to an embodiment of the present application. The eyeball tracking apparatus 300 is applied to a head-mounted display device, and the eyeball tracking apparatus 300 includes a data acquisition module 301, an imaging acquisition module 302, a first calculation module 303 and a second calculation module 304, as follows:
the data acquisition module 301 is configured to acquire event stream data that is output by the dynamic vision sensor module when the eyes of the user are photographed;
an imaging acquisition module 302, configured to acquire imaging information of an eyeball in the dynamic vision sensor module according to the event stream data;
the first calculating module 303 is configured to calculate coordinates of an imaging position of an eyeball on the dynamic vision sensor module according to the imaging information;
and a second calculating module 304, configured to calculate, according to the coordinate mapping relationship between the display screen and the dynamic vision sensor module and the imaging position coordinate, a gaze position coordinate of the eyeball on the display screen.
In some embodiments, the imaging acquisition module 302 is further configured to determine a target tracking time point from the event stream data;
and acquiring a plurality of activated pixel points corresponding to the target tracking time point from the event stream data, and forming imaging information of eyeballs on the dynamic data sensor by the plurality of activated pixel points.
In some embodiments, the first calculating module 303 is further configured to perform feature point identification on the imaging information, and obtain target activated pixel points forming an iris outline from the plurality of activated pixel points; and calculating the central position coordinates of the target activation pixel points, and taking the central position coordinates as the imaging position coordinates of the eyeballs.
In some embodiments, the eye tracking apparatus 300 further comprises:
and the display control module is used for controlling the display of the display screen according to the target display content corresponding to the fixation position coordinates on the display screen.
In some embodiments, the eye tracking apparatus 300 further comprises:
the coordinate correction module is used for displaying calibration information on the display screen and outputting prompt information, and the prompt information is used for prompting a user to watch the calibration information;
acquiring a first position coordinate of an eyeball when a user watches the calibration information according to the dynamic vision sensor module;
acquiring a plurality of first position coordinates according to the calibration information of a plurality of different positions;
and calculating to obtain a coordinate mapping relation between the display screen and the dynamic vision sensor module according to the respective second position coordinates of the plurality of calibration information and the plurality of first position coordinates.
It should be noted that the eyeball tracking device provided in the embodiment of the present application and the eyeball tracking method in the above embodiment belong to the same concept, and any method provided in the embodiment of the eyeball tracking method can be implemented by the eyeball tracking device, and the specific implementation process thereof is described in the embodiment of the eyeball tracking method, and is not described herein again.
As can be seen from the above, the eyeball tracking apparatus provided in the embodiment of the present application captures an eye of a user through the dynamic vision sensor module, collects event stream data output by the dynamic vision sensor, obtains imaging information of the eyeball in the dynamic vision sensor module according to the event stream data, calculates an imaging position coordinate of the eyeball according to the imaging information, calculates a gaze position coordinate of the eyeball on the display screen according to a coordinate mapping relationship between the display screen and the dynamic vision sensor and the imaging position coordinate, and according to the scheme, when the user gazes at the display screen and changes a gaze direction, the dynamic vision sensor can image according to a movement of the eyeball, determines a change of the gaze direction of the user according to a change of the imaging position, and further calculates the gaze position coordinate of the eyeball on the display screen, so that the scheme captures only a characteristic of a change of a target movement, the method greatly reduces the redundancy of information, reduces the data volume, has the characteristic of low time delay, can efficiently acquire the fixation position coordinate, and improves the instantaneity of eyeball tracking.
The embodiment of the application also provides head-mounted display equipment. The head mounted display device may be AR glasses or the like. Referring to fig. 4, fig. 4 is a schematic view illustrating a first structure of a head-mounted display device according to an embodiment of the present disclosure. The head mounted display device 400 comprises a processor 401 and a memory 402, and a display screen 403 connected thereto. The processor 401 and the memory 402 are electrically connected to the display 403, and the processor 401 is electrically connected to the memory 402.
The processor 401 is a control center of the head-mounted display apparatus 400, connects various parts of the entire head-mounted display apparatus using various interfaces and lines, and performs various functions of the head-mounted display apparatus and processes data by running or calling a computer program stored in the memory 402 and calling data stored in the memory 402, thereby performing overall monitoring of the head-mounted display apparatus.
Memory 402 may be used to store computer programs and data. The memory 402 stores computer programs containing instructions executable in the processor. The computer program may constitute various functional modules. The processor 401 executes various functional applications and data processing by calling a computer program stored in the memory 402.
The display screen 403 may be used to display information input by or provided to the user as well as various graphical user interfaces of the head mounted display device, which may be comprised of images, text, icons, video, and any combination thereof. Wherein, in some embodiments, the head mounted display device may be AR glasses and the display screen 403 is a lens of the AR glasses.
In this embodiment, the processor 401 in the head-mounted display device 400 loads instructions corresponding to one or more processes of the computer program into the memory 402 according to the following steps, and the processor 401 runs the computer program stored in the memory 402, so as to implement various functions:
acquiring event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user;
acquiring imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data;
calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information;
and calculating to obtain the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
In some embodiments, please refer to fig. 5, and fig. 5 is a second structural diagram of a head-mounted display device according to an embodiment of the present disclosure. The head mounted display device 400 further includes: radio frequency circuit 404, display screen 403, control circuit 405, input unit 406, audio circuit 407, sensor 408, power supply 409 and dynamic vision sensor module 410. The processor 401 is electrically connected to the rf circuit 404, the display 403, the control circuit 405, the input unit 406, the audio circuit 407, the sensor 408, and the power source 409.
The radio frequency circuit 404 is used for transceiving radio frequency signals to communicate with a network device or other head-mounted display device through wireless communication.
The control circuit 405 is electrically connected to the display 403, and is configured to control the display 403 to display information.
The input unit 406 may be used to receive input numbers, character information, or user characteristic information (e.g., fingerprint), and to generate keyboard, mouse, joystick, optical, or trackball signal inputs related to user settings and function control. The input unit 406 may include a fingerprint recognition module.
Audio circuitry 407 may provide an audio interface between the user and the head mounted display device through speakers, microphones. Wherein the audio circuit 407 comprises a microphone. The microphone is electrically connected to the processor 401. The microphone is used for receiving voice information input by a user.
The sensor 408 is used to collect external environmental information. The sensors 408 may include one or more of ambient light sensors, acceleration sensors, gyroscopes, etc.
The power supply 409 is used to power the various components of the head mounted display device 400. In some embodiments, the power source 409 may be logically connected to the processor 401 through a power management system, so that functions of managing charging, discharging, and power consumption are implemented through the power management system.
The dynamic vision sensor module 410 is used for collecting the eye image of the wearer, and when the user wears the head-mounted display device, the eyes of the user are located in the field of view of the dynamic vision sensor module. Therein, in some embodiments, the dynamic vision sensor module 410 may be mounted at the lower edge of the display screen. Alternatively, in some other embodiments, the head-mounted display device is AR glasses, the AR glasses include a frame, and the dynamic vision sensor module 40 may be mounted on a lower frame or a side frame of the frame.
Although not shown in the drawings, the head-mounted display device 400 may further include a bluetooth module or the like, which is not described in detail herein.
In this embodiment, the processor 401 in the head-mounted display device 400 loads instructions corresponding to one or more processes of the computer program into the memory 402 according to the following steps, and the processor 401 runs the computer program stored in the memory 402, so as to implement various functions:
acquiring event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user;
acquiring imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data;
calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information;
and calculating to obtain the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
In some embodiments, when acquiring imaging information of an eye on the dynamic data sensor from the event stream data, the processor 401 performs:
determining a target tracking time point from the event stream data;
and acquiring a plurality of activated pixel points corresponding to the target tracking time point from the event stream data, and forming imaging information of eyeballs on the dynamic data sensor by the plurality of activated pixel points.
In some embodiments, when calculating the imaging position coordinate of the eyeball on the dynamic vision sensor module according to the imaging information, the processor 401 executes:
carrying out feature point identification on the imaging information, and obtaining target activation pixel points forming an iris outline from the plurality of activation pixel points;
and calculating the central position coordinates of the target activation pixel points, and taking the central position coordinates as the imaging position coordinates of the eyeballs.
In some embodiments, after the calculating the gaze location coordinates of the eyeball on the display screen, the processor 401 further performs:
and controlling the display of the display screen according to the target display content corresponding to the fixation position coordinates on the display screen.
In some embodiments, before acquiring the event stream data output by the dynamic vision sensor module for shooting the eyes of the user, the processor 401 further performs:
displaying calibration information on the display screen, and outputting prompt information, wherein the prompt information is used for prompting a user to watch the calibration information;
acquiring a first position coordinate of an eyeball when a user watches the calibration information according to the dynamic vision sensor module;
acquiring a plurality of first position coordinates according to the calibration information of a plurality of different positions;
and calculating to obtain a coordinate mapping relation between the display screen and the dynamic vision sensor module according to the respective second position coordinates of the plurality of calibration information and the plurality of first position coordinates.
From the above, the embodiment of the present application provides a head-mounted display device, where the head-mounted display device photographs eyes of a user through a dynamic vision sensor module, collects event stream data output by the dynamic vision sensor, obtains imaging information of eyeballs in the dynamic vision sensor module according to the event stream data, calculates imaging position coordinates of the eyeballs according to the imaging information, calculates gaze position coordinates of the eyeballs on a display screen according to a coordinate mapping relationship between the display screen and the dynamic vision sensor and the imaging position coordinates, and by using the scheme, when the user gazes at the display screen and changes a gaze direction, the dynamic vision sensor can image according to movement of the eyeballs, determines a change of the gaze direction of the user according to the change of the imaging position, and further calculates the gaze position coordinates of the eyeballs on the display screen, therefore, the scheme only captures the characteristics of target motion change, greatly reduces the redundancy of information, reduces the data volume, has the characteristic of low time delay, can efficiently acquire the fixation position coordinates, and improves the real-time performance of eyeball tracking.
An embodiment of the present application further provides a storage medium, where a computer program is stored, and when the computer program runs on a computer, the computer executes the eyeball tracking method according to any one of the above embodiments.
It should be noted that, all or part of the steps in the methods of the above embodiments may be implemented by hardware related to instructions of a computer program, which may be stored in a computer-readable storage medium, which may include, but is not limited to: read Only Memory (ROM), Random Access Memory (RAM), magnetic or optical disks, and the like.
Furthermore, the terms "first", "second", and "third", etc. in this application are used to distinguish different objects, and are not used to describe a particular order. Furthermore, the terms "include" and "have," as well as any variations thereof, are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or modules is not limited to only those steps or modules listed, but rather, some embodiments may include other steps or modules not listed or inherent to such process, method, article, or apparatus.
The eyeball tracking method, device, storage medium and head-mounted display device provided by the embodiment of the application are described in detail above. The principle and the implementation of the present application are explained herein by applying specific examples, and the above description of the embodiments is only used to help understand the method and the core idea of the present application; meanwhile, for those skilled in the art, according to the idea of the present application, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present application.

Claims (10)

1. An eye tracking method, comprising:
acquiring event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user;
acquiring imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data;
calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information;
and calculating to obtain the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
2. The eye tracking method according to claim 1, wherein said obtaining imaging information of an eye on said dynamic data sensor based on said event stream data comprises:
determining a target tracking time point from the event stream data;
and acquiring a plurality of activated pixel points corresponding to the target tracking time point from the event stream data, and forming imaging information of eyeballs on the dynamic data sensor by the plurality of activated pixel points.
3. The method for tracking an eyeball according to claim 2, wherein the calculating of coordinates of an imaging position of the eyeball on the dynamic vision sensor module according to the imaging information comprises:
carrying out feature point identification on the imaging information, and obtaining target activation pixel points forming an iris outline from the plurality of activation pixel points;
and calculating the central position coordinates of the target activation pixel points, and taking the central position coordinates as the imaging position coordinates of the eyeballs.
4. The eye tracking method of claim 1, wherein after said calculating the gaze location coordinates of the eye on the display screen, further comprising:
and controlling the display of the display screen according to the target display content corresponding to the fixation position coordinates on the display screen.
5. The eye tracking method according to claim 1, wherein before acquiring the event stream data that is output by the dynamic vision sensor module for capturing the eye of the user, the method further comprises:
displaying calibration information on the display screen, and outputting prompt information, wherein the prompt information is used for prompting a user to watch the calibration information;
acquiring a first position coordinate of an eyeball when a user watches the calibration information according to the dynamic vision sensor module;
acquiring a plurality of first position coordinates according to the calibration information of a plurality of different positions;
and calculating to obtain a coordinate mapping relation between the display screen and the dynamic vision sensor module according to the respective second position coordinates of the plurality of calibration information and the plurality of first position coordinates.
6. An eye tracking device, comprising:
the data acquisition module is used for acquiring event stream data which is shot and output by the dynamic vision sensor module to the eyes of the user;
the imaging acquisition module is used for acquiring the imaging information of the eyeballs in the dynamic vision sensor module according to the event stream data;
the first calculation module is used for calculating the imaging position coordinates of the eyeballs on the dynamic vision sensor module according to the imaging information;
and the second calculation module is used for calculating and obtaining the fixation position coordinate of the eyeball on the display screen according to the coordinate mapping relation between the display screen and the dynamic vision sensor module and the imaging position coordinate.
7. The eye tracking device of claim 6, wherein said imaging acquisition module is further configured to:
determining a target tracking time point from the event stream data;
and acquiring a plurality of activated pixel points corresponding to the target tracking time point from the event stream data, and forming imaging information of eyeballs on the dynamic data sensor by the plurality of activated pixel points.
8. An eye tracking device according to claim 7, further comprising:
the coordinate correction module is used for displaying calibration information on the display screen and outputting prompt information, and the prompt information is used for prompting a user to watch the calibration information;
acquiring a first position coordinate of an eyeball when a user watches the calibration information according to the dynamic vision sensor module;
acquiring a plurality of first position coordinates according to the calibration information of a plurality of different positions;
and calculating to obtain a coordinate mapping relation between the display screen and the dynamic vision sensor module according to the respective second position coordinates of the plurality of calibration information and the plurality of first position coordinates.
9. A storage medium having stored thereon a computer program for causing a computer to perform the eye tracking method according to any one of claims 1 to 5 when the computer program runs on the computer.
10. A head-mounted display device comprises a display screen, a processor and a memory, wherein the processor and the memory are connected with the display screen, and the memory stores computer programs; the processor is adapted to perform the eye tracking method according to any one of claims 1 to 5 by calling the computer program.
CN202010664802.8A 2020-07-10 2020-07-10 Eyeball tracking method and device, storage medium and head-mounted display equipment Pending CN111831119A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010664802.8A CN111831119A (en) 2020-07-10 2020-07-10 Eyeball tracking method and device, storage medium and head-mounted display equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010664802.8A CN111831119A (en) 2020-07-10 2020-07-10 Eyeball tracking method and device, storage medium and head-mounted display equipment

Publications (1)

Publication Number Publication Date
CN111831119A true CN111831119A (en) 2020-10-27

Family

ID=72900532

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010664802.8A Pending CN111831119A (en) 2020-07-10 2020-07-10 Eyeball tracking method and device, storage medium and head-mounted display equipment

Country Status (1)

Country Link
CN (1) CN111831119A (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112148200A (en) * 2020-11-27 2020-12-29 亿昇(深圳)眼镜科技有限公司 Intelligent glasses system
CN112669344A (en) * 2020-12-24 2021-04-16 北京灵汐科技有限公司 Method and device for positioning moving object, electronic equipment and storage medium
CN113262464A (en) * 2021-04-21 2021-08-17 青岛小鸟看看科技有限公司 Dynamic change method and device of virtual reality scene and electronic equipment
CN113467619A (en) * 2021-07-21 2021-10-01 腾讯科技(深圳)有限公司 Picture display method, picture display device, storage medium and electronic equipment
CN113891002A (en) * 2021-11-12 2022-01-04 维沃移动通信有限公司 Shooting method and device
CN114646457A (en) * 2022-05-20 2022-06-21 南昌虚拟现实研究院股份有限公司 Eye tracking accuracy testing method and system
CN114760440A (en) * 2022-03-22 2022-07-15 浙江大学 Visual monitoring glasses
CN114791673A (en) * 2021-01-26 2022-07-26 精工爱普生株式会社 Display method, display device, and recording medium
CN114954265A (en) * 2022-05-25 2022-08-30 中国第一汽车股份有限公司 Intelligent visual tracking vehicle-mounted display screen device and control method
WO2022247646A1 (en) * 2021-05-25 2022-12-01 北京芯海视界三维科技有限公司 Timing controllers and display device
CN115984950A (en) * 2022-12-28 2023-04-18 北京字跳网络技术有限公司 Sight line detection method and device, electronic equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105812777A (en) * 2015-01-21 2016-07-27 成都理想境界科技有限公司 Binocular AR head-mounted display device and information display method therefor
CN110196640A (en) * 2019-05-31 2019-09-03 维沃移动通信有限公司 A kind of method of controlling operation thereof and terminal
US20200026349A1 (en) * 2018-07-19 2020-01-23 Samsung Electronics Co., Ltd System and method for hybrid eye tracker

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105812777A (en) * 2015-01-21 2016-07-27 成都理想境界科技有限公司 Binocular AR head-mounted display device and information display method therefor
US20200026349A1 (en) * 2018-07-19 2020-01-23 Samsung Electronics Co., Ltd System and method for hybrid eye tracker
CN110196640A (en) * 2019-05-31 2019-09-03 维沃移动通信有限公司 A kind of method of controlling operation thereof and terminal

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112148200A (en) * 2020-11-27 2020-12-29 亿昇(深圳)眼镜科技有限公司 Intelligent glasses system
CN112669344A (en) * 2020-12-24 2021-04-16 北京灵汐科技有限公司 Method and device for positioning moving object, electronic equipment and storage medium
CN112669344B (en) * 2020-12-24 2024-05-28 北京灵汐科技有限公司 Method and device for positioning moving object, electronic equipment and storage medium
WO2022135511A1 (en) * 2020-12-24 2022-06-30 北京灵汐科技有限公司 Method and apparatus for positioning moving object, and electronic device and storage medium
CN114791673B (en) * 2021-01-26 2023-12-15 精工爱普生株式会社 Display method, display device, and recording medium
CN114791673A (en) * 2021-01-26 2022-07-26 精工爱普生株式会社 Display method, display device, and recording medium
CN113262464A (en) * 2021-04-21 2021-08-17 青岛小鸟看看科技有限公司 Dynamic change method and device of virtual reality scene and electronic equipment
US11782505B1 (en) 2021-04-21 2023-10-10 Qingdao Pico Technology Co., Ltd. Dynamic changing method and apparatus for virtual reality scene, and electronic device
WO2022247646A1 (en) * 2021-05-25 2022-12-01 北京芯海视界三维科技有限公司 Timing controllers and display device
CN113467619A (en) * 2021-07-21 2021-10-01 腾讯科技(深圳)有限公司 Picture display method, picture display device, storage medium and electronic equipment
CN113467619B (en) * 2021-07-21 2023-07-14 腾讯科技(深圳)有限公司 Picture display method and device, storage medium and electronic equipment
CN113891002A (en) * 2021-11-12 2022-01-04 维沃移动通信有限公司 Shooting method and device
CN114760440A (en) * 2022-03-22 2022-07-15 浙江大学 Visual monitoring glasses
CN114646457B (en) * 2022-05-20 2022-08-12 南昌虚拟现实研究院股份有限公司 Eye movement tracking accuracy testing method and system
CN114646457A (en) * 2022-05-20 2022-06-21 南昌虚拟现实研究院股份有限公司 Eye tracking accuracy testing method and system
CN114954265A (en) * 2022-05-25 2022-08-30 中国第一汽车股份有限公司 Intelligent visual tracking vehicle-mounted display screen device and control method
CN115984950A (en) * 2022-12-28 2023-04-18 北京字跳网络技术有限公司 Sight line detection method and device, electronic equipment and storage medium
CN115984950B (en) * 2022-12-28 2024-03-12 北京字跳网络技术有限公司 Sight line detection method, device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN111831119A (en) Eyeball tracking method and device, storage medium and head-mounted display equipment
CN110647865B (en) Face gesture recognition method, device, equipment and storage medium
CN106873778B (en) Application operation control method and device and virtual reality equipment
CN108170279B (en) Eye movement and head movement interaction method of head display equipment
US12041380B2 (en) Event-based IR camera
CN111263066B (en) Composition guiding method, composition guiding device, electronic equipment and storage medium
CN108076290B (en) Image processing method and mobile terminal
CN115167723A (en) Wearable device capable of eye tracking
CN112911182A (en) Game interaction method, device, terminal and storage medium
KR20170100641A (en) Virtual representations of real-world objects
CN108431667A (en) Information processing apparatus, information processing method, and program
CN115209057B (en) Shooting focusing method and related electronic equipment
KR20160046495A (en) Method and device to display screen in response to event related to external obejct
CN111698564B (en) Information recommendation method, device, equipment and storage medium
US20210409626A1 (en) Patch tracking image sensor
CN113467619B (en) Picture display method and device, storage medium and electronic equipment
CN107174824B (en) Special effect information processing method and device, electronic equipment and storage medium
CN112257552B (en) Image processing method, device, equipment and storage medium
CN110263617B (en) Three-dimensional face model obtaining method and device
CN109978996B (en) Method, device, terminal and storage medium for generating expression three-dimensional model
CN112000221A (en) Method for automatically detecting skin, method for automatically guiding skin care and makeup and terminal
CN109189225A (en) Display interface method of adjustment, device, wearable device and storage medium
US11328187B2 (en) Information processing apparatus and information processing method
CN111589138B (en) Action prediction method, device, equipment and storage medium
CN112907702A (en) Image processing method, image processing device, computer equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20201027