KR20140061009A - Hybrid augmented reality using voice recognition and method the same - Google Patents
Hybrid augmented reality using voice recognition and method the same Download PDFInfo
- Publication number
- KR20140061009A KR20140061009A KR1020120128083A KR20120128083A KR20140061009A KR 20140061009 A KR20140061009 A KR 20140061009A KR 1020120128083 A KR1020120128083 A KR 1020120128083A KR 20120128083 A KR20120128083 A KR 20120128083A KR 20140061009 A KR20140061009 A KR 20140061009A
- Authority
- KR
- South Korea
- Prior art keywords
- augmented reality
- information
- image
- voice
- reality information
- Prior art date
Links
- 230000003190 augmentative effect Effects 0.000 title claims abstract description 98
- 238000000034 method Methods 0.000 title claims abstract description 25
- 239000003550 marker Substances 0.000 claims description 21
- 238000010586 diagram Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 238000007796 conventional method Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Computer Hardware Design (AREA)
- Computer Graphics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A method of operating a hybrid augmented reality system according to an embodiment of the present invention includes: acquiring an image including an object; Recognizing at least one object included in the acquired image; Detecting augmented reality information related to the recognized object; Inputting voice information and comparing the voice information with the augmented reality information; And a user interface for separately displaying the augmented reality information matched with the voice information.
Description
The present invention relates to an augmented reality service providing system, and more particularly, to a system and method for providing an augmented reality service providing system, in which an image of a marker is received from a camera, 3D content is output on a screen, To an augmented reality system.
AR (Augmented Reality) is one of the virtual reality in which the user views the real world viewed by the user and the virtual world having the additional information as one image. AR, a concept that complements the real world with a virtual system, uses a virtual environment created by computer graphics, but the protagonist is a real environment. Computer graphics serve to provide additional information needed for the real world. It means that the distinction between the real environment and the virtual screen is blurred by superimposing the 3D virtual image on the real image that the user is viewing. Such augmented reality is applied to various fields such as medicine, industry, amusement and military field because it provides users with improved reality and awareness.
It is very important to accurately estimate the motion of the camera or tracked object to realize the augmented reality. Conventional methods for realizing augmented reality include a marker based method for realizing augmented reality using AR (Augmented Reality) markers and augmented reality using feature points or 3D models collected from an object existing in the real world. And the marker-less method. The marker here refers to providing size, direction, and position information to be drawn on a three-dimensional graphic model that is actually present on a two-dimensional plane and connected to itself.
The present invention provides a 3D content corresponding to a current camera view using a smart phone equipped with an application (an application program of some sort) that implements such an augmented reality technology. However, the conventional execution of the application by the general hand operation is inconvenient when the number of applications is large, it may be troublesome for the hand or the eye inconvenient person, and it is applied to the layer which is vulnerable to the function of the smart phone such as a child or a student There was a difficult problem.
In order to solve the above problems, the present invention has been made to solve the above problems by using a smart phone equipped with an operating system capable of executing an augmented reality application, and by adding a technique of matching a voice of a smartphone user, For that purpose.
A method of operating a hybrid augmented reality system according to the present invention includes: acquiring an image including an object; Recognizing at least one object included in the acquired image; Detecting augmented reality information related to the recognized object; Inputting voice information and comparing the voice information with the augmented reality information; And a user interface for separately displaying the augmented reality information matched with the voice information.
According to another aspect of the present invention, there is provided a hybrid AR system including: an image acquiring unit acquiring an image including an object; A display unit for outputting the image or augmented reality information output from the image acquisition unit; A recognition unit for recognizing at least one object included in the image acquired from the image acquiring unit and for detecting augmented reality information related to the recognized object, And a control unit for generating a user interface for separately displaying the augmented reality information by matching the information and the voice information recognized by the voice recognition unit.
According to the embodiment of the present invention, it is possible to provide a service for providing personalized contents together with three-dimensional contents in a smart phone by combining speech recognition and a markerless augmented reality.
In addition, there is an advantage that the matched three-dimensional content is recognized by recognizing the photographed marker or voice, and the narration is expressed, thereby realizing an enhanced sense of reality, thereby helping children or students to learn by themselves .
In addition, the present invention can use the voice recognition technology to allow a user to search for desired information after listing contents to be searched with only a voice command without a manual operation, and to easily provide information to a child or a student who has an uncomfortable hand There is an advantage.
1 is a block diagram of an augmented reality user interface device according to an embodiment of the present invention;
2 is a diagram showing a speech recognition flow according to an embodiment of the present invention.
3 is a view schematically showing a configuration of an augmented reality user interface according to an embodiment of the present invention.
4 is a flowchart for explaining a method of providing an augmented reality user interface according to an embodiment of the present invention
Generally, the augmented reality information (AR information) can be obtained by a location based (GPS based) method, a marker recognition based method, or the like.
In the case of the location-based method, the mobile terminal uses the GPS information and the geomagnetic sensor information (direction, tilt information) to determine the augmented reality information about the object viewed by the mobile terminal (for example, And display the acquired augmented reality information on the photographed image.
In the case of the marker recognition based method, the mobile terminal finds a marker displayed on the image, and recognizes the size of the marker and the distance between the marker and the mobile terminal, thereby determining the three-dimensional position or distance of the marker. The mobile terminal can obtain the augmented reality information directly from the augmented reality marker or obtain the augmented reality information associated with the augmented reality marker from the server and display the acquired augmented reality information at the image or marker position.
For example, the augmented reality markers may be implemented in the form of a two-dimensional code, in which case various data such as letters, numbers, symbols, control codes, etc. may be included in the augmented reality markers themselves. The mobile terminal can acquire the augmented reality information by reading the augmented reality marker encoded with the augmented reality information and decoding the read image or the two-dimensional code. The concrete method for constructing the augmented reality markers with the two-dimensional code can be understood in a similar manner to the known two-dimensional codes (e.g., QR code, PDF417, DataMatrix, MaxiCode, etc.), and therefore detailed description thereof will be omitted.
Alternatively, identification information (e.g., a combination of numbers or characters, etc.) capable of identifying each augmented reality information may be encoded in the augmented reality marker. In this case, the mobile terminal can read the augmented reality marker encoded with the identification information, and decode the read image or the two-dimensional code to obtain the identification information. The mobile terminal can query the server for the identification information and acquire corresponding augmented reality information.
The augmented reality information may be obtained from a single augmented reality information server or a plurality of augmented reality information servers, and an object to which the augmented reality information is displayed may be acquired from all the augmented reality information servers that can provide guide information such as buildings, goods, Includes objects.
According to an embodiment of the present invention, when a plurality of augmented reality information about objects photographed through a camera of a mobile terminal is displayed on a screen, augmented reality information that is not necessary for a user is displayed on the screen By displaying the augmented reality information only on the object, the screen area and the layer desired by the user according to the voice input of the user so as not to cover the screen, the user can easily recognize only the augmented reality information he wants.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. In the following description of the present invention, a detailed description of known functions and configurations incorporated herein will be omitted when it may make the subject matter of the present invention rather unclear. Prior to describing the present invention, terms used throughout the specification are defined. These terms are defined in consideration of the functions in the embodiments of the present invention and can be sufficiently modified according to the intention, customs, etc. of the user or operator. Therefore, the definitions of these terms are based on the contents of the specification of the present invention It should be reduced.
1 is a configuration diagram of an augmented reality system apparatus according to an embodiment of the present invention.
1, an augmented reality system apparatus according to the present invention includes an
The
The
The
The
The user
Like the
FIG. 2 is a diagram illustrating a speech recognition flow according to an embodiment of the present invention. FIG.
Referring to FIG. 2, voice features are extracted from voice data recognized by command 1, and a reference pattern is generated. This process can be performed not only 10 times, but also the number of times the actual speech recognition is executed. When the user sends an input signal for inputting actual voice to the terminal, the reference pattern corresponding to the number of times of execution is generated, and the end point signal for ending the voice is transmitted to extract the feature of the voice. An input pattern is generated based on the extracted features and compared with the reference pattern to obtain a speech recognition result with an improved accuracy.
FIG. 3 is a schematic view showing an augmented reality system according to the present invention.
Referring to FIG. 3, a marker included in a smartphone terminal may be photographed with a camera, and a 3D image-based content may be recognized by the terminal. Then, the augmented reality information (3D contents) can be outputted on the screen of the terminal through the application for implementing the augmented reality system executed in the terminal. Also, voice input can be performed by executing a voice recognition application supported by the terminal, and the input voice is compared with the current image to display the augmented reality information (3D content) matching the condition.
The
The
The
The
The user
4 is a flowchart illustrating a method for providing an augmented reality system according to a preferred embodiment of the present invention.
Referring to FIG. 4, a user executes an augmented reality application for implementing an augmented reality system. Then, when the object recognition mode is set by the user's key input, the
The
In
In
If it is determined in
That is, the
As described above, the augmented reality system according to the present embodiment is a hybrid augmented reality system that matches object information and voice, and recognizes the photographed marker or voice, displays the matched three-dimensional content, This has the advantage of helping children and students to learn and immerse themselves.
Further, according to the present invention, a user can search for desired information by listing contents to be searched by voice commands alone without using a hand operation by using a speech recognition technology, and can provide a user-driven and sensible-type content.
Claims (8)
Recognizing at least one object included in the acquired image;
Detecting augmented reality information related to the recognized object;
Inputting voice information and comparing the voice information with the augmented reality information; And
And generating a user interface for separately displaying the augmented reality information matched with the voice information.
Wherein the object is a marker or markerless method.
Wherein the voice input uses a voice recognition application provided in a smart phone.
Wherein the image is a digital camera provided in a smartphone.
A display unit for outputting the image or augmented reality information output from the image acquisition unit;
A voice recognition unit for recognizing a voice associated with the image;
Recognizing at least one object included in the image acquired from the image acquisition unit, detecting augmented reality information related to the recognized object, and matching the detected augmented reality information and the voice information recognized by the voice recognition unit And a controller for generating a user interface for separating and displaying the augmented reality information.
Wherein the image acquiring unit uses a digital camera provided in a smart phone.
Wherein the object uses a marker or a markerless method.
Wherein the speech recognition unit is linked to a control unit using a speech recognition application embedded in a smart phone.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020120128083A KR20140061009A (en) | 2012-11-13 | 2012-11-13 | Hybrid augmented reality using voice recognition and method the same |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020120128083A KR20140061009A (en) | 2012-11-13 | 2012-11-13 | Hybrid augmented reality using voice recognition and method the same |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20140061009A true KR20140061009A (en) | 2014-05-21 |
Family
ID=50890169
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020120128083A KR20140061009A (en) | 2012-11-13 | 2012-11-13 | Hybrid augmented reality using voice recognition and method the same |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20140061009A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101990284B1 (en) * | 2018-12-13 | 2019-06-18 | 주식회사 버넥트 | Intelligent cognitive technology based augmented reality system using speech recognition |
WO2019117494A1 (en) * | 2017-12-15 | 2019-06-20 | 코디소프트 주식회사 | Content-providing device and method |
KR20220118672A (en) * | 2021-02-19 | 2022-08-26 | 대우조선해양 주식회사 | System and method for supporting augmented reality with text display function and computer-readable recording medium including the same |
-
2012
- 2012-11-13 KR KR1020120128083A patent/KR20140061009A/en not_active Application Discontinuation
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019117494A1 (en) * | 2017-12-15 | 2019-06-20 | 코디소프트 주식회사 | Content-providing device and method |
KR101990284B1 (en) * | 2018-12-13 | 2019-06-18 | 주식회사 버넥트 | Intelligent cognitive technology based augmented reality system using speech recognition |
KR20220118672A (en) * | 2021-02-19 | 2022-08-26 | 대우조선해양 주식회사 | System and method for supporting augmented reality with text display function and computer-readable recording medium including the same |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11080885B2 (en) | Digitally encoded marker-based augmented reality (AR) | |
CN109145680B (en) | Method, device and equipment for acquiring obstacle information and computer storage medium | |
KR101357260B1 (en) | Apparatus and Method for Providing Augmented Reality User Interface | |
US10225506B2 (en) | Information processing apparatus and information processing method | |
KR101469398B1 (en) | Text-based 3d augmented reality | |
US9946963B2 (en) | Barcode visualization in augmented reality | |
US11842514B1 (en) | Determining a pose of an object from rgb-d images | |
US10147399B1 (en) | Adaptive fiducials for image match recognition and tracking | |
TWI506563B (en) | A method and apparatus for enhancing reality of two - dimensional code | |
US20140210857A1 (en) | Realization method and device for two-dimensional code augmented reality | |
US10621787B2 (en) | Method and apparatus for overlaying a picture of a real scene with a virtual image, and mobile device | |
US11263818B2 (en) | Augmented reality system using visual object recognition and stored geometry to create and render virtual objects | |
KR20150039252A (en) | Apparatus and method for providing application service by using action recognition | |
KR20120010875A (en) | Apparatus and Method for Providing Recognition Guide for Augmented Reality Object | |
CN111640193A (en) | Word processing method, word processing device, computer equipment and storage medium | |
CN112598805A (en) | Prompt message display method, device, equipment and storage medium | |
KR20140061009A (en) | Hybrid augmented reality using voice recognition and method the same | |
KR102026475B1 (en) | Processing visual input | |
CN113867875A (en) | Method, device, equipment and storage medium for editing and displaying marked object | |
KR101582225B1 (en) | System and method for providing interactive augmented reality service | |
KR20130079962A (en) | Method of displaying post-it contents using augmented reality and apparatus using the same | |
US20220076465A1 (en) | Electronic device and method for editing content of external device | |
US12008813B2 (en) | Photo instructions and confirmation | |
KR102158751B1 (en) | User equipment, control method thereof and computer readable medium having computer program recorded therefor | |
KR20230049904A (en) | Method and system for providing contents based on identifier included in 3d printout |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WITN | Withdrawal due to no request for examination |