[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN109669662A - A kind of pronunciation inputting method, device, storage medium and mobile terminal - Google Patents

A kind of pronunciation inputting method, device, storage medium and mobile terminal Download PDF

Info

Publication number
CN109669662A
CN109669662A CN201811575128.5A CN201811575128A CN109669662A CN 109669662 A CN109669662 A CN 109669662A CN 201811575128 A CN201811575128 A CN 201811575128A CN 109669662 A CN109669662 A CN 109669662A
Authority
CN
China
Prior art keywords
voice
input
press event
content
mobile terminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811575128.5A
Other languages
Chinese (zh)
Inventor
洪旭杰
洪俞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huizhou TCL Mobile Communication Co Ltd
Original Assignee
Huizhou TCL Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huizhou TCL Mobile Communication Co Ltd filed Critical Huizhou TCL Mobile Communication Co Ltd
Priority to CN201811575128.5A priority Critical patent/CN109669662A/en
Publication of CN109669662A publication Critical patent/CN109669662A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Telephone Function (AREA)

Abstract

The embodiment of the present application provides a kind of pronunciation inputting method, device, storage medium and mobile terminal, which comprises detection screen press event;If the corresponding user interface components in position that the screen press event occurs are input controls, start voice and listen to, to receive the voice content of user's input;To the end of the screen press event, speech recognition is carried out to the voice content, obtains the corresponding content of text of the voice content;The content of text is inserted into the input control.Pronunciation inputting method provided herein, by pressure touch technical application into voice input, the voice that input control in user interface can be triggered according to pressing operation of the user on mobile terminal screen inputs, it solves the problems, such as that user needs to toggle phonitic entry method and inputting method when modifying the content of voice input, effectively increases the input efficiency of voice input, improves the user experience of mobile terminal.

Description

A kind of pronunciation inputting method, device, storage medium and mobile terminal
Technical field
The application belongs to voice processing technology field, and in particular to a kind of pronunciation inputting method, device, storage medium and shifting Dynamic terminal.
Background technique
With the fast development of mobile terminal technology, the popularity of phonitic entry method on mobile terminals is more and more wider, Especially on smart phone and tablet computer, phonitic entry method is because of its rapid and convenient, interaction naturally, becoming more and more users Input method selection.
But at present there is an experience in existing phonitic entry method.Usual user using voice when being inputted, meeting Because the accuracy rate of speech recognition not enough needs to modify the content of input, to need to switch to keyboard input, modify again Switch back into voice input.The inconvenience toggled affects the user experience of input efficiency and mobile terminal.
Summary of the invention
The embodiment of the present application provides a kind of pronunciation inputting method, and be able to solve user needs when modifying the content of voice input The problem of toggling phonitic entry method and inputting method promotes the input efficiency of phonitic entry method and the use of mobile terminal Family experience.
In a first aspect, the embodiment of the present application provides a kind of pronunciation inputting method, it is applied to mobile terminal, comprising:
Detect screen press event;
If the corresponding user interface components in position that the screen press event occurs are input controls, start voice and detect It listens, to receive the voice content of user's input;
To the end of the screen press event, speech recognition is carried out to the voice content, obtains the voice content pair The content of text answered;
The content of text is inserted into the input control.
Second aspect, the embodiment of the present application provide a kind of speech input device, are applied to mobile terminal, comprising:
Detection module, for detecting screen press event;
Voice input module, the corresponding user interface components in position for occurring when the screen press event are inputs When control, starting voice is listened to, to receive the voice content of user's input;
Speech recognition module carries out speech recognition to the voice content after to the screen press event, Obtain the corresponding content of text of the voice content;
Execution module, for the content of text to be inserted the input control.
The third aspect, the embodiment of the present application provide a kind of storage medium, are stored thereon with computer program, when the calculating When machine program executes on computers, so that the voice that the computer executes the offer of the embodiment of the present application first aspect is defeated Enter method.
Fourth aspect, the embodiment of the present application provide a kind of mobile terminal, and the mobile terminal includes processor and memory, The memory is for storing executable program code, and the processor is by reading the executable journey stored in the memory Sequence code is to execute following steps:
Detect screen press event;
If the corresponding user interface components in position that the screen press event occurs are input controls, start voice and detect It listens, to receive the voice content of user's input;
To the end of the screen press event, speech recognition is carried out to the voice content, obtains the voice content pair The content of text answered;
The content of text is inserted into the input control.
Pronunciation inputting method provided herein, by pressure touch technical application to voice input in, can according to The voice input of family input control in the pressing operation triggering user interface on mobile terminal screen, solves user in modification language The problem of needing to toggle phonitic entry method and inputting method when the content of sound input effectively increases the defeated of voice input Enter efficiency, improve the user experience of mobile terminal.
Detailed description of the invention
In order to more clearly explain the technical solutions in the embodiments of the present application, make required in being described below to embodiment Attached drawing is briefly described.It should be evident that the drawings in the following description are only some examples of the present application, for For those skilled in the art, without creative efforts, it can also be obtained according to these attached drawings other attached Figure.
Fig. 1 is a kind of flow diagram of pronunciation inputting method provided by the embodiments of the present application;
Fig. 2 is a usage scenario schematic diagram of pronunciation inputting method provided by the embodiments of the present application;
Fig. 3 is another usage scenario schematic diagram of pronunciation inputting method provided by the embodiments of the present application;
Fig. 4 is another flow diagram of pronunciation inputting method provided by the embodiments of the present application;
Fig. 5 is the structural schematic diagram of speech input device provided by the embodiments of the present application;
Fig. 6 is the structural schematic diagram of mobile terminal provided by the embodiments of the present application.
Specific embodiment
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application carries out clear, complete Site preparation description, it is clear that described embodiments are only a part of embodiments of the present application, instead of all the embodiments.It is based on Embodiment in the application, those skilled in the art's every other implementation obtained without creative efforts Example, shall fall in the protection scope of this application.
The embodiment of the present application provides a kind of pronunciation inputting method, device, storage medium and mobile terminal, carries out individually below It is described in detail.
Referring to Fig. 1, Fig. 1 is a kind of flow diagram of pronunciation inputting method provided by the embodiments of the present application.The voice Input method can be applied to mobile terminal, and the process of the pronunciation inputting method may include:
Step 101, screen press event is detected.
In the embodiment of the present application, the voice input of mobile terminal is to press some element on screen by user to be triggered , therefore needing to detect on the screen of the mobile terminal whether have pressing operation, this step is mainly by pressure touch module (Force Touch it) completes.
In one embodiment, pressure touch module can be by touch screen (Touch Panel) and pressure sensor (Force Touch Sensor) composition.When user presses some element on the touch screen, which can be obtained The intensity for getting user's pressing, by the event system of the mobile terminal, the position of user's pressing operation and will be by by Compressive Strength It is reported to mobile terminal.
Step 102, if the corresponding user interface components in position that the screen press event occurs are input controls, start Voice is listened to, to receive the voice content of user's input.
In the embodiment of the present application, user can press any position on mobile terminal screen, and each position has its right User interface (UI, the User Interface) component answered, but be not that each user interface components support voice input.Than Such as, if user is carrying out wechat communication using tablet computer at this time, in Chat page, in addition to the input frame of page bottom, Remaining part point is not support voice to input.If the position that user presses on the screen of the mobile terminal is input frame, that is, should The corresponding user interface components in position that screen press event occurs are input controls, then the voice that can trigger the mobile terminal is defeated Enter, which listens to voice is started, to receive the voice content of user's input.
Step 103, to the end of the screen press event, speech recognition is carried out to the voice content, obtains the voice content Corresponding content of text.
In the embodiment of the present application, if after the voice that user triggers mobile terminal inputs, the screen of the Continued depression mobile terminal Curtain, then the mobile terminal default user does not complete voice input also, and lasting unlatching voice is listened to.When the mobile terminal detects The screen press event terminates, then the mobile terminal judges that voice input is completed in user, will close voice and listens to, to user The identification that voice content through inputting performs the next step operates and obtains recognition result.
In one embodiment, in order to save the power consumption of mobile terminal, the mobile terminal can be set a screen by Press the upper limit threshold of event duration duration, such as two minutes.When the screen press event duration be more than the upper limit threshold, Then think the screen press event belong to user accidentally touch, rather than user wish to carry out triggering mobile terminal voice input Operation.
Step 104, text content is inserted into the input control.
In the embodiment of the present application, the voice content that the speech recognition process of mobile terminal can input user is identified, Obtain the corresponding content of text of the voice content, text content is only user and desires to input into be sent in input frame Content, so when the mobile terminal identifies the voice content, it, will after obtaining the corresponding content of text of the voice content Text content inserts corresponding input control, then completes the text input of the corresponding input frame of the input control.
Fig. 2 to Fig. 3 is please referred to, Fig. 2 to Fig. 3 is a usage scenario of pronunciation inputting method provided by the embodiments of the present application Schematic diagram.
User wants that sending short message about friend Zhang San using smart phone watches a ball game to home, can send out in the short message to Zhang San The page is sent to press Text Entry, as shown in Figure 2.Smart phone detects screen press event at this time, and the screen presses The corresponding user interface components in position that event occurs are input controls, then the smart phone is opened voice and listened to.User says The voice content of " coming my family at night to watch a ball game together ", is listened to by the smart phone, and by identification, which will be to the language Sound content " coming my family at night to watch a ball game together " carries out being recognized and converted into content of text, inserts in Text Entry, such as Fig. 3 institute Show.User only need to send page click send button in short message and carry out confirmation operation.
Pronunciation inputting method provided herein, by pressure touch technical application to voice input in, can according to The voice input of family input control in the pressing operation triggering user interface on mobile terminal screen, solves user in modification language The problem of needing to toggle phonitic entry method and inputting method when the content of sound input effectively increases the defeated of voice input Enter efficiency, improve the user experience of mobile terminal.
In one embodiment, speech recognition is carried out to the voice content, comprising: whether judge mobile terminal local There are offline speech recognition modelings;If so, the offline speech recognition modeling is called to carry out speech recognition to the voice content.
In one embodiment, it after judging the mobile terminal locally and whether there is offline speech recognition modeling, also wraps It includes: speech recognition being carried out to the voice content if it is not, then connecting cloud speech recognition server.
In one embodiment, after detection screen press event, further includes: judge whether the screen press event is full Sufficient preset condition;If the screen press event meets the corresponding user in position that the preset condition and the screen press event occur Interface assembly is input control, then starts voice and listen to, to receive the voice content of user's input.
In one embodiment, preset condition, comprising: it is default that the duration of the screen press event reaches first Value.
In one embodiment, preset condition, comprising: it is default that the pressure value of the screen press event reaches second Value.
In one embodiment, preset condition, comprising: it is default that the duration of the screen press event reaches first The pressure value of value and the screen press event reaches the second preset value.
Referring to Fig. 4, Fig. 4 is another flow diagram of pronunciation inputting method provided by the embodiments of the present application.The language Phoneme input method can be applied to mobile terminal, and the process of the pronunciation inputting method may include:
Step 401, screen press event is detected.
In the embodiment of the present application, the voice input of mobile terminal is to press some element on screen by user to be triggered , therefore needing to detect on the screen of the mobile terminal whether have pressing operation, this step is mainly by pressure touch module (Force Touch it) completes.
In one embodiment, pressure touch module can be by touch screen (Touch Panel) and pressure sensor (Force Touch Sensor) composition.When user presses some element on the touch screen, which can be obtained The intensity for getting user's pressing, by the event system of the mobile terminal, the position of user's pressing operation and will be by by Compressive Strength It is reported to mobile terminal.
Step 402, judge whether the screen press event meets preset condition.It is to then follow the steps 403, otherwise returns to step Rapid 401.
In the embodiment of the present application, missed to avoid mobile terminal that user is missed the clicking operation of touch screen curtain and user on the screen It is judged to the trigger signal of voice input, a preset condition can be set.If the screen press event meets the preset condition, The screen press event is identified as the trigger signal of voice input by the mobile terminal, performs the next step operation;If the screen is pressed Pressure event is unsatisfactory for the preset condition, then the mobile terminal is not believed the triggering that the screen press event is identified as voice input Number, continue to test next screen press event.
In one embodiment, the duration which can be set as to the screen press event reaches first Preset value is also duration one the first lower threshold of setting of the screen press event.When user presses movement end When the time of the screen at end is more than first lower threshold, which will confirm that user needs to input using voice.
In one embodiment, the pressure value which can be set as to the screen press event reaches second in advance It is also pressure value one the second lower threshold of setting of the screen press event if value.When user presses the mobile terminal When the dynamics of screen is more than second lower threshold, which will confirm that user needs to input using voice.
It should be noted that the preset condition in the embodiment of the present application is not limited to this, it can also be other modes, such as The duration that the preset condition can be set as to the screen press event reaches the first preset value and the screen press event Pressure value reaches the second preset value.It is more than first lower threshold (first when the time that user presses the screen of the mobile terminal Preset value) and user press the screen of the mobile terminal dynamics be more than the second lower threshold (second preset value) when, the shifting Dynamic terminal will confirm that user needs to input using voice.
Step 403, whether the corresponding user interface components in position for judging that the screen press event occurs are input controls. It is to then follow the steps 404, otherwise return step 401.
In the embodiment of the present application, user can press any position on mobile terminal screen, and each position has its right User interface (UI, the User Interface) component answered, but be not that each user interface components support voice input.Than Such as, if user is carrying out wechat communication using tablet computer at this time, in Chat page, in addition to the input frame of page bottom, Remaining part point is not support voice to input.If the position that user presses on the screen of the mobile terminal is input frame, that is, should The corresponding user interface components in position that screen press event occurs are input controls, then the voice that can trigger the mobile terminal is defeated Enter, which will perform the next step operation;If the position that user presses on the screen of the mobile terminal is not input frame, also The corresponding user interface components in position that i.e. the screen press event occurs are not input controls, then will not trigger the mobile terminal Voice input, which will continue to test next screen press event.
Step 404, starting voice is listened to receive the voice content of user's input.
In the embodiment of the present application, if the screen press event meets the position that preset condition and the screen press event occur Corresponding user interface components are input controls, then mobile terminal starting voice is listened to, and receive the voice content of user's input.
Step 405, judge whether the screen press event terminates.It is to then follow the steps 406, otherwise return step 404.
In the embodiment of the present application, if after the voice that user triggers mobile terminal inputs, the screen of the Continued depression mobile terminal Curtain, then the mobile terminal default user does not complete voice input also, and lasting unlatching voice is listened to.When the mobile terminal detects The screen press event terminates, then the mobile terminal judges that voice input is completed in user, will close voice and listens to, to user Voice content through inputting performs the next step operation.
In one embodiment, in order to save the power consumption of mobile terminal, the mobile terminal can be set a screen by Press the upper limit threshold of event duration duration, such as two minutes.When the screen press event duration be more than the upper limit threshold, Then think that the screen press event belongs to user and accidentally touches, rather than the voice input for triggering the mobile terminal that user wishes to carry out Operation.
Step 406, speech recognition is carried out to the voice content of user's input, obtained in the corresponding text of the voice content Hold.
In the embodiment of the present application, after screen press event, voice input is completed in mobile terminal default user, will It closes voice to listen to, speech recognition is carried out to the voice content of user's input, obtains the corresponding content of text of the voice content.
In one embodiment, which can be complete by the offline speech recognition modeling of mobile terminal local At;In one embodiment, which can also be completed by cloud speech recognition server.
Step 407, text content is inserted into the input control.
In the embodiment of the present application, the voice content that the speech recognition process of mobile terminal can input user is identified, Obtain the corresponding content of text of the voice content, text content is only user and desires to input into be sent in input frame Content, so when the mobile terminal identifies the voice content, it, will after obtaining the corresponding content of text of the voice content Text content inserts corresponding input control, then completes the text input of the corresponding input frame of the input control.
Referring to Fig. 5, Fig. 5 is the structural schematic diagram of speech input device provided by the embodiments of the present application.Voice input Device can be applied to mobile terminal, which may include: detection module 501, voice input module 502, language Sound identification module 503 and execution module 504.
Detection module 501, for detecting screen press event;
Voice input module 502, the corresponding user interface components in position for occurring when the screen press event are defeated When entering control, starting voice is listened to, to receive the voice content of user's input;
Speech recognition module 503 carries out speech recognition to the voice content after to the screen press event, Obtain the corresponding content of text of the voice content;
Execution module 504, for text content to be inserted the input control.
In one embodiment, when carrying out speech recognition to the voice content, speech recognition module 503 can be used In: judge the mobile terminal locally and whether there is offline speech recognition modeling;If so, calling the offline speech recognition modeling pair The voice content carries out speech recognition.
In one embodiment, after judging the mobile terminal locally and whether there is offline speech recognition modeling, language Sound identification module 503 can be also used for: carry out voice knowledge to the voice content if it is not, then connecting cloud speech recognition server Not.
In one embodiment, after detecting screen press event, detection module 501 be can be also used for: judgement should Whether screen press event meets preset condition;If the screen press event meets the preset condition and the screen press event is sent out The corresponding user interface components in raw position are input controls, then voice input module 502 starts voice and listens to, to receive user The voice content of input.
In one embodiment, preset condition, comprising: it is default that the duration of the screen press event reaches first Value.
In one embodiment, preset condition, comprising: it is default that the pressure value of the screen press event reaches second Value.
In one embodiment, preset condition, comprising: it is default that the duration of the screen press event reaches first The pressure value of value and the screen press event reaches the second preset value.
When it is implemented, the above modules can be used as independent entity to realize, any combination can also be carried out, is made It is realized for same or several entities, the specific implementation of the above modules can be found in the embodiment of the method for front, herein not It repeats again.
The embodiment of the present application provides a kind of computer-readable storage medium, is stored thereon with computer program, when the meter When calculation machine program executes on computers, so that the computer executes the voice input provided such as the embodiment of the present application first aspect Step in method.
The embodiment of the present application also provides a kind of mobile terminal, including memory, and processor, the processor is by calling this to deposit The computer program stored in reservoir, for executing the step in pronunciation inputting method provided by the embodiments of the present application.
Referring to Fig. 6, Fig. 6 is the structural schematic diagram of mobile terminal provided by the embodiments of the present application.Mobile terminal can wrap Include the processor 601 of one or more than one processing core, one or more computer readable storage mediums are deposited Reservoir 602, input unit 603, output unit 604, sensor 605, voicefrequency circuit 606, Wireless Fidelity (WiFi, Wireless Fidelity) module 607, radio frequency (RF, Radio Frequency) components such as circuit 608 and power supply 609.Those skilled in the art Member is appreciated that mobile terminal structure shown in Fig. 6 does not constitute the restriction to the mobile terminal, may include than illustrating more More or less component perhaps combines certain components or different component layouts.Wherein:
Processor 601 is the control centre of mobile terminal, utilizes each of various interfaces and the entire mobile terminal of connection A part by running or execute the application program and/or module that are stored in memory 602, and calls and is stored in storage Data in device 602 execute the various functions and processing data of mobile terminal, to carry out integral monitoring to mobile terminal.It can Choosing, processor 601 may include one or more processing cores;Preferably, processor 601 can integrate application processor and modulation Demodulation processor.Wherein, the main processing operation system of application processor, user interface and application program etc., modulation /demodulation processing Device mainly handles wireless communication.It is understood that above-mentioned modem processor can not also be integrated into processor 601.
Memory 602 can be used for storing application program and data.Including in the application program that memory 602 stores can Code is executed, application program can form various functional modules.Processor 601 is stored in answering in memory 602 by operation With program, thereby executing various function application and data processing.Memory 602 can mainly include storing program area and storage number According to area.Wherein, storing program area can application program needed for storage program area, at least one function (for example sound plays function Energy, image player function etc.) etc.;Storage data area, which can be stored, uses created data (such as audio according to mobile terminal Data, phone directory etc.) etc..In addition, memory 602 may include high-speed random access memory, it can also include non-volatile Memory, for example, at least a disk memory, flush memory device or other volatile solid-state parts.Correspondingly, it stores Device 602 can also include Memory Controller, to provide the access of processor 601 and input unit 603 to memory 602.
Input unit 603 can be used for receiving the number of input, character information or user's characteristic information (such as fingerprint, vocal print, Face information or iris information etc.), and generate keyboard related with user setting and function control, mouse, operating stick, light Or trackball signal input etc..Specifically, in a specific embodiment, input unit 603 may include touch sensitive surface And other input equipments.Touch sensitive surface, also referred to as touch display screen or Trackpad collect user on it or nearby Touch operation (for example user uses any suitable objects or attachment such as finger, stylus on touch sensitive surface or in touch sensitive surface Neighbouring operation), and corresponding attachment device is driven according to preset formula.Optionally, touch sensitive surface may include touching Two parts of detection device and touch controller.Wherein, the touch orientation of touch detecting apparatus detection user, and detect touch behaviour Make bring signal, transmits a signal to touch controller;Touch controller receives touch information from touch detecting apparatus, and It is converted into contact coordinate, then gives processor 601, and order that processor 601 is sent can be received and executed.This Outside, touch sensitive surface can be realized using multiple types such as resistance-type, condenser type, infrared ray and surface acoustic waves.In addition to touch-sensitive table Face, input unit 603 can also include other input equipments.Specifically, other input equipments can include but is not limited to physics One of keyboard, function key (such as volume control button, switch key etc.), trace ball, mouse, operating stick etc. are a variety of.
Output unit 604 can be used for exporting information input by user or be supplied to user information and terminal it is various Graphical user interface, these graphical user interface can be made of figure, text, icon, video and any combination thereof.Output Unit 604 may include that loudspeaker and display panel etc. optionally can use liquid crystal display (LCD, Liquid Crystal Display), the forms such as Organic Light Emitting Diode (OLED, Organic Light-Emitting Diode) configure display surface Plate.Further, touch sensitive surface can cover display panel, after touch sensitive surface detects touch operation on it or nearby, Processor 601 is sent to determine the type of touch event, is followed by subsequent processing device 601 according to the type of touch event in display panel It is upper that corresponding visual output is provided.Although touch sensitive surface and display panel are realized as two independent components in Fig. 6 Output and input function, but in some embodiments it is possible to it is touch sensitive surface is integrated with display panel and realize input and defeated Function out.
Terminal may also include at least one sensor 605, such as optical sensor, motion sensor and other sensors. Specifically, optical sensor may include ambient light sensor and proximity sensor, wherein ambient light sensor can be according to ambient light Light and shade adjust the brightness of display panel, proximity sensor can close display panel and/or back when terminal is moved in one's ear Light.As a kind of motion sensor, gravity accelerometer can detect (generally three axis) acceleration in all directions Size can detect that size and the direction of gravity when static, can be used to identify mobile phone posture application (such as horizontal/vertical screen switching, Dependent game, magnetometer pose calibrating), Vibration identification correlation function (such as pedometer, tap) etc.;It can also configure as terminal The other sensors such as gyroscope, barometer, hygrometer, thermometer, infrared sensor, details are not described herein.
Voicefrequency circuit 606 may include loudspeaker and microphone, it is possible to provide the audio interface between user and terminal.Audio Electric signal after the audio data received conversion can be transferred to loudspeaker, be converted to voice signal by loudspeaker by circuit 606 Output;On the other hand, the voice signal of collection is converted to electric signal by microphone, is converted to sound after being received by voicefrequency circuit 606 Frequency evidence, then by after the processing of audio data output processor 601, such as another terminal is sent to through RF circuit 608, or will Audio data exports to memory 602 to be further processed.Voicefrequency circuit 606 is also possible that earphone jack, outer to provide If the communication of earphone and terminal.
WiFi belongs to short range wireless transmission technology, and mobile terminal can help user to receive and dispatch electricity by WiFi module 607 Sub- mail, browsing webpage and access streaming video etc., it provides wireless broadband internet access for user.Although Fig. 6 shows Go out WiFi module 607, but it is understood that, and it is not belonging to must be configured into for mobile terminal, it completely can be according to need It to omit within the scope of not changing the essence of the invention.
RF circuit 608 can be used for receiving and sending messages or communication process in signal send and receive, particularly, will be under base station After row information receives, one or the processing of more than one processor 601 are transferred to;In addition, the data for being related to uplink are sent to base It stands.In general, RF circuit 608 includes but is not limited to antenna, at least one amplifier, tuner, one or more oscillators, user Identity module (SIM, Subscriber Identity Module) card, transceiver, coupler, low-noise amplifier (LNA, Low Noise Amplifier), duplexer etc..In addition, RF circuit 608 can also by wireless communication with network and other set Standby communication.The wireless communication can be used any communication standard or agreement, including but not limited to global system for mobile communications (GSM, Global System of Mobile communication), general packet radio service (GPRS, General Packet Radio Service), CDMA (CDMA, Code Division Multiple Access), wideband code division multiple access (WCDMA, Wideband Code Division Multiple Access), long term evolution (LTE, Long Term Evolution), Email, short message service (SMS, Short Messaging Service) etc..
Mobile terminal further includes the power supply 609 (such as battery) powered to all parts, it is preferred that power supply 609 can lead to Cross power-supply management system and processor 601 be logically contiguous, thus by power-supply management system realize management charging, electric discharge and The functions such as power managed.Power supply 609 can also include one or more direct current or AC power source, recharging system, electricity The random components such as source fault detection circuit, power adapter or inverter, power supply status indicator.
Although being not shown, mobile terminal can also include camera, bluetooth module etc., and details are not described herein.
In the embodiment of the present application, processor 601 in mobile terminal can according to following instruction, by one or one with On the corresponding executable code of process of application program be loaded into memory 602, and be stored in by processor 601 to run Application program in memory 602, thereby executing:
Detect screen press event;
If the corresponding user interface components in position that the screen press event occurs are input controls, start voice and detect It listens, to receive the voice content of user's input;
To the end of the screen press event, speech recognition is carried out to the voice content, obtains the corresponding text of the voice content This content;
Text content is inserted into the input control.
In one embodiment, when carrying out speech recognition to the voice content, processor 601 can be executed: judgement The mobile terminal locally whether there is offline speech recognition modeling;If so, calling the offline speech recognition modeling to the voice Content carries out speech recognition.
In one embodiment, after judging the mobile terminal locally and whether there is offline speech recognition modeling, place Reason device 601 can also be performed: carry out speech recognition to the voice content if it is not, then connecting cloud speech recognition server.
In one embodiment, after detecting screen press event, processor 601 be can also be performed: judge the screen Whether curtain press event meets preset condition;If the screen press event meets the preset condition and the screen press event occurs The corresponding user interface components in position be input control, then start voice and listen to, with receive user input voice content.
In one embodiment, preset condition, comprising: it is default that the duration of the screen press event reaches first Value.
In one embodiment, preset condition, comprising: it is default that the pressure value of the screen press event reaches second Value.
In one embodiment, preset condition, comprising: it is default that the duration of the screen press event reaches first The pressure value of value and the screen press event reaches the second preset value.
Pronunciation inputting method in mobile terminal provided by the embodiments of the present application and foregoing embodiments belongs to same design, Either offer method, specific implementation process in the pronunciation inputting method embodiment can be provided on the mobile terminal and are detailed in this Pronunciation inputting method embodiment, details are not described herein again.
In the above-described embodiments, it all emphasizes particularly on different fields to the description of each embodiment, there is no the portion being described in detail in some embodiment Point, it may refer to the detailed description above with respect to pronunciation inputting method, details are not described herein again.
It should be noted that for pronunciation inputting method provided by the embodiments of the present application, those of ordinary skill in the art It is understood that realize all or part of the process of pronunciation inputting method provided by the embodiments of the present application, being can be by computer journey Sequence is completed to control relevant hardware, which can be stored in a computer-readable storage medium, such as stores In memory, and by least one processor it executes, in the process of implementation may include the embodiment such as the pronunciation inputting method Process.Wherein, which can be magnetic disk, CD, read-only memory (ROM, Read Only Memory), arbitrary access Memory body (RAM, Random Access Memory) etc..
For the speech input device of the embodiment of the present application, each functional module be can integrate in a processing chip In, it is also possible to modules and physically exists alone, can also be integrated in two or more modules in a module.It is above-mentioned Integrated module both can take the form of hardware realization, can also be realized in the form of software function module.This is integrated If module is realized and when sold or used as an independent product in the form of software function module, also can store at one In computer-readable storage medium, which is for example read-only memory, disk or CD etc..
Above to a kind of pronunciation inputting method, device, storage medium and mobile terminal provided by the embodiment of the present application It is described in detail, specific examples are used herein to illustrate the principle and implementation manner of the present application, the above reality The explanation for applying example is merely used to help understand the present processes and its core concept;Meanwhile for those skilled in the art, According to the thought of the application, there will be changes in the specific implementation manner and application range, in conclusion in this specification Hold the limitation that should not be construed as to the application.

Claims (10)

1. a kind of pronunciation inputting method is applied to mobile terminal characterized by comprising
Detect screen press event;
If the corresponding user interface components in position that the screen press event occurs are input controls, start voice and listens to, To receive the voice content of user's input;
To the end of the screen press event, speech recognition is carried out to the voice content, it is corresponding to obtain the voice content Content of text;
The content of text is inserted into the input control.
2. the method according to claim 1, wherein described carry out speech recognition to the voice content, comprising:
Judge the mobile terminal locally and whether there is offline speech recognition modeling;
If so, the offline speech recognition modeling is called to carry out speech recognition to the voice content.
3. according to the method described in claim 2, it is characterized in that, the judgement mobile terminal locally whether there is offline After speech recognition modeling, further includes:
Speech recognition is carried out to the voice content if it is not, then connecting cloud speech recognition server.
4. the method according to claim 1, wherein after the detection screen press event, further includes:
Judge whether the screen press event meets preset condition;
If the screen press event meets the corresponding user in position that the preset condition and the screen press event occur Interface assembly is input control, then starts voice and listen to, to receive the voice content of user's input.
5. according to the method described in claim 4, it is characterized in that, the preset condition, comprising:
The duration of the screen press event reaches the first preset value.
6. according to the method described in claim 4, it is characterized in that, the preset condition, comprising:
The pressure value of the screen press event reaches the second preset value.
7. according to the method described in claim 4, it is characterized in that, the preset condition, comprising:
The duration of the screen press event reaches the first preset value and the pressure value of the screen press event reaches Two preset values.
8. a kind of speech input device is applied to mobile terminal characterized by comprising
Detection module, for detecting screen press event;
Voice input module, the corresponding user interface components in position for occurring when the screen press event are input controls When, starting voice is listened to, to receive the voice content of user's input;
Speech recognition module carries out speech recognition to the voice content, obtains after to the screen press event The corresponding content of text of the voice content;
Execution module, for the content of text to be inserted the input control.
9. a kind of storage medium, which is characterized in that the storage medium is stored with a plurality of instruction, described instruction be suitable for processor into Row load requires the step in 1 to 7 described in any item pronunciation inputting methods with perform claim.
10. a kind of mobile terminal, the mobile terminal includes processor and memory;Wherein, the memory can for storing Execute program code, the processor run by reading the executable program code stored in the memory with it is described can The corresponding program of program code is executed, to execute following steps:
Detect screen press event;
If the corresponding user interface components in position that the screen press event occurs are input controls, start voice and listens to, To receive the voice content of user's input;
To the end of the screen press event, speech recognition is carried out to the voice content, it is corresponding to obtain the voice content Content of text;
The content of text is inserted into the input control.
CN201811575128.5A 2018-12-21 2018-12-21 A kind of pronunciation inputting method, device, storage medium and mobile terminal Pending CN109669662A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811575128.5A CN109669662A (en) 2018-12-21 2018-12-21 A kind of pronunciation inputting method, device, storage medium and mobile terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811575128.5A CN109669662A (en) 2018-12-21 2018-12-21 A kind of pronunciation inputting method, device, storage medium and mobile terminal

Publications (1)

Publication Number Publication Date
CN109669662A true CN109669662A (en) 2019-04-23

Family

ID=66145916

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811575128.5A Pending CN109669662A (en) 2018-12-21 2018-12-21 A kind of pronunciation inputting method, device, storage medium and mobile terminal

Country Status (1)

Country Link
CN (1) CN109669662A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110069200A (en) * 2019-04-24 2019-07-30 努比亚技术有限公司 Wearable device input control method, wearable device and storage medium
CN111459288A (en) * 2020-04-23 2020-07-28 捷开通讯(深圳)有限公司 Method and device for realizing voice input by using head control
CN112199033A (en) * 2020-09-30 2021-01-08 北京搜狗科技发展有限公司 Voice input method and device and electronic equipment
CN113946229A (en) * 2021-08-31 2022-01-18 安徽咪鼠科技有限公司 Non-input method type cross-application voice typing method facing android operating system
CN114281182A (en) * 2020-09-17 2022-04-05 华为技术有限公司 Man-machine interaction method, device and system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103257824A (en) * 2013-04-02 2013-08-21 东莞宇龙通信科技有限公司 Method, system and mobile terminal for selecting operation object displayed on screen
CN103399709A (en) * 2013-08-02 2013-11-20 北京小米科技有限责任公司 Screen unlocking method and device
CN105487809A (en) * 2015-12-28 2016-04-13 魅族科技(中国)有限公司 Terminal control method and device
CN106775349A (en) * 2016-11-29 2017-05-31 珠海市魅族科技有限公司 A kind of speech modification method and device of word content
CN106990887A (en) * 2016-01-21 2017-07-28 阿里巴巴集团控股有限公司 Data inputting method, device and terminal device
CN107645598A (en) * 2017-09-20 2018-01-30 联想(北京)有限公司 A kind of message display method and electronic equipment

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103257824A (en) * 2013-04-02 2013-08-21 东莞宇龙通信科技有限公司 Method, system and mobile terminal for selecting operation object displayed on screen
CN103399709A (en) * 2013-08-02 2013-11-20 北京小米科技有限责任公司 Screen unlocking method and device
CN105487809A (en) * 2015-12-28 2016-04-13 魅族科技(中国)有限公司 Terminal control method and device
CN106990887A (en) * 2016-01-21 2017-07-28 阿里巴巴集团控股有限公司 Data inputting method, device and terminal device
CN106775349A (en) * 2016-11-29 2017-05-31 珠海市魅族科技有限公司 A kind of speech modification method and device of word content
CN107645598A (en) * 2017-09-20 2018-01-30 联想(北京)有限公司 A kind of message display method and electronic equipment

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110069200A (en) * 2019-04-24 2019-07-30 努比亚技术有限公司 Wearable device input control method, wearable device and storage medium
CN110069200B (en) * 2019-04-24 2024-02-09 努比亚技术有限公司 Wearable device input control method, wearable device and storage medium
CN111459288A (en) * 2020-04-23 2020-07-28 捷开通讯(深圳)有限公司 Method and device for realizing voice input by using head control
CN114281182A (en) * 2020-09-17 2022-04-05 华为技术有限公司 Man-machine interaction method, device and system
CN112199033A (en) * 2020-09-30 2021-01-08 北京搜狗科技发展有限公司 Voice input method and device and electronic equipment
CN113946229A (en) * 2021-08-31 2022-01-18 安徽咪鼠科技有限公司 Non-input method type cross-application voice typing method facing android operating system

Similar Documents

Publication Publication Date Title
CN104243671B (en) Volume adjusting method, device and electronic equipment
CN104636047B (en) The method, apparatus and touch screen terminal operated to the object in list
CN108549519B (en) Split screen processing method and device, storage medium and electronic equipment
CN104978176B (en) Application programming interfaces call method, device and computer readable storage medium
CN109669662A (en) A kind of pronunciation inputting method, device, storage medium and mobile terminal
CN105471704B (en) A kind of method, apparatus and system for realizing more people's calls
CN105022552B (en) A kind of method and apparatus for showing messaging list
CN104252341B (en) The client device of the skin change method of application program, device and application program
CN103473011A (en) Mobile terminal, and performance detecting method and performance detecting device for mobile terminal
CN107168609A (en) status bar display method, device and terminal
CN106200970B (en) The method and terminal of a kind of split screen display available
CN104077184B (en) A kind of course control method and computer system of application program
CN103312597B (en) A kind of speech message generation method and device
CN103197944B (en) Method for determining performance of process of application program and computer system
CN104660769B (en) A kind of methods, devices and systems for adding associated person information
CN105162930B (en) The way of recording and device of mobile terminal
CN103488421A (en) Method and device for processing call record and mobile terminal
CN106708501B (en) Activity monitoring method and device
CN105700620A (en) Control instruction execution method and apparatus
CN105991405B (en) A kind of method for building up and device of instant messaging
CN106371896B (en) Virutal machine memory configuration method and device, mobile terminal
CN106255146B (en) A kind of electricity-saving control method of terminal, device and terminal
CN109062643A (en) A kind of display interface method of adjustment, device and terminal
WO2015043239A1 (en) Method and device for playing media data on a terminal
CN111897916A (en) Voice instruction recognition method and device, terminal equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190423