CN106921560B - Voice communication method, device and system - Google Patents

Voice communication method, device and system Download PDF

Info

Publication number
CN106921560B
CN106921560B CN201710111296.8A CN201710111296A CN106921560B CN 106921560 B CN106921560 B CN 106921560B CN 201710111296 A CN201710111296 A CN 201710111296A CN 106921560 B CN106921560 B CN 106921560B
Authority
CN
China
Prior art keywords
voice
voice information
additional
information
terminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710111296.8A
Other languages
Chinese (zh)
Other versions
CN106921560A (en
Inventor
王柯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Priority to CN201710111296.8A priority Critical patent/CN106921560B/en
Publication of CN106921560A publication Critical patent/CN106921560A/en
Priority to US15/893,697 priority patent/US10728196B2/en
Priority to EP18158998.7A priority patent/EP3367379B1/en
Application granted granted Critical
Publication of CN106921560B publication Critical patent/CN106921560B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/10Multimedia information
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/08Annexed information, e.g. attachments

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Telephonic Communication Services (AREA)
  • Telephone Function (AREA)

Abstract

The disclosure relates to a voice communication method, device and system. The method comprises the following steps: acquiring a first operation instruction, wherein the first operation instruction is used for indicating to acquire additional voice content of first voice information; acquiring additional voice content of the first voice information according to the first operation instruction; generating second voice information according to the additional voice content and the additional identification, wherein the additional identification is used for indicating that the voice content included in the second voice information is the additional voice content; and sending the second voice information to the server. In the disclosure, the terminal may add the additional voice content on the basis of the transmitted voice information according to the user instruction, and by sending the additional voice content carrying the additional identifier to the server and instructing the server to send the additional voice content to the opposite terminal, the situation that the voice received by the opposite terminal is incomplete due to a network environment or human reasons and the like is effectively avoided.

Description

Voice communication method, device and system
Technical Field
The present disclosure relates to the field of communications technologies, and in particular, to a voice communication method, apparatus, and system.
Background
With the development of communication technology, mobile phones are widely used, and the social range of people is greatly expanded. Meanwhile, based on the expansion of the social contact range, the situation that the user uses the instant messaging software of the mobile phone to communicate is more and more. In the related art, a user can send characters, pictures and voice through instant messaging software. Because the voice information does not need to be typed by the user and is sent more conveniently and quickly, the voice information becomes the preferred form for the user to carry out instant messaging.
Disclosure of Invention
To overcome the problems in the related art, embodiments of the present disclosure provide a voice communication method, apparatus, and system. The technical scheme is as follows:
according to a first aspect of the embodiments of the present disclosure, there is provided a voice communication method, including:
acquiring a first operation instruction, wherein the first operation instruction is used for indicating to acquire additional voice content of first voice information;
acquiring additional voice content of the first voice information according to the first operation instruction;
generating second voice information according to the additional voice content and the additional identification, wherein the additional identification is used for indicating that the voice content included in the second voice information is the additional voice content;
and sending the second voice information to the server.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: in this embodiment, the terminal may add the additional voice content on the basis of the voice information that has been sent according to the user instruction, and by sending the additional voice content carrying the additional identifier to the server, may instruct the server to send the additional voice content to the opposite terminal. By the embodiment, when the terminal sends the voice information to the opposite terminal, the situation that the voice received by the opposite terminal is incomplete due to network environment or human reasons and the like is effectively avoided; and moreover, the opportunity of modifying the sent voice information is provided for the terminal user, and the user experience is favorably improved.
In one embodiment, the obtaining the first operation instruction comprises:
determining whether the first voice information is complete;
and when the first voice information is incomplete, acquiring a first operation instruction.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: after first voice information sends for the server, whether terminal can judge first voice information at first is complete, when first voice information is incomplete, acquire the first operating instruction of user input, and then acquire first voice information's additional information, it is unclear to have avoided adding the semanteme that additional voice content leads to under the complete circumstances of first voice information, cause misleading or puzzlement for the contralateral terminal user who receives first voice information, user experience has been promoted.
In one embodiment, the method further comprises:
when the first voice information is incomplete, prompt information is displayed and used for prompting a user that the first voice information is incomplete.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: when the terminal determines that the first voice information is incomplete, prompt information can be displayed on a screen for prompting that the first voice information input by a user is incomplete, so that the user can timely indicate the terminal to add the additional voice content of the first voice information, the integrity of the voice information sent to the opposite side terminal is guaranteed, misleading or puzzling caused by the incomplete voice information to the opposite side terminal user is avoided, and user experience is improved.
In one embodiment, the determining whether the first speech information is complete comprises:
determining whether a first time length occupied by effective voice content in the first voice information is smaller than or equal to a preset proportion of a second time length, wherein the second time length is the time length of starting a microphone when the first voice information is received;
and when the first time length occupied by the effective voice content in the first voice information is less than or equal to the preset proportion of the second time length, determining that the first voice information is incomplete.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: whether the first voice information is complete or not is determined according to the time occupied by the effective voice content in the first voice information, so that the accuracy of determining whether the voice information is complete or not by the terminal is improved, and the misjudgment of the terminal is avoided.
In one embodiment, the method further comprises:
and storing the additional voice content at the storage address corresponding to the first voice information.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: and storing the additional voice content in a storage address corresponding to the first voice information, so that the relevance between the first voice information and the additional voice content is ensured, and the user can read the complete voice information conveniently.
In one embodiment, the obtaining the first operation instruction comprises:
and acquiring a first operation instruction through a voice icon corresponding to the first voice information on the user interface.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: when a user needs to input the additional voice content of the first voice information, a first operation instruction can be input to the terminal by operating the voice icon of the first voice information on the user interface, so that the terminal takes the acquired voice as the additional voice content of the first voice information. The realization mode carries out voice additional speaking aiming at the sent voice information for the terminal side user, and the operation mode is more visual and convenient.
According to a second aspect of the embodiments of the present disclosure, there is provided a voice communication method, including:
receiving third voice information sent by the first terminal;
detecting whether the third voice information carries an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the third voice information is additional voice content;
and if the third voice information carries the additional identification, sending the voice content included in the third voice information to the second terminal as the additional voice content of fourth voice information, wherein the fourth voice information is the voice information sent to the second terminal by the first terminal before the third voice information is sent.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: the server can send the voice content of one voice message to the terminal as the additional voice content of another voice message, so that the integrity of the voice message received by the terminal is ensured, the problems of misleading or puzzlement to a terminal user and the like caused by incomplete voice message received by the terminal due to network environment or human reasons and the like are avoided, and the user experience is improved.
In an embodiment, if the third voice information carries the additional identifier, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information includes:
if the third voice information carries the additional identification, detecting whether the third voice information carries a second terminal identification, wherein the second terminal identification is used for uniquely identifying the second terminal;
and if the third voice information carries the second terminal identification, the voice content included in the third voice information is used as the additional voice content of the fourth voice information and is sent to the second terminal.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: the server can determine which terminal the additional voice content is sent to through the terminal identification carried by the voice information, so that the accuracy of judgment of the sending object is ensured, and the server is prevented from misjudging.
In an embodiment, if the third voice information carries the additional identifier, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information includes:
if the third voice information carries an additional identifier, detecting whether the third voice information carries a fourth voice information identifier, wherein the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and if the third voice information carries the fourth voice information identifier, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: the server can determine the additional voice content to be attached to the voice information through the voice information identification carried by the voice information, so that the accuracy of judging the voice additional object is ensured, and the server is prevented from misjudging.
In one embodiment, the sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information includes:
instructing the second terminal to store the voice content included in the third voice information to the storage address of the fourth voice information and withdraw the fourth voice information; or
And instructing the second terminal to add the voice content included in the third voice information to the voice content included in the fourth voice information.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: by the embodiment, the additional voice content can replace the voice information at the same storage address, and under the situation, the situation that the sending end wants to re-express due to the wrong voice expression is supported, so that the opportunity of modifying the sent voice information is provided for the terminal user, and the user experience is effectively improved; or the additional voice content and the voice information can be stored according to the receiving sequence, under the situation, the situation that the information sent by the terminal is incomplete and the receiving end cannot receive the complete voice due to the network and the like is supported, and the integrity of the voice received by the receiving end can be ensured.
In one embodiment, said revoking said fourth voice information comprises:
determining whether read feedback information of fourth voice information sent by the second terminal is received;
and if the read feedback information of the fourth voice information sent by the second terminal is not received, withdrawing the fourth voice information.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: the embodiment supports the withdrawal of the voice information under the condition that the voice information is not read by the user, fully considers the feeling of the user at the information receiving end, saves the time for the user to read the information and improves the reading efficiency.
According to a third aspect of the embodiments of the present disclosure, there is provided a voice communication method, including:
receiving a voice additional request sent by a server, wherein the voice additional request is used for requesting that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is the voice information received before the third voice information is received, the voice additional request carries the third voice information, an additional identifier and a fourth voice information identifier, the additional identifier is used for indicating that the voice content included in the third voice information is the additional voice content, and the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and according to the voice additional request, taking the voice content included in the third voice information as the additional voice content of the fourth voice information.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects: the terminal can add the additional voice content to the user after the incomplete voice information when determining that the received voice information is the additional voice information of another voice information, so that the problems of misleading or puzzlement to the terminal user and the like caused by incomplete voice information received by the terminal due to network environment or human reasons and the like are avoided, and the user experience is improved.
According to a fourth aspect of the embodiments of the present disclosure, there is provided a voice communication apparatus including:
the first acquisition module is used for acquiring a first operation instruction, and the first operation instruction is used for indicating to acquire additional voice content of first voice information;
the second acquisition module is used for acquiring the additional voice content of the first voice message according to the first operation instruction;
a generating module, configured to generate second voice information according to the additional voice content and an additional identifier, where the additional identifier is used to indicate that the voice content included in the second voice information is additional voice content;
and the first sending module is used for sending the second voice information to a server.
In one embodiment, the first obtaining module comprises:
the determining submodule is used for determining whether the first voice information is complete;
and the first acquisition submodule is used for acquiring the first operation instruction when the first voice information is incomplete.
In one embodiment, the apparatus further comprises:
and the prompt module is used for displaying prompt information when the first voice information is incomplete, and the prompt information is used for prompting a user that the first voice information is incomplete.
In one embodiment, the determining sub-module includes:
a first determining unit, configured to determine whether a first duration occupied by valid voice content in the first voice message is less than or equal to a preset proportion of a second duration, where the second duration is a duration for which a microphone is turned on when the first voice message is received;
and the second determining unit is used for determining that the first voice information is incomplete when the first time length occupied by the effective voice content in the first voice information is less than or equal to the preset proportion of the second time length.
In one embodiment, the apparatus further comprises:
and the storage module is used for storing the additional voice content in a storage address corresponding to the first voice message.
In one embodiment, the first obtaining module comprises:
and the second obtaining submodule is used for obtaining the first operation instruction through a voice icon corresponding to the first voice information on a user interface.
According to a fifth aspect of the embodiments of the present disclosure, there is provided a voice communication apparatus including:
the first receiving module is used for receiving third voice information sent by the first terminal;
a detection module, configured to detect whether the third voice information carries an additional identifier, where the additional identifier is used to indicate that a voice content included in the third voice information is an additional voice content;
and a second sending module, configured to send, when the third voice information carries the additional identifier, a voice content included in the third voice information to a second terminal as an additional voice content of fourth voice information, where the fourth voice information is the voice information sent by the first terminal to the second terminal before the third voice information is sent.
In one embodiment, the second sending module comprises:
a first detection submodule, configured to detect whether the third voice information carries a second terminal identifier when the third voice information carries the additional identifier, where the second terminal identifier is used to uniquely identify the second terminal;
and a first sending submodule, configured to send, when the third voice information carries the second terminal identifier, the voice content included in the third voice information to the second terminal as an additional voice content of the fourth voice information.
In one embodiment, the second sending module further comprises:
a second detection submodule, configured to detect whether the third voice information carries a fourth voice information identifier when the third voice information carries the additional identifier, where the fourth voice information identifier is used to uniquely identify the fourth voice information;
and a second sending submodule, configured to send, when the third voice information carries the fourth voice information identifier, the voice content included in the third voice information to the second terminal as an additional voice content of the fourth voice information.
In one embodiment, the second sending module comprises:
the first indication submodule is used for indicating the second terminal to store the voice content included in the third voice message into the storage address of the fourth voice message and withdraw the fourth voice message; or
And the second indicating submodule is used for indicating the second terminal to add the voice content included in the third voice message to the voice content included in the fourth voice message.
In one embodiment, the first indication sub-module includes:
a third determining unit, configured to determine whether read feedback information of the fourth voice information sent by the second terminal is received;
and the withdrawing unit is used for withdrawing the fourth voice message when the read feedback information of the fourth voice message sent by the second terminal is not received.
According to a sixth aspect of the embodiments of the present disclosure, there is provided a voice communication apparatus including:
a second receiving module, configured to receive a voice additional request sent by a server, where the voice additional request is used to request that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is voice information received before the third voice information is received, the voice additional request carries the third voice information, an additional identifier, and a fourth voice information identifier, the additional identifier is used to indicate that the voice content included in the third voice information is an additional voice content, and the fourth voice information identifier is used to uniquely identify the fourth voice information;
and the processing module is used for taking the voice content included in the third voice information as the additional voice content of the fourth voice information according to the voice additional request.
According to a seventh aspect of the embodiments of the present disclosure, there is provided a voice communication system including:
any one of the voice communication apparatuses provided by the fourth aspect, any one of the voice communication apparatuses provided by the fifth aspect, and the voice communication apparatus provided by the sixth aspect.
According to an eighth aspect of the embodiments of the present disclosure, there is provided a voice communication apparatus including:
a first processor;
a first memory for storing first processor-executable instructions;
wherein the first processor is configured to:
acquiring a first operation instruction, wherein the first operation instruction is used for indicating to acquire additional voice content of first voice information;
acquiring additional voice content of the first voice message according to the first operation instruction;
generating second voice information according to the additional voice content and an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the second voice information is the additional voice content;
and sending the second voice information to a server.
According to a ninth aspect of the embodiments of the present disclosure, there is provided a voice communication apparatus including:
a second processor;
a second memory for storing second processor-executable instructions;
wherein the second processor is configured to:
receiving third voice information sent by the first terminal;
detecting whether the third voice information carries an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the third voice information is additional voice content;
and if the third voice information carries the additional identifier, sending the voice content included in the third voice information to a second terminal as the additional voice content of fourth voice information, wherein the fourth voice information is the voice information sent to the second terminal by the first terminal before the third voice information is sent.
According to a tenth aspect of the embodiments of the present disclosure, there is provided a voice communication apparatus including:
a third processor;
a third memory for storing third processor-executable instructions;
wherein the third processor is configured to:
receiving a voice additional request sent by a server, wherein the voice additional request is used for requesting that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is the voice information received before the third voice information is received, the voice additional request carries the third voice information, an additional voice identifier and a fourth voice information identifier, the additional voice identifier is used for indicating that the voice content included in the third voice information is the additional voice content, and the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and according to the voice additional request, taking the voice content included in the third voice information as the additional voice content of the fourth voice information.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure.
FIG. 1a is a flow chart illustrating a method of voice communication according to an example embodiment.
FIG. 1b is an interface diagram of a terminal shown in accordance with an example embodiment.
FIG. 1c is a flow chart illustrating a method of voice communication according to an example embodiment.
FIG. 1d is a flow chart illustrating a method of voice communication according to an example embodiment.
FIG. 1e is a flow chart illustrating a method of voice communication according to an example embodiment.
Fig. 2a is a flow chart illustrating a method of voice communication according to an example embodiment.
Fig. 2b is a flow chart illustrating a method of voice communication according to an example embodiment.
Fig. 2c is a flow chart illustrating a method of voice communication according to an example embodiment.
Fig. 3 is a flow chart illustrating a method of voice communication according to an example embodiment.
Fig. 4 is an interaction diagram illustrating a voice communication method according to an example embodiment.
Fig. 5 is an interaction diagram illustrating a voice communication method according to an example embodiment.
Fig. 6a is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 6b is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 6c is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 6d is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 6e is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 6f is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 7a is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 7b is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 7c is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 7d is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 7e is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 7f is a schematic diagram illustrating the structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 8 is a schematic diagram illustrating a structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 9 is a block diagram illustrating a structure of a voice communication apparatus according to an exemplary embodiment.
Fig. 10 is a block diagram illustrating another voice communication apparatus according to an example embodiment.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
The technical scheme provided by the embodiment of the disclosure relates to a server and a terminal. The terminal can be a mobile phone, a tablet personal computer, an intelligent watch and other devices capable of carrying out voice communication; the server may be a server provided by an operator or a server provided by a third-party platform, which is not limited in this disclosure. In the related technology, when a user sends voice information, the user can click a voice icon on a chat interface, at the moment, a microphone is started by a sending end, the voice content input by the user is received and sent to a server, and the server sends the voice information to a receiving end, so that information interaction between the sending end and the receiving end is realized.
In practical application, situations sometimes occur that the voice received by the receiving end is incomplete due to network or human reasons, or the voice sent by the sending end is not the voice content really intended to be sent by the user, or the voice content sent by the user is misspoken and cannot be recovered. For example, if the user mistakenly touches the first terminal to turn on the microphone, and the first terminal user does not input any voice content, no valid information exists in the transmitted voice information; or due to a network reason, the first terminal only sends a part of voice content of the first voice message, and the other part of voice content is lost, so that the voice message received by the receiving end is incomplete.
The above situation will cause misleading or puzzlement to the user of the receiving end or the sending end, and the user experience is not good. In the embodiment of the disclosure, a terminal can add additional voice content on the basis of the transmitted voice information according to the instruction of a user, and can instruct a server to transmit the additional voice content to an opposite terminal by transmitting the additional voice content carrying an additional identifier to the server. By the embodiment, when the terminal sends the voice information to the opposite terminal, the situation that the voice received by the opposite terminal is incomplete due to poor network environment or user reasons and the like is effectively avoided; and moreover, the opportunity of modifying the sent voice information is provided for the terminal user, and the user experience is favorably improved.
Fig. 1a is a flowchart illustrating a voice communication method according to an exemplary embodiment, as shown in fig. 1a, the voice communication method includes the following steps 101 to 104:
in step 101, a first operation instruction is obtained, wherein the first operation instruction is used for indicating that the additional voice content of the first voice information is obtained.
For example, the terminal may send the voice information in various ways, for example, in an instant messaging application, or in a short message application, or in a game application with a voice chat function, and the like, which is not limited in this disclosure.
In the embodiment of the present disclosure, taking the instant messaging application as an example, after the user sends the first voice message, for example, if the user finds that the first voice message is incomplete, or the user finds that the voice content included in the first voice message is expressed incorrectly, the user may perform further operation on the terminal to instruct the terminal to acquire the additional voice content of the first voice message. Optionally, the additional voice content may be displayed on the terminal in a manner that the additional voice content corresponds to the same voice icon as the first voice information. It should be noted that, it is assumed that the first voice message can be read on the terminal by clicking a "first voice bubble" on the chat interface, and the corresponding same voice icon means that after the voice addition is successful, the user can read the additional voice content by clicking the first voice bubble.
It should be noted that the displaying manner of the additional voice content includes, but is not limited to, the above-mentioned manner, and the disclosure does not specifically limit this. For example, the additional voice contents can be sequentially displayed on the chat interface according to the occurrence time sequence of the voice contents which have already occurred. For example, the additional voice content may occupy a separate voice bubble and show the association between the additional voice content and the attached voice message (the first voice message) on the chat interface. For example, the voice bubble corresponding to the additional voice content and the voice bubble corresponding to the first voice information may be connected by a wire. For another example, after receiving the additional voice content of the first voice message, the voice bubble corresponding to the first voice message and the voice bubble corresponding to the additional voice content "flash" at the same time to prompt the user of the association between the two.
Optionally, the process of obtaining the first operation instruction may be implemented by a user clicking an icon (e.g., a voice bubble) corresponding to the first voice message on the chat interface. And after receiving the clicking operation, the terminal opens the microphone for receiving the additional voice content input by the user. For example, if the user holds down a voice bubble, the interface will present a number of actionable options, including an "add voice" function; after clicking the "add voice" function, the user will automatically open the microphone to receive the voice input by the user, i.e., to enable receiving additional voice content.
For example, as shown in fig. 1b, when the user presses the icon b01 of the first voice information for a long time, the terminal pops up the operation menu b02, and the operation menu b02 is provided with a plurality of operable options b03, which include "copy", "favorite", "withdraw", "delete", and "add voice", etc. The user may click the "add voice" option, at which point the terminal instructs the microphone to turn on, receives additional voice content of the first voice message input by the user, and when the user releases his hand, the terminal instructs the microphone to turn off. Note that the long press method may be replaced by a single click, a double click, or the like.
Or, when the user presses the icon of the first voice message for a long time, the terminal can pop up a prompt box, the prompt box displays the character of 'whether to continue speaking', and if the user selects to continue speaking, the terminal indicates that the microphone is turned on to receive the voice content input by the user.
It should be noted that, the manner of receiving the first operation instruction in the present embodiment includes, but is not limited to, the foregoing manner, and the present disclosure does not specifically limit this. For example, a button is set on the terminal, after the user clicks the button, each piece of chat information on the interface is displayed in an optional state (for example, a click can be performed), the user determines that the piece of chat information is information to be subjected to voice addition by selecting the target chat content (that is, the first voice information), and controls the microphone to be turned on to receive the additional voice content. The first operation instruction is used for instructing the terminal to acquire the additional voice content of the first voice message.
Optionally, when the terminal user suspects that the voice information may be incomplete and may affect reading by others, the terminal user may click a voice icon corresponding to the first voice information (e.g., click), turn on a speaker to play the voice content of the first voice information, and manually detect whether the first voice information is complete. If the first voice information is incomplete, the user can input a first operation instruction to input the additional voice content. It should be noted that, the judgment of whether the voice content is complete may be made manually by the user or by the terminal, and this disclosure does not specifically limit this.
Optionally, the expression form of the first operation instruction may be a preset type of operation, for example, the first operation instruction may be in a double-click mode, a single-click mode, or a long-press mode, which is not limited in this disclosure.
In step 102, according to the first operation instruction, additional voice content of the first voice information is acquired.
For example, after determining that the first operation instruction is received, in response to the first operation instruction, the terminal may instruct the microphone to turn on, where the voice content collected by the microphone is the additional voice content of the first voice information input by the user.
In step 103, second voice information is generated according to the additional voice content and the additional identifier, wherein the additional identifier is used for indicating that the voice content included in the second voice information is the additional voice content.
For example, after acquiring the additional voice content of the first voice information, the terminal may package the additional voice content and the additional identifier to generate the second voice information. The additional identification is used for identifying that the second voice information comprises additional voice information.
An additional identification field may be generally set in an extension field of a header of the second voice information, and by writing a preset number in the additional identification field, an additional identification is added to the second voice information where the additional voice content is located. For example, the header field of regular voice information (excluding additional voice content) is the number 0; the header field of the voice information including the additional voice content is the number 1.
In step 104, the second voice information is sent to the server.
For example, the terminal may access a hotspot or Wi-Fi (Wireless-Fidelity), send the second voice information to the server through the internet, and may also send the second voice information to the server through the cellular data network, which is not limited in this disclosure.
Optionally, the user may input different additional voice contents of the first voice message for multiple times, and the terminal may generate multiple different voice messages according to the different additional voice contents input by the user, and send the multiple different voice messages to the server.
In this embodiment, after receiving the second voice information sent by the terminal, the server may parse the additional voice content contained therein and send it to the opposite terminal of the terminal, so as to implement communication between the terminal and the opposite terminal.
In the technical scheme provided by the embodiment of the disclosure, a terminal can add additional voice content on the basis of the sent voice information according to the instruction of a user, and the server can be instructed to send the additional voice content to the opposite terminal by sending the additional voice content carrying the additional identifier to the server. By the embodiment, when the terminal sends the voice information to the opposite terminal, the situation that the voice received by the opposite terminal is incomplete due to network environment or human reasons and the like is effectively avoided; and moreover, the opportunity of modifying the sent voice information is provided for the terminal user, and the user experience is favorably improved.
In one embodiment, as shown in fig. 1c, in step 101, acquiring a first operation instruction may be implemented by steps 1011 and 1012:
in step 1011, it is determined whether the first speech information is complete.
In step 1012, when the first voice message is not complete, a first operation instruction is obtained.
In this embodiment, it is determined whether the first voice information is complete, which may be determined by the user or determined by the terminal. For example, the user may turn on the speaker to listen to the transmitted voice information to determine whether the voice is complete. Alternatively, the terminal may perform speech analysis on the transmitted content to determine whether the transmitted speech is complete.
For example, after the terminal has sent the first voice message, the terminal may obtain a first duration occupied by the valid voice content in the first voice message, and then determine whether the first duration is less than or equal to a preset proportion of a second duration, where the second duration is a time when the microphone is turned on when the first voice message is received. And when the first time length occupied by the effective voice content in the first voice information is less than or equal to the preset proportion of the second time length, determining that the first voice information is incomplete. The preset proportion can be set according to actual conditions. For example, the terminal may determine whether the first duration is less than or equal to 30% of the second duration, and determine that the first voice message is incomplete when the first duration occupied by the valid voice content in the first voice message is less than or equal to 30% of the second duration. Whether the first voice information is complete or not is determined according to the time occupied by the effective voice content in the first voice information, so that the accuracy of determining whether the voice information is complete or not by the terminal is improved, the misjudgment of the terminal is avoided, and the user experience is improved.
It should be noted that, the manner for determining whether the first speech information is complete in the present disclosure includes, but is not limited to, the manner described above. For example, the terminal may divide the voice content included in the first voice information into a plurality of voice segments. Taking the first voice segment as an example, the terminal may detect a sound wave frequency of the first voice segment, indicate that the first voice segment includes valid voice content if the sound wave frequency is greater than or equal to a preset threshold, and indicate that the first voice segment is invalid voice content if the sound wave frequency is less than the preset threshold. The sum of the time occupied by the plurality of voice segments including the effective voice content is the first duration.
It should be noted that, in different application scenarios, the process of determining whether the first speech information is complete may be replaced with a process of determining other content. For example, whether the voice content contained in the first voice information meets the intention of the user is judged; or judging whether the voice content contained in the first voice information contains the uncivilized phrase or not; or, whether the voice content included in the first voice information is too long or too short is determined. Step 1012 may be adaptively adjusted for different application scenarios.
In the technical scheme provided by the embodiment of the disclosure, after first voice information is sent to the server, the terminal can firstly judge whether the first voice information is complete, when the first voice information is incomplete, the first operation instruction input by the user is acquired, and then the additional information of the first voice information is acquired, so that the situation that the added additional voice content leads to unclear semantics under the condition that the first voice information is complete is avoided, misleading or puzzling is caused to the opposite side terminal user receiving the first voice information, and the user experience is improved.
In one embodiment, as shown in fig. 1d, the method further comprises step 105:
in step 105, when the first voice message is incomplete, a prompt message for prompting the user that the first voice message is incomplete is displayed.
For example, after the user finishes sending the first voice message, the user may not know whether the first voice message is completely sent, and when the terminal detects that the first voice message is incomplete, prompt information may be displayed on the display screen, for example, a word "the first voice message is incomplete" is displayed on the display screen, and the user is prompted to input additional voice content of the first voice message in time.
In the technical scheme provided by the embodiment of the disclosure, when the terminal determines that the first voice information is incomplete, the prompt information can be displayed on the screen, the prompt information is used for prompting that the first voice information input by the user is incomplete, so that the user can timely indicate the terminal to add the additional voice content of the first voice information, the integrity of the voice information sent to the opposite side terminal is ensured, misleading or puzzlement caused by the incomplete voice information to the opposite side terminal user is avoided, and the user experience is improved.
In one embodiment, as shown in fig. 1e, the method further comprises step 106:
in step 106, the additional voice content is stored in the storage address corresponding to the first voice information.
For example, the additional voice content of the first voice information included in the first voice information and the second voice information may be sequentially stored in the storage address corresponding to the first voice information according to the acquired time sequence. When the terminal transmits the voice information, the voice information can be sequentially transmitted according to the stored sequence, namely, the voice information stored firstly is transmitted firstly, and then the voice information stored later is transmitted.
Alternatively, in order to save the memory of the terminal, when the first voice message is already transmitted, the additional voice content of the first voice message included in the second voice message may be used to overwrite the first voice message.
Accordingly, the opposite terminal, upon receiving the additional voice content transmitted by the terminal (via the server), stores the acquired additional voice content in the same storage address as the first voice information that has been acquired. On the interface of the opposite terminal, it may appear that the first voice information and the additional voice content correspond to the same voice icon (e.g., voice bubble). In fact, in this case, for the opposite end user, the first voice information and the additional voice information have merged into one piece of information (or, the additional voice content replaces the voice content contained in the first voice information).
The above embodiments are equally applicable to the solutions shown in fig. 1c and 1 d.
In the technical scheme provided by the embodiment of the disclosure, the second voice information including the additional voice content of the first voice information is stored in the storage address corresponding to the first voice information, so that the relevance between the first voice information and the second voice information is ensured, and the user can conveniently read the complete voice information.
Fig. 2a is a flowchart illustrating a voice communication method according to an exemplary embodiment, and as shown in fig. 2a, the voice communication method includes the following steps 201 to 203:
in step 201, third voice information sent by the first terminal is received.
For example, the server may receive the third voice information sent by the first terminal through the internet, and may also receive the third voice information sent by the first terminal through the cellular data network, which is not limited in this disclosure.
In step 202, it is detected whether the third voice information carries an additional identifier, where the additional identifier is used to indicate that the voice content included in the third voice information is an additional voice content.
For example, after receiving the third voice information, the terminal may parse the third voice information to determine whether the third voice information includes the additional identifier. The additional identifier may be set in an extension field of a header of the third voice information, and the additional identifier is added to the third voice information where the additional voice content is located by writing a preset number in the additional identifier field. For example, the header field of regular voice information (excluding additional voice content) is the number 0; the header field of the voice information including the additional voice content is the number 1. When the server analyzes that the header field of the third voice message is number 1, it is determined that the third voice message carries the additional identifier, that is, it is determined that the voice content included in the third voice message is the additional voice content.
It should be noted that, the carrying manner of the additional identifier includes, but is not limited to, the above manner, and the manners that can be used to identify a voice message as the additional voice content are within the protection scope of the present disclosure, and the present disclosure is not particularly limited thereto.
In step 203, if the third voice message carries the additional identifier, the voice content included in the third voice message is sent to the second terminal as the additional voice content of the fourth voice message, and the fourth voice message is the voice message sent by the first terminal to the second terminal before the third voice message is sent.
For example, if the third voice information carries the additional identifier, the server may send the additional voice content to the second terminal, so that the additional voice content is attached to the voice content of the fourth voice information.
The voice content included in the third voice information is used as the additional voice content of the fourth voice information, which includes but is not limited to: (1) replacing the voice content included in the fourth voice message with the voice content included in the third voice message (i.e., the additional voice content), i.e., withdrawing the fourth voice message; optionally, the additional voice content is stored in the storage address where the original fourth voice information is located; (2) after the voice content included in the third voice message is added to the voice content included in the fourth voice message, the third voice message and the fourth voice message correspond to the same voice icon (e.g., a voice bubble), that is, are stored in the same storage address (e.g., the storage address where the fourth voice message is located), and the newly formed voice content includes the voice content included in the third voice message and the voice content included in the fourth voice message; (3) after the voice content included in the third voice message is deployed in the voice content included in the fourth voice message, the third voice message and the fourth voice message correspond to different voice icons, namely, correspond to different storage addresses and are stored in the different storage addresses; a connection may be established between the two voice icons, such as a dashed connection, to give an end user a prompt, etc.
Optionally, if the third voice information carries the additional identifier, the server may withdraw the fourth voice information sent to the second terminal, then obtain the voice content of the fourth voice information, add the additional voice content of the third voice information to the voice content of the fourth voice information, package the additional voice content of the third voice information to generate fifth voice information, and send the fifth voice information to the second terminal. The embodiment can ensure the consistency of voice information and improve the listening effect of users.
Or, if the third voice information carries the additional identifier, the server may also directly obtain the voice content of the fourth voice information from the local storage without withdrawing the fourth voice information, and then add the additional voice content of the third voice information to the voice content of the fourth voice information, package the additional voice content of the third voice information to generate fifth voice information, and send the fifth voice information to the second terminal.
Or, if the third voice information carries the additional identifier, the server may directly send the additional voice content of the third voice information to the second terminal without withdrawing the fourth voice information, and instruct the second terminal to add the additional voice content to the voice content of the fourth voice information.
According to the technical scheme, the server can send the voice content of the third voice information to the second terminal as the additional voice content of the fourth voice information, the integrity of the voice information sent to the second terminal is guaranteed, misdirection or trouble caused to a second terminal user due to incomplete voice information is avoided, and user experience is improved.
In an embodiment, as shown in fig. 2b, in step 203, if the third voice information carries an additional identifier, the voice content included in the third voice information is sent to the second terminal as an additional voice content of the fourth voice information, which may be implemented by step 2031 and step 2032:
in step 2031, if the third voice information carries the additional identifier, it is detected whether the third voice information carries a second terminal identifier, where the second terminal identifier is used to uniquely identify the second terminal.
In step 2032, if the third voice information carries the second terminal identifier, the voice content included in the third voice information is sent to the second terminal as the additional voice content of the fourth voice information.
Illustratively, the second terminal identifier in this embodiment is used to indicate that the third voice information includes voice content for sending to the second terminal. For example, the server attaches the received additional voice content to the last piece of information received before by default (that is, the fourth voice information described above), and when it is detected that the third voice information carries the additional identifier and the second terminal identifier, the server sends the voice content included in the third voice information to the second terminal and attaches the voice content to the last piece of voice information received by the second terminal.
In the embodiment, the server can determine which terminal the additional voice content is sent to through the terminal identification carried by the voice information, so that the accuracy of judgment of the sending object is ensured, and the server is prevented from misjudging.
In an embodiment, as shown in fig. 2c, in step 203, if the third voice information carries an additional identifier, the voice content included in the third voice information is sent to the second terminal as an additional voice content of the fourth voice information, which may be implemented by step 2033 and step 2034:
in step 2033, if the third voice information carries an additional identifier, it is detected whether the third voice information carries a fourth voice information identifier, where the fourth voice information identifier is used to uniquely identify the fourth voice information.
In step 2034, if the third voice information carries the fourth voice information identifier, the voice content included in the third voice information is sent to the second terminal as the additional voice content of the fourth voice information.
Illustratively, the fourth voice information identifier described in this embodiment is used to indicate that the third voice information includes voice content for attaching to the fourth voice information. For example, the server defaults to send the voice group to all terminals (or default terminals) that have sent messages within a preset time period, and when it is detected that the third voice information carries the additional identifier and the fourth voice information identifier, sends the additional voice content group included in the third voice information to all terminals (or default terminals) and attaches the additional voice content group to the fourth voice information of the corresponding terminal.
According to the technical scheme provided by the embodiment of the disclosure, the server can determine the voice information to which the additional voice content is attached through the voice information identifier carried by the voice information, so that the accuracy of judging the voice additional object is ensured, and the server is prevented from misjudging.
In one embodiment, when the server withdraws the fourth voice message, it may first determine whether read feedback information of the fourth voice message sent by the second terminal is received, and if the read feedback information of the fourth voice message sent by the second terminal is not received, withdraw the fourth voice message.
For example, after the user clicks to listen to the voice message, the terminal may send read feedback information of the voice message to the server to inform the server that the voice message has been read. The embodiment supports the withdrawal of the voice information under the condition that the voice information is not read by the user, fully considers the feeling of the user at the information receiving end, saves the time for the user to read the information and improves the reading efficiency.
In addition, optionally, when the server determines that the third voice message includes the additional voice content of the fourth voice message, if the read feedback information of the fourth voice message sent by the second terminal has been received, the server may send incomplete prompt information to the second terminal, where the incomplete prompt information is used to prompt the user that the fourth voice message is incomplete voice message, and then the server may send the processed complete voice content to the second terminal.
The embodiment supports the withdrawal of the voice information under the condition that the voice information is not read by the user, fully considers the feeling of the user at the information receiving end, saves the time for the user to read the information and improves the reading efficiency.
Fig. 3 is a flowchart illustrating a voice communication method according to an exemplary embodiment, and as shown in fig. 3, the voice communication method includes the following steps 301 to 302:
in step 301, a voice attachment request transmitted by a server is received.
The voice additional request is used for requesting that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is the voice information received before the third voice information is received, the voice additional request carries the third voice information, an additional identifier and a fourth voice information identifier, the additional identifier is used for indicating that the voice content included in the third voice information is the additional voice content, and the fourth voice information identifier is used for uniquely identifying the fourth voice information.
For example, after receiving a voice addition request sent by a server, a terminal first parses the voice addition request, obtains information included in the voice addition request, and then determines, according to an additional identifier and a fourth voice information identifier included in the voice addition request, that a voice content included in third voice information included in the voice addition request is an additional voice content of fourth voice information. At this time, the terminal may store the voice content included in the third voice information to the storage address where the fourth voice information is located.
In step 302, according to the voice addition request, the voice content included in the third voice information is used as the additional voice content of the fourth voice information.
For example, after the terminal determines that the voice content included in the third voice message included in the voice addition request is the additional voice content of the fourth voice message, if the user does not read the fourth voice message, the terminal may combine the voice content of the fourth voice message and the voice content included in the third voice message into a complete voice message, and display the complete voice message on the screen of the terminal through the voice icon; if the user has read the fourth voice message, the terminal can combine the voice content of the fourth voice message and the voice content included in the third voice message into a complete voice message, and then display the voice message on the screen of the terminal through the original voice icon of the fourth voice message, and mark the voice icon as unread; or after combining the voice content of the fourth voice message and the voice content included in the third voice message into a complete voice message, displaying the complete voice message on the screen of the terminal through the new voice icon, and simultaneously displaying the relationship between the new voice icon and the voice icon of the fourth voice message, for example, displaying a dotted line connecting the new voice icon and the voice icon of the fourth voice message, so that the user can conveniently know the relationship between the two voice messages.
In the technical scheme provided by the embodiment of the disclosure, when the terminal determines that the received voice information is the additional voice information of another voice information, the additional voice content is added to the incomplete voice information and then fed back to the user, so that the problems of misleading or puzzlement to the terminal user and the like caused by incomplete voice information received by the terminal due to network environment or human reasons and the like are avoided, and the user experience is improved.
The implementation is described in detail below by way of several embodiments.
Fig. 4 is an interaction diagram of a voice communication method according to an exemplary embodiment, which is applicable to a system consisting of a terminal and a server, where the terminal may be a mobile phone, a tablet computer, a smart watch, or other devices capable of performing voice communication; the server may be a server provided by an operator or a server provided by a third-party platform, which is not limited in this disclosure. As shown in fig. 4, the voice communication method includes the following steps 401 to 409:
in step 401, a first terminal acquires a first operation instruction input by a user.
In step 402, the first terminal acquires additional voice content of the first voice information in response to a first operation instruction input by a user.
In step 403, the first terminal generates second voice information according to the additional voice content and the additional identifier.
In step 404, the first terminal stores the second voice message at the storage address corresponding to the first voice message.
In step 405, the first terminal sends the second voice information to the server.
In step 406, the server determines whether the voice content included in the second voice message is the additional voice content of the first voice message according to the additional identifier included in the second voice message.
In step 407, when the server determines that the second voice message includes additional voice content whose voice content is the first voice message, the server withdraws the first voice message transmitted to the second terminal.
When the second voice information includes additional voice content whose voice content is not the first voice information, the server transmits the second voice information to the second terminal.
In step 408, the server generates third voice information by appending the additional voice content of the second voice information to the voice content of the first voice information.
In step 409, the server transmits the third voice information to the second terminal.
In the embodiment of the disclosure, the terminal may add the additional voice content on the basis of the voice information that has been sent according to the user instruction, and by sending the additional voice content carrying the additional identifier to the server, the server may be instructed to send the additional voice content to the opposite terminal. By the embodiment, when the terminal sends the voice information to the opposite terminal, the situation that the voice received by the opposite terminal is incomplete due to network environment or human reasons and the like is effectively avoided; and moreover, the opportunity of modifying the sent voice information is provided for the terminal user, and the user experience is favorably improved.
Fig. 5 is an interaction diagram of a voice communication method according to an exemplary embodiment, which is applicable to a system consisting of a terminal and a server, where the terminal may be a mobile phone, a tablet computer, a smart watch, or other devices capable of performing voice communication; the server may be a server provided by an operator or a server provided by a third-party platform, which is not limited in this disclosure. As shown in fig. 5, the voice communication method includes the following steps 501 to 515:
in step 501, the first terminal generates first voice information according to the received voice content.
In step 502, the first terminal sends the first voice message to the server.
In step 503, the server forwards the first voice information to the second terminal.
In step 504, the first terminal determines whether a first duration occupied by the valid voice content in the first voice message is less than or equal to a preset proportion of a second duration, where the second duration is a time when the microphone is turned on when the first voice message is received.
In step 505, if the first duration occupied by the valid voice content in the first voice message is less than or equal to the preset ratio of the second duration, the first terminal displays the prompt message.
If the first duration occupied by the effective voice content in the first voice message is less than or equal to the preset proportion of the second duration, the first terminal determines that the first voice message is incomplete.
And if the first duration occupied by the effective voice content in the first voice message is greater than the preset proportion of the second duration, the first terminal confirms that the first voice message is complete.
In step 506, the first terminal obtains a first operation instruction input by a user.
In step 507, the first terminal acquires the additional voice content of the first voice information in response to the first operation instruction.
In step 508, the first terminal generates second voice information according to the additional voice content, the second terminal identifier and the additional identifier.
In step 509, the first terminal stores the second voice message at the storage address corresponding to the first voice message.
In step 510, the first terminal sends the second voice information to the server.
In step 511, the server determines whether the second voice information includes the additional identity and the second terminal identity.
In step 512, if the second voice message includes the additional identifier and the second terminal identifier, the server determines whether the read feedback information of the first voice message sent by the second terminal is received.
In step 513, if the read feedback information of the first voice message sent by the second terminal is not received, the server withdraws the first voice message sent to the second terminal.
In step 514, the server adds the additional voice content of the second voice information to the voice content of the first voice information, and generates third voice information.
In step 515, the server transmits the third voice information to the second terminal.
In the embodiment of the disclosure, the terminal may add the additional voice content on the basis of the voice information that has been sent according to the user instruction, and by sending the additional voice content carrying the additional identifier to the server, the server may be instructed to send the additional voice content to the opposite terminal. By the embodiment, when the terminal sends the voice information to the opposite terminal, the situation that the voice received by the opposite terminal is incomplete due to network environment or human reasons and the like is effectively avoided; and moreover, the opportunity of modifying the sent voice information is provided for the terminal user, and the user experience is favorably improved.
The following are embodiments of the disclosed apparatus that may be used to perform embodiments of the disclosed methods.
Fig. 6a is a schematic diagram illustrating a structure of a voice communication apparatus 60 according to an exemplary embodiment, where the apparatus 60 may be implemented as part of or all of an electronic device through software, hardware or a combination of both. As shown in fig. 6a, the voice communication apparatus 60 includes a first obtaining module 601, a second obtaining module 602, a generating module 603 and a first sending module 604.
The first obtaining module 601 is configured to obtain a first operation instruction, where the first operation instruction is used to instruct to obtain additional voice content of first voice information;
a second obtaining module 602, configured to obtain additional voice content of the first voice message according to the first operation instruction;
a generating module 603, configured to generate second voice information according to the additional voice content and an additional identifier, where the additional identifier is used to indicate that the voice content included in the second voice information is an additional voice content;
a first sending module 604, configured to send the second voice information to a server.
In one embodiment, as shown in fig. 6b, the first obtaining module 601 includes a determining submodule 6011 and a first obtaining submodule 6012.
The determining submodule 6011 is configured to determine whether the first voice information is complete.
The first obtaining submodule 6012 is configured to obtain the first operation instruction when the first voice information is incomplete.
In one embodiment, as shown in fig. 6c, the apparatus 60 further comprises a prompting module 605.
The prompting module 605 is configured to display a prompting message when the first voice message is incomplete, where the prompting message is used to prompt a user that the first voice message is incomplete.
In one embodiment, as shown in fig. 6d, the determining submodule 6011 includes a first determining unit 6011a and a second determining unit 6011 b.
The first determining unit 6011a is configured to determine whether a first time duration occupied by an effective voice content in the first voice information is less than or equal to a preset ratio of a second time duration, where the second time duration is a time duration for turning on a microphone when the first voice information is received.
A second determining unit 6011b, configured to determine that the first voice information is incomplete when a first duration occupied by the valid voice content in the first voice information is less than or equal to a preset ratio of the second duration.
The above embodiments are equally applicable to the voice communication apparatus 60 shown in fig. 6 b.
In one embodiment, as shown in fig. 6e, the apparatus 60 further comprises a storage module 606.
The storage module 606 is configured to store the additional voice content at a storage address corresponding to the first voice information.
The above embodiments are equally applicable to the voice communication apparatus 60 shown in fig. 6b, 6c and 6 d.
In one embodiment, as shown in fig. 6f, the first obtaining module 601 includes:
the second obtaining sub-module 6013 is configured to obtain the first operation instruction through a voice icon corresponding to the first voice information on a user interface.
Embodiments of the present disclosure provide a voice communication apparatus that can add additional voice content on the basis of voice information that has been transmitted according to a user instruction, and can instruct a server to transmit the additional voice content to an opposite-side apparatus by transmitting the additional voice content carrying an additional identification to the server. By the embodiment, when the device sends the voice information to the opposite side device, the situation that the voice received by the opposite side device is incomplete due to network environment or human reasons and the like is effectively avoided; and moreover, the opportunity of modifying the sent voice information is provided for the device user, and the user experience is favorably improved.
Fig. 7a is a schematic structural diagram illustrating a voice communication apparatus 70 according to an exemplary embodiment, where the apparatus 70 may be implemented as part of or all of an electronic device through software, hardware or a combination of both. As shown in fig. 7a, the voice communication apparatus 70 includes a first receiving module 701, a detecting module 702 and a second sending module 703.
The first receiving module 701 is configured to receive third voice information sent by the first terminal.
A detecting module 702, configured to detect whether the third voice information carries an additional identifier, where the additional identifier is used to indicate that a voice content included in the third voice information is an additional voice content.
A second sending module 703, configured to send, when the third voice information carries the additional identifier, the voice content included in the third voice information to a second terminal as an additional voice content of fourth voice information, where the fourth voice information is the voice information sent by the first terminal to the second terminal before sending the third voice information.
In one embodiment, as shown in fig. 7b, the second sending module 703 includes a first detecting sub-module 7031 and a first sending sub-module 7032.
The first detecting sub-module 7031 is configured to detect whether the third voice information carries a second terminal identifier when the third voice information carries the additional identifier, where the second terminal identifier is used to uniquely identify the second terminal.
A first sending sub-module 7032, configured to send, when the third voice information carries the second terminal identifier, the voice content included in the third voice information as the additional voice content of the fourth voice information to the second terminal.
In one embodiment, as shown in fig. 7c, the second sending module 703 further includes a second detecting sub-module 7033 and a second sending sub-module 7034.
The second detecting submodule 7033 is configured to detect whether the third voice information carries a fourth voice information identifier when the third voice information carries the additional identifier, where the fourth voice information identifier is used to uniquely identify the fourth voice information.
A second sending sub-module 7034, configured to send, when the third voice information carries the fourth voice information identifier, the voice content included in the third voice information as an additional voice content of the fourth voice information to the second terminal.
The above embodiments are equally applicable to the voice communication apparatus 70 shown in fig. 7 b.
In one embodiment, as shown in fig. 7d, the second sending module 703 includes a first indication sub-module 7035.
The first instruction submodule 7035 is configured to instruct the second terminal to store the voice content included in the third voice information in the storage address where the fourth voice information is located, and withdraw the fourth voice information;
alternatively, as shown in fig. 7e, the second sending module 703 includes a second indicating sub-module 7036.
The second indicating sub-module 7036 is configured to instruct the second terminal to add the voice content included in the third voice information to the voice content included in the fourth voice information.
The above embodiments are equally applicable to the voice communication apparatus 70 shown in fig. 7b and 7 c.
In one embodiment, as shown in fig. 7f, the first indication submodule 7035 comprises a third determining unit 7035a and a revoking unit 7035 b.
Wherein, the third determining unit 7035a is configured to determine whether read feedback information of the fourth voice information sent by the second terminal is received.
A withdrawing unit 7035b, configured to withdraw the fourth voice message when the read feedback information of the fourth voice message sent by the second terminal is not received.
The embodiment of the disclosure provides a voice communication device, which can send the voice content of one voice message to a terminal as the additional voice content of another voice message, so as to ensure the integrity of the voice message received by the terminal, avoid the problem that the voice message received by the terminal is incomplete due to network environment or human reasons, etc., and cause misleading or puzzling to a terminal user, and improve user experience.
Fig. 8 is a schematic structural diagram of a voice communication apparatus 80 according to an exemplary embodiment, where the apparatus 80 may be implemented as part or all of an electronic device through software, hardware or a combination of the two. As shown in fig. 8, the voice communication apparatus 80 includes a second receiving module 801 and a processing module 802.
A second receiving module 801, configured to receive a voice additional request sent by a server, where the voice additional request is used to request that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is voice information received before the third voice information is received, the voice additional request carries the third voice information, an additional identifier, and a fourth voice information identifier, the additional identifier is used to indicate that the voice content included in the third voice information is an additional voice content, and the fourth voice information identifier is used to uniquely identify the fourth voice information.
A processing module 802, configured to use, according to the voice additional request, the voice content included in the third voice information as the additional voice content of the fourth voice information.
The embodiment of the disclosure provides a voice communication device, which can add additional voice content to a user after incomplete voice information when determining that received voice information is additional voice information of another voice information, thereby avoiding the problem of misleading or puzzling to a terminal user due to incomplete voice information received by a terminal caused by network environment or human reasons, and the like, and improving user experience.
The disclosed embodiment provides a system for voice communication, the system includes:
any of the voice communication devices 60 shown in fig. 6a to 6f described above, any of the voice communication devices 70 shown in fig. 7a to 7f described above, and any of the voice communication devices 80 shown in fig. 8.
The disclosed embodiment provides a voice communication apparatus, which includes:
a first processor;
a first memory for storing first processor-executable instructions;
wherein the first processor is configured to:
acquiring a first operation instruction, wherein the first operation instruction is used for indicating to acquire additional voice content of first voice information;
acquiring additional voice content of the first voice message according to the first operation instruction;
generating second voice information according to the additional voice content and an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the second voice information is the additional voice content;
and sending the second voice information to a server.
The disclosed embodiment provides a voice communication apparatus, which includes:
a second processor;
a second memory for storing second processor-executable instructions;
wherein the second processor is configured to:
receiving third voice information sent by the first terminal;
detecting whether the third voice information carries an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the third voice information is additional voice content;
and if the third voice information carries the additional identifier, sending the voice content included in the third voice information to a second terminal as the additional voice content of fourth voice information, wherein the fourth voice information is the voice information sent to the second terminal by the first terminal before the third voice information is sent.
The disclosed embodiment provides a voice communication apparatus, which includes:
a third processor;
a third memory for storing third processor-executable instructions;
wherein the third processor is configured to:
receiving a voice additional request sent by a server, wherein the voice additional request is used for requesting that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is the voice information received before the third voice information is received, the voice additional request carries the third voice information, an additional voice identifier and a fourth voice information identifier, the additional voice identifier is used for indicating that the voice content included in the third voice information is the additional voice content, and the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and according to the voice additional request, taking the voice content included in the third voice information as the additional voice content of the fourth voice information.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
Fig. 9 is a block diagram illustrating a structure of a voice communication apparatus 90, which is suitable for a terminal device, according to an exemplary embodiment. For example, the apparatus 90 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, an exercise device, a personal digital assistant, and the like.
The apparatus 90 may include one or more of the following components: processing component 902, memory 904, power component 906, multimedia component 908, audio component 910, input/output (I/O) interface 912, sensor component 914, and communication component 916.
The processing component 902 generally controls overall operation of the device 90, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. Processing component 902 may include one or more processors 920 to execute instructions to perform all or a portion of the steps of the methods described above. Further, processing component 902 can include one or more modules that facilitate interaction between processing component 902 and other components. For example, the processing component 902 can include a multimedia module to facilitate interaction between the multimedia component 908 and the processing component 902.
The memory 904 is configured to store various types of data to support operation at the apparatus 90. Examples of such data include instructions for any application or method operating on the device 90, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 904 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power supply component 906 provides power to the various components of the device 90. The power components 906 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the device 90.
The multimedia component 908 comprises a screen providing an output interface between the device 90 and a user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 908 includes a front facing camera and/or a rear facing camera. The front camera and/or the rear camera may receive external multimedia data when the device 90 is in an operating mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The audio component 910 is configured to output and/or input audio signals. For example, the audio component 910 includes a Microphone (MIC) configured to receive external audio signals when the apparatus 90 is in an operating mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signals may further be stored in the memory 904 or transmitted via the communication component 916. In some embodiments, audio component 910 also includes a speaker for outputting audio signals.
I/O interface 912 provides an interface between processing component 902 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The sensor assembly 914 includes one or more sensors for providing various aspects of status assessment for the device 90. For example, the sensor assembly 914 may detect the open/closed status of the device 90, the relative positioning of the components, such as the display and keypad of the device 90, the sensor assembly 914 may also detect a change in the position of the device 90 or a component of the device 90, the presence or absence of user contact with the device 90, the orientation or acceleration/deceleration of the device 90, and a change in the temperature of the device 90. The sensor assembly 914 may include a proximity sensor configured to detect the presence of a nearby object in the absence of any physical contact. The sensor assembly 914 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 914 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 916 is configured to facilitate wired or wireless communication between the apparatus 90 and other devices. The device 90 may access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component 916 receives a broadcast signal or broadcast associated information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 916 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the apparatus 90 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the above-described methods.
In an exemplary embodiment, a non-transitory computer readable storage medium comprising instructions, such as the memory 904 comprising instructions, executable by the processor 920 of the apparatus 90 to perform the method of fig. 1a, 1 c-1 e, or 3, described above, is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
Fig. 10 is a block diagram illustrating an apparatus for voice communication 100 according to an example embodiment. For example, the apparatus 100 may be provided as a server. The apparatus 100 includes a processing component 1002 that further includes one or more processors, and memory resources, represented by memory 1003, for storing instructions, such as applications, that are executable by the processing component 1002. The application programs stored in memory 1003 may include one or more modules that each correspond to a set of instructions. Further, the processing component 1002 is configured to execute instructions to perform the methods illustrated in fig. 2a, 2b and 2c described above.
The device 100 may also include a power component 1006 configured to perform power management of the device 100, a wired or wireless network interface 1005 configured to connect the device 100 to a network, and an input/output (I/O) interface 1008. The apparatus 100 may operate based on an operating system stored in the memory 1003, such as Windows Server, Mac OSXTM, UnixTM, LinuxTM, FreeBSDTM, or the like.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (26)

1. A method of voice communication, comprising:
acquiring a first operation instruction, wherein the first operation instruction is used for indicating to acquire additional voice content of first voice information;
acquiring additional voice content of the first voice message according to the first operation instruction;
generating second voice information according to the additional voice content and an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the second voice information is the additional voice content;
sending the second voice information to a server;
wherein the obtaining the first operation instruction comprises:
determining whether the first voice information is complete;
and when the first voice information is incomplete, acquiring the first operation instruction.
2. The method of claim 1, further comprising:
and when the first voice information is incomplete, displaying prompt information, wherein the prompt information is used for prompting a user that the first voice information is incomplete.
3. The method of claim 1, wherein the determining whether the first speech information is complete comprises:
determining whether a first time length occupied by effective voice content in the first voice information is smaller than or equal to a preset proportion of a second time length, wherein the second time length is the time length of starting a microphone when the first voice information is received;
and when the first time length occupied by the effective voice content in the first voice information is less than or equal to the preset proportion of the second time length, determining that the first voice information is incomplete.
4. The method according to any one of claims 1 to 3, further comprising:
and storing the additional voice content at a storage address corresponding to the first voice information.
5. The method of claim 1, wherein the obtaining the first operation instruction comprises:
and acquiring the first operation instruction through a voice icon corresponding to the first voice information on a user interface.
6. A method of voice communication, comprising:
receiving third voice information sent by a first terminal, wherein the third voice information is sent by the first terminal when the first terminal detects that fourth voice information is incomplete, and the fourth voice information is the voice information sent by the first terminal to a second terminal before the third voice information is sent;
detecting whether the third voice information carries an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the third voice information is additional voice content;
and if the third voice information carries the additional identifier, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information.
7. The method of claim 6, wherein if the third voice message carries the additional identifier, sending the voice content included in the third voice message to the second terminal as the additional voice content of the fourth voice message comprises:
if the third voice information carries the additional identifier, detecting whether the third voice information carries a second terminal identifier, wherein the second terminal identifier is used for uniquely identifying the second terminal;
and if the third voice information carries the second terminal identification, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information.
8. The method of claim 6, wherein if the third voice message carries the additional identifier, sending the voice content included in the third voice message to the second terminal as the additional voice content of the fourth voice message comprises:
if the third voice information carries the additional identifier, detecting whether the third voice information carries a fourth voice information identifier, wherein the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and if the third voice information carries the fourth voice information identifier, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information.
9. The method according to any one of claims 6 to 8, wherein the sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information comprises:
instructing the second terminal to store the voice content included in the third voice message to the storage address of the fourth voice message, and withdrawing the fourth voice message; or
And instructing the second terminal to add the voice content included in the third voice information to the voice content included in the fourth voice information.
10. The method of claim 9, wherein said revoking the fourth voice information comprises:
determining whether read feedback information of the fourth voice information sent by the second terminal is received;
and if the read feedback information of the fourth voice information sent by the second terminal is not received, withdrawing the fourth voice information.
11. A method of voice communication, comprising:
receiving a voice additional request sent by a server, wherein the voice additional request is used for requesting that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is the voice information received before the third voice information is received, the third voice information is the voice information received when the fourth voice information is incomplete, the voice additional request carries the third voice information, an additional identifier and a fourth voice information identifier, the additional identifier is used for indicating that the voice content included in the third voice information is the additional voice content, and the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and according to the voice additional request, taking the voice content included in the third voice information as the additional voice content of the fourth voice information.
12. A voice communication apparatus, comprising:
the first acquisition module is used for acquiring a first operation instruction, and the first operation instruction is used for indicating to acquire additional voice content of first voice information;
the second acquisition module is used for acquiring the additional voice content of the first voice message according to the first operation instruction;
a generating module, configured to generate second voice information according to the additional voice content and an additional identifier, where the additional identifier is used to indicate that the voice content included in the second voice information is additional voice content;
the first sending module is used for sending the second voice information to a server;
wherein the first obtaining module comprises:
the determining submodule is used for determining whether the first voice information is complete;
and the first acquisition submodule is used for acquiring the first operation instruction when the first voice information is incomplete.
13. The apparatus of claim 12, further comprising:
and the prompt module is used for displaying prompt information when the first voice information is incomplete, and the prompt information is used for prompting a user that the first voice information is incomplete.
14. The apparatus of claim 12, wherein the determination submodule comprises:
a first determining unit, configured to determine whether a first duration occupied by valid voice content in the first voice message is less than or equal to a preset proportion of a second duration, where the second duration is a duration for which a microphone is turned on when the first voice message is received;
and the second determining unit is used for determining that the first voice information is incomplete when the first time length occupied by the effective voice content in the first voice information is less than or equal to the preset proportion of the second time length.
15. The apparatus of any one of claims 12 to 14, further comprising:
and the storage module is used for storing the additional voice content in a storage address corresponding to the first voice message.
16. The apparatus of claim 12, wherein the first obtaining module comprises:
and the second obtaining submodule is used for obtaining the first operation instruction through a voice icon corresponding to the first voice information on a user interface.
17. A voice communication apparatus, comprising:
the first receiving module is used for receiving third voice information sent by a first terminal, wherein the third voice information is sent by the first terminal when the first terminal detects that fourth voice information is incomplete, and the fourth voice information is the voice information sent by the first terminal to a second terminal before the third voice information is sent;
a detection module, configured to detect whether the third voice information carries an additional identifier, where the additional identifier is used to indicate that a voice content included in the third voice information is an additional voice content;
and a second sending module, configured to send, when the third voice information carries the additional identifier, a voice content included in the third voice information to the second terminal as an additional voice content of the fourth voice information.
18. The apparatus of claim 17, wherein the second sending module comprises:
a first detection submodule, configured to detect whether the third voice information carries a second terminal identifier when the third voice information carries the additional identifier, where the second terminal identifier is used to uniquely identify the second terminal;
and a first sending submodule, configured to send, when the third voice information carries the second terminal identifier, the voice content included in the third voice information to the second terminal as an additional voice content of the fourth voice information.
19. The apparatus of claim 17, wherein the second sending module further comprises:
a second detection submodule, configured to detect whether the third voice information carries a fourth voice information identifier when the third voice information carries the additional identifier, where the fourth voice information identifier is used to uniquely identify the fourth voice information;
and a second sending submodule, configured to send, when the third voice information carries the fourth voice information identifier, the voice content included in the third voice information to the second terminal as an additional voice content of the fourth voice information.
20. The apparatus according to any one of claims 17 to 19, wherein the second sending module comprises:
the first indication submodule is used for indicating the second terminal to store the voice content included in the third voice message into the storage address of the fourth voice message and withdraw the fourth voice message; or
And the second indicating submodule is used for indicating the second terminal to add the voice content included in the third voice message to the voice content included in the fourth voice message.
21. The apparatus of claim 20, wherein the first indication submodule comprises:
a third determining unit, configured to determine whether read feedback information of the fourth voice information sent by the second terminal is received;
and the withdrawing unit is used for withdrawing the fourth voice message when the read feedback information of the fourth voice message sent by the second terminal is not received.
22. A voice communication apparatus, comprising:
a second receiving module, configured to receive a voice additional request sent by a server, where the voice additional request is used to request that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is voice information received before the third voice information is received, the third voice information is voice information received when the fourth voice information is incomplete, the voice additional request carries the third voice information, an additional identifier, and a fourth voice information identifier, the additional identifier is used to indicate that the voice content included in the third voice information is an additional voice content, and the fourth voice information identifier is used to uniquely identify the fourth voice information;
and the processing module is used for taking the voice content included in the third voice information as the additional voice content of the fourth voice information according to the voice additional request.
23. A voice communication system, comprising: the voice communication device of any one of claims 12 to 16, the voice communication device of any one of claims 17 to 21, and the voice communication device of claim 22.
24. A voice communication apparatus, comprising:
a first processor;
a first memory for storing first processor-executable instructions;
wherein the first processor is configured to:
acquiring a first operation instruction, wherein the first operation instruction is used for indicating to acquire additional voice content of first voice information;
acquiring additional voice content of the first voice message according to the first operation instruction;
generating second voice information according to the additional voice content and an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the second voice information is the additional voice content;
sending the second voice information to a server;
wherein the obtaining the first operation instruction comprises:
determining whether the first voice information is complete;
and when the first voice information is incomplete, acquiring the first operation instruction.
25. A voice communication apparatus, comprising:
a second processor;
a second memory for storing second processor-executable instructions;
wherein the second processor is configured to:
receiving third voice information sent by a first terminal, wherein the third voice information is sent by the first terminal when the first terminal detects that fourth voice information is incomplete, and the fourth voice information is the voice information sent by the first terminal to a second terminal before the third voice information is sent;
detecting whether the third voice information carries an additional identifier, wherein the additional identifier is used for indicating that the voice content included in the third voice information is additional voice content;
and if the third voice information carries the additional identifier, sending the voice content included in the third voice information to the second terminal as the additional voice content of the fourth voice information.
26. A voice communication apparatus, comprising:
a third processor;
a third memory for storing third processor-executable instructions;
wherein the third processor is configured to:
receiving a voice additional request sent by a server, wherein the voice additional request is used for requesting that a voice content included in third voice information is used as an additional voice content of fourth voice information, the fourth voice information is the voice information received before the third voice information is received, the third voice information is the voice information received when the fourth voice information is incomplete, the voice additional request carries the third voice information, an additional voice identifier and a fourth voice information identifier, the additional voice identifier is used for indicating that the voice content included in the third voice information is the additional voice content, and the fourth voice information identifier is used for uniquely identifying the fourth voice information;
and according to the voice additional request, taking the voice content included in the third voice information as the additional voice content of the fourth voice information.
CN201710111296.8A 2017-02-28 2017-02-28 Voice communication method, device and system Active CN106921560B (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN201710111296.8A CN106921560B (en) 2017-02-28 2017-02-28 Voice communication method, device and system
US15/893,697 US10728196B2 (en) 2017-02-28 2018-02-12 Method and storage medium for voice communication
EP18158998.7A EP3367379B1 (en) 2017-02-28 2018-02-27 Method, device and system for voice communication

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710111296.8A CN106921560B (en) 2017-02-28 2017-02-28 Voice communication method, device and system

Publications (2)

Publication Number Publication Date
CN106921560A CN106921560A (en) 2017-07-04
CN106921560B true CN106921560B (en) 2020-06-02

Family

ID=59454543

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710111296.8A Active CN106921560B (en) 2017-02-28 2017-02-28 Voice communication method, device and system

Country Status (3)

Country Link
US (1) US10728196B2 (en)
EP (1) EP3367379B1 (en)
CN (1) CN106921560B (en)

Families Citing this family (71)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9811314B2 (en) 2016-02-22 2017-11-07 Sonos, Inc. Metadata exchange involving a networked playback system and a networked microphone system
US10095470B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Audio response playback
US9947316B2 (en) 2016-02-22 2018-04-17 Sonos, Inc. Voice control of a media playback system
US9965247B2 (en) 2016-02-22 2018-05-08 Sonos, Inc. Voice controlled media playback system based on user profile
US9826306B2 (en) 2016-02-22 2017-11-21 Sonos, Inc. Default playback device designation
US10264030B2 (en) 2016-02-22 2019-04-16 Sonos, Inc. Networked microphone device control
US9978390B2 (en) 2016-06-09 2018-05-22 Sonos, Inc. Dynamic player selection for audio signal processing
US10134399B2 (en) 2016-07-15 2018-11-20 Sonos, Inc. Contextualization of voice inputs
US10115400B2 (en) 2016-08-05 2018-10-30 Sonos, Inc. Multiple voice services
US9942678B1 (en) 2016-09-27 2018-04-10 Sonos, Inc. Audio playback settings for voice interaction
US10181323B2 (en) 2016-10-19 2019-01-15 Sonos, Inc. Arbitration-based voice recognition
US11183181B2 (en) 2017-03-27 2021-11-23 Sonos, Inc. Systems and methods of multiple voice services
US10475449B2 (en) 2017-08-07 2019-11-12 Sonos, Inc. Wake-word detection suppression
US10048930B1 (en) 2017-09-08 2018-08-14 Sonos, Inc. Dynamic computation of system response volume
US10531157B1 (en) * 2017-09-21 2020-01-07 Amazon Technologies, Inc. Presentation and management of audio and visual content across devices
US10446165B2 (en) 2017-09-27 2019-10-15 Sonos, Inc. Robust short-time fourier transform acoustic echo cancellation during audio playback
US10621981B2 (en) 2017-09-28 2020-04-14 Sonos, Inc. Tone interference cancellation
US10051366B1 (en) 2017-09-28 2018-08-14 Sonos, Inc. Three-dimensional beam forming with a microphone array
US10482868B2 (en) 2017-09-28 2019-11-19 Sonos, Inc. Multi-channel acoustic echo cancellation
US10466962B2 (en) 2017-09-29 2019-11-05 Sonos, Inc. Media playback system with voice assistance
KR102460491B1 (en) * 2017-12-06 2022-10-31 삼성전자주식회사 Electronic apparatus and controlling method of thereof
US10880650B2 (en) 2017-12-10 2020-12-29 Sonos, Inc. Network microphone devices with automatic do not disturb actuation capabilities
US10818290B2 (en) 2017-12-11 2020-10-27 Sonos, Inc. Home graph
CN107861398A (en) * 2017-12-21 2018-03-30 重庆金鑫科技产业发展有限公司 The system and intelligent domestic system of a kind of voice control electric appliance
US11343614B2 (en) 2018-01-31 2022-05-24 Sonos, Inc. Device designation of playback and network microphone device arrangements
US10600408B1 (en) * 2018-03-23 2020-03-24 Amazon Technologies, Inc. Content output management based on speech quality
US11175880B2 (en) 2018-05-10 2021-11-16 Sonos, Inc. Systems and methods for voice-assisted media content selection
US10959029B2 (en) 2018-05-25 2021-03-23 Sonos, Inc. Determining and adapting to changes in microphone performance of playback devices
US10681460B2 (en) 2018-06-28 2020-06-09 Sonos, Inc. Systems and methods for associating playback devices with voice assistant services
US11076035B2 (en) 2018-08-28 2021-07-27 Sonos, Inc. Do not disturb feature for audio notifications
US10461710B1 (en) 2018-08-28 2019-10-29 Sonos, Inc. Media playback system with maximum volume setting
US10587430B1 (en) 2018-09-14 2020-03-10 Sonos, Inc. Networked devices, systems, and methods for associating playback devices based on sound codes
US11024331B2 (en) 2018-09-21 2021-06-01 Sonos, Inc. Voice detection optimization using sound metadata
US10811015B2 (en) 2018-09-25 2020-10-20 Sonos, Inc. Voice detection optimization based on selected voice assistant service
US11100923B2 (en) 2018-09-28 2021-08-24 Sonos, Inc. Systems and methods for selective wake word detection using neural network models
US10692518B2 (en) 2018-09-29 2020-06-23 Sonos, Inc. Linear filtering for noise-suppressed speech detection via multiple network microphone devices
CN111063344B (en) * 2018-10-17 2022-06-28 青岛海信移动通信技术股份有限公司 Voice recognition method, mobile terminal and server
US11899519B2 (en) 2018-10-23 2024-02-13 Sonos, Inc. Multiple stage network microphone device with reduced power consumption and processing load
EP3654249A1 (en) 2018-11-15 2020-05-20 Snips Dilated convolutions and gating for efficient keyword spotting
CN109614470B (en) * 2018-12-07 2023-08-08 北京小米移动软件有限公司 Method and device for processing answer information, terminal and readable storage medium
US11183183B2 (en) 2018-12-07 2021-11-23 Sonos, Inc. Systems and methods of operating media playback systems having multiple voice assistant services
US11132989B2 (en) 2018-12-13 2021-09-28 Sonos, Inc. Networked microphone devices, systems, and methods of localized arbitration
US10602268B1 (en) 2018-12-20 2020-03-24 Sonos, Inc. Optimization of network microphone devices using noise classification
US10867604B2 (en) 2019-02-08 2020-12-15 Sonos, Inc. Devices, systems, and methods for distributed voice processing
US11315556B2 (en) 2019-02-08 2022-04-26 Sonos, Inc. Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification
CN109817203B (en) * 2019-02-19 2021-07-27 广东小天才科技有限公司 A method and system for voice interaction
EP3709194A1 (en) 2019-03-15 2020-09-16 Spotify AB Ensemble-based data comparison
CN110061910B (en) * 2019-04-30 2021-11-30 上海掌门科技有限公司 Method, device and medium for processing voice short message
US11120794B2 (en) 2019-05-03 2021-09-14 Sonos, Inc. Voice assistant persistence across multiple network microphone devices
US11200894B2 (en) 2019-06-12 2021-12-14 Sonos, Inc. Network microphone device with command keyword eventing
US10586540B1 (en) 2019-06-12 2020-03-10 Sonos, Inc. Network microphone device with command keyword conditioning
US11361756B2 (en) 2019-06-12 2022-06-14 Sonos, Inc. Conditional wake word eventing based on environment
US10871943B1 (en) 2019-07-31 2020-12-22 Sonos, Inc. Noise classification for event detection
US11138969B2 (en) 2019-07-31 2021-10-05 Sonos, Inc. Locally distributed keyword detection
US11138975B2 (en) 2019-07-31 2021-10-05 Sonos, Inc. Locally distributed keyword detection
US11094319B2 (en) 2019-08-30 2021-08-17 Spotify Ab Systems and methods for generating a cleaned version of ambient sound
US11189286B2 (en) 2019-10-22 2021-11-30 Sonos, Inc. VAS toggle based on device orientation
US11200900B2 (en) 2019-12-20 2021-12-14 Sonos, Inc. Offline voice control
US11562740B2 (en) 2020-01-07 2023-01-24 Sonos, Inc. Voice verification for media playback
US11556307B2 (en) 2020-01-31 2023-01-17 Sonos, Inc. Local voice data processing
US11308958B2 (en) 2020-02-07 2022-04-19 Sonos, Inc. Localized wakeword verification
US11328722B2 (en) * 2020-02-11 2022-05-10 Spotify Ab Systems and methods for generating a singular voice audio stream
US11308959B2 (en) 2020-02-11 2022-04-19 Spotify Ab Dynamic adjustment of wake word acceptance tolerance thresholds in voice-controlled devices
CN111613207A (en) * 2020-05-11 2020-09-01 国网内蒙古东部电力有限公司呼伦贝尔供电公司 A transmission line inspection system
US11308962B2 (en) * 2020-05-20 2022-04-19 Sonos, Inc. Input detection windowing
US11482224B2 (en) 2020-05-20 2022-10-25 Sonos, Inc. Command keywords with input detection windowing
US11727919B2 (en) 2020-05-20 2023-08-15 Sonos, Inc. Memory allocation for keyword spotting engines
US11698771B2 (en) 2020-08-25 2023-07-11 Sonos, Inc. Vocal guidance engines for playback devices
US12283269B2 (en) 2020-10-16 2025-04-22 Sonos, Inc. Intent inference in audiovisual communication sessions
US11984123B2 (en) 2020-11-12 2024-05-14 Sonos, Inc. Network device interaction by range
CN115167734B (en) * 2022-07-01 2024-06-04 广州华数云计算有限公司 Message control method and device based on man-machine interaction interface

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1759383A (en) * 2003-01-09 2006-04-12 雅虎公司 Voice and video greeting system for personal advertisement and method
CN1957585A (en) * 2004-03-18 2007-05-02 诺基亚公司 System and associated terminal, method and computer program product for uploading content
US8428227B2 (en) * 2010-05-18 2013-04-23 Certicall, Llc Certified communications system and method
US8683355B1 (en) * 2008-06-24 2014-03-25 Sprint Communications Company L.P. Chat space system and method
CN104144097A (en) * 2013-05-07 2014-11-12 百度在线网络技术(北京)有限公司 Voice message transmission system, sending end, receiving end and voice message transmission method
US9036618B1 (en) * 2005-03-11 2015-05-19 Hewlett-Packard Development Company, L.P. Method and system for providing voice assisted configuration on an internet protocol (IP) telephone
CN106375182A (en) * 2016-08-22 2017-02-01 腾讯科技(深圳)有限公司 Voice communication method and device based on instant messaging application

Family Cites Families (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100636106B1 (en) * 1999-10-26 2006-10-18 삼성전자주식회사 Voice mail box management method using short message service in voice mail system and computer-readable medium therefor
DE10153526B4 (en) * 2001-10-30 2013-10-24 Ipcom Gmbh & Co. Kg Method for signaling a call originating for a telecommunication terminal, telecommunication terminal and network unit
US7272216B2 (en) * 2003-09-26 2007-09-18 Comverse Ltd. Incomplete call notification
US20050100145A1 (en) * 2003-10-01 2005-05-12 Spencer Bradford L. Multi-user intelligent call screening
US7295660B1 (en) * 2003-10-23 2007-11-13 Aol Llc Telemarketer screening
US7532710B2 (en) * 2003-11-21 2009-05-12 Verizon Business Global Llc Systems and methods for providing voicemail services
US20070192427A1 (en) * 2006-02-16 2007-08-16 Viktors Berstis Ease of use feature for audio communications within chat conferences
JP4537987B2 (en) * 2006-10-31 2010-09-08 株式会社東芝 Communication system and voice mail apparatus
US7877084B2 (en) * 2006-12-18 2011-01-25 International Business Machines Corporation Method and system for automatic call filtering based on user selectable parameters
US8306509B2 (en) * 2007-08-31 2012-11-06 At&T Mobility Ii Llc Enhanced messaging with language translation feature
CA2727951A1 (en) * 2008-06-19 2009-12-23 E-Lane Systems Inc. Communication system with voice mail access and call by spelling functionality
US8199888B2 (en) * 2008-12-04 2012-06-12 At&T Intellectual Property I, L.P. System and method for automatically transcribing voicemail
US8577543B2 (en) * 2009-05-28 2013-11-05 Intelligent Mechatronic Systems Inc. Communication system with personal information management and remote vehicle monitoring and control features
US8428562B2 (en) * 2009-11-19 2013-04-23 At&T Mobility Ii Llc Systems and methods for retrieving voicemail account information
US20130253971A1 (en) * 2010-10-19 2013-09-26 ClearCare, Inc. System and apparatus for generating work schedules
US20130304533A1 (en) * 2011-07-11 2013-11-14 ClearCare, Inc. System and apparatus for generating work schedules
US20140039962A1 (en) * 2010-10-19 2014-02-06 ClearCare, Inc. System and Apparatus for Generating Work Schedules
US20130191145A1 (en) * 2010-10-19 2013-07-25 ClearCare, Inc. System and apparatus for generating work schedules
US8824681B2 (en) * 2011-12-06 2014-09-02 Motorola Solutions, Inc. Method and device for link layer decrypting and/or encrypting a voice message stream already supporting end to end encryption
US8699677B2 (en) * 2012-01-09 2014-04-15 Comcast Cable Communications, Llc Voice transcription
US20140172539A1 (en) * 2012-12-14 2014-06-19 Apple Inc. Media station with custom bumper
US20150256679A1 (en) * 2014-03-08 2015-09-10 J. Stephen Burnett Electronic Message Aggregation and Sharing System and Apparatus
CN103986698A (en) * 2014-05-04 2014-08-13 苏州乐聚一堂电子科技有限公司 Karaoke mobile phone song query system with sound special effect
US20160219153A1 (en) * 2015-01-22 2016-07-28 John Wiley Honea Method for Providing Personalized Voicemails
US9538000B1 (en) * 2015-06-24 2017-01-03 Mast Mobile, Inc. Incoming call management in a communication environment with unified communication interfaces

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1759383A (en) * 2003-01-09 2006-04-12 雅虎公司 Voice and video greeting system for personal advertisement and method
CN1957585A (en) * 2004-03-18 2007-05-02 诺基亚公司 System and associated terminal, method and computer program product for uploading content
US9036618B1 (en) * 2005-03-11 2015-05-19 Hewlett-Packard Development Company, L.P. Method and system for providing voice assisted configuration on an internet protocol (IP) telephone
US8683355B1 (en) * 2008-06-24 2014-03-25 Sprint Communications Company L.P. Chat space system and method
US8428227B2 (en) * 2010-05-18 2013-04-23 Certicall, Llc Certified communications system and method
CN104144097A (en) * 2013-05-07 2014-11-12 百度在线网络技术(北京)有限公司 Voice message transmission system, sending end, receiving end and voice message transmission method
CN106375182A (en) * 2016-08-22 2017-02-01 腾讯科技(深圳)有限公司 Voice communication method and device based on instant messaging application

Also Published As

Publication number Publication date
CN106921560A (en) 2017-07-04
EP3367379A3 (en) 2018-10-31
EP3367379A2 (en) 2018-08-29
EP3367379B1 (en) 2023-04-19
US10728196B2 (en) 2020-07-28
US20180248823A1 (en) 2018-08-30

Similar Documents

Publication Publication Date Title
CN106921560B (en) Voice communication method, device and system
CN107908351B (en) Application interface display method and device and storage medium
US9967811B2 (en) Method and device for displaying WIFI list
CN109951379B (en) Message processing method and device
CN109245997B (en) Voice message playing method and device
CN107566892B (en) Video file processing method and device and computer readable storage medium
CN107038214A (en) Expression information processing method and processing device
KR101735755B1 (en) Method and apparatus for prompting device connection
WO2017088247A1 (en) Input processing method, device and apparatus
CN110704647B (en) Content processing method and device
CN106603381B (en) Method and device for processing chat information
CN111510556A (en) Method, device and computer storage medium for processing call information
CN108270661B (en) Information reply method, device and equipment
CN109302341B (en) Instant messaging method, instant messaging device, electronic equipment and storage medium
CN108011990B (en) Contact management method and device
CN107493366B (en) Address book information updating method and device and storage medium
CN105516457A (en) Communication message processing method and apparatus
CN105101121B (en) A kind of method and device that information is sent
CN109842543B (en) Instant messaging method and device and instant messaging message storage method and device
CN106506808B (en) Method and device for prompting communication message
CN109120499B (en) Information processing method and device
CN113286218B (en) Translation method and device and earphone equipment
CN113127613B (en) Chat information processing method and device
CN110196747B (en) Information processing method and device
CN104317480B (en) Character keys display methods, device and terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant