Disclosure of Invention
In view of the above, an object of the embodiments of the present invention is to provide a method and an apparatus for human-machine interaction, and a robot, so as to solve the above problems.
In a first aspect, an embodiment of the present invention provides a method for human-computer interaction, including:
the method comprises the following steps of collecting multimedia information of a user, wherein the multimedia information comprises: user image information and/or user voice information;
determining the identity and the interaction mode of the user according to multimedia information stored in a preset database and the collected multimedia information; wherein, the interaction mode comprises: voice interaction and/or action interaction;
and calling interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user.
With reference to the first aspect, an embodiment of the present invention provides a first possible implementation manner of the first aspect, where the acquiring the multimedia information of the user includes:
when monitoring that a user enters an image acquisition area, automatically acquiring multimedia information of the user;
or when the condition that the user triggers the acquisition option function is monitored, acquiring the multimedia information of the user.
With reference to the first aspect, an embodiment of the present invention provides a second possible implementation manner of the first aspect, where the acquiring the multimedia information of the user includes: when a phone ring is identified, automatically answering the phone and collecting voice information of a user in a call;
the determining the identity and the interaction mode of the user according to the multimedia information stored in the preset database and the collected multimedia information comprises the following steps: recognizing a voice keyword according to the collected voice information of the user in the call; determining the identity and the current interaction mode of the user as action interaction according to the voice keywords and voice recognition information stored in a preset database, wherein the voice keywords comprise the name of the user and/or the name of a person to which the user needs to connect;
the calling of the interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user comprises: and calling a telephone switching number corresponding to the identity of the user, and switching the call to a terminal corresponding to the telephone switching number.
With reference to any one of the first aspect to the second possible implementation manner of the first aspect, an embodiment of the present invention provides a third possible implementation manner of the first aspect, where the invoking of the interactive content corresponding to the identity of the user according to the determined interactive manner to interact with the user includes:
and recognizing the language type corresponding to the voice according to the voice information of the user, and calling interactive content corresponding to the identity of the user according to the determined interactive mode and the recognized language type to interact with the user.
With reference to the first aspect, an embodiment of the present invention provides a fourth possible implementation manner of the first aspect, where the method for human-computer interaction further includes: when the doorbell sound is identified, the automatic door is started to open.
With reference to the first aspect, an embodiment of the present invention provides a fifth possible implementation manner of the first aspect, where the invoking of the interactive content corresponding to the identity of the user according to the determined interactive manner to interact with the user includes:
searching interactive content corresponding to the identity of the user from a preset database according to a determined interactive mode, wherein one or more of the following are stored in the database in advance: the information of visiting clients and corresponding interactive contents, the information of staff and corresponding prompting instructions, the identities of identified strangers and corresponding image information are pre-prompted;
and calling the searched interactive content and interacting with the user.
With reference to the first aspect, an embodiment of the present invention provides a sixth possible implementation manner of the first aspect, where the method for human-computer interaction further includes:
when multimedia information in the inspection process is collected, judging the abnormal condition of the collected multimedia information;
when the multimedia information is identified to be in a burst or abnormal condition, alarming, wherein the alarming at least comprises one of the following conditions: the alarm is given by voice or light on site, and the alarm is given to the customer service center or related personnel.
In a second aspect, an embodiment of the present invention further provides a method and an apparatus for human-computer interaction, where the apparatus includes:
the information acquisition module is used for acquiring multimedia information of a user, wherein the multimedia information comprises: user image information and/or user voice information;
the identity and interaction determining module is used for determining the identity and the interaction mode of the user according to the multimedia information stored in a preset database and the collected multimedia information; wherein, the interaction mode comprises: voice interaction and/or action interaction;
and the interactive content calling module is used for calling the interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user.
With reference to the second aspect, an embodiment of the present invention provides a first possible implementation manner of the second aspect, where the information acquisition module includes:
the first acquisition unit is used for automatically acquiring multimedia information of a user when the situation that the user enters an image acquisition area is monitored; or,
and the second acquisition unit is used for acquiring the multimedia information of the user when the condition that the user triggers the acquisition option function is monitored.
With reference to the second aspect, an embodiment of the present invention provides a second possible implementation manner of the second aspect, where the information acquisition module includes:
the call voice acquisition unit is used for automatically answering the call when a call ringtone is identified, and acquiring voice information of a user in the call;
the identity and interaction determination module comprises:
the keyword recognition unit is used for recognizing voice keywords according to the collected voice information of the user in the call;
the identity determining unit of the user is used for determining the identity of the user and the current interaction mode as action interaction according to the voice keywords and voice recognition information stored in a preset database, wherein the voice keywords comprise the name of the user and/or the name of a person to which the user needs to connect;
the interactive content calling module comprises:
and the transfer number calling unit is used for calling a telephone transfer number corresponding to the identity of the user and transferring the call to a terminal corresponding to the telephone transfer number.
With reference to any one of the second aspect to the second possible implementation manner of the second aspect, an embodiment of the present invention provides a third possible implementation manner of the second aspect, where the interactive content retrieving module includes:
the language type identification unit is used for identifying the language type corresponding to the voice according to the voice information of the user;
and the first interactive content calling unit is used for calling the interactive content corresponding to the identity of the user according to the determined interactive mode and the identified language type to interact with the user.
With reference to the second aspect, an embodiment of the present invention provides a fourth possible implementation manner of the second aspect, where the human-computer interaction apparatus further includes:
and the automatic door starting module is used for starting the automatic door to open when the doorbell sound is identified.
With reference to the second aspect, an embodiment of the present invention provides a fifth possible implementation manner of the second aspect, where the interactive content retrieving module includes:
the interactive content searching unit is used for searching interactive content corresponding to the identity of the user from a preset database according to a determined interactive mode, wherein one or more of the following are stored in the database in advance: the information of visiting clients and corresponding interactive contents, the information of staff and corresponding prompting instructions, the identities of identified strangers and corresponding image information are pre-prompted;
and the second interactive content calling unit is used for calling the searched interactive content and interacting with the user.
With reference to the second aspect, an embodiment of the present invention provides a sixth possible implementation manner of the second aspect, where the human-computer interaction apparatus further includes:
the abnormal condition judgment module is used for judging the abnormal condition of the collected multimedia information when the multimedia information in the inspection process is collected;
the alarm module is used for alarming when the multimedia information is identified to be in a burst or abnormal condition, wherein the alarm at least comprises one of the following: the alarm is given by voice or light on site, and the alarm is given to the customer service center or related personnel.
In a third aspect, an embodiment of the present invention further provides a robot, including: the man-machine interaction device.
In the method and the device for man-machine interaction provided by the embodiment of the invention, firstly, multimedia information of a user is collected, wherein the multimedia information comprises: user image information and/or user voice information; then, determining the identity and the interaction mode of the user according to the multimedia information stored in a preset database and the collected multimedia information; wherein, the interactive mode includes: voice interaction and/or action interaction; and finally, calling the interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user, calling the interactive content matched with the identified user identity according to the identified user identity, and further interacting with the user, so that the diversity of man-machine interaction is realized, and a man-machine interaction mode is realized in a targeted manner.
In order to make the aforementioned and other objects, features and advantages of the present invention comprehensible, preferred embodiments accompanied with figures are described in detail below.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
Considering that most of robots mainly achieve a question or guide function at present, the functions of the robots are single, and intelligence is not high, so that a human-computer interaction effect is poor. Based on this, the embodiment of the invention provides a method, a device and a robot for human-machine interaction, which are described below by embodiments.
As shown in fig. 1, an embodiment of the present invention provides a method for human-computer interaction, where the method includes steps S102 to S106, which are specifically as follows:
step S102: collecting multimedia information of a user, wherein the multimedia information can be user image information, user voice information, or user image information and user voice information which are collected simultaneously;
the method can acquire the multimedia information of the user in the following ways:
one mode is that the multimedia information of a user entering an image acquisition area is actively acquired by real-time monitoring, namely, the multimedia information of the user is automatically acquired when the situation that the user enters the image acquisition area is monitored;
the other mode is to passively collect the multimedia information of the user entering the image collecting area, namely when it is monitored that the user triggers the collecting option function, the multimedia information of the user is collected, and if it is detected that the user presses a collecting key or clicks a check box on a touch screen, the collecting action is started.
Step S104: determining the identity and the interaction mode of the user according to multimedia information stored in a preset database and the collected multimedia information; wherein, the above-mentioned interactive mode includes: voice interaction and/or action interaction;
specifically, the user image information collected in step S102 is compared with image information stored in a preset database, so as to determine the identity of the user; meanwhile, identifying keywords according to the collected voice information of the user, determining an interaction mode according to the identified keywords, and if the user is identified as a visitor and needs to find the working hours of a company XXX, the interaction mode is action interaction and voice interaction, and then calling corresponding interaction content and interacting through the following step S106: first asked in a fashion to receive guests and then the user is directed to a workstation of an employee of company XXX.
Step S106: and calling interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user.
After the user identity and the interaction mode are determined in step S104, step S106 is executed to retrieve the corresponding interaction content according to the user identity and the interaction mode and perform interaction, that is, retrieve the interaction content corresponding to the user identity in a preset database according to the identified user identity and the determined interaction mode for communication, for example: if the user is judged to be a company employee, then question XXX morning good/noon good; if the identity is not recognized, the user can be considered as a visitor, and then a question is asked, namely 'welcome to XXX company and ask what you need help'; and then, identifying a keyword according to the voice information of the user, retrieving a relevant answer from a preset database, communicating and interacting with the user, and in addition, under the condition that a matched answer is not retrieved, transferring to a customer service center, continuously serving visitors or people in a telephone by manual service, and simultaneously realizing face-to-face communication and interaction in a video call mode.
In the human-computer interaction method provided by the embodiment of the invention, firstly, multimedia information of a user is collected, wherein the multimedia information comprises: user image information and/or user voice information; then, determining the identity and the interaction mode of the user according to the multimedia information stored in a preset database and the collected multimedia information; wherein, the interactive mode includes: voice interaction and/or action interaction; and finally, calling interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user, wherein in the embodiment of the invention, the identity of the user is identified by using a face identification technology, the interactive content matched with the identified user is called according to the identified user identity, meanwhile, the keyword is identified by using a voice keyword identification technology, the interactive content is determined according to the keyword and the user identity, and then the interactive content interacts with the user, so that the diversity of man-machine interaction is realized, the flexibility of man-machine interaction is improved, different interactive contents are called for different users, and a mode of performing man-machine interaction in a targeted manner is realized.
Further, considering that there may be a guest calling the company at any time during the working hours or the working hours, based on this, fig. 2 shows another flow chart of the man-machine interaction method, which is as follows:
step S202: when a phone ring is identified, automatically answering the phone and collecting voice information of a user in a call;
step S204: recognizing a voice keyword according to the collected voice information of the user in the call;
step S206: determining the identity and the current interaction mode of the user as action interaction according to the voice keywords and voice recognition information stored in a preset database, wherein the voice keywords comprise the name of the user and/or the name of a person to which the user needs to connect;
step S208: and calling a telephone switching number corresponding to the identity of the user, and switching the call to a terminal corresponding to the telephone switching number.
The telephone switching number corresponding to the identity of the user can comprise a called mobile phone number, an extension number and a customer service center number, wherein if the identity of the user is XX for needing to contact company employees, the XX telephone is used as the telephone switching number, and if the XX telephone for company employees is not inquired in a preset database, the customer service center number is called as the telephone switching number.
In addition, the telephone answering process can interact with the user in the call to directly answer the consultation questions of the user, specifically, when the call is identified, the telephone can be automatically answered, keywords are identified according to the voice of the user in the call, then relevant answers are retrieved from a preset database to be communicated and interacted with the user in the call, and in addition, under the condition that matched answers are not retrieved, the user can be transferred to a customer service center to continue to serve the personnel in the call through manual service.
In the embodiment of the invention, by setting the function of automatically answering the call and combining the voice keyword recognition technology, the interactive content corresponding to the identity of the user can be output, and the call can be automatically transferred to the call transfer number required by the user in the call, so that the diversity of man-machine interaction is further realized, and the flexibility of man-machine interaction is improved.
Further, since there may be a plurality of language types of the collected users, in order to better improve the user experience, based on this, invoking the interactive content corresponding to the identity of the user according to the determined interactive mode to interact with the user includes:
and recognizing the language type corresponding to the voice according to the voice information of the user, and calling interactive content corresponding to the identity of the user according to the determined interactive mode and the recognized language type to interact with the user.
The method can also identify the appearance characteristics of the user according to the image information of the user, call the interactive content corresponding to the identity of the user according to the determined interactive mode and the language type corresponding to the identified appearance characteristics to interact with the user, can determine the language type in the two modes, can ensure that the determined language type is more accurate if the two modes are simultaneously adopted, and can determine the language type according to the multimedia information of the user in both modes of identifying the language type corresponding to the voice according to the voice information of the user and identifying the appearance characteristics of the user according to the image information of the user and then determining the language type according to the appearance characteristics, so that the multi-language environment communication can be realized.
In the embodiment of the invention, the language type corresponding to the voice is identified according to the collected language information of the user or the appearance characteristic of the user is identified according to the collected image information of the user, then the language type is determined according to the appearance characteristic, the corresponding language type is adopted to communicate and interact with the visitor, and the flexibility of man-machine interaction is further increased and the experience of the visitor is improved by adopting a multi-language environment communication mode.
Further, the method further comprises: when the doorbell sound is identified, the automatic door is started to open.
In the embodiment of the present invention, the automatic door can be automatically opened, specifically including: when the sound of the doorbell is identified, the automatic door is automatically opened, the collected image information of the user is recorded, and a targeted interaction mode is adopted, so that real-time supervision and subsequent tracking are facilitated.
In order to further increase the flexibility of human-computer interaction, the calling the interactive content corresponding to the identity of the user according to the determined interaction mode to interact with the user comprises:
searching interactive content corresponding to the identity of the user from a preset database according to a determined interactive mode, wherein one or more of the following are stored in the database in advance: the information of visiting clients and corresponding interactive contents, the information of staff and corresponding prompting instructions, the identities of identified strangers and corresponding image information are pre-prompted;
and calling the searched interactive content and interacting with the user.
Specifically, the information of the visiting client and the corresponding interactive content may be stored in a preset database in advance, for example: when a certain leader may be observed in the near future, the identity recognition data (face image and/or voice information) of the leader can be temporarily added, a corresponding identity recognition mode (face recognition and/or voice recognition) is added to the robot, and a man-machine interaction mode (such as a welcome mode specially set for the robot) corresponding to the identity recognition mode is added; if the user is judged to be an expert/professor for clinical pre-stored designs, then a question of "strongly welcome XXX expert/professor to guide XXX company" is asked; therefore, the robot can be controlled to perform some special interaction modes, and the flexibility and the practicability of the man-machine interaction are further improved;
the information of the staff to be prompted and the corresponding prompting instruction can be stored in a preset database in advance, for example, for internal staff, the required interactive service can be set on the robot, after the identity of the staff is recognized, some corresponding interactive modes can be provided for the staff with the identity, if the internal staff needs to be in an X meeting in the next day, meeting-opening reminding service can be set on the robot in advance, and after the robot stores the setting, the set man-machine interactive operation for reminding can be executed after the staff is recognized in the next day; for another example, the reminding operation can also be to remind whether to take a work card or not, or whether to arrive late according to the current time, and the like, and the flexibility and the practicability of man-machine interaction can be further enhanced by combining the identity recognition and the setting of a man-machine interaction mode;
the identity of the identified stranger and the corresponding image information can also be stored in a preset database in real time, and the method specifically comprises the following steps: the method comprises the steps of recording images of a person who enters a company for the first time and storing the identity of the person, automatically identifying the identity of the person who enters the company again, determining and outputting a communication statement according to the stored identity of the person, and carrying out communication and interaction corresponding to the identity of the person, wherein the communication statement comprises the following steps: the identity of the person is identified as the courier through the first communication, and the courier asks the company to ask the courier in an active manner for the second time.
In the embodiment of the invention, the user information and the corresponding interactive content are preset in the preset database, and the human-computer interaction mode is flexibly set, and meanwhile, the human face identity memory function is added, so that the flexibility of human-computer interaction is further increased, and the user experience is improved.
In addition, the method further comprises: when multimedia information in the polling process is collected, judging the abnormal condition of the collected multimedia information;
when the multimedia information is identified to be in a burst or abnormal condition, alarming, wherein the alarming at least comprises one of the following conditions: the alarm is given by voice or light on site, and the alarm is given to the customer service center or related personnel.
Specifically, when the emergency or abnormal condition is identified, the alarm module can be used for alarming, wherein the alarm mode can be on-site voice alarm, light flicker alarm, switching to a customer service center or switching to a preset mobile phone, so that real-time monitoring is carried out through a camera acquisition device on a robot, 24-hour day and night monitoring is realized, if a person breaks into the robot outside the working hours, an alarm system can be triggered in time, the robot can be pushed to a company foreground system and an APP, the building property security is informed in time, and the loss of company properties is avoided; meanwhile, the monitoring data are stored in the cloud server, the cloud server is communicated with the client in a wireless communication mode, and a user on the corresponding client can acquire required information from the cloud server at any time.
In summary, in the method for human-computer interaction provided in the embodiment of the present invention, first, multimedia information of a user is collected, where the multimedia information includes: user image information and/or user voice information; then, determining the identity and the interaction mode of the user according to the multimedia information stored in a preset database and the collected multimedia information; wherein, this interactive mode includes: voice interaction and/or action interaction; finally, the interactive content corresponding to the identity of the user is called to interact with the user according to the determined interactive mode, the embodiment of the invention identifies the identity of the user by using a face recognition technology, calls the interactive content matched with the user according to the identified user identity, simultaneously identifies the keywords by using a voice keyword recognition technology, determines the interactive content according to the keywords and the user identity, and further interacts with the user, so that the diversity of man-machine interaction is realized, the flexibility of man-machine interaction is improved, different interactive contents are called for different users, and a mode of performing man-machine interaction in a targeted manner is realized; furthermore, by setting the function of automatically answering the call and combining the voice keyword recognition technology, the interactive content corresponding to the identity of the user can be output, and the call can be automatically transferred to the call transfer number required by the user in the call, so that the diversity of man-machine interaction is further realized, and the flexibility of man-machine interaction is improved; furthermore, the language type corresponding to the voice is identified according to the collected language information of the user or the appearance characteristics of the user are identified according to the collected image information of the user, then the language type is determined according to the appearance characteristics, the corresponding language type is adopted to communicate and interact with the visitor, and the flexibility of man-machine interaction is further improved and the experience degree of the visitor is improved by adopting a multi-language environment communication mode; furthermore, the user information and the corresponding interactive content are preset in the preset database, the man-machine interaction mode is flexibly set, the face identity memory function is added, the flexibility of man-machine interaction is further improved, and the user experience is improved.
As shown in fig. 3, an embodiment of the present invention provides a human-computer interaction apparatus, including:
an information collecting module 402, configured to collect multimedia information of a user, where the multimedia information includes: user image information and/or user voice information;
an identity and interaction determining module 404, configured to determine an identity and an interaction manner of the user according to multimedia information stored in a preset database and the collected multimedia information; wherein, the above-mentioned interactive mode includes: voice interaction and/or action interaction;
and an interactive content retrieving module 406, configured to retrieve, according to the determined interactive manner, interactive content corresponding to the identity of the user to interact with the user.
Further, the information collecting module 402 includes:
the first acquisition unit is used for automatically acquiring the multimedia information of the user when the situation that the user enters an image acquisition area is monitored; or,
and the second acquisition unit is used for acquiring the multimedia information of the user when the condition that the user triggers the acquisition option function is monitored.
Further, as shown in fig. 4, the information collecting module 402 includes:
the call voice acquisition unit 4021 is used for automatically answering a call and acquiring voice information of a user in a call when a call ringtone is identified;
the identity and interaction determination module 404 includes:
a keyword recognition unit 4041, configured to recognize a voice keyword according to the collected voice information of the user in the call;
a user identity determining unit 4042, configured to determine, according to the voice keyword and voice recognition information stored in a preset database, that the identity of the user and the current interaction mode are action interactions, where the voice keyword includes a name of the user and/or a name of a person to whom the user needs to connect a call;
the interactive content retrieving module 406 includes:
the transfer number retrieving unit 4061 is configured to retrieve a call transfer number corresponding to the identity of the user, and transfer the call to a terminal corresponding to the call transfer number.
Further, as shown in fig. 5, the interactive content retrieving module 406 includes:
a language type recognition unit 4062, configured to recognize a language type corresponding to the voice according to the voice information of the user;
the first interactive content retrieving unit 4063 is configured to retrieve the interactive content corresponding to the identity of the user according to the determined interaction manner and the identified language type to interact with the user.
Further, the above human-computer interaction device further comprises:
and the automatic door starting module is used for starting the automatic door to open when the doorbell sound is identified.
Further, the interactive content retrieving module 406 includes:
the interactive content searching unit is used for searching interactive content corresponding to the identity of the user from a preset database according to a determined interactive mode, wherein one or more of the following are stored in the database in advance: the information of visiting clients and corresponding interactive contents, the information of staff and corresponding prompting instructions, the identities of identified strangers and corresponding image information are pre-prompted;
and the second interactive content calling unit is used for calling the searched interactive content and interacting with the user.
Further, the above human-computer interaction device further comprises:
the abnormal condition judging module is used for judging the abnormal condition of the collected multimedia information when the multimedia information in the polling process is collected;
the alarm module is used for alarming when the multimedia information is identified to be in a burst or abnormal condition, wherein the alarm at least comprises one of the following: the alarm is given by voice or light on site, and the alarm is given to the customer service center or related personnel.
An embodiment of the present invention further provides a robot, including: the man-machine interaction device.
Considering that most of robots mainly achieve a question or guide function at present, the functions of the robots are single and not high in intelligence, the functions which can be achieved by the existing robots are not enough to meet the requirements of supervision of foreground work and cannot replace foreground personnel to complete related work of a foreground, the robot provided by the embodiment of the invention is embedded with a face recognition function, the identity of the personnel entering a company is judged by acquiring image information of the personnel, then the image information is compared with image information stored in a database in advance, communication is carried out by using a preset communication mode, a voice keyword recognition function is embedded at the same time, the voice information of the personnel entering the company is acquired, then keywords are recognized, and interactive contents are called according to the keywords and the personnel identities, so that the robot provided by the embodiment of the invention can be used for replacing the personnel to accurately communicate with visitors, Interaction can better meet the actual requirements of foreground robots; and other scenes requiring man-machine interaction diversity and flexibility can be applied.
Further, a card punching system can be embedded into the robot, and an attendance record can be automatically generated, specifically including: through identifying the images of the personnel who enter and exit, the time of the personnel who enter and exit the company is recorded, and an attendance record list is generated, wherein the record items comprise the date, the working time and the working time, and the working time can be recorded by the company.
Further, a predetermined module may be embedded in the robot: such as meeting room reservations and the like.
Furthermore, a sound source positioning function can be embedded into the robot, and the robot can adjust steering in time according to the position of a sound source and always face the human face.
Based on the above analysis, compared with the robot in the related art, the human-computer interaction device and the robot provided in the embodiment of the present invention first collect multimedia information of a user, where the multimedia information includes: user image information and/or user voice information; then, determining the identity and the interaction mode of the user according to the multimedia information stored in a preset database and the collected multimedia information; wherein, this interactive mode includes: voice interaction and/or action interaction; finally, the interactive content corresponding to the identity of the user is called to interact with the user according to the determined interactive mode, the embodiment of the invention identifies the identity of the user by using a face recognition technology, calls the interactive content matched with the user according to the identified user identity, simultaneously identifies the keywords by using a voice keyword recognition technology, determines the interactive content according to the keywords and the user identity, and further interacts with the user, so that the diversity of man-machine interaction is realized, the flexibility of man-machine interaction is improved, different interactive contents are called for different users, and a mode of performing man-machine interaction in a targeted manner is realized; furthermore, by setting the function of automatically answering the call and combining the voice keyword recognition technology, the interactive content corresponding to the identity of the user can be output, and the call can be automatically transferred to the call transfer number required by the user in the call, so that the diversity of man-machine interaction is further realized, and the flexibility of man-machine interaction is improved; furthermore, the language type corresponding to the voice is identified according to the collected language information of the user or the appearance characteristics of the user are identified according to the collected image information of the user, then the language type is determined according to the appearance characteristics, the corresponding language type is adopted to communicate and interact with the visitor, and the flexibility of man-machine interaction is further improved and the experience degree of the visitor is improved by adopting a multi-language environment communication mode; furthermore, the user information and the corresponding interactive content are preset in the preset database, the man-machine interaction mode is flexibly set, the face identity memory function is added, the flexibility of man-machine interaction is further improved, and the user experience is improved.
The man-machine interaction device provided by the embodiment of the invention can be specific hardware on the equipment or software or firmware installed on the equipment. The device provided by the embodiment of the present invention has the same implementation principle and technical effect as the method embodiments, and for the sake of brief description, reference may be made to the corresponding contents in the method embodiments without reference to the device embodiments. It is clear to those skilled in the art that, for convenience and brevity of description, the specific working processes of the foregoing systems, apparatuses and units may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units is only one logical division, and there may be other divisions when actually implemented, and for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some communication interfaces, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments provided by the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit.
The functions, if implemented in the form of software functional units and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: various media capable of storing program codes, such as a usb disk, a removable hard disk, a Read-only memory (ROM), a Random Access Memory (RAM), a magnetic disk, or an optical disk.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus once an item is defined in one figure, it need not be further defined and explained in subsequent figures, and moreover, the terms "first", "second", "third", etc. are used merely to distinguish one description from another and are not to be construed as indicating or implying relative importance.
Finally, it should be noted that: the above-mentioned embodiments are only specific embodiments of the present invention, which are used for illustrating the technical solutions of the present invention and not for limiting the same, and the protection scope of the present invention is not limited thereto, although the present invention is described in detail with reference to the foregoing embodiments, those skilled in the art should understand that: any person skilled in the art can modify or easily conceive the technical solutions described in the foregoing embodiments or equivalent substitutes for some technical features within the technical scope of the present disclosure; such modifications, changes or substitutions do not depart from the spirit and scope of the present invention in its spirit and scope. Are intended to be covered by the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.