This application claims priority to Chinese Patent Application No. 201710752403.5 filed on Aug. 28, 2017, the contents of which are incorporated by reference herein.
The subject matter herein generally relates to an interactive robot.
Interactive robots are currently limited in the ways they can interact with people.
Implementations of the present disclosure will now be described, by way of example only, with reference to the attached figures.
It will be appreciated that for simplicity and clarity of illustration, where appropriate, reference numerals have been repeated among the different figures to indicate corresponding or analogous elements. In addition, numerous specific details are set forth in order to provide a thorough understanding of the embodiments described herein. However, it will be understood by those of ordinary skill in the art that the embodiments described herein can be practiced without these specific details. In other instances, methods, procedures and components have not been described in detail so as not to obscure the related relevant feature being described. The drawings are not necessarily to scale and the proportions of certain parts may be exaggerated to better illustrate details and features. The description is not to be considered as limiting the scope of the embodiments described herein.
Several definitions that apply throughout this disclosure will now be presented.
The term “coupled” is defined as connected, whether directly or indirectly through intervening components, and is not necessarily limited to physical connections. The connection can be such that the objects are permanently connected or releasably connected. The term “substantially” is defined to be essentially conforming to the particular dimension, shape, or other word that “substantially” modifies, such that the component need not be exact. For example, “substantially cylindrical” means that the object resembles a cylinder, but can have one or more deviations from a true cylinder. The term “comprising” means “including, but not necessarily limited to”; it specifically indicates open-ended inclusion or membership in a so-described combination, group, series and the like.
In general, the word “module” as used hereinafter refers to logic embodied in hardware or firmware, or to a collection of software instructions, written in a programming language such as, for example, Java, C, or assembly. One or more software instructions in the modules may be embedded in firmware such as in an erasable-programmable read-only memory (EPROM). It will be appreciated that the modules may comprise connected logic units, such as gates and flip-flops, and may comprise programmable units, such as programmable gate arrays or processors. The modules described herein may be implemented as either software and/or hardware modules and may be stored in any type of computer-readable medium or other computer storage unit.
The input module 11 can include, but is not limited to, an image input element 111, an audio input element 112, an olfactory input element 113, a pressure input element 114, an infrared input element 115, a temperature input element 116, and a touch input element 117.
The image input element 111 is used for capturing images from around the interactive robot 1. For example, the image input element 111 can capture images of a person or an object. In at least one embodiment, the image input element 111 can be a camera.
The audio input element 112 is used for capturing audio from around the interactive robot 1. In at least one embodiment, the audio input element 112 can be a microphone array.
The olfactory input element 113 is used for capturing smells from around the interactive robot 1.
The pressure input element 114 is used for detecting an external pressure on the interactive robot 1.
The infrared input element 115 is used for detecting heat signatures of people around the interactive robot 1.
The temperature input element 116 is used for detecting a temperature around the interactive robot 1.
The touch input element 117 is used for receiving touch input from a user. In at least one embodiment, the touch input element 117 can be a touch screen.
The output module 12 can include, but is not limited to, an audio output element 121, a facial expression output element 122, a display output element 123, and a movement output element 124.
The audio output element 121 is used for outputting audio. In at least one embodiment, the audio output element 121 can be a loudspeaker.
The facial expression output element 122 is used for outputting a facial expression. In at least one embodiment, the facial expression output element 122 can include eyes, eyelids, and a mouth of the interactive robot 1.
The display output element 123 is used for outputting text, images, or videos. In other embodiments, the display output element 123 can display a facial expression. In other embodiments, the touch input element 117 and the display output element 123 can be the same display screen.
The movement output element 124 is used for moving the interactive robot 1. The movement output element 124 can include a first driving element 1241, two second driving elements 1242, and a third driving element 1243. Referring to
The communication unit 13 is used for providing communication between the interactive robot 1 and the server 2. In at least one embodiment, the communication unit 13 can use WIFI, ZIGBEE, BLUETOOTH, or other wireless communication method.
The storage unit 15 can store a plurality of instructions of the interactive system 3, and the interactive system 3 can be executed by the processor 14. In another embodiment, the interactive system 3 can be embedded in the processor 14. The image acquisition system 100 can be divided into a plurality of modules, which can include one or more software programs in the form of computerized codes stored in the storage unit 15. The computerized codes can include instructions executed by the processor 14 to provide functions for the modules. The storage device 20 can be a read-only memory, random access memory, or an external storage device such as a magnetic disk, a hard disk, a smart media card, a secure digital card, a flash card, or the like.
The processor 14 can be a central processing unit, a microprocessing unit, or other data processing chip.
Referring to
The establishing module 31 can establish at least one standby input element of the input module 11 and establish at least one standby output element of the output module 12.
In at least one embodiment, the establishing module 31 provides an interface 40 (shown in
The obtaining module 32 can obtain input information from the at least one standby input element. For example, when the image input element 111 is established as the standby input element, the obtaining module 32 can obtain images captured by the image input element 111. When the audio input element 112 is established as the standby input element, the obtaining module 32 can obtain audio input from the audio input element 112.
The analyzing module 33 can analyze the input information obtained by the obtaining module 32 and generate a control command according to the input information.
The executing module 34 can execute the control command to generate an output and output the output through the at least one standby output element.
In at least one embodiment, the audio input element 112 is established as the standby input element and the display output element 123 is established as the standby output element. The obtaining module 32 obtains the input information in the form of audio input, and the analyzing module 33 analyzes the audio input to recognize words to generate the control command according to the audio input. In at least one embodiment, the storage unit 15 stores a first relationship table S1 (shown in
In at least one embodiment, the audio input element 112 is established as the standby input element, and the audio output element 121 is established as the standby output element. The obtaining module 32 obtains the input information in the form of audio input, and the analyzing module 33 analyzes the audio input to recognize words to generate the control command according to the audio input. The first relationship table S1 can include the words “play the song . . . ” and the control command “play the song . . . ”. When the words “play the song . . . ” are recognized by the analyzing module 33, the analyzing module 33 generates the control command “play the song . . . ” according to the first relationship table S1. For example, the storage unit 15 can store a plurality of songs, and the analyzing module 33 can determine the song mentioned in the words of the input information. The executing module 34 executes the control command by controlling the audio output element 121 to play the corresponding song. In detail, the executing module 34 opens a stored music library (not shown) and searches for the song according to the audio input and controls the audio output element 121 to play the song.
In at least one embodiment, the audio input element 112 and the image input element 111 are established as the standby input elements, and the audio output element 121, the facial expression output element 122, the display output element 123, and the movement output element 124 are established as the standby output elements. The obtaining module 32 obtains the input information from the audio input element 112 and the image input element 111. The analyzing module 33 analyzes the input information to recognize a target. In at least one embodiment, the analyzing module 33 recognizes the target according to voiceprint characteristics and facial features of the target. The target can be a person or an animal. In at least one embodiment, the storage unit 15 stores a second relationship table (not shown). The second relationship table defines a preset relationship among the target and the recognized voiceprint characteristics and facial features.
The analyzing module 33 analyzes the input information from the audio input element 112 and the image input element 111 to obtain key information. In detail, the key information of the input information from the audio input element 112 is obtained by converting the input information from the audio input element 112 into text data. The key information of the input information from the image input element 111 is obtained by determining facial expression parameters and limb movement parameters.
The analyzing module 33 searches a preset public knowledge library according to the key information and uses a deep learning algorithm on the public knowledge library to determine a response. The response is a control command for controlling the standby output elements. For example, the audio output element 121 is controlled to output an audio response, the facial expression output element 122 is controlled to output a facial expression response, the display output element 123 is controlled to output a display response, and the movement output element 124 is controlled to output a movement response. In such a way, the interactive robot 1 can interact with the target.
In at least one embodiment, the public knowledge library can include information related to, but not limited to, human ethics, laws and regulations, moral sentiment, religion, astronomy, and geography. The public knowledge library can be stored in the storage unit 15. In other embodiments, the public knowledge library can be stored in the server 2. In at least one embodiment, the deep learning algorithm can include, but is not limited to, a neuro-bag model, a recurrent neural network, and a convolutional neural network.
The executing module 34 executes the control commands for controlling the corresponding standby output elements. The executing module 34 controls the audio output element 121 to output audio and the facial expression output element 122 to display a facial expression. For example, if a user smiles toward the interactive robot 1 and says, “these flowers are beautiful!”, the analyzing module 33 can identify the user as the target and determine the key information of the words to be “flowers”, “beautiful”, determine the key information of the images to be “smile”, search the public knowledge library according to the key information, and use the deep learning algorithm on the public knowledge library to determine the response. The response can control the audio output element 121 to output “These flowers are really beautiful, I also like them!” and control the facial expression output element 122 to display a smiling face by controlling the eyelids, eyes, and mouth.
In another embodiment, the executing module 34 can control the movement output element 124 to control the interactive robot 1 to move and control the display output element 123 to display a facial expression. For example, when the user smiles at the interactive robot 1 and says, “these flowers are really pretty!”, the executing module 34 can control the first driving element 1241 of the movement output element 124 to rotate the head 101 360 degrees, control the third driving element 1243 to drive the wheels 105 to rotate the interactive robot 1 in a circle, and control the display output element 123 to output a preset facial expression.
The embodiments shown and described above are only examples. Even though numerous characteristics and advantages of the present technology have been set forth in the foregoing description, together with details of the structure and function of the present disclosure, the disclosure is illustrative only, and changes may be made in the detail, including in matters of shape, size and arrangement of the parts within the principles of the present disclosure up to, and including, the full extent established by the broad general meaning of the terms used in the claims.
| Number | Date | Country | Kind |
|---|---|---|---|
| 201710752403.5 | Aug 2017 | CN | national |