Individuals live busy lives that can be very stressful. For example, parents can be stressed from the daily grind of getting children ready for school, commuting to work, problems that arise at work, shuttling children to and from activities, and the like. Individuals can cope with stress in a variety of different ways. For example, some individuals may get a massage, some individuals may meditate, and so forth.
Examples described herein provide an apparatus that generates images based on emotions of a user. The images can be used by individuals as a coloring image to provide relaxation and stress relief. In one example, the image may be generated based on identified emotions of a user.
Examples herein include a printer (e.g., a two dimensional printer or a three dimensional printer) that is enabled with voice detection and a network connection. The printer may prompt a user to speak a word or a phrase. The phrase may be captured and transmitted to a server in a network that generates emotional resonance images based on analysis of the phrase spoken by the user. The emotional resonance image may be a Mandala image that can be used as an adult coloring page to provide relaxation.
In one example, the IP network 102 may include an image generator 104, a web-based voice assistant 108, and a web-based application service 106. In one example, the web-based voice assistant 108, the web-based application service 106, and the image generator 104 may be communicatively coupled to one another over the IP network 102.
In one example, the web-based voice assistant 108 may be a service that works in coordination with a voice assistant application that is executed on an endpoint device. The web-based voice assistant 108 may receive voice commands for execution from the endpoint device that are sent over the IP network 102 to the web-based voice assistant 108 for execution. Examples of the web-based voice assistant 108 may include Google™ Asisstant, Siri™ Cortana™, Amazon Alexa™, and the like.
In one example, the web-based application service 106 may provide services to connect web-based applications (e.g., the web-based voice assistant 108) with third party applications or services. In one example, the web-based application service 106 may include services such as Macedon™ Web Services. The web-based application services 106 may allow the web-based voice assistant 108 to work with the image generator 104.
In one example, the image generator 104 may be a remotely located server or computing device located in the IP network 102. The image generator 104 may generate an image 114 based on emotions of a user 118. The image 114 may be a Mandala image that can be used to relieve stress of the user 118. For example, the image 114 may be colored for relaxation.
In one example, the system 100 may include a printer 110 that is connected to the IP network 102. The printer 110 may be connected to the IP network 102 via a wired or wireless connection.
The printer 110 may include a microphone 116. The microphone 116 may be integrated with the printer 110. In another example, the microphone 116 may be external to the printer 110. For example, the microphone 116 may be connected via universal serial bus (USB) connection to printer 110 or may be a mobile endpoint device wirelessly connected to the printer 110. For example, an application executed on a smart phone of the user 118 may record a phrase spoken by the user and wirelessly transmit the recording of the phrase to the printer 110.
In one example, the printer 110 may be communicatively coupled to a display 112. In one example, the display 112 may be part of the printer 110. In another example, the display 112 may be an external display. For example, the display 112 may be a monitor or part of a computing system (e.g., a desktop computer, a laptop computer, and the like).
In one example, the user 118 may activate the web-based voice assistant 108 by speaking a wake command that is captured by the microphone 116. The web-based voice assistant 108 may then prompt the user 118 to speak a phrase via a speaker on the printer 110 or a computing device connected to the printer 110. The phrase may be any sentence, group of words, a word, and the like, that the user 118 wishes to speak.
In one example, the phrase spoken by the user 118 may be captured by the microphone 116 and recorded by the printer 110. A recording of the phrase may then be transmitted to the web-based voice assistant 108 over the IP network 102. The phrase may then be transmitted to the image generator 104 via the web-based application service 106.
The image generator 104 may then analyze the phrase to identify an emotion of the user 118. Based on an emotion of the user 118, the image generator 104 may generate the image 114. Further details of the analysis are described below.
The image 114 may be transmitted to the printer 110. The printer 110 may display the image 114 via the display 112 for a preview. The user 118 may then accept the image 114 to be printed or reject the image 114 and request a new image 114 to be generated. If the image 114 is accepted, the image 114 may be printed by the printer 110 on a print media 120. The print media 120 may be paper.
If the image 114 is rejected, the image generator 104 may attempt to generate a new image 114. In another example, the user 118 may be prompted to speak a new phrase and the new image 114 may be generated by the image generator 104 based on the emotion identified in the newly spoken phrase.
In one example, the communication interface 204 may be a wired or wireless interface. For example, the communication interface 204 may be an Ethernet interface, a Wi-Fi radio, and the like. The image generator 104 may receive a phrase spoken by the user 118 via the communication interface 204. The phrase may be recorded and the data packets associated with the recording may be transmitted to the image generator 104 via the communication interface 204.
In one example, the sentiment analysis component 206 may identify an emotion of the user 118 based on the phrase. In one example, the emotion may be based on a score rating calculated from the phrase by the sentiment analysis component 206.
In one example, the sentiment analysis component 206 may calculate the score rating as a value of (−10 to +10). The score rating may be calculated based on a comparison of the words in the phrase spoken by the user 118 and a value assigned to words in a pre-defined list of possible emotion based words. In one example, the pre-defined list may be an AFINN sentiment lexicon list. In one example, the image generator 104 may include a computer readable memory that stores a table that includes the pre-defined list of words and the associated integer values or respective scores of each word.
The AFINN list provides a list of English words rated for valence with an integer. Positive words can be assigned positive integers and negative words can be assigned negative integers. The more positive the value of the score rating, the more positive the emotion of the user 118 (e.g., happy, excited, etc.). The more negative the value of the score rating, the more negative the emotion of the user 118 (e.g., angry, mad, unhappy, etc.). A score rating of 0 may be identified as a neutral emotion (e.g., content).
The score rating may be provided to the sentiment analysis component 206 to identify the emotion of the user 118. For example, if the score rating is determined to be +8, the emotion of the user 118 may be identified as positive or happy. The emotion may be also referred to as a mood value. The mood value may be used to select from three groups of predefined images. Each group of predefined images may include two images. Each group may be associated with a different emotion or mood value. For example, one set of images may be associated with a negative emotion or mood value, one set of images may be associated with a positive emotion or mood value, and one set of images may be associated with a neutral emotion or mood value.
The images selected based on the emotion may be placed on an X-Y coordinate plane associated with the print media 120. For example, the dimensions of the print media 120 may determine the size of the X-Y coordinate plane. The X-Y coordinate plane may be the largest area of 36-degree pie-slice areas that can fit onto the dimensions of the print media 120.
As noted above, the sentiment analysis component 206 may calculate the score rating. The sentiment analysis component 206 may also calculate a value for a comparative rating. The values for the score rating and the comparative rating may be converted into X-Y coordinates that determine a portion of the selected image that may be defined to generate the image 114.
In one example, the comparative rating may be calculated based on a sum of the integer values for each word in the phrase and a total number of words. For example, the phrase may include five words. The words in the phrase may be compared to the AFINN list and the sum of the values of the integers of each word may be determined to be +10. The comparative rating may be calculated to be 2 (e.g., 10/5=2).
In one example, the voice analysis component may calculate values of a frequency and a midi signature by converting the phrase into a musical note. The frequency and the midi signature may be converted into X-Y coordinates with the score rating and the comparative rating.
In one example, to convert the phrase into a musical note, a step (e.g., a musical step), an alteration (e.g., a musical alteration), and an octave may be calculated. The step may be calculated by a total length of the phrase (e.g., a total number of letters) then modulus of a maximum step value. The alteration may be calculated by converting each character or letter in the phrase into Unicode then modulus a maximum alteration value. The octave may be calculated by converting each character or letter into a hexadecimal then modulus a maximum octave value. In one example, the step, the alteration, and the octave of the phrase may be delivered to a note parser to return a frequency value and a midi signature value.
Thus, through the vocal analysis component 208 and the sentiment analysis component 206, five values can be calculated (e.g., the score rating, the mood/emotion, the comparative score, the frequency, and the midi signature). The score rating, the comparative rating, the frequency, and the midi signature can then be converted into X-Y coordinates to select a portion of the pre-defined image that was selected based on the mood/emotion of the user 118.
In one example, the score rating, the comparative rating, the frequency, and the midi signature may be divided by a maximum value of the score rating, the comparative rating, the frequency, and the midi signature and multiplied by a maximum size of the X-Y coordinate plane based on the print media 120. The first X-Y coordinate pair may be the resulting value based on the frequency and the score rating (e.g., the frequency may be the X coordinate and the score rating may be the Y coordinate). The second X-Y coordinate pair may be the resulting value based on the midi signature and the comparative rating (e.g., the midi signature may be the X coordinate and the comparative rating may be the Y coordinate).
The first X-Y coordinate pair may be used to set the location of the first image of the two images that are selected based on the mood value/emotion of the user 118. The first X-Y coordinate pair may be used to set the upper left corner of the first image at the first X-Y coordinate pair on the X-Y coordinate plane.
The second X-Y coordinate pair may be used to set the location of the second image of the two images that are selected based on the mood value/emotion of the user 118. The second X-Y coordinate pair may be used to set the upper left corner of the second image at the second X-Y coordinate pair on the X-Y coordinate plane. The first image and the second image may overlap one another on the X-Y coordinate plane.
In one example, the image generator 104 may also include a multiply filter. The processor 202 may control the multiply filter to combine the overlapping first image and second image into a single composite image. For example, the multiply filter may make the top layer (e.g., the second image) translucent or partially translucent to allow the lower layer (e.g., the first image) to show through.
In one example, the pre-defined area of the 36-degree pie-slice of the X-Y coordinate plane may be extracted with the portion of the composite image that is located within the area of the 36-degree pie-slice. The area of the 36-degree pie-slice with the portion of the composite image may be copied ten times in a circular fashion to form the Mandala image generated in the image 114. For example, each pie-slice may include a vertex and the vertex of each pie-slice containing the portion of the composite image may be connected to form the circular image.
The image 114 may then be transmitted back to the printer 110 via the communication interface 204. For example, the image 114 may be transmitted to the web-based application service 106 via the IP network 102. The web-based application service 106 may then transmit the image 114 to the printer 110. The printer 110 may display the image 114 on the display 112 to allow the user 118 to preview the image 114 as described above. The user 118 may then select a command of accept, cancel, redo, or no (e.g., such as by physically interacting with a user interface, or issuing a verbal command, by way of non-limiting example).
“Accept” may cause the image 114 to be printed on the print media 120 by the printer 110. “Cancel” may exit the application. “Redo” may request the image 114 to be regenerated. In one example, the image 114 may be regenerated by adding a random value to each X-Y coordinate value calculated above. For example, adding the random value may cause a different pie-slice of the composite image to be captured. The different pie-slice may be repeated in a circular fashion to generate a new image 114 based on the emotion of the user 118. The new image 114 may then be transmitted back to the printer 110. “No” may cause the entire process to be repeated beginning with speaking a new phrase.
Thus, the system 100 of the present disclosure with the image generator 104 may generate an emotional resonance image based on a phrase that is spoken by a user. The system 100 may analyze the phrase (e.g., the words chosen by a user for the spoken phrase) and detect an emotion. An image may be generated based on the emotion that is detected. The image may be printed by the printer 110 and provide a coloring image for the user 118 to help or enhance an emotion felt by the user (e.g., negative, neutral, or positive).
At block 302, the method 300 begins. At block 304, the method 300 receives a phrase spoken by a user. For example, a voice assistant application on an endpoint device of a user that is connected to a printer, or the printer itself, may execute the voice assistant application. The voice assistant application may work with a web-based voice assistant server. In one example, the user may speak a “wake” word to activate the voice assistant application followed by a command. The voice assistant application may then prompt the user to speak a phrase. The spoken phrase may be captured by a microphone and temporarily stored for transmission. For example, the spoken phrase may be temporarily stored as an audio file that can be transmitted.
At block 306, the method 300 identifies an emotion based on the phrase. In one example, the spoken phrase may be transmitted to an image generator that can analyze the phrase. A vocal analysis component in the image generator may identify a variety of different parameters such as a score rating, a comparative score, a frequency, and a midi signature. A sentiment analysis component may then identify an emotion based on the parameters. For example, the score rating may be used to determine a mood/emotion (e.g., negative, neutral, or positive).
At block 308, the method 300 converts the phrase into X-Y coordinates of a coordinate plane based on the phrase and the emotion that is identified. As described above, the phrase may be converted into a musical note by converting the phrase into a step, an alteration, and an octave. The musical note can be analyzed by a note parser to extract a frequency and a midi signature. The score rating, the comparative score, the frequency, and the midi signature may be converted into pairs of X-Y coordinates. The emotion that is detected can be used to select a set of pre-defined images associated with the emotion that is detected.
At block 310, the method 300 generates an image based on the emotion and the X-Y coordinates. As described above, a first set of X-Y coordinates may be used to place a first one of the pre-defined images associated with the emotion on an X-Y coordinate plane. The X-Y coordinate plane may be determined by a size or dimensions of a print media that is used. A second set of X-Y coordinates may be used to place a second one of the pre-defined images associated with the emotion on the X-Y coordinate plane.
The images may be overlaid on top of one another and blended. For example, a multiply filter may be used to make the second image that is on top of the first image translucent or partially translucent to allow the lower image to show through. In one example, an area of the layered images can be captured and repeated into a pattern that can form an image that is printed by the printer.
In one example, a pre-defined area of a pie-slice (e.g., a 36-degree pie-slice) may be applied over the composite image. The portion of the composite image that is located within the pre-defined area of the pie-slice may then be repeated ten times in a circular fashion to form the image. For example, each pie-slice may have a vertex and the vertex of each pie-slice having a copy of the portion of the composite image may be connected to form a circular image. The image that is formed may be a Mandala image.
At block 312, the method 300 transmits the image to a printer to be printed. For example, the Mandala image may be transmitted to the printer. The Mandala image may be shown in a display (e.g., an external display or a display associated with the printer). The user may then select one of the following options: accept, cancel, redo, or no. The actions associated with the options accept, cancel, redo, and no, are described above. At block 314, the method 300 ends.
In one example, the instructions 406 may include instructions to identify a mood based on a phrase spoken by a user. The instructions 408 may include instructions to convert the phrase into a first set of X-Y coordinates and a second set of X-Y coordinates. The instructions 410 may include instructions to select a set of images based on the mood. The instructions 412 may include instructions to place a first image of the set of images at the first set of X-Y coordinates on a coordinate plane. The instructions 414 may include instructions to place a second image of the set of images at the second set of X-Y coordinates on the coordinate plane. The instructions 416 may include instructions to generate a composite image of the first image and the second image. The instructions 418 may include instructions to capture a pie slice of the composite image. The instructions 420 may include instructions to generate an emotion based image formed by repeating the pie slice of the composite image in a circular fashion. The instructions 422 may include instructions to transmit the emotion based image to a printer to be printed.
In one example, the instructions 506 may include instructions to prompt a user to speak a phrase. The instructions 508 may include instructions to record the phrase. The instructions 510 may include instructions to transmit the phrase to an emotion based image generator. The instructions 512 may include instructions to receive an image generated by the emotion based image generator, wherein the image is generated by the emotion based image generator based on an emotion detected from the phrase and X-Y coordinates calculated from the phrase. The instructions 514 may include instructions to display the image to be printed.
It will be appreciated that variants of the above-disclosed and other features and functions, or alternatives thereof, may be combined into many other different systems or applications. Various presently unforeseen or unanticipated alternatives, modifications, variations, or improvements therein may be subsequently made by those skilled in the art which are also intended to be encompassed by the following claims.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2018/045336 | 8/6/2018 | WO | 00 |