The present disclosure relates generally to electronic teleconferencing systems and more specifically to measuring and responding to user attention in teleconference systems.
Teleconference systems may utilize communication networks, including but not limited to the internet, to connect communication systems and communication devices such as computers, tablet computers, and/or smartphones. Teleconference systems may permit communication systems to share visual imagery and audio data associated with a speaking user with other communication systems. However, teleconference systems may not be able to detect actual user participation in the teleconference, and may misinterpret a communication device connecting to the teleconference as a user paying attention to the teleconference. Furthermore, teleconference systems may fail to provide mechanisms to prompt user attention and encourage user engagement when user participation falters.
The following presents a simplified summary of the disclosure in order to provide a basic understanding to the reader. This summary is not an extensive overview of the disclosure and it does not identify key/critical elements of the invention or delineate the scope of the invention. Its sole purpose is to present some concepts disclosed herein in a simplified form as a prelude to more detailed descriptions presented below.
In embodiments of the disclosed subject matter, the unique systems and methods described herein make use of an exemplary system and method to measure and respond to attention levels in group teleconferences. Embodiments of the disclosed subject matter include two or more communication devices, including but not limited to tablet computers or smartphones, and a computer coupled with a database and comprising a processor and memory. The computer generates a teleconference space and transmits requests to join the teleconference space to the two or more communication devices. The computer stores in memory identification information for each of the two or more communication devices. Each of the two or more communication devices stores audiovisual data pertaining to one or more users associated with each of the two or more communication devices.
In embodiments of the disclosed subject matter, each communication device converts the audiovisual data into facial expressions data, generates emotions data from the facial expressions data, generates attention data from the emotions data, and reacts to the attention data, such as but not limited to generating one or more alert messages when attention data drops below a defined threshold.
These and other features of the disclosed subject matter are described in greater detail below.
A more complete understanding of the present invention may be derived by referring to the detailed description when considered in connection with the following illustrative figures. In the figures, like reference numbers refer to like elements or acts throughout the figures.
Aspects and applications of the invention presented herein are described below in the drawings and detailed description of the invention. Unless specifically noted, it is intended that the words and phrases in the specification and the claims be given their plain, ordinary, and accustomed meaning to those of ordinary skill in the applicable arts.
In the following description, and for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the various aspects of the invention. It will be understood, however, by those skilled in the relevant arts, that the present invention may be practiced without these specific details. In other instances, known structures and devices are shown or discussed more generally in order to avoid obscuring the invention. In many cases, a description of the operation is sufficient to enable one to implement the various forms of the invention, particularly when the operation is to be implemented in software. It should be noted that there are many different and alternative configurations, devices and technologies to which the disclosed inventions may be applied. The full scope of the inventions is not limited to the examples that are described below.
As described more fully below, embodiments of the following disclosure relate to measuring and responding to the attention levels of users participating in a group teleconference. Embodiments of the following disclosure generate a teleconference space including a plurality of communication systems and communication devices, each of which is operated by an individual user or group of users. Embodiments of the teleconference space include a visual component, which may include video imagery, and an audio component, which may comprise audio from a speaking user associated with one or more communication systems. Embodiments transmit the visual and audio components as a single outbound teleconference stream to the plurality of communication systems, each of which displays the outbound teleconference stream to one or more associated users. Each communication system measures and analyzes the attention level of one or more associated users viewing the outbound teleconference stream, and takes actions to improve the user's attention level when the user's attention begins to waver, and/or when the user leaves the vicinity of the associated communication system.
Embodiments of the following disclosure promote user engagement in group teleconferences by automatically prompting inattentive users to reengage and pay attention as the teleconference progresses using a variety of attention-promoting mechanisms.
In one embodiment, cloud system 110 comprises administrator 112 and database 114. Administrator 112 generates teleconference space 150 in which one or more communication systems 120 may participate. Database 114 comprises one or more databases 114 or other data storage arrangements at one or more locations local to, or remote from, cloud system 110. In one embodiment, one or more databases 114 is coupled with the one or more administrators 112 using one or more local area networks (LANs), metropolitan area networks (MANs), wide area networks (WANs), or network 130, such as, for example, the Internet, or any other appropriate wire line, wireless link, or any other communication links 140-144. One or more databases 114 stores data that is made available to and may be used by one or more administrators 112 according to the operation of teleconference system 100 described below. According to embodiments, administrator 112 hosts and runs one or more runtime processes associated with cloud system 110.
According to embodiments, one or more users may be associated with each of one or more communication systems 120. Each of the one or more users may comprise, for example, an individual person or customer, one or more employees or teams of employees within a business, or any other individual, person, group of persons, business, or enterprise which communicates or otherwise interacts with one or more separate communication systems 120. Although an exemplary number of communication systems 120 are shown and described, embodiments contemplate any number of communication systems 120 interacting with network 130 and one or more cloud systems 110 according to particular needs. By way of an example only and not by way of limitation, teleconference system 100 may allow up to 50, 100, 500, or 1,000 separate communication systems 120 to join and participate in teleconference space 150 simultaneously.
Each of one or more communication systems 120 comprises one or more communication devices 122, such as, for example, cellular phones or smartphones, desktop computers, laptop computers, notebook computers, tablet-type devices, terminals, or any other communication device 122 capable of receiving, transmitting, and displaying audiovisual information through network 130. In an embodiment, each of one or more communication devices 122 may comprise an audiovisual recording device, such as a computer camera and microphone, and an audiovisual display device, such as an electronic display screen and one or more speakers. The audiovisual display devices permit each of the one or more users interacting with each of one or more communication devices 122 to see and hear visual component 152 and audio component 154 of teleconference space 150. The audiovisual recording devices record audiovisual information regarding the one or more users associated with one or more communication devices 122. In addition, each of one or more communication devices 122 may comprise an input device, such as a keyboard, mouse, or touchscreen.
Each of one or more communication devices 122 that comprise each of one or more communication systems 120 may be coupled with other communication devices 122, as well as one or more cloud systems 110, by network 130 via communication link 142. Although communication links 142a-142n are shown connecting each of communication systems 120a-120n, respectively, to network 130, embodiments contemplate any number of communication links 140-144 connecting any number of communication systems 120 or communication devices 122 with network 130, according to particular needs. In addition, or as an alternative, communication links 140-144 may connect one or more communication systems 120 and/or communication devices 122 directly to one or more cloud systems 110 and/or one or more separate communication systems 120 and/or communication devices 122. According to embodiments, two or more communication devices 122 may be associated with each of one or more users.
According to embodiments, one or more communication links 140-144 couple one or more cloud systems 110, including each cloud system 110 administrator 112 and database 114, and one or more communication systems 120 with network 130. Each communication link 140-144 may comprise any wireline, wireless, or other link suitable to support data communications between one or more cloud systems 110 and one or more communication systems 120 and network 130 and/or teleconference space 150. Although communication links 140-144 are shown as generally coupling one or more cloud systems 110 and one or more communication systems 120 with network 130, one or more cloud systems 110 and one or more communication systems 120 may communicate directly with each other according to particular needs.
According to embodiments, network 130 includes the Internet, telephone lines, any appropriate LANs, MANs, or WANs, and any other communication network 130 coupling one or more cloud systems 110 and one or more communication systems 120. For example, data may be maintained by one or more cloud systems 110 at one or more locations external to one or more cloud systems 110, and made available to one or more cloud systems 110 or one or more communication systems 120 using network 130, or in any other appropriate manner.
According to embodiments, one or more cloud systems 110 and/or one or more communication systems 120 may each operate on one or more computers that are integral to or separate from the hardware and/or software that supports teleconference system 100. In addition, or as an alternative, the one or more users may be associated with teleconference system 100 including one or more cloud systems 110 and/or one or more communication systems 120. These one or more users may include, for example, one or more computers programmed to generate teleconference space 150 and measure and respond to the attention levels of users participating in teleconference space 150. As used herein, the computer, the term “computer,” and “computer system” comprise an input device and an output device. The computer input device includes any suitable input device, such as a keypad, mouse, touch screen, microphone, or other device to input information. The computer output device comprises any suitable output device that may convey information associated with the operation of teleconference system 100, including digital or analog data, visual information, or audio information. Furthermore, the one or more computers include any suitable fixed or removable non-transitory computer-readable storage media, such as magnetic computer disks, CD-ROMs, or other suitable media to receive output from and provide input to teleconference system 100. The one or more computers also include one or more processors and associated memory to execute instructions and manipulate information according to the operation of teleconference system 100.
Embodiments contemplate one or more cloud systems 110 generating teleconference space 150. Each of one or more communication devices 122 may connect to one or more cloud systems 110 using network 130 and communication links 140-144, and may participate in teleconference space 150. Teleconference space 150 allows one or more communication devices 122 to conduct and participate in an audiovisual teleconference. According to embodiments, teleconference space 150 may comprise visual component 152 and/or audio component 154. Although teleconference space 150 is shown and described as comprising single visual component 152 and audio component 154, embodiments contemplate teleconference space 150 comprising any number of components or related information, according to particular needs. Visual component 152 may comprise video imagery of one or more users associated with one or more communication devices 122. Audio component 154 may comprise audio from one or more currently-speaking users associated with one or more communication devices 122.
According to embodiments, cloud system 110 administrator 112 generates an outbound teleconference stream, comprising visual component 152 and/or audio component 154 of teleconference space 150, and transmits the outbound teleconference stream to each of one or more communication devices 122 participating in teleconference space 150. Each communication device 122 uses an associated audiovisual display device to display the outbound teleconference stream. Each communication device 122 uses an audiovisual recording device (such as, for example, a camera associated with communication device 122) to record the facial expression of one or more users associated with each communication device 122. Each communication device 122 analyzes the facial expression, assess the emotional content of the facial expression, and assign a qualitative attention value that measures one or more qualities of the facial expression in real time.
Each communication devices 122 continuously monitors the qualitative attention value assigned to each of the one or more users associated with communication device 122. When communication device 122 determines that the qualitative attention value of a particular user has decreased below a specified value, communication device 122 takes one or more alert actions, such as but not limited to generating an alert message and displaying the alert message on communication device 122 audiovisual display device, to increase the attention the user pays to the outbound teleconference stream.
Each communication device 122 also continuously monitors whether the one or more users associated with communication device 122 are facing communication device 122. If communication device 122 detects that a particular user has left the vicinity of or has turned away from communication device 122 for a defined period of time, communication device 122 may transmit an absence notification to cloud system 110 that the user has disengaged from communication device 122. Cloud system 110 may transmit a notification message to other communication devices 122 associated with the disengaged user, as described in greater detail below, to prompt the user's attention and to encourage the user to reengage with his or her communication device 122 and the outbound teleconference stream displayed thereon.
According to embodiments, administrator 112 comprises administration module 202, graphical user interface module 204, and notification module 206. Although a particular configuration of administrator 112 is illustrated and described, embodiments contemplate any suitable number or combination of administration modules 202, graphical user interface modules 204, notification modules 206, and/or other modules located at one or more locations, local to, or remote from, cloud system 110, according to particular needs. In addition, or as an alternative, administration module 202, graphical user interface module 204, and notification module 206 may be located on multiple administrators 112 or computers at any location in teleconference system 100.
Database 114 may comprise communication systems data 210, teleconference stream data 212, and notification data 214. Although database 114 is illustrated and described as comprising communication systems data 210, teleconference stream data 212, and notification data 214, embodiments contemplate any suitable number or combination of communication systems data 210, teleconference stream data 212, notification data 214, and/or other data pertaining to teleconference system 100 located at one or more locations, local to, or remote from, cloud system 110, according to particular needs.
Administration module 202 of administrator 112 may configure, update, and/or manage the operation of cloud system 110. That is, administration module 202 may configure, update, and/or manage the broader operation of teleconference system 100 and change which data is executed and/or stored on one or more cloud systems 110 and/or one or more communication devices 122. Teleconference system 100 may comprise a user-configurable system, such that cloud system 110 administrator 112 may store communication systems data 210, teleconference stream data 212, and/or notification data 214 either singularly or redundantly in cloud system 110 database 114 and/or one or more communication devices 122, according to particular needs. According to other embodiments, administration module 202 monitors, processes, updates, creates, and stores communication systems data 210, teleconference stream data 212, and/or notification data 214 in cloud system 110 database 114, as discussed in greater detail below.
According to embodiments, administration module 202 of administrator 112 may generate teleconference space 150, which one or more communication devices 122 may join. When communication device 122 joins teleconference space 150, administration module 202 may record unique identifying information regarding communication device 122, such as by assigning each communication device 122 a unique ID or by recording the IP or MAC address of each communication device, in communication systems data 210 of database 114, as is further described below.
Graphical user interface module 204 of administrator 112 generates the outbound teleconference stream, which administration module 202 transmits to one or more communication devices 122 using network 130 and one or more communication links 140-144. More specifically, graphical user interface module 204 accesses teleconference stream data 212 stored in database 114, and uses teleconference stream data 212 to generate an outbound teleconference stream, which administration module 202 transmits to one or more communication devices 122 participating in teleconference space 150. Graphical user interface module 204 stores and retrieves data from cloud system 110 database 114 including communication systems data 210 and outbound teleconference stream data 212, in the process of generating the outbound teleconference stream. Graphical user interface module 204 may generate different graphical user interface displays conveying different types of information for different communication devices 122, as discussed in greater detail below.
According to embodiments, notification module 206 of administrator 112 generates one or more communication device 122 notifications. As described in greater detail below, each communication device 122 participating in teleconference space 150 may continuously monitor whether the one or more users associated with communication device 122 are facing communication device 122. Upon detecting that one or more associated users have left the vicinity of communication device 122 and/or have turned away from facing communication device 122 for a defined period of time, communication device 122 may transmit an absence notification to notification module 206 of administrator 112, using network 130 and communication links 140-144, indicating the one or more users' disengagement. In response, notification module 206 accesses notification data 214 stored in cloud system 110 database 114, and generates a notification message to be sent to the one or more separate communication devices 122 associated with each disengaged user. Notification module 206 transmits the notification message to administration module 202. Administration module 202 transmits the notification message to one or more separate communication devices 122 associated with each disengaged user to prompt the user's attention and to encourage the user to reengage and pay attention to the outbound teleconference stream.
By way of example only and not by way of limitation, in an embodiment, cloud system 110 may register and associate two separate communication devices 122 (in this example, a computer and a smartphone) with a particular user and user account. Administration module 202 of cloud system 110 administrator 112 may store information regarding the user's account, and the two communication devices 122 associated with the user, in communication systems data 210 of database 114, as discussed in greater detail below. Continuing the example, the user connects to and participates in an audiovisual teleconference using the computer. At a later point in the ongoing teleconference, the computer determines that the user has stepped away from the computer and is no longer engaged with the teleconference. The computer transmits an absence notification to notification module 206 of cloud system 110. Notification module 206 generates a notification message in the form of the text message “Are you still participating in the teleconference?”, which in this example administration module 202 transmits to the user's smartphone. The user, who had disengaged from the computer, sees the notification message on her smartphone, and reengages with the computer to continue participating in the teleconference. Although this exemplary embodiment comprises particular users, communication devices 122, and notification messages, embodiments contemplate teleconference system 100 comprising any configuration or type of users, communication devices 122, and/or notification messages, as described in greater detail below.
Communication systems data 210 of database 114 comprises the identification information of one or more communication devices 122, such as, for example, names and addresses of the one or more users associated with each of one or more communication devices 122, company contact information, telephone numbers, email addresses, IP addresses, and the like. According to embodiments, identification information may also comprise information regarding the operating systems of each of one or more communication systems 120, internet browser information regarding each of one or more communication devices 122 associated with each of one or more communication systems 120, or system specifications (such as, for example, processor speed, available memory, hard drive space, and the like) for each of one or more communication devices 122 associated with each of one or more communication systems 120.
Communication systems data 210 may also include end user ID information, end user account information (comprising one or more communication devices 122 associated with each user), end user personal identification number (PIN) information, communication device 122 ID information, communication device 122 MAC address information, or any other type of information which cloud system 110 may use to identify and track each of one or more communication systems 120 participating in teleconference system 100. Communication systems data 210 may further comprise identification data that identifies and tracks each of one or more communication devices 122 which comprise each of one or more communication systems 120. Although particular communication systems data 210 are described, embodiments contemplate any type of communication systems data 210 associated with one or more communication devices 122 or communication devices 122, according to particular needs. In one embodiment, cloud system 110 uses communication systems data 210 to identify one or more participating communication devices 122 in teleconference system 100 in order to aid the selection of one or more communication device 122 streams to comprise the outbound teleconference stream, such as by prioritizing communication device 122 streams of predetermined very important person (VIP) communication devices 122. In another embodiment, cloud system 110 uses communication systems data 210 to generate teleconference space 150 which specifically includes only particular identified communication devices 122, such as in the case of a private teleconference space 150.
Teleconference stream data 212 of database 114 comprises data related to the outbound teleconference stream, which cloud system 110 transmits to one or more communication devices 122. As described in greater detail below, one or more communication devices 122 may transmit audiovisual information regarding one or more speaking users to administration module 202, which may store this information in teleconference stream data 212. Graphical user interface module 204 may access teleconference stream data 212 and use it to generate an outbound teleconference stream, comprising visual component 152 and audio component 154, which administration module 202 transmits to one or more communication devices 122 participating in teleconference space 150.
Notification data 214 of database 114 may comprise one or more notification messages. As described above, notification module 206 may access the one or more notification messages stored in notification data 214, and may transmit one or more notification messages to administration module 202. Notification data 214 may comprise any form of notification messages, including SMS and/or text messages (such as, for example, a “Please respond to the teleconference in progress” text message), auditory notification messages (such as, for example, an alert chime that may be played by communication device 122 audiovisual display device), visual notification messages (such as, for example, a red-colored notification message that is displayed on communication device 122's audiovisual display device), email notification messages sent to one or more email accounts associated with one or more users, haptic notification messages, or any other notification message. In an embodiment, before generating teleconference space 150 and conducting an audiovisual teleconference, one or more communication devices 122 may transmit one or more sample notification messages, which are to be used with teleconference space 150 in the event one or more users disengage from their associated communication devices 122 for a defined period of time, to administration module 202 of cloud system 110. In this exemplary embodiment, administration module 202 stores the transmitted sample notification messages in notification data 214 of cloud system 110 database 114.
According to embodiments, one or more communication devices 122 may transmit a request to administration module 202 of cloud system 110, using network 130 and communication links 140-144, for administration module 202 to generate teleconference space 150. In response, administration module 202 may generate teleconference space 150, and transmit requests to join teleconference space 150 to one or more other communication devices 122 using network 130 and the communication links 140-144. A plurality of communication devices 122 may accept the requests and join and participate in teleconference space 150. Embodiments contemplate any number of communication devices 122 joining and participating in teleconference space 150.
In an embodiment, a user associated with one of one or more communication devices 122 uses teleconference space 150 to deliver a teleconference presentation to the users associated with the one or more other communication devices 122 (the presenting user is henceforth referred to as the “host,” and the particular communication device 122 associated with the host as the “host device”). The audiovisual recording device of the host device records audiovisual information regarding the host speaking. The host device transmits the audiovisual information to administration module 202 using network 130 and communication links 140-144. Administration module 202 stores this audiovisual information in teleconference stream data 212.
Continuing the above example, graphical user interface module 204 accesses teleconference stream data 212, comprising visual component 152 and audio component 154 of the host's audiovisual information. Graphical user interface module 204 generates an outbound teleconference stream, comprising visual component 152 displaying the host and audio component 154 comprising the host's spoken audio, which administration module 202 transmits to the other one or more communication devices 122 participating in teleconference space 150. Each of one or more communication devices 122 displays the audiovisual content of the outbound teleconference stream using one or more associated audiovisual display devices.
Communication device 122 processor 302 may comprise audiovisual recording module 310, facial analysis module 312, emotions analysis module 314, and alert module 316. Although processor 302 is described as comprising a single audiovisual recording module 310, facial analysis module 312, emotions analysis module 314, and alert module 316, embodiments contemplate any suitable number of audiovisual recording modules 310, facial analysis modules 312, emotions analysis modules 314, alert modules 316, or other modules, internal to or externally coupled with communication device 122. Processor 302 may execute an operating system program stored in memory 304 to control the overall operation of communication device 122. For example, processor 302 may control the reception of signals and the transmission of signals within teleconference system 100. Processor 302 may execute other processes and programs resident in memory 304, such as, for example, registration, identification or communication over network 130 and communication links 140-144.
Communication device 122 memory 304 may comprise audiovisual data 320, facial expressions data 322, emotions data 324, attention data 326, and alert data 328. Although memory 304 is described as comprising audiovisual data 320, facial expressions data 322, emotions data 324, attention data 326, and alert data 328, embodiments contemplate any suitable number of audiovisual data 320, facial expressions data 322, emotions data 324, attention data 326, alert data 328, or other data, internal to or externally coupled with communication device 122, according to particular needs.
In an embodiment, audiovisual recording module 310 may be operatively associated with, and may monitor and facilitate the operation of, communication device 122 audiovisual recording device. By way of example only and not by way of limitation, audiovisual recording module 310 may activate the audiovisual recording device of a host user's communication device 122, and may record audiovisual information regarding the host user speaking to the one or more other communication devices 122 participating in teleconference space 150. In an embodiment, audiovisual recording module 310 may transmit the host user audiovisual information to cloud system 110 administration module 202, using network 130 and one or more communication links 140-144.
Audiovisual recording module 310 may also store audiovisual information pertaining to one or more users in audiovisual data 320 of communication device 122 memory 304. According to embodiments, audiovisual data 320 may comprise visual information, such as a video file or real-time visual stream, or one or more individual image snapshots, of one or more users associated with communication device 122. Audiovisual data 320 may store time entry information with the video file, real-time visual stream, or one or more individual image snapshots, enabling communication device 122 processor 302 to determine when audiovisual recording module 310 captured and stored the associated visual information in audiovisual data 320. Audiovisual data 320 may also comprise audio information, such as recorded audio of one or more speaking users. Although particular audiovisual data 320 are described herein, embodiments contemplate audiovisual recording module 310 storing any form of audiovisual data 320, including but not limited to data that is exclusively visual in nature or data that is exclusively audio in nature, in audiovisual data 320.
Facial analysis module 312 of communication device 122 processor 302 may analyze audiovisual data 320 to determine the facial expression of one or more users associated with communication device 122. Facial analysis module 312 may access audiovisual data 320, determine whether one or multiple users are currently associated with communication device 122, and may store information related to each of the one or more user facial expressions in facial expressions data 322. In an embodiment, facial analysis module 312 may use facial recognition techniques to separately identify each of the one or more users currently associated with communication device 122, and may separately store information related each user's facial expression in facial expressions data 322.
According to embodiments and as discussed in greater detail below, facial analysis module 312 may determine the status of each user's facial expression by, for example: (1) assigning one or more data points to the facial structure of individual snapshots or a real-time visual stream of a user stored in audiovisual data 320, and (2) interpreting these assigned data points in accordance with one or more facial expression templates which may be stored in facial expressions data 322. Although particular procedures by which facial analysis module 312 may analyze user facial expression information and store such information in facial expressions data 322 are shown and described, embodiments contemplate facial analysis module 312 utilizing any analysis technique to review information stored in audiovisual data 320 and to convert this information into facial expressions information stored in facial expressions data 322, according to particular needs.
Facial expressions data 322 of communication device 122 memory 304 stores information regarding the current facial expression of each of the one or more users associated with communication device 122, according to an embodiment. Facial expressions data 322 may further comprise one or more facial expression templates, which facial analysis module 312 may use to interpret data points which facial analysis module 312 has assigned to the facial structure of each user. In an embodiment, one or more cloud systems 110 and/or one or more communication devices 122 may transmit one or more facial expression templates to facial expressions data 322. In an embodiment, cloud system 110 may transmit, to facial expressions data 322, facial expression templates comprising exemplary emotional templates for the following emotions: attentiveness, anger, disgust, fear, sadness, surprise, and happiness.
Facial analysis module 312 may analyze each of the one or more users' facial expressions stored in facial expressions data 322, utilizing one or more facial expression templates stored in facial expressions data 322, to interpret the presence of one or more emotions associated with each user's facial expressions. For example, facial analysis module 312 may analyze a particular user's facial and/or micro expressions for the presence of specific assigned data points which suggest the user is happy (such as, for example, by determining that a cluster of assigned data points around the user's mouth suggests the user is smiling), sad, surprised, neutral, angry, or unfocused. Although particular emotions are described herein, embodiments contemplate facial analysis module 312 analyzing a user's facial expression to detect the presence of one or more of any possible emotions, according to embodiments. Having assessed the presence of one or more emotions in the user's facial and/or micro expression, facial analysis module 312 stores this emotion information in emotions data 324 of communication device 122 memory 304.
In an embodiment, facial analysis module 312 may analyze audiovisual data 320 stored in communication device 122 memory 304, including the time entry information associated with audiovisual data 320, and determine that audiovisual data 320 does not comprise one or more facial expressions. This may indicate that the one or more users associated with communication device 122 have left the vicinity of communication device 122 and/or have turned away from facing communication device 122. Facial analysis module 312 may store information regarding the absence of one or more facial expressions detectable in audiovisual data 320 (hereinafter referred to as an “absence notification”), and the duration of time for which facial analysis module 312 could not detect one or more facial expressions in audiovisual data 320, in attention data 326 of communication device 122 memory 304.
According to embodiments, emotions data 324 of communication device 122 memory 304 stores information regarding one or more emotions associated with each of the one or more users' facial expression. In an embodiment, emotions data 324 may store separate variables for one or more of any possible emotions, assigned by facial analysis module 312. By way of an example only and not by way of limitation, facial analysis module 312 may analyze a particular user facial expression stored in facial expressions data 322 and assign separate emotion scores representing a plurality of separate emotions (in this example: happy 78%; sad 21%; surprised 44%; neutral 0%; angry, 5%; unfocused 10%). Facial analysis module 312 may store each of these separate emotion scores in emotions data 324. Although particular emotions and emotion scores are shown and described, embodiments contemplate emotions data 324 storing score information regarding any number of separate defined emotions, according to particular needs.
Emotions analysis module 314 of communication device 122 processor 302 may access data regarding emotions and emotion scores stored in emotions data 324, and may use data regarding emotions and emotion scores to assign a qualitative attention value indicating whether each particular user of the one or more users associated with communication device 122 is attentive to and following teleconference space 150. Emotions analysis module 314 may utilize one or more attention criteria, stored in attention data 326, to assign a qualitative attention value. For example, an exemplary attention criteria might specify that if a user's assessed happiness emotion is greater than 50%, and the user's assessed unfocused emotion is also less than 30%, that user is engaged and is attentively participating in teleconference space 150. Although specific attention criteria are described, embodiments contemplate the emotional analysis module utilizing any attention criteria to analyze the emotions and emotion scores stored in emotions data 324 in order to assign a qualitative attention value. In an embodiment, one or more cloud systems 110 or one or more other communication devices 122 may transmit information to emotions analysis module 314, using network 130 and communication links 140-144, directing which attention criteria emotions analysis module 314 should use to assign a qualitative attention value. Having assigned a qualitative attention value, emotions analysis module 314 stores the qualitative attention value in attention data 326 of communication device 122 memory 304.
According to embodiments, attention data 326 may store an assigned qualitative attention value pertaining to the attentiveness of each of one or more users. Attention data 326 may also store one or more attention criteria, which may be transmitted to communication device 122 by one or more cloud systems 110 and/or one or more other communication devices 122, and which emotions analysis module 314 may use to generate a qualitative attention value for each user based on emotions data 324 and the emotion scores stored in emotions data 324 of communication device 122 memory 304. Although particular examples of attention data 326 are described herein, embodiments contemplate attention data 326 comprising any number or type of attention criteria or qualitative attention values, according to particular needs. In an embodiment, emotions analysis module 314 may store a separate binary qualitative attention value (such as, for example, “attentive” or “inattentive”) for each of the one or more users associated with communication device 122 in attention data 326. Emotions analysis module 314 may also store time entry information associated with the qualitative attention value (such as, for example, the length of time for which emotions analysis module 314 assigns an “inattentive” qualitative attention value to a particular user, measured in seconds, minutes, or any other unit of time) in attention data 326.
According to embodiments, alert module 316 of communication device 122 processor 302 generates one or more communication device 122 alerts. Alert module 316 may access the qualitative attention values, stored in attention data 326, of each of the one or more users associated with communication device 122. If alert module 316 determines that the qualitative attention value associated with one or more users has been “inattentive” for a defined period of time (such as, for example, thirty seconds, one minute, three minutes, or any other defined period of time), alert module 316 generates one or more alerts to prompt the user's attention and to encourage the user to pay attention to the outbound teleconference stream, as described in greater detail below.
To generate an alert, alert module 316 accesses alert data 328 of memory 304. Alert data 328 may comprise any form of one or more alert messages, including SMS and/or text messages (such as, for example, a “Please respond to the teleconference in progress” text message), auditory alert messages (such as, for example, an alert chime that may be played by communication device 122 audiovisual display device), visual alert messages (such as, for example, a red-colored notification message that is displayed on communication device 122's audiovisual display device), email notification messages sent to one or more email accounts associated with one or more users, haptic notification messages, or any other notification message. In an embodiment, before generating teleconference space 150 and conducting an audiovisual teleconference, one or more communication devices 122 may select and/or transmit to cloud system 110 and/or other communication devices 122 one or more sample alert messages, which are to be used with teleconference space 150 in the event alert module 316 of communication device 122 detects an “inattentive” qualitative attention value.
Having generated an alert, alert module 316 displays the alert on communication device 122 audiovisual display device. In an embodiment, alert module 316 may continuously monitor the qualitative attention values associated with each associated user of communication device 122, and may display an alert using communication device 122 audiovisual display device until alert module 316 determines that all users' qualitative attention values meet or exceed a predetermined value. In another embodiment, alert module 316 may display an alert on communication device 122 audiovisual display device at any point at which alert module 316 determines that any users associated with communication device 122 have “inattentive” qualitative attention values.
In an embodiment, alert module 316 may access attention data 326 and determine that facial analysis module 312 has associated an absence notification with one or more users associated with communication device 122. Facial analysis module 312 may store an absence notification in attention data 326 when facial analysis module 312 determines that audiovisual data 320 does not comprise one or more current facial expressions, indicating that one or more users associated with communication device 122 have left the vicinity of communication device 122 and/or have turned away from facing communication device 122. Alert module 316 may transmit the absence notification to notification module 206 of administrator 112, using network 130 and communication links 140-144. Notification module 206 may generate and transmit a notification message to one or more other communication devices 122 associated with each absent or disengaged user.
At action 402 of method 400, teleconference system 100 generates teleconference space 150. Communication device 122 transmits a request to administration module 202, using network 130 and communication links 140-144, to generate teleconference space 150. Administration module 202 generates teleconference space 150 and transmits, using network 130, requests to join teleconference space 150 to one or more separate communication devices 122 that will participate in teleconference space. Each of one or more separate communication devices 122 accepts the request to join teleconference space 150 and transmits acceptance to administration module 202. Administration module 202 records unique identifying information regarding each of the one or more communication devices 122, such as by assigning each communication device 122 a unique ID and/or by recording IP or MAC address of each communication device 122 in communication systems data 210. In an embodiment, communication device 122 that transmitted the initial request to generate teleconference space 150 to administration module 202 uses teleconference space 150 to deliver a teleconference presentation to one or more separate communication devices 122 that joined teleconference space 150. Communication device 122 that transmitted the initial request to generate teleconference space 150 to administration module 202 is henceforth referred to as the “host device,” and the user associated with the host device is referred to as the “host.”
Continuing action 402, the audiovisual recording device of the host device records audiovisual information regarding the host speaking in the form of visual component 152 and audio component 154. The host device transmits visual component 152 and audio component 154 to administration module 202 using network 130 and communication links 140-144. Administration module 202 stores visual component 152 and audio component 154 in teleconference stream data 212. Graphical user interface module 204 accesses teleconference stream data 212, which comprises visual component 152 and audio component 154 of the host's audiovisual information. Graphical user interface module 204 uses visual component 152 and audio component 154 to generate an outbound teleconference stream, comprising visual component 152 displaying the host and audio component 154 comprising the host's spoken audio. Administration module 202 transmits the outbound teleconference stream to one or more communication device 122 participating in teleconference space 150. Each of one or more communication devices 122 displays the audiovisual content of the outbound teleconference stream as teleconference display 602, illustrated by
At action 404, teleconference system 100 chooses relevant user facial expressions. In an embodiment, the host device selects one or more relevant user facial expressions by which to measure user attention. Embodiments contemplate host devices selecting any number of user facial expressions or emotions to measure user attention, according to particular needs. The host device transmits the host's selection of one or more relevant user facial expressions by which to measure user attention to administration module 202. Administration module 202 transmits the host device's selection of one or more relevant user facial expressions by which to measure user attention to each of one or more communication devices 122 participating in teleconference space 150. Each communication device 122 stores the selection of one or more relevant user facial expressions by which to measure user attention in communication device 122 facial expressions data 322.
At action 406, each communication device 122 participating in teleconference space 150 converts audiovisual data 320 pertaining to one or more users associated with each communication device 122 into facial expressions data 322. Audiovisual recording module 310 of each communication device 122 activates the associated audiovisual recording device of each communication device 122 and captures at least visual information, such as but not limited to a real-time visual stream and/or individual visual snapshots, of a user associated with communication device 122. Audiovisual recording module 310 stores the visual information in audiovisual data 320. Communication device 122 facial analysis module 312 accesses audiovisual data 320 and uses audiovisual data 320 to generate facial expression data 322 pertaining to one or more facial expressions of one or more associated users. In an embodiment, to generate facial expression data 322, facial analysis module 312 (1) assigns data points 702, illustrated by
At action 408, teleconference system 100 generates emotions data 324. In an embodiment, facial analysis module 312 accesses facial expressions data 322 and interprets the presence of one or more emotions associated with the one or more user facial expressions stored in facial expressions data 322. Facial analysis module 312 may compare facial expressions with one or more facial expression templates, stored in facial expressions data 322, to interpolate emotions associated with one or more facial expressions and to store the one or more emotions in emotions data 324. Other embodiments contemplate facial analysis module 312 utilizing any method to analyze facial expressions data 322 and to assign emotions data 324 based on facial expressions data 322, according to particular needs.
At action 410, teleconference system 100 generates attention data 326 from emotions data 324. In an embodiment, emotions analysis module 314 accesses emotions data 324 and assigns attention data 326, in the form of a qualitative attention value, to the emotion scores stored in emotions data 324. According to embodiments, emotions analysis module 314 may use any process, including but not limited to combining one or more emotion scores assigned to emotions data 342 into a single Boolean value (such as, for example, “attentive” or “inattentive”), to generate a qualitative attention value. Emotions analysis module 314 stores the qualitative attention value in attention data 326.
At action 412, one or more communication devices 122 respond to attention data 326. In an embodiment, alert module 316 of each communication device 122 participating in teleconference space 150 accesses qualitative attention values stored in attention data 326. According to embodiments, if qualitative attention values stored in attention data 326 indicate one or more users associated with communication device 122 is not paying attention to teleconference space 150, and/or has stepped away from communication device 122, alert module 316 may respond by generating an alert. Alert module 316 accesses alert data 328, generates an alert, and displays the alert on communication device 122 audiovisual display device, as illustrated by
In order to illustrate the operation of method 400, an example is now given. In the following example, exemplary teleconference system 100 comprises cloud system 110, five communication devices 122 (comprising, in this example, computers 502-510), network 130, and six communication links 140-142e. Although a particular number of cloud systems 110, communication devices 122, networks 130, and communication links 140-142e are shown and described, embodiments contemplate any number of cloud systems 110, communication devices 122, networks 130, or communication links 140-144, according to particular needs.
At action 402 of method 400, host computer 502 transmits a request to administration module 202, using network 130 and communication links 140-142a, to generate teleconference space 150. Administration module 202 generates teleconference space 150 and transmits, using network 130, requests to join teleconference space 150 to each of computers 502-510. Each of computers 502-510 transmits the computer's acceptance of the request to join teleconference space 150 to administration module 202. As discussed above, administration module 202 records unique identifying information regarding each of computers 502-510, such as by assigning each computer a unique ID and by recording the computer's IP or MAC address, in communication systems data 210.
Continuing the example, the audiovisual recording device of host computer 502 records audiovisual information regarding the host speaking. Host computer 502 transmits this audiovisual information to administration module 202 using network 130 and communication links 140-142a. Administration module 202 stores the audiovisual information in teleconference stream data 212. Graphical user interface module 204 accesses teleconference stream data 212, which comprises visual component 152 and audio component 154 of the audiovisual information transmitted by host computer 502. Graphical user interface module 204 generates an outbound teleconference stream, comprising visual component 152 displaying the host and audio component 154 comprising the host's spoken audio, which administration module 202 transmits to computers 502-510 participating in teleconference space 150. Each of computers 502-510 displays the audiovisual content of the outbound teleconference stream as teleconference display 602 using an associated audiovisual display device.
According to embodiments, participant panel 606 on the right side of teleconference display 602 displays a visual representation of communication devices 122 currently participating in teleconference space 150. Participant panel 606 may identify participating communication devices 122 (in this example, computers 502-510) by the names of the users associated with communication devices 122, or by identifying communication devices 122 themselves (such as “Mini Android,” “Acer One,” and the like). In an embodiment, administration module 202 may assign names to communication devices 122 displayed in participant panel 606 using information contained in communication systems data 210. Continuing the example, participant panel 606 of exemplary teleconference stream 602 lists computers 502-510. Although a specific configuration of participant panel 606 is shown and described, embodiments contemplate teleconference displays 602 displaying participant panels in any configuration, according to particular needs.
Continuing the example, at action 404, host computer 502 selects a combination of “happy,” “angry,” “sad,” “surprised,” “neutral,” and “inattentive” as the relevant user facial expressions by which to measure user attention. Although in this example host computer 502 selects six particular user facial expression by which to measure user attention, embodiments contemplate hosts selecting any other user facial expressions, emotions, or any number of user facial expressions or emotions to measure, according to various needs. Host computer 502 transmits the host's selection of “happy,” “angry,” “sad,” “surprised,” “neutral,” and “inattentive” as the relevant user facial expressions to administration module 202, which transmits this selection to each of computers 502-510 participating in teleconference space 150. Each computer 502-510 stores the selection of “happy,” “angry,” “sad,” “surprised,” “neutral,” and “inattentive” as the relevant user facial expressions in the facial expression data of memory 304.
Continuing the example, at action 406, each of computers 504-510 (excluding in this example host computer 502) converts audiovisual data 320 pertaining to a user associated with each computer 504-510 into facial expressions data 322. To accomplish action 406, audiovisual recording module 310 of each computer 504-510 uses the audiovisual recording device associated with each computer 504-510 to capture visual information, in the form of a real-time visual stream, of a user associated with each computer 504-510. For each computer 504-510, audiovisual recording module 310 stores the real-time visual stream in audiovisual data 320 of memory 304. Facial analysis module 312 analyzes the real-time visual stream, stored in audiovisual data 320, to generate facial expressions data 322. In this example, facial analysis module 312 analyzes the real-time visual stream by assigning seventy-one data points 702 to the facial structure of the user recorded in the real-time visual stream, illustrated by
Continuing the example, at action 408, facial analysis module 312 generates emotions data 324 from facial expressions data 322. Facial analysis module 312 accesses facial expressions data 322 and interprets the presence of one or more emotions associated with the facial expression stored in facial expressions data 322. In this example, facial analysis module 312 compares facial expressions stored in facial expressions data 322 to facial expression templates, also stored as data in facial expressions data 322, to generate emotions data 324. Other embodiments contemplate facial analysis module 312 utilizing any method to analyze facial expressions data 322 and to assign emotions data 324 based on facial expressions data 322, according to particular needs.
Continuing the example, facial analysis module 312 analyzes facial structure data points stored in facial expressions data 322 and compares the data points to facial expression templates, also stored in facial expressions data 322, to interpret the presence of one or more emotions. As illustrated in
Continuing the example, at action 410, emotions analysis module 314 accesses emotions data 324 and assigns attention data 326, in the form of a qualitative attention value, to the emotion scores stored in emotions data 324.
Continuing the example, emotions analysis module 314 accesses the emotion scores stored in emotions data 324 emotions data box 802, and compares the emotion scores to the relevant user facial expressions selected at action 404. In this example, emotions analysis module 314 of computer 504, executing action 410, weights the average values of the six selected emotions, and determines that the user associated with computer 504 is currently inattentive. In this example, emotions analysis module 314 of computer 504 stores a qualitative attention value of “inattentive” in attention data 326 of computer 504 memory 304. In alternative embodiments, emotions analysis module 314 may use any analysis procedure to average one or more emotion scores into one or more qualitative attention values.
Continuing the example, at action 412, alert module 316 of computer 504 accesses attention data 326 and the “inattentive” qualitative attention value stored therein. In response, alert module 316 of computer 504 accesses alert data 328 of memory 304 and generates alert message 1002 to prompt the user associated with computer 504 to engage in teleconference space 150. Alert module 316 displays alert message 1002 on computer 504 audiovisual display device, as illustrated in
Reference in the foregoing specification to “one embodiment”, “an embodiment”, or “some embodiments” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
While the exemplary embodiments have been shown and described, it will be understood that various changes and modifications to the foregoing embodiments may become apparent to those skilled in the art without departing from the spirit and scope of the present invention.
The present disclosure is related to that disclosed in the U.S. Provisional Application No. 62/876,412, filed Jul. 19, 2019, entitled “Measuring and Responding to Attention Levels in Group Teleconferences.” U.S. Provisional Application No. 62/876,412 is assigned to the assignee of the present application. The subject matter disclosed in U.S. Provisional Application No. 62/876,412 is hereby incorporated by reference into the present disclosure as if fully set forth herein. The present invention hereby claims priority under 35 U.S.C. § 119(e) to U.S. Provisional Application No. 62/876,412.
Number | Date | Country | |
---|---|---|---|
62876412 | Jul 2019 | US |