The present application relates generally to systems and methods for providing a video analysis and specifically to utilizing emotional state data of a subject in a video and providing an interface to convey that data.
Companies often perform market research to determine a consumer reaction to various everyday problems, as well as the effectiveness of a stimulus in solving those problems. While this market research has historically taken the form of consumer surveys, live test groups, etc., oftentimes these traditional market research tools are inefficient or do not provide enough data to accurately capture the desired data.
Included are embodiments of a system with a memory component that stores logic that, when executed by the system, causes the system to receive video content of a subject, identify a mannerism of the subject at a point in time in the video content, and determine an emotional state associated with the mannerism. In some embodiments, the logic causes the system to receive a file that correlates the mannerism with the point in time and provide a user interface that provides the video content with data from the file, such that the mannerism is provided with the data.
Also included are embodiments of a method. Some embodiments of the method include receiving video content of a subject, the video content comprising the subject interacting with a stimulus, identifying a mannerism of the subject at a point in time in the video content when the subject interacts with the stimulus, and determining an emotional state associated with the mannerism. In some embodiments, the method includes receiving a file that correlates the mannerism with the point in time and providing a user interface that provides the video content with data from the file, such that the mannerism is provided with the data, wherein the user interface further provides an emotional state value.
Also included are embodiments of a non-transitory computer-readable medium. Some embodiments of the non-transitory computer-readable medium include logic that, when executed by a computing device, causes the computing device to receive video content of a subject, the video content comprising the subject interacting with a stimulus and identify a facial expression of the subject at a point in time in the video content when the subject interacts with the stimulus. In some embodiment, the logic causes the computing device to determine an emotional state associated with the facial expression, receive a file that correlates the facial expression with the point in time, and provide a user interface that provides the video content with data from the file, such that the facial expression is provided with the data, wherein the user interface further provides an emotional state value.
It is to be understood that both the foregoing general description and the following detailed description describe various embodiments and are intended to provide an overview or framework for understanding the nature and character of the claimed subject matter. The accompanying drawings are included to provide a further understanding of the various embodiments, and are incorporated into and constitute a part of this specification. The drawings illustrate various embodiments described herein, and together with the description serve to explain the principles and operations of the claimed subject matter.
Embodiments disclosed herein include systems and methods for performing video analysis. Specifically, embodiments disclosed herein may be configured to determine an emotional state of a subject at one or more points in time in a video. The emotional state and the video content may additionally be provided within a user interface such that the user can play the video content and view the identified emotional states via indicators that are provided at the point in time or a plurality of points in time where the emotional state is detected. Embodiments disclosed herein also provide options for bookmarking various portions of the video content to easily recall the identified emotional states at a later time.
Additionally, some embodiments may be configured to create summary statistics, such as data tables and graphs from single or multiple time periods from a video and/or multiple time points from multiple videos and/or subjects. This allows the system to quickly summarize responses from the same subject in response to a stimulus and/or to different stimuli (e.g., products, digital images, physical images, still images, animated images, a discussion of an experience, etc.) and/or a plurality of subjects to the same stimuli. Embodiments may also be configured to store data from sharing among different computing devices.
Referring now to the drawings,
Accordingly, the user computing device 102 may include a personal computer, laptop computer, tablet, mobile computing device, mobile communications device, database, and/or other computing device operated by a corporate employee or other entity. The user computing device 102 includes a memory component 140, which stores state logic 144a and interface logic 144b. The state logic 144a may be configured to determine an emotional state of a subject in a video, as well as determine other information, such as determining a point in time that the emotional state occurred in the video, and/or other determinations. As discussed in more detail below, the state logic 144a may cause the user computing device 102 to accomplish these tasks internally or may accomplish these tasks by receiving the data from an external source, such as from the remote computing device 104 or other source. The interface logic 144b may cause the user computing device 102 to provide an interface that correlates the video content with the emotional state data, such that the user can determine the stimuli for creating the emotional state.
The remote computing device 104 may be configured to determine the emotional state of a subject in the video content. While in some embodiments the user computing device 102 can make this determination, in some embodiments, the video content is sent by the user computing device 102 to the remote computing device 104 for this purpose. In such embodiments, the remote computing device 104 may return a text file or other electronic document that the user computing device 102 can utilize for assembling the user interfaces and correlating the video content with the text file.
It should be understood that while the user computing device 102 is depicted as a single personal computer and the remote computing device 104 is depicted as a single server, these are merely examples. Specifically, the user computing device 102 may include one or more personal computers, servers, laptops, tablets, mobile computing devices, data storage devices, etc. that are configured for providing information to a user. Similarly, the remote computing device 104 may include one or more servers, personal computers, laptops, tablets, mobile computing devices, data storage devices, etc. for determining the emotional state and other data discussed herein.
Additionally, the memory component 140 may be configured to store operating logic 242, the state logic 144a, and the interface logic 144b, each of which may be embodied as a computer program, firmware, and/or hardware, as an example. A local communications interface 246 is also included in
The processor 230 may include any processing component operable to receive and execute instructions (such as from the data storage component 236 and/or memory component 140). The input/output hardware 232 may include and/or be configured to interface with a monitor, keyboard, mouse, printer, camera, microphone, speaker, and/or other device for receiving, sending, and/or presenting data. The network interface hardware 234 may include and/or be configured for communicating with any wired or wireless networking hardware, a satellite, an antenna, a modem, LAN port, wireless fidelity (Wi-Fi) card, WiMax card, mobile communications hardware, and/or other hardware for communicating with other networks and/or devices. From this connection, communication may be facilitated between the user computing device 102 and other computing devices.
Similarly, it should be understood that the data storage component 236 may reside local to and/or remote from the user computing device 102 and may be configured to store one or more pieces of data for access by the user computing device 102 and/or other components. In some embodiments, the data storage component 236 may be located remotely from the user computing device 102 and thus accessible via the network 100. In some embodiments however, the data storage component 236 may merely be a peripheral device, but external to the user computing device 102.
Included in the memory component 140 are the operating logic 242, the state logic 144a, and the interface logic 144b. The operating logic 242 may include an operating system and/or other software for managing components of the user computing device 102. Similarly, the state logic 144a may be configured to cause the user computing device 102 to utilize the video data 238a for determining an emotional state of the subject. The determined emotional state may be stored as the state date 238b, which may then be provided as an interface via the interface logic 144b. The interface logic 144b may cause the user computing device 102 to provide the user interfaces, as well as receive and respond to inputs from the user.
It should be understood that the components illustrated in
Also included are a yellow threshold option 336 and a green threshold option 338. As discussed in more detail, below, the user computing device 102 may be configured to correlate the video content with an accuracy determination of the determined emotional states of a user in the video content. Additionally, a probability of accuracy of the determined emotional states may be visually signified with a color indicator to provide the user with a mechanism for quickly and easily identifying whether the determined emotional state is most likely accurate, neutrally, accurate, or not likely accurate. Accordingly, the yellow threshold option 336 may be utilized to identify the boundary between the probability that a determined emotional state being neutrally accurate or not likely accurate. The green threshold option 338 may be utilized to identify the boundary between the determined emotional state being likely accurate or neutrally accurate.
Also included in the user interface 330 is an a min rating option 342 to identify the minimum emotional rating value that will be recorded and a max rating option 344 to identify the maximum emotional rating value that will be recorded. Specifically, once a data file has been selected, the min rating option 342 and the max rating option 344 will populate to show the range of emotion exhibited by the subject in the video content.
Also is a secondary data file option 340 for identifying a secondary data file that may be compared with the data file identified in the data file option 334. As an example, the user computing device 102 may create a data file that determines emotional states exhibited by the subject in the video content and correlates those determined emotional states with corresponding points in time in the video content. Additionally, the video content may be sent to the remote computing device 104 for making a similar determination. Accordingly, in such an example, the data file as determined by the user computing device 102 may be identified in the data file option 334, while the data file as determined by the remote computing device 104 may be identified in the secondary data file option 340. The two data files may thus be provided in the user interface 430, depicted in
The video content section 432 may provide video content of a subject that is interacting with a stimulus. As an example, a subject may be recorded viewing or interacting with a stimulus. The stimulus may be given to the subject as part of a consumer test or may simply be part of a shopping experience for the subject. Regardless, after the video content is captured, the video content may be analyzed to identify a mannerism, such as a facial expression, hand gesture, body position, posture, etc. of the subject and a point in time that the mannerism occurred. Additionally, an emotional state may be determined from the identified mannerism.
Accordingly, the video content section 432 may provide the captured video content to the user. Additionally, a graphical representation of the emotional response of the subject may be provided in the data section 440. Specifically, upon determining the emotional state of the subject, an emotional state value may be determined and graphically (and/or numerically) provided as to whether the emotional state is a positive, negative, or neutral state. Specifically, the emotional state may be determined as a probability of being one of the archetypal emotions (plus true happiness versus social smile versus neutrality). Thus, as the video content is played, an indicator in the data section 440 may identify the ranking may be provided. Similarly, as the user makes a user selection or manipulates the scroll option or indicates a selected point of the video content to view, video may skip to the selected point and the indicator in the data section 440 may advance with the video scroll option 446.
Similarly, the frame information section 442 includes the frame state indicator 444. As defined by the yellow threshold option 336 and the green threshold option 338 in
The user interface 430 also includes the bookmark section 434. Specifically, if the user desires to create a bookmarked point for later use, the user may input an annotation into the annotation section 438 and select the bookmark option 436a. This will create a marker for that point in time of the video content to retrieve later and direct the user to additional options, as depicted in
Specifically, the user may enter a title of the bookmark in the title prompt 532. The user may additionally input any notes or comments in the notes prompt 534. The start field 536 and the stop field 538 may identify when the bookmark starts and stops. While the user may graphically identify start and stop points of the bookmark in the user interface 430 (
It should be understood that in some embodiments, the user computing device 102 (
The dimensions and values disclosed herein are not to be understood as being strictly limited to the exact numerical values recited. Instead, unless otherwise specified, each such dimension is intended to mean both the recited value and a functionally equivalent range surrounding that value. For example, a dimension disclosed as “40 mm” is intended to mean “about 40 mm.”
Every document cited herein, including any cross referenced or related patent or application, is hereby incorporated herein by reference in its entirety unless expressly excluded or otherwise limited. The citation of any document is not an admission that it is prior art with respect to any invention disclosed or claimed herein or that it alone, or in any combination with any other reference or references, teaches, suggests or discloses any such invention. Further, to the extent that any meaning or definition of a term in this document conflicts with any meaning or definition of the same term in a document incorporated by reference, the meaning or definition assigned to that term in this document shall govern.
While particular embodiments of the present invention have been illustrated and described, it would be understood to those skilled in the art that various other changes and modifications can be made without departing from the spirit and scope of the invention. It is therefore intended to cover in the appended claims all such changes and modifications that are within the scope of this invention.