This U.S. patent application claims priority under 35 U.S.C. §119 to India Application No. 3252/CHE/2015, filed Jun. 27, 2015. The entire contents of the aforementioned application are incorporated herein by reference.
The present subject matter is related, in general to human behavior detection, and more particularly, but not exclusively to emotion detection system and method for determining emotions of a user using a camera.
An individual may undergo various emotions, such as anger, anxiety, happiness, sadness, fear, anger, surprise and disgust. Nowadays, sensing and understanding the individual's emotions have become crucial for one or more scenarios of society. Considering a business scenario, where sensing and understanding emotions of customer while dealing with the customer helps in offering right products or services to the customer. In such a way, in any commercial business transactions, the customer and a service provider can be benefitted in an effective manner. In a scenario of public breach like dacoit, terrorism etc., sensing and understanding of the individual's emotions desiring to cause such public breach is significant to stop the individual from committing such public breach. Other scenario can be interviewing candidate, where an interviewer can ask particular level and type of questions as per he emotions of the candidate.
In one conventional method, an individual's emotions may be detected based on facial expressions of the individual. The facial expressions are generally observed through video feeds captured using a camera. However, such a way of determining emotions of the individual from the facial expressions is error prone because the individual can manipulate the facial expression for suppressing actual emotions, mood and intent of the individual. Hence, such a conventional method determining the emotions from the facial expressions may not indicate the actual emotions of the individual.
In another conventional method, an individual's physiological parameters may be measured using wearable devices, that is, the wearable devices are placed on the individual to measure the physiological parameters. However, such wearable devices are expensive and the user may not be able to afford such devices or may not be willing to spend extra on the wearable devices. Further, measurement of the physiological parameters using the wearable devices may not help in determine the emotions accurately of the individual since some physiological parameters may get erroneously captured by the wearable devices.
One or more shortcomings of the prior art are overcome and additional advantages are provided through the present disclosure. Additional features and advantages are realized through the techniques of the present disclosure. Other embodiments and aspects of the disclosure are described in detail herein and are considered a part, of the claimed disclosure.
In one embodiment, the present disclosure relates to a method for determining emotions of a user using a camera. The method comprises receiving at least one image of the user from the camera. The method further comprises detecting at least one region of interest of the user in the at least one image. The method further comprises generating a video plethysmographic waveform by analyzing the at least one region of interest. The method further comprises determining at least one physiological characteristic based on the video plethysmographic waveform. The method further comprises determining the emotions of the user by comparing the at least one physiological characteristic with predefined physiological characteristics defined for each emotion.
In another embodiment, the present disclosure relates to an emotion detection system for determining emotions of a user using a camera. The system further comprises a processor and a memory communicatively coupled to the processor, wherein the memory stores processor-executable instructions, which, on execution, cause the processor to perform operations comprising receiving at least one image of the user from the camera. The operations further comprise detecting at least one region of interest of the user in the at least one image. The operations further comprise generating a video plethysmographic waveform by analyzing the at least one region of interest. The operations further comprise determining at least one physiological characteristic based on the video plethysmographic waveform. The operations further comprise determining the emotions of the user by comparing the at least one physiological characteristic with predefined physiological characteristics defined for each emotion.
In another embodiment, the present disclosure relates to a non-transitory computer readable medium including instructions stored thereon that when processed by at least one processor causes an emotion detection system to perform the act of receiving at least one image of the user from the camera. The act further comprises detecting at least one region of interest of the user in the at least one image. The act further comprises generating a video plethysmographic waveform by analyzing the at least one region of interest. The act further comprises determining at least one physiological characteristic based on the video plethysmographic waveform. The act further comprises determining emotions of the user by comparing the at least one physiological characteristic with predefined physiological characteristics defined for each emotion.
The foregoing summary is illustrative only and is not intended to be in any way limiting. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features will become apparent by reference to the drawings and the following detailed description.
The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate exemplary embodiments and, together with the description, serve to explain the disclosed principles. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the figures to reference like features and components. Some embodiments of system and/or methods in accordance with embodiments of the present subject matter are now described, by way of example only, and with reference to the accompanying figures, in which:
It should be appreciated by those skilled in the art that any block diagrams herein represent conceptual views of illustrative systems embodying the principles of the present subject matter. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudo code, and the like represent various processes which may be substantially represented in computer readable medium and executed by a computer or processor, whether or not such computer or processor is explicitly shown.
In the present document, the word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any embodiment or implementation of the present subject matter described herein as “exemplary” not necessarily to be construed as preferred or advantageous over other embodiments.
While the disclosure is susceptible to various modifications and alternative forms, specific embodiment thereof has been shown by way of example in the drawings and will be described in detail below. It should be understood, however that it is not intended to limit the disclosure to the particular forms disclosed, but on the contrary, the disclosure is to cover all modifications, equivalents, and alternative falling within the scope of the disclosure.
The terms “comprises”, “comprising”, or any other variations thereof, are intended to cover a nonexclusive inclusion, such that a setup, device or method that comprises a list of components or steps does not include only those components or steps but may include other components or steps not expressly listed or inherent to such setup or device or method. In other words, one or more elements in a system or apparatus proceeded by “comprises . . . a” does not, without more constraints, preclude the existence of other elements or additional elements in the system or apparatus.
In the following detailed description of the embodiments of the disclosure, reference is made to the accompanying drawings that form a part hereof, and in which are shown by way of illustration specific embodiments in which the disclosure may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the disclosure, and it is to be understood that other embodiments may be utilized and that changes may be made without departing from the scope of the present disclosure. The following description is, therefore, not to be taken in a limiting sense.
The present disclosure relates to a method and an emotion detection system for determining emotions of a user using a camera. The method comprises receiving at least one image of the user from the camera which is communicatively connected to the emotion detection system. In an embodiment, the at least one image is a series of images taken from a video frame. In an embodiment, the camera is configured in the emotion detection system and/or coupled to the emotion detection system. In an embodiment, the camera can be connected to the emotion detection system over a network comprising wired or wireless network. From the received at least one image of the user, regions of interest of the user is detected. More particularly, the regions of interest are the uncovered body parts of the user. Then, Video Plethysmographic Waveforms (VPW) are generated by analyzing the regions of the interest. Particularly, the video plethysmographic waveforms are generated based on pixel variations of the at least one image of the user corresponding to the detected regions of the interest.
In one implementation, the emotion detection system 200 may be implemented in a variety of computing systems, such as a laptop computer, a desktop computer, a Personal Computer (PC), a notebook, a smartphone, a tablet, e-book readers (e.g., Kindles and Nooks), a server, a network server, and the like. In one example, the emotion detection system 200 is configured to determine emotions of the user non-invasively. Particularly, the emotions of the user are detectable from video frames captured. Thus, any wearable device and/or contact of the user with the emotion detection system 200 and/or any device connected to the emotion detection system 200 is excluded. The components of the emotion detection system 200 are explained in detail below sections of the description.
In an embodiment, the emotion detection system 200 is communicatively connected to at least one camera 208. In one example, the at least one camera 208 includes, but is not limited to, a video camera, digital camera, Charged Couple Device (CCD) camera, an image camera, Universal Serial Bus (USB) camera, video cards with composite or S-video devices and other such camera which is capable of capturing video frames of users. Here, the users are persons or subjects whose at least one age is captured and emotions are to be determined, In one implementation, the at least one camera 208 is a separate device which is coupled to the emotion detection system 200 and/or connected to the emotion detection system 200 over a network. In one implementation, the at least one camera 208 is configured in one or more user devices (not shown) used by the users. The one or more user devices include, but are not limited to, computing systems, such as a laptop computer, a desktop computer, a Personal Computer (PC), a notebook, a smartphone, a smartwatch, a wearable device, a tablet, e-book readers (e.g., Kindles and Nooks). In such a case, the emotion detection system 200 may be communicatively connected to the one or more user devices and one or more servers (not shown). In one implementation, the at least one image of the user can be received from the one or more user devices and/or the one or more servers (not shown) by the emotion detection system 200. In such a case, the at least one image of the user may be stored in files/library of the one or more user devices, and/or memory chip, USBs, hard disks, and/or the one or more servers. Here, the one or more servers are a server associated with the emotion detection system 200 and/or third party servers accessible by the emotion detection system 200. In one implementation, the at least one image of the user can be downloaded from Internet.
In the illustrated
The I/O interface 202 is a medium through which the at least one image of the user can be received from the at least one camera 208, and/or the one or more user devices and/or the one or more servers. Further, the I/O interface 202 is configured to receive a feedback from the users and/or operators who are capable of operating the emotion detection system 200. The I/O interface 202 provides results on determining the emotions of the user. Particularly, the emotions of the user are provided to a display unit (not shown in
The processor 204 may comprise at least one data processor for executing program components for processing system-generated video plethysmographic waveforms of corresponding regions of interest of the user from the at least one image of the user. The processor 204 is configured to detect at least one region of interest of the user. In an exemplary embodiment, the region of interest of the user is uncovered body part of the user. The processor 202 analyzes the at least one region of interest of the user and forms colored histogram i.e. red channel, and/or green channel and/or blue channel of the at least region of interest. In an embodiment, only green channel of the at least one region of interest is analyzed. Then, the processor 202 generates Video Plethysmographic Waveforms (VPW) for the colored histogram of the at least one region of interest of the user. The processor 202 generates the VPW based on pixel variations of the at least one image of the user of the region of interest. The processor 202 determines at least one physiological characteristic of the user based one the VPW. The processor 202 compares the at least one physiological characteristic of the user with predefined physiological characteristics defined for each emotion. The processor 202 is configured to determine the emotions of the user based on the comparison. In an embodiment, the processor 202 is configured to process the feedback received from the user and/or the operator for updating the predefined physiological characteristics.
The memory 206 stores instructions which are executable by the at least one processor 204. In an embodiment, the memory 206 stores image information, region of interest data, VPW data, physiological characteristic data, predefined physiological data and an emotion list. In an embodiment, the image information, the region of interest data, the VPW data, the physiological characteristic data, the predefined physiological data and the emotion list are stored as one or more data required for determining the emotions of the user as described in the following description of the disclosure.
In an embodiment, the one or more data 300 may include, for example, the image information 302, the region of interest data 304, the VPW data 306, the physiological characteristic data 308, the predefined physiological data 310 and the emotion list 312, and other data 314 for determining the emotions of the user. In an embodiment, the data 300 including the image information 302, the region of interest data 304, the VPW data 306, the physiological characteristic data 308 are the data which are detected and/or calculated in real-time. Particularly, the image information 302, the region of interest data 304, the VPW data 306, the physiological characteristic data 308 are not predefined or preconfigured beforehand in the emotion detection system 200. The predefined physiological data 310 and the emotion list 312 are defined beforehand and stored in the emotion detection system 200.
The image information 302 refers to information of pixels of the at least one image of the user. The information of the pixels of the at least image may include, but is not limiting to, pixel size, pixel color and number of pixels of the at least one image.
The region of interest data 304 refers to the at least one region of interest of the user. For example, face and hands may be the at least one region of interest of the user.
The VPW data 306 refers to the VPW of the corresponding at least one region of interest, which is being generated based on pixel variations of the at least one image of the user. The VPW data 306 includes details of the VPW including depth, width, altitude, distortion/noise of the waveforms being generated.
The physiological characteristic data 308 refers to the at least one physiological characteristic being determined from the VPW. The physiological characteristic data 308 may include, but is not limited to, SPO2, respiratory rate and heart rate of the user, which are being measured from the corresponding generated VPW.
The predefined physiological characteristics data 310 includes the physiological characteristics which includes, but is not limited to, SPO2, respiratory rate and heart rate that are defined for each corresponding emotion during configuration. In an example, the predefined physiological characteristics data 310 may also comprise at least one of training sequence waveforms. The training sequence waveforms are implemented using machine learning techniques and correspond to various emotions the user may undergo. The training sequence waveforms may be used for identifying emotions by comparing with the VPW.
The emotion list 312 refers to list of all the emotions of user in general. The emotion list 312 may include, but not limited to, happiness, sadness, fear, anger, surprise and disgust and other emotions of the user may be exhibited in a human being. In an embodiment, the predefined physiological characteristics data 310 and the emotion list 312 may be charted together in the memory 206. Particularly, for each physiological characteristics data 310, a corresponding emotion is defined and stored.
The other data 314 may refer to such data which can be referred for determining the emotions of the user.
In an embodiment, the one or more data 300 in the memory 206 are processed by the one or more modules 316 of the emotion detection system 200. The one or more modules 316 may be stored within the memory 206 as shown in
In one implementation, the one or more modules 316 may include, for example, a receiving module 318, a detection module 320, a VPW generation module 322, an emotion detection module 324, and an output module 326. The memory 206 may also comprise other modules 328 to perform various miscellaneous functionalities of the emotion detection system 200. It will be appreciated that such aforementioned modules may be represented as a single module or a combination of different modules.
The receiving module 318 receives the at least one image of the user from the at least one camera 208 and/or the one or more user devices and/or the one or more servers. For example, consider the camera 208 in a public place such as railway station, shopping malls, bus terminus, highways etc. In such a case, the at least one image of each of the users or persons is captured by the camera which is in turn received by the receiving module 318.
The detection module 320 detects the at least one region of interest of the user from the received at least one image of the user. In an embodiment, the detection module 320 detects the at least one region of interest which is body part of the user that is not covered. For example, the detection module 320 can detect face, lower palm and upper palm as the region of interest from the at least one image of the user.
The Video Plethysmographic Waveform (VPW) generation module 322 analyzes the at least one region of interest of the user being detected. The VPW generation module 322 generates a color histogram i.e., video plethysmographic gradient histogram of the at least one region of interest. In an embodiment, the color histogram includes, but is not limited to, red channel, green channel and blue channel. Then, the waveforms of the at least one region of interest is generated from the color histogram in a form of corresponding trace, for example, red trace, green trace and blue trace. Then, the VPW generation module 322 generates the VPW of the corresponding trace. In an example, since emotions of the users are dynamic and may change over a short time, the VPW generation module 322 may generate the VPW by monitoring the at least one regions of interest for a predefined time period. For instance, the VPW generation module 322 may monitor the region of interest and receive video feed of the region of interest for 30 seconds. Thereafter, the VPW generation module 322 may generate the VPW for the 30 seconds to identify the emotion of the user.
The emotion detection module 324 determines at least one physiological characteristic of the user based on the VPW being generated. The at least one physiological characteristic includes, but is not limited to SPO2, the respiratory rate and the heart rate of the user. Then, the emotion detection module 324 compares the at least one physiological characteristic with the predefined physiological characteristics. If the at least one physiological characteristic matches with the predefined physiological characteristic, then the emotion corresponding to the predefined physiological characteristic is determined for the user. In one implementation, the emotion detection module 324 may compare the VPW generated with the at least one of training sequence waveforms to identifying the emotion the user is going through. In an example, upon determining the emotion by comparing the VPW and the at least one of training sequence waveforms, the emotion detection module 324 may validate the emotion of the user by matching the at least one physiological characteristic with the predefined physiological characteristics. In case, the emotion detection module 324 identifies that the at least one physiological characteristic doesn't match with the predefined physiological characteristic, the emotion detection module 324 may use the at least one physiological characteristic and the emotion identified for training the emotion detection system 200.
The output module 326 provides the detected emotions of the user to at least one of the display unit, the one or more user devices and the one or more servers. In an embodiment, the feedback from the user is received from the user by the receiving module 318 and the predefined physiological characteristics is updated by the output module 326 based on the feedback.
As illustrated in
The order in which the method 500 is described is not intended to be construed as a limitation, and any number of the described method blocks can be combined in any order to implement the method. Additionally, individual blocks may be deleted from the methods without departing from the scope of the subject matter described herein. Furthermore, the method can be implemented in any suitable hardware, software, firmware, or combination thereof.
At block 502, the at least one image of the user is received from the camera, and/or the one or more user devices and/or the one or more servers.
At block 504, the at least one region of interest of the user is detected. In an embodiment, uncovered body part of the user is detected as the at least one region of interest.
At block 506, the VPW of the corresponding at least one region of interest is generated by analyzing the corresponding at least one region of interest. In an embodiment, the video plethysmographic waveform is generated based on pixel variations of the image, corresponding to each of the at least one region of interest.
At block 508, the at least one physiological characteristic of the user is determined based on the VPW. In an embodiment, the at least one physiological characteristic comprises SPO2, the respiratory rate, and the heart rate of the user.
At block 510, the at least one physiological characteristic is compared with the predefined physiological characteristics defined for the emotions. If the at least one physiological characteristic matches with the predefined physiological characteristics then process goes to block 512 via “Yes”.
At block 512, the emotions of the user are determined based on the matching. If the at least one physiological characteristic does not match with the predefined physiological characteristics then process goes to block 514 via “No” where the process is ended. In an embodiment, the feedback is received from the user/any other persons to update the predefined physiological characteristics.
The processor 602 may be disposed in communication with one or more input/output (I/O) devices (not shown) via I/O interface 601. The I/O interface 601 may employ communication protocols/methods such as, without limitation, audio, analog, digital, monoaural, RCA, stereo, IEEE-1394, serial bus, universal serial bus (USB), infrared, PS/2, BNC, coaxial, component, composite, digital visual interface (DVI), high-definition multimedia interface (HDMI), RF antennas, S-Video, VGA, IEEE 802.n/b/g/n/x, Bluetooth, cellular (e.g., code-division multiple access (CDMA), high-speed packet access (HSPA+), global system for mobile communications (GSM), long-term evolution (LTE), WiMax, or the like), etc.
Using the I/O interface 601, the computer system 600 may communicate with one or more I/O devices. For example, the input device may be an antenna, keyboard, mouse, joystick, (infrared) remote control, camera, card reader, fax machine, dongle, biometric reader, microphone, touch screen, touchpad, trackball, stylus, scanner, storage device, transceiver, video device/source, etc. The output device may be a printer, fax machine, video display (e.g., cathode ray tube (CRT), liquid crystal display (LCD), light-emitting diode (LED), plasma, Plasma display panel (PDP), Organic light-emitting diode display (OLED) or the like), audio speaker, etc.
In some embodiments, the computer system 600 is connected to the one or more user devices 611a, . . . , 611n, the one or more servers 610a, . . . , 610n and the camera 614 through a communication network 609. The processor 602 may be disposed in communication with the communication network 609 via a network interface 603. The network interface 603 may communicate with the communication network 609. The network interface 603 may employ connection protocols including, without limitation, direct connect, Ethernet (e.g., twisted pair 10/100/1000 Base T), transmission control protocol/internet protocol (TCP/IP), token ring, IEEE 802.11a/b/g/n/x, etc. The communication network 609 may include, without limitation, a direct interconnection, local area network (LAN), wide area network (WAN), wireless network (e.g., using Wireless Application Protocol), the Internet, etc. Using the network interface 603 and the communication network 609, the computer system 600 may communicate with the one or more user devices 611a, . . . , 611n, the one or more servers 610a, . . . , 610n and the camera 614. The network interface 603 may employ connection protocols include, but not limited to, direct connect, Ethernet (e.g., twisted pair 10/100/1000 Base T), transmission control protocol/internet protocol (TCP/IP), token ring, IEEE 802.11a/b/g/n/x etc.
The communication network 609 includes, but is not limited to, a direct interconnection, an e-commerce network, a peer to peer (P2P) network, local area network (LAN), wide area network (WAN), wireless network (e.g., using Wireless Application Protocol), the Internet, Wi-Fi and such. The first network and the second network may either be a dedicated network or a shared network, which represents an association of the different types of networks that use a variety of protocols, for example, Hypertext Transfer Protocol (HTTP), Transmission Control Protocol/Internet Protocol (TCP/IP), Wireless Application Protocol (WAP), etc., to communicate with each other. Further, the first network and the second network may include a variety of network devices, including routers, bridges, servers, computing devices, storage devices, etc.
In some embodiments, the processor 602 may be disposed in communication with a memory 605 (e.g., RAM, ROM, etc. not shown in
The memory 605 may store a collection of program or database components, including, without limitation, user interface 606, an operating system 607, web server 608 etc. In some embodiments, computer system 600 may store user/application data 606, such as the data, variables, records, etc. as described in this disclosure. Such databases may be implemented as fault-tolerant, relational, scalable, secure databases such as Oracle or Sybase.
The operating system 607 may facilitate resource management and operation of the computer system 600. Examples of operating systems include, without limitation, Apple Macintosh OS X, Unix, Unix-like system distributions (e.g., Berkeley Software Distribution (BSD), FreeBSD, NetBSD, OpenBSD, etc.), Linux distributions (e.g., Red Hat, Ubuntu, Kubuntu, etc.), IBM OS/2, Microsoft Windows (XP, Vista/7/8, etc.), Apple iOS, Google Android, Blackberry OS, or the like.
In some embodiments, the computer system 600 may implement a web browser 607 stored program component. The web browser 608 may be a hypertext viewing application, such as Microsoft Internet Explorer, Google Chrome, Mozilla Firefox, Apple Safari, etc. Secure web browsing may be provided using Secure Hypertext Transport Protocol (HTTPS), Secure Sockets Layer (SSL), Transport Layer Security (TLS), etc. Web browsers 608 may utilize facilities such as AJAX, DHTML, Adobe Flash, JavaScript, Java, Application Programming Interfaces (APIs), etc, In some embodiments, the computer system 600 may implement a mail server stored program component, The mail server may be an Internet mail server such as Microsoft Exchange, or the like. The mail server may utilize facilities such as ASP, ActiveX, ANSI C++/C#, Microsoft.NET, CGI scripts, Java, JavaScript, PERL, PHP, Python, WebObjects, etc. The mail server may utilize communication protocols such as Internet Message Access Protocol (IMAP), Messaging Application Programming Interface (MAPI), Microsoft Exchange, Post Office Protocol (POP), Simple Mail Transfer Protocol (SMTP), or the like. In some embodiments, the computer system 600 may implement a mail client stored program component. The mail client may be a mail viewing application, such as Apple Mail, Microsoft Entourage, Microsoft Outlook, Mozilla Thunderbird, etc.
Furthermore, one or more computer-readable storage media may be utilized in implementing embodiments consistent with the present disclosure. A computer-readable storage medium refers to any type of physical memory on which information or data readable by a processor may be stored. Thus, a computer-readable storage medium may store instructions for execution by one or more processors, including instructions for causing the processor(s) to perform steps or stages consistent with the embodiments described herein. The term “computer-readable medium” should be understood to include tangible items and exclude carrier waves and transient signals, i.e., be non-transitory. Examples include Random Access Memory (RAM), Read-Only Memory (ROM), volatile memory, nonvolatile memory, hard drives, CD ROMs, DVDs, flash drives, disks, and any other known physical storage media.
Advantages of the embodiment of the present disclosure are illustrated herein.
Embodiments of the present disclosure are capable of determining emotions of the user using just a camera, that is, from image clips or video clips of the user. In such a way, usage of wearable devices is eliminated, user contact is eliminated and also the application of the present disclosure is cost effective.
Embodiments of the present disclosure reduce measuring physiological characteristics erroneously by generating video plethysmographic waveforms which actually indicates the actual emotions of the user.
Embodiments of the present disclosure provide dynamic technique of determining the emotions of the user using the images of the user in real-time and the video plethysmographic waveforms corresponding to the region of interest of the user from the image.
Embodiments of the present disclosure can determine the actual hidden emotions of the user without reading the facial expressions.
The described operations may be implemented as a method, system or article of manufacture using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof. The described operations may be implemented as code maintained in a “non-transitory computer readable medium”, where a processor may read and execute the code from the computer readable medium. The processor is at least one of a microprocessor and a processor capable of processing and executing the queries. A non-transitory computer readable medium may comprise media such as magnetic storage medium (e.g., hard disk drives, floppy disks, tape, etc.), optical storage (CD-ROMs, DVDs, optical disks, etc.), volatile and non-volatile memory devices (e.g., EEPROMs, ROMs, PROMs, RAMs, DRAMs, SRAMs, Flash Memory, firmware, programmable logic, etc.), etc. Further, non-transitory computer-readable media comprise all computer-readable media except for a transitory. The code implementing the described operations may further be implemented in hardware logic (e.g., an integrated circuit chip, Programmable Gate Array (PGA), Application Specific Integrated Circuit (ASIC), etc.).
Still further, the code implementing the described operations may be implemented in “transmission signals”, where transmission signals may propagate through space or through a transmission media, such as an optical fiber, copper wire, etc. The transmission signals in which the code or logic is encoded may further comprise a wireless signal, satellite transmission, radio waves, infrared signals, Bluetooth, etc. The transmission signals in which the code or logic is encoded is capable of being transmitted by a transmitting station and received by a receiving station, where the code or logic encoded in the transmission signal may be decoded and stored in hardware or a non-transitory computer readable medium at the receiving and transmitting stations or devices. An “article of manufacture” comprises non-transitory computer readable medium, hardware logic, and/or transmission signals in which code may be implemented. A device in which the code implementing the described embodiments of operations is encoded may comprise a computer readable medium or hardware logic. Of course, those skilled in the art will recognize that many modifications may be made to this configuration without departing from the scope of the invention, and that the article of manufacture may comprise suitable information bearing medium known in the art.
The terms “an embodiment”, “embodiment”, “embodiments”, “the embodiment”, “the embodiments”, “one or more embodiments”, “some embodiments”, and “one embodiment” mean “one or more (but not all) embodiments of the invention(s)” unless expressly specified otherwise.
The terms “including”, “comprising”, “having” and variations thereof mean “including but not limited to”, unless expressly specified otherwise.
The enumerated listing of items does not imply that any or all of he items are mutually exclusive, unless expressly specified otherwise.
The terms “a”, and “the” mean “one or more”, unless expressly specified otherwise.
A description of an embodiment with several components in communication with each other does not imply that all such components are required. On the contrary a variety of optional components are described to illustrate the wide variety of possible embodiments of the invention.
When a single device or article is described herein, it will be readily apparent that more than one device/article (whether or not they cooperate) may be used in place of a single device/article. Similarly, where more than one device or article is described herein (whether or not they cooperate), it will be readily apparent that a single device/article may be used in place of the more than one device or article or a different number of devices/articles may be used instead of the shown number of devices or programs. The functionality and/or the features of a device may be alternatively embodied by one or more other devices which are not explicitly described as having such functionality/features. Thus, other embodiments of the invention need not include the device itself.
The illustrated operations of
Finally, the language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the invention be limited not by this detailed description, but rather by any claims that issue on an application based here on. Accordingly, the disclosure of the embodiments of the invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.
While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.
Number | Date | Country | Kind |
---|---|---|---|
3252/CHE/2015 | Jun 2015 | IN | national |