1. Field of the Disclosure
The present invention relates generally to systems and methods for deterring theft in a retail store. In particular, examples of the present invention are related to recording evidence of theft using an augmented reality device.
2. Background
Some retail stores extend across tens of thousands of feet and offer thousands of items for sale. Many customers visit such retail stores when shopping for a diverse set of items such as groceries, office supplies, and household wares. Typically, these stores can have dozens of aisles and/or departments. Accordingly, monitoring every portion of the store to prevent theft can be a challenging task. Merchants who sell products including groceries, office supplies, and household wares employ personnel and implement systems and policies to deal with the problem of theft. Eyewitness accounts of theft provide strong evidence used to convict thieves yet in many cases the eyewitness testimony cannot be trusted. It is the policy of many merchants that only security guards are trusted eyewitnesses to theft.
Non-limiting and non-exhaustive embodiments of the present invention are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various views unless otherwise specified.
Corresponding reference characters indicate corresponding components throughout the several views of the drawings. Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of various embodiments of the present disclosure. Also, common but well-understood elements that are useful or necessary in a commercially feasible embodiment are often not depicted in order to facilitate a less obstructed view of these various embodiments of the present disclosure.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present disclosure. It will be apparent, however, to one having ordinary skill in the art that the specific detail need not be employed to practice the present invention. In other instances, well-known materials or methods have not been described in detail in order to avoid obscuring the present disclosure.
Reference throughout this specification to “one embodiment”, “an embodiment”, “one example” or “an example” means that a particular feature, structure or characteristic described in connection with the embodiment or example is included in at least one embodiment of the present disclosure. Thus, appearances of the phrases “in one embodiment”, “in an embodiment”, “one example” or “an example” in various places throughout this specification are not necessarily all referring to the same embodiment or example. Furthermore, the particular features, structures or characteristics may be combined in any suitable combinations and/or sub-combinations in one or more embodiments or examples. In addition, it is appreciated that the figures provided herewith are for explanation purposes to persons ordinarily skilled in the art and that the drawings are not necessarily drawn to scale.
Embodiments in accordance with the present disclosure may be embodied as an apparatus, method, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.), or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “module” or “system.” Furthermore, the present invention may take the form of a computer program product embodied in any tangible medium of expression having computer-usable program code embodied in the medium.
It is desirable to have evidence of theft when prosecuting a suspected thief. A video of a theft occurring can be used as evidence. Eye witness testimony can be used as evidence. However, many merchants consider only security guards as reliable eyewitnesses.
Embodiments of the present disclosure can help merchants prevent theft and prosecute perpetrators recording evidence of theft. Some embodiments of the present disclosure can also allow a security guard to witness a theft in real-time. For example, a system according to an embodiment of the disclosure can include a monitoring server receiving signals from an augmented reality device such as a head mountable unit worn by a store employee as he goes about his duties in the retail store. When the employee witnesses suspicious customer behavior, the augmented reality device worn by the employee can transmit a theft alert signal. The monitoring server can receive and process the theft alert signal. In response to the theft alert signal, the monitoring server can link the augmented reality device with an electronic computing device operated by a second employee, such as a security guard. The security guard can be located at the retail store or at a remote location.
The theft alert signal can be communicated to the monitoring server 12 with an augmented reality device such as a head mountable unit 14. The head mountable unit 14 can be worn by an employee while the employee is performing his duties within the retail store. In the illustrated embodiment of
Signals transmitted by the head mountable unit 14 and received by the monitoring server 12, and vice-versa, can be communicated over a network 20. As used herein, the term “network” can include, but is not limited to, a Local Area Network (LAN), a Metropolitan Area Network (MAN), a Wide Area Network (WAN), the Internet, or combinations thereof. Embodiments of the present disclosure can be practiced with a wireless network, a hard-wired network, or any combination thereof.
The monitoring server 12 can determine that the theft alert signal contains data indicative of an alert or warning that a theft may be occurring. The first employee can reach this conclusion while observing the behavior of a person in the retail store and use the head mountable unit 14 to convey this suspicion/conclusion to the security guard. For example, the signal can be an audio signal containing the first employee's voice stating a theft is occurring. In response to receiving the theft alert signal, the monitoring server 12 can link the head mountable unit 14 worn by the first employee with an electronic computing device 22 that is physically remote from the head mountable unit 14. The monitoring server 12 can link the head mountable unit 14 and the electronic computing device 22 to permit communication between the first employee and a security guard operating the electronic computing device 22. In some embodiments of the present disclosure, the electronic computing device 22 can be located in the same retail store with the first employee. In some embodiments of the present disclosure, the electronic computing device 22 can be remote from the retail store occupied by the first employee.
The operator of the electronic computing device 22 is a security guard operable to assist the first employee in gathering evidence of a theft. For example, the first employee can verbally state the circumstance giving rise to the suspicion that a theft is occurring. The statements of the first employee can be captured by a microphone 44 of the head mountable unit 14 and transmitted by the head mountable unit 14 to the monitoring server 12. The initial signal from the first employee can be denoted as a theft alert signal. Subsequent signals originating from the first employee during the interaction with the security guard can be denoted as monitoring communication signals, as the first employee is monitoring the suspected perpetrator's behavior in the retail store.
The monitoring server 12 can receive the theft alert signal and one or more subsequent monitoring communication signals from the first employee. The monitoring server 12 can transmit the theft alert and monitoring communication signals to the security guard operating the electronic computing device 22. The verbal statements of the first employee can be emitted through a speaker 24 of the electronic computing device 22, allowing the security guard to hear the first employee's statements.
The security guard can verbally respond to the first employee's statements. The statements of the security guard can be captured by a microphone 26 of the electronic computing device 22 and transmitted by the electronic computing device 22 as one or more directing communication signals to the monitoring server 12, as the security is directing the actions of the first employee. Directing communication signals provide guidance to the first employee in gathering evidence of theft. The monitoring server 12 can receive the directing communication signals from the security guard and transmit the directing communication signals to the first employee wearing the head mountable unit 14. The verbal statements of the security guard can be emitted through a speaker 52 of the head mountable unit 14, allowing the first employee to hear the security guard's statements.
The security guard can also receive video signals corresponding to the first employee's field of view, so that the security guard can see what the first employee is seeing. The field of view of the first employee can be captured by a camera 42 of the head mountable unit 14 and transmitted by the head mountable unit 14 as a monitoring communication signal to the monitoring server 12. The monitoring server 12 can receive a monitoring communication signal containing video data from the first employee and transmit the monitoring communication signal to the security guard operating the electronic computing device 22. The video feed corresponding to the first employee's field of view can be displayed on a display 28 of the electronic computing device 22, allowing the security guard to see what the first employee is seeing in real-time. The security guard can use the video feed to direct the first employee's gaze to a particular location to better gather evidence of theft. In some embodiments of the present disclosure, the video feed generated by the first employee can be “backdated” by some length of time, such as by way of example and not limitation one minute. This feature can be desirable since a theft may be witnessed before the first employee can speak or gesture to prompt the transmission of the theft alert signal. In some embodiments, the augmented reality device or the monitoring server can store a predetermined number of minutes of video.
The exchange of video and audio information can facilitate the first employee's usefulness in gathering evidence of theft within the retail store. In addition, the security guard can transmit textual data and information to the first employee with the electronic computing device 22. For example, the security guard can transmit textual directions to the first employee instead of verbal statements to prevent sound from being emitted by the speaker 52. The first employee can view the instructions on a display 46 of the head mountable unit 14.
The processor 40 can be operable to receive signals generated by the other components of the communications unit 18. The processor 40 can also be operable to control the other components of the communications unit 18. The processor 40 can also be operable to process signals received by the head mount unit 14. While one processor 40 is illustrated, it should be appreciated that the term “processor” can include two or more processors that operate in an individual or distributed manner.
The head mount unit 14 can include one or more cameras 42. Each camera 42 can be configured to generate a video signal. One of the cameras 42 can be oriented to generate a video signal that approximates the field of view of the first employee wearing the head mountable unit 14. Each camera 42 can be operable to capture single images and/or video and to generate a video signal based thereon. The video signal may be representative of the field of view of the first employee wearing the head mountable unit 14.
In some embodiments of the disclosure, cameras 42 may be a plurality of forward-facing cameras 42. The cameras 42 can be a stereo camera with two or more lenses with a separate image sensor or film frame for each lens. This arrangement allows the camera to simulate human binocular vision and thus capture three-dimensional images. This process is known as stereo photography. The cameras 42 can be configured to execute computer stereo vision in which three-dimensional information is extracted from digital images. In such embodiments, the orientation of the cameras 42 can be known and the respective video signals can be processed to triangulate an object with both video signals. This processing can be applied to determine the distance that the first employee is spaced from the object. Determining the distance that the first employee is spaced from the object can be executed by the processor 40 or by the monitoring server 12 using known distance calculation techniques.
Processing of the one or more, forward-facing video signals can also be applied to determine the identity of the object. Determining the identity of the object, such as the identity of an item in the retail store, can be executed by the processor 40 or by the monitoring server 12. If the processing is executed by the monitoring server 12, the processor 40 can modify the video signals limit the transmission of data back to the monitoring server 12. For example, the video signal can be parsed and one or more image files can be transmitted to the monitoring server 12 instead of a live video feed. Further, the video can be modified from color to black and white to further reduce transmission load and/or ease the burden of processing for either the processor 40 or the monitoring server 12. Also, the video can cropped to an area of interest to reduce the transmission of data to the monitoring server 12.
In some embodiments of the present disclosure, the cameras 42 can include one or more inwardly-facing camera 42 directed toward the first employee's eyes. A video signal revealing the first employee's eyes can be processed using eye tracking techniques to determine the direction that the first employee is viewing. In one example, a video signal from an inwardly-facing camera can be correlated with one or more forward-facing video signals to determine the object the first employee is viewing.
The microphone 44 can be configured to generate an audio signal that corresponds to sound generated by and/or proximate to the first employee. The audio signal can be processed by the processor 40 or by the monitoring server 12. For example, verbal signals can be processed by the monitoring server 12 such as “this item appears interesting.” Such audio signals can be correlated to the video recording.
The display 46 can be positioned within the first employee's field of view. Video content can be shown to the first employee with the display 46. The display 52 can be configured to display text, graphics, images, illustrations and any other video signals to the first employee. The display 46 can be transparent when not in use and partially transparent when in use to minimize the obstruction of the first employee's field of view through the display 46.
The transmitter 48 can be configured to transmit signals generated by the other components of the communications unit 18 from the head mountable unit 14. The processor 40 can direct signals generated by components of the communications unit 18 to the commerce sever 12 through the transmitter 48. The transmitter 48 can be an electrical communication element within the processor 40. In one example, the processor 40 is operable to direct the video and audio signals to the transmitter 40 and the transmitter 48 is operable to transmit the video signal and/or audio signal from the head mountable unit 14, such as to the monitoring server 12 through the network 20.
The receiver 50 can be configured to receive signals and direct signals that are received to the processor 40 for further processing. The receiver 50 can be operable to receive transmissions from the network 20 and then communicate the transmissions to the processor 40. The receiver 50 can be an electrical communication element within the processor 40. In some embodiments of the present disclosure, the receiver 50 and the transmitter 48 can be an integral unit.
The transmitter 48 and receiver 50 can communicate over a Wi-Fi network, allowing the head mountable device 14 to exchange data wirelessly (using radio waves) over a computer network, including high-speed Internet connections. The transmitter 48 and receiver 50 can also apply Bluetooth® standards for exchanging data over short distances by using short-wavelength radio transmissions, and thus creating personal area network (PAN). The transmitter 48 and receiver 50 can also apply 3G or 4G, which is defined by the International Mobile Telecommunications-2000 (IMT-2000) specifications promulgated by the International Telecommunication Union.
The head mountable unit 14 can include one or more speakers 52. Each speaker 52 can be configured to emit sounds, messages, information, and any other audio signal to the first employee. The speaker 52 can be positioned within the first employee's range of hearing. Audio content transmitted by the monitoring server 12 can be played for the first employee through the speaker 52. The receiver 50 can receive the audio signal from the monitoring server 12 and direct the audio signal to the processor 40. The processor 40 can then control the speaker 52 to emit the audio content.
The direction sensor 54 can be configured to generate a direction signal that is indicative of the direction that the first employee is facing. The direction signal can be processed by the processor 40 or by the monitoring server 12. For example, the direction sensor 54 can electrically communicate the direction signal containing direction data to the processor 40 and the processor 40 can control the transmitter 48 to transmit the direction signal to the monitoring server 12 through the network 20. By way of example and not limitation, the direction signal can be useful in determining the identity of an item(s) visible in the video signal, as well as the location of the first employee within the retail store.
The direction sensor 54 can include a compass or another structure for deriving direction data. For example, the direction sensor 54 can include one or more Hall effect sensors. A Hall effect sensor is a transducer that varies its output voltage in response to a magnetic field. For example, the sensor operates as an analog transducer, directly returning a voltage. With a known magnetic field, its distance from the Hall plate can be determined. Using a group of sensors disposing about a periphery of a rotatable magnetic needle, the relative position of one end of the needle about the periphery can be deduced. It is noted that Hall effect sensors can be applied in other sensors of the head mountable unit 14.
The position sensor 56 can be configured to generate a position signal indicative of the position of the first employee within the retail store. The position sensor 56 can be configured to detect an absolute or relative position of the first employee wearing the head mountable unit 14. The position sensor 56 can electrically communicate a position signal containing position data to the processor 40 and the processor 40 can control the transmitter 48 to transmit the position signal to the monitoring server 12 through the network 20.
Identifying the position of the first employee can be accomplished by radio, ultrasound or ultrasonic, infrared, or any combination thereof. The position sensor 56 can be a component of a real-time locating system (RTLS), which is used to identify the location of objects and people in real time within a building such as a retail store. The position sensor 56 can include a tag that communicates with fixed reference points in the retail store. The fixed reference points can receive wireless signals from the position sensor 56. The position signal can be processed to assist in determining one or more items that are proximate to the first employee and are visible in the video signal. The monitoring server 12 can receive position data and identify the location of the first employee in some embodiments of the present disclosure.
The orientation sensor 58 can be configured to generate an orientation signal indicative of the orientation of the first employee's head, such as the extent to which the first employee is looking downward, upward, or parallel to the ground. A gyroscope can be a component of the orientation sensor 58. The orientation sensor 58 can generate the orientation signal in response to the orientation that is detected and communicate the orientation signal to the processor 40. The orientation of the first employee's head can indicate whether the first employee is viewing a lower shelf, an upper shelf, or a middle shelf.
The accelerometer 60 can be configured to generate an acceleration signal indicative of the motion of the first employee. The acceleration signal can be processed to assist in determining if the first employee has slowed or stopped, tending to indicate that the first employee is evaluating one or more items for purchase. The accelerometer 60 can be a sensor that is operable to detect the motion of the first employee wearing the head mountable unit 14. The accelerometer 60 can generate a signal based on the movement that is detected and communicate the signal to the processor 40. The motion that is detected can be the acceleration of the first employee and the processor 40 can derive the velocity of the first employee from the acceleration. Alternatively, the monitoring server 12 can process the acceleration signal to derive the velocity and acceleration of the first employee in the retail store.
The proximity sensor 62 can be operable to detect the presence of nearby objects without any physical contact. The proximity sensor 62 can apply an electromagnetic field or a beam of electromagnetic radiation such infrared and assess changes in the field or in the return signal. Alternatively, the proximity sensor 62 can apply capacitive photoelectric principles or induction. The proximity sensor 62 can generate a proximity signal and communicate the proximity signal to the processor 40. The proximity sensor 62 can be useful in determining when a first employee has grasped and is inspecting an item.
The distance sensor 64 can be operable to detect a distance between an object and the head mountable unit 14. The distance sensor 64 can generate a distance signal and communicate the signal to the processor 40. The distance sensor 64 can apply a laser to determine distance. The direction of the laser can be aligned with the direction that the first employee is facing. The distance signal can be useful in determining the distance to an object in the video signal generated by one of the cameras 42, which can be useful in determining the first employee's location in the retail store.
Any combination of one or more computer-usable or computer-readable media may be utilized. For example, a computer-readable medium may include one or more of a portable computer diskette, a hard disk, a random access memory (RAM) device, a read-only memory (ROM) device, an erasable programmable read-only memory (EPROM or Flash memory) device, a portable compact disc read-only memory (CDROM), an optical storage device, and a magnetic storage device. Computer program code for carrying out operations of the present invention may be written in any combination of one or more programming languages.
The theft incident database 216 can include memory containing data associated with interactions between first employees and security guards. The data associated with a particular interaction between a first employee and a security guard can include audio data, video data, textual data, or other forms of data. For example, verbal conversations between the first employee and security guard can be stored as data associated with a particular interaction in the theft incident database 216. A video signal that is generated by an augmented reality device worn by the first employee during the interaction can also be stored as data associated with a particular interaction in the theft incident database 216. The identity of the first employee who detected theft can also be stored as data associated with a particular interaction in the theft incident database 216. The identity of the security guard who assisted the first employee can also be stored as data associated with a particular interaction in the theft incident database 216. The data in the sale help interaction database 216 can be organized based on one or more tables that may utilize one or more algorithms and/or indexes.
The processing device 218 can communicate with the database 216 and can receive one or more signals from the head mountable unit 14 and from the electronic computing device 22. The processing device 218 can include computer readable memory storing computer readable instructions and one or more processors executing the computer readable instructions.
The receiving module 220 can be operable to receive signals over the network 20, assess the signals, and communicate the signals or the data contained in the signals to other components of the monitoring server 212. The receiving module 220 can be configured to receive theft alert signals and monitoring communication signals from one or more first employees wearing respective augmented reality devices. The receiving module 220 can also be configured to receive one or more directing communication signals from one or more security guards operating respective electronic computing devices.
The receiving module 220 can receive a signal containing audio data such as the voice of a first employee. A signal containing audio data can be directed to the audio processing module 222 for further processing. Speech by a first employee can be captured by the microphone 44 and transmitted to the monitoring server 212 by the head mountable unit 14. The voice of the first employee can be continuously monitored as the first employee shops in some embodiments of the present disclosure.
The audio processing module 222 can analyze the audio data contained in a first employee signal, such as verbal statements made by a first employee. The audio processing module 222 can implement known speech recognition techniques to identify speech in an audio signal. The first employee's speech can be encoded into a compact digital form that preserves its information. The encoding can occur at the head mountable unit 14 or at the monitoring server 212. The audio processing module 222 can be loaded with a series of models honed to comprehend language. When encoded locally, the speech can be evaluated locally, on the head mountable unit 14. A recognizer installed on the head mountable unit 14 can communicate with the monitoring server 212 to gauge whether the voice contains a command can be best handled locally or if the monitoring server is better suited to execute the command. The audio processing module 222 can compare the first employee's speech against a statistical model to estimate, based on the sounds spoken and the order in which the sounds were spoken, what letters might be contained in the speech. At the same time, the local recognizer can compare the speech to an abridged version of that statistical model applied by the audio processing module 222. For both the monitoring server 212 and the head mountable unit 14, the highest-probability estimates are accepted as the letters contained in the first employee's speech. Based on these estimations, the first employee's speech, now embodied as a series of vowels and consonants, is then run through a language model, which estimates the words of the speech. Given a sufficient level of confidence, the audio processing module 222 can then create a candidate list of interpretations for what the sequence of words in your speech might mean. If there is enough confidence in this result, the audio processing module 222 can determine the first employee's intent.
In a first example, a first employee can state “I see a theft in progress” in an embodiment of the present disclosure. This statement can be contained in a signal received by the monitoring server 212. The signal can be processed and the statement can be recognized by the audio processing module 222. In response, the audio processing module 222 can communicate the indication that a theft is occurring to the linking module 224 for further processing, as will be set forth in greater detail below. Thus, the signal containing the first employee's voice expressing a theft is occurring can define a theft alert signal.
The receiving module 220 can receive a signal containing video data such as video containing the field of view of the first employee. A signal containing video data can be directed to the video processing module 223 for further processing. The field of view of the first employee can be captured by the camera 52 and transmitted to the monitoring server 212 by the head mountable unit 14. The video showing the field of view of the first employee can be continuously monitored as the first employee works within the retail store in some embodiments of the present disclosure.
The video processing sub-module 223 can receive a video signal generated by the camera 42 of the head mountable unit 14 from the receiving module 222. The display 46 of the head mountable unit 14 can overlap the field of view of the camera 42. Thus, the view of the first employee can also define the field of view of a video signal generated by the camera 42 and communicated to the monitoring server 212.
The video processing sub-module 223 can implement known video recognition/analysis techniques and algorithms to identify hand gestures by the first employee in the field of view of the camera 42. For example, the video processing sub-module 223 can identify the first employee's hand moving, such as movement in one rectilinear direction, rotation motion, and side-to-side or up-down movement. Any form of movement can be recognized as a theft alert signal by the commerce server in various embodiments of the present disclosure. The video signal can be processed and the images showing movement of the first employee's hand can be recognized by the video processing module 223. In response, the video processing module 223 can communicate the indication that a theft is occurring to the linking module 224 for further processing, as will be set forth in greater detail below. Thus, the signal containing the first employee's hand gesturing in the field of view can define a theft alert signal.
The linking module 224 can be configured to act on theft alerts contained in signals received from first employees. In response to the detection of a theft alert by the audio processing module 222 or video processing module 223, the linking module 224 can direct the transmission module 226 to transmit a signal to the electronic computing device 22. The initial signal transmitted to the electronic computing device 22 can include the data in the theft alert signal itself, such the voice of the first employee. In some embodiments of the present disclosure, the initial signal transmitted to the electronic computing device 22 can also contain the identity of the first employee (based on the identity of the head mountable unit 14), the location of the retail store occupied first employee, and/or some other data that may be useful in assisting the security guard. Subsequent monitoring communication signals can also be directed to the electronic computing device 22, unaltered or supplemented.
The electronic computing device 22 can respond to the initial theft alert signal received from the monitoring server 212 and subsequent monitoring communication signals by transmitting one or more directing communication signals back to the monitoring server. The receiving module 220 can be configured to pass directing communication signals to the linking module 224, bypassing the audio processing module 222 and the video processing module 223. The linking module 224 can direct the transmission module 226 to transmit directing communication signals to the head mountable unit 14. Thus, the linking module 224 can facilitate continuous and real-time communication between the first employee and the security guard.
After receiving an initial theft alert signal from the first employee, the linking module 224 can direct the receiving module 222 to direct audio and video signals received from the head mountable unit 14 directly to the linking module 224 and bypass the audio processing module 222 and the video processing module 223. The linking module 224 can then direct the transmission module 226 to transmit these signals, monitoring communication signals, to the electronic computing device 22.
The linking module 224 can also be configured to direct data associated with the interaction between the first employee and the security guard to the theft incident database 216 for storage. In response to the detection of a theft alert by the audio processing module 222, the linking module 224 can access the theft incident database 216 and establish an entry for the current interaction. Subsequent signals that are received from either the first employee or the security guard can be transmitted to the other party and also stored in the theft incident database 216. Thus, the theft incident database 216 can contain a record of each first employee-security guard interaction. Each record or entry in the theft incident database 216 can include data identifying the first employee, the security guard, the date and time of the interaction, and/or the location of the retail store occupied by the first employee in some embodiments of the present disclosure.
After a theft detection interaction has ended, the security guard can control the electronic computing device 22 to transmit a termination signal to the monitoring server 212. The termination signal can contain data directing the linking module 224 to terminate the link. The linking module 224 can direct the receiving module 220 to again direct audio signals from the head mountable unit 14 to the audio processing module 222 and direct video signals from the head mountable unit 14 to the video processing module 223.
It is noted that the various processing functions set forth above can be executed differently than described above in order to enhance the efficiency of an embodiment of the present disclosure in a particular operating environment. The processor 40 can assume a greater role in processing some of the signals in some embodiments of the present disclosure. For example, in some embodiments, the processor 40 of the head mountable unit 14a could modify the video signal to require less bandwidth. The processor 40 could convert a video signal containing color to black and white in order to reduce the bandwidth required for transmitting the video signal. In some embodiments, the processor 40 could crop the video, or sample the video and display frames of interest. A frame of interest could be a frame that is significantly different from other frames, such as a generally low quality video having an occasional high quality frame. Thus, in some embodiments, the processor 40 could selectively extract video or data of interest from a video signal containing data of interest and other data.
A portion of the first employee's field of view is overlapped by the display 46 of the head mountable unit 14. In
It is noted that the terms “employee” and security guard have been used to distinguish two parties from one another for clarity. Embodiments of the present disclosure can be practiced in which neither the “first employee” or the security guard are employees of the retail store in legal sense, both are employees of the retail store, or one of the “first employee” or the security guard are employees of the retail store. The parties interacting to capture theft can be third party contractors are have some other relationship with respect to the retail store.
Embodiments may also be implemented in cloud computing environments. In this description and the following claims, “cloud computing” may be defined as a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned via virtualization and released with minimal management effort or service provider interaction, and then scaled accordingly. A cloud model can be composed of various characteristics (e.g., on-demand self-service, broad network access, resource pooling, rapid elasticity, measured service, etc.), service models (e.g., Software as a Service (“SaaS”), Platform as a Service (“PaaS”), Infrastructure as a Service (“IaaS”)), and deployment models (e.g., private cloud, community cloud, public cloud, hybrid cloud, etc.).
The above description of illustrated examples of the present disclosure, including what is described in the Abstract, are not intended to be exhaustive or to be limitation to the precise forms disclosed. While specific embodiments of, and examples for, the present disclosure are described herein for illustrative purposes, various equivalent modifications are possible without departing from the broader spirit and scope of the present disclosure. Indeed, it is appreciated that the specific example voltages, currents, frequencies, power range values, times, etc., are provided for explanation purposes and that other values may also be employed in other embodiments and examples in accordance with the teachings of the present disclosure.
Number | Name | Date | Kind |
---|---|---|---|
3863245 | Swinamer et al. | Jan 1975 | A |
6502749 | Snyder | Jan 2003 | B1 |
7248161 | Spoltore | Jul 2007 | B2 |
7262690 | Heaton | Aug 2007 | B2 |
7724131 | Chen | May 2010 | B2 |
8203603 | Harbert | Jun 2012 | B2 |
8489065 | Green et al. | Jul 2013 | B2 |
8493210 | Maziak et al. | Jul 2013 | B2 |
8547401 | Mallinson | Oct 2013 | B2 |
8559030 | Tsongas | Oct 2013 | B2 |
8633869 | Devereaux | Jan 2014 | B2 |
8686851 | Davis | Apr 2014 | B2 |
8803687 | Valiulis et al. | Aug 2014 | B2 |
20070076095 | Tomaszewski | Apr 2007 | A1 |
20070080806 | Lax et al. | Apr 2007 | A1 |
20080272910 | Anderson | Nov 2008 | A1 |
20090224875 | Rabinowitz et al. | Sep 2009 | A1 |
20090265106 | Bearman et al. | Oct 2009 | A1 |
20100185529 | Chesnut | Jul 2010 | A1 |
20110057797 | Parker et al. | Mar 2011 | A1 |
20110149078 | Fan et al. | Jun 2011 | A1 |
20120062380 | Bird et al. | Mar 2012 | A1 |
20120233032 | Calman | Sep 2012 | A1 |
20120282974 | Green et al. | Nov 2012 | A1 |
20130031202 | Mick | Jan 2013 | A1 |
20130035581 | Vesto | Feb 2013 | A1 |
20130050496 | Jeong | Feb 2013 | A1 |
20130136242 | Ross et al. | May 2013 | A1 |
20130142494 | Valiulis et al. | Jun 2013 | A1 |
20130187952 | Berkovich | Jul 2013 | A1 |
20130227681 | Hong | Aug 2013 | A1 |
20130249942 | Green | Sep 2013 | A1 |
20130278631 | Border | Oct 2013 | A1 |
20140096220 | Da Cruz Pinto | Apr 2014 | A1 |
20140100997 | Mayerle | Apr 2014 | A1 |
20140118140 | Amis | May 2014 | A1 |
20140167917 | Wallace et al. | Jun 2014 | A2 |
20140211017 | Argue et al. | Jul 2014 | A1 |
Number | Date | Country |
---|---|---|
2645667 | Oct 2013 | EP |
WO2004066590 | Aug 2004 | WO |
WO2012162799 | Dec 2012 | WO |
WO2014063724 | May 2014 | WO |
Number | Date | Country | |
---|---|---|---|
20140210621 A1 | Jul 2014 | US |