The present invention generally relates to merchandising solutions, and more specifically relates to an intelligent marketing and advertising platform which employs digital “smart” screens on retail product containers thereby providing an innovative advertising solution.
Although more and more people are purchasing items online, brick and mortar retail sales are still substantial and are, in fact, growing in certain market segments such as grocery, drug and convenience stores. Over 75 percent of shopping in brick and mortar stores is impulse purchasing. Research shows that impulse decisions are highly influenced by in-store messaging, and that most of that in-store impulse shopping occurs at retail product containers, such as coolers and freezers.
If impulse shopping behavior can be disrupted by what is displayed on a retail product container, such as cooler or freezer doors, this would present an enormous value creation opportunity for both consumer packaged goods brands and retailers. Arguably, coolers are the most overlooked, least promoted, least technology sophisticated and one of the most operationally challenged sections of a retail store.
Conventional retail product containers such as coolers, freezers, vending machines, etc. have a glass panel or door through which a customer views products stored in the retail product container and which are available for purchase.
Problems with this approach include the fact that oftentimes products stored therein are not neatly presented, products are blocked by other products such that a customer cannot easily see what products are available for purchase, etc. Still further, store employees are required to maintain the appearance of the products stored therein, keep track of stock, install and swap out promotional tags and other displays (such as displays relating to pricing and promotions) in order to keep those items current.
It is a well-understood problem that coolers and refrigerators in grocery stores are one area that is difficult for brands to conduct effective “on-the-spot” advertisements and promotions. Typically, brands rely on other advertising medias (e.g. print, billboards, online and TV advertisements, etc.) to promote products that are retailed on the shelves of coolers/freezers and/or refrigerators at retail and convenience stores. It is also widely acknowledged that current advertising efforts are not effective and impactful.
Recognizing these inefficiencies several display manufacturers explored opportunities to convert the “see-thru” glass windows on the cooler doors into transparent displays that can enable running ads while the shoppers still see the products and continue their usual shopping experience.
Transparent display technology that was being developed started to be targeted for these applications and were eventually integrated into cooler doors by several manufacturers. These systems allowed the conversion of these “passive” doors into an advertising platform while minimally disrupting a typical consumer shopping experience. The consumer could still see through and select the items they plan to pick; while the store owners or media platform owners can run promotional videos and advertisements on transparent display media. However, over the last few years, it was evident that the market penetration of such systems has been very slow. There are a number of factors that have worked to limit adoption, such as:
Brands always seek media platforms that enable them to reach the critical mass of their target customer. For digital signage media platforms, this translates to a need for a very significant installed base.
An object of an embodiment of the present invention is to provide an intelligent marketing and advertising platform.
Briefly, an embodiment of the present invention provides an intelligent marketing and advertising platform which comprises a plurality of retail product containers (such as coolers, freezers and/or vending machines). Each retail product container has internal storage volume, and each comprises at least one non-transparent display which prohibits viewing of the internal storage volume of the retail product container. Each retail product container further comprises customer-detecting hardware and inventory-taking hardware.
The platform also comprises a controller/data collector which is in communication with the retail product containers. The controller/data collector is configured to control the at least one non-transparent display of each retail product container such that each non-transparent display provides a planogram relating to retail products physically contained in the internal storage volume of the retail product container based on inventory taken by the controller/data collector using the inventory-taking hardware of the retail product container.
The controller/data collector is preferably configured to control the at least one non-transparent display of each retail product container to display current pricing information regarding products physically contained in the internal storage volume of the retail container, wherein the pricing is dictated by the controller/data collector.
The controller/data collector is preferably configured to detect customers using the customer-detecting hardware, and is configured to display promotions on the at least one non-transparent display of each retail product container based on what is detected by the controller/data collector using the customer-detecting hardware.
The intelligent marketing and advertising platform provides an innovative merchandising solution for retailers by effectively transforming the glass surface of retail product containers (such as cooler doors) into a non-transparent display of planograms. The merchandising solution disclosed herein provides for digital planograms and pricing management, real time promotional updates and sales data, etc. The advertising method innovation for in-store retail signage disclosed herein is effectively accomplished by converting/transforming the simple glass surface of a retail product container (such as cooler/freezer doors) into digital “smart” screens that provide for innovative advertising solutions.
The organization and manner of the structure and operation of the invention, together with further objects and advantages thereof, may best be understood by reference to the following description taken in connection with the accompanying drawings wherein like reference numerals identify like elements in which:
While this invention may be susceptible to embodiment in different forms, there is shown in the drawings and will be described herein in detail, a specific embodiment with the understanding that the present disclosure is to be considered an exemplification of the principles of the invention, and is not intended to limit the invention to that as illustrated.
The retail product containers 14 need not be the same, but preferably each is a device such as a cooler, freezer and/or vending machine. Regardless, preferably each has internal storage volume in which products for purchase can be stored.
A shown in
Preferably, each retail product container 14 further comprises customer-detecting hardware 18, such as one or more proximity sensors (such as heat maps), cameras, facial sensors or scanners, and eye-sensors (i.e., iris-tracking sensors).
Assuming cameras are employed, preferably cameras are mounted on doors of the retail product containers 14. Preferably, the cameras have a depth of field of view of twenty feet or more, and have a range of field of view of 170 degrees with preferably 150 degree of facial recognition ability. Preferably, software is employed in association with the cameras to monitor shopper interactions, serve up relevant advertisement content on the displays 16, and track advertisement engagement in-store. The controller/data collector 12 may be configured to independently control each camera, as well as collectively control a plurality of displays 16, to serve up single-screen and/or multiple screen content and interactions. Preferably, the controller/data collector 12 and displays 16 are configured to allow for both banner advertisements and full-screen advertisements to be displayed at the same time on a single display 16.
Preferably, each retail product container 14 further comprises inventory-taking hardware 20 such as additional cameras and/or sensors that are disposed inside the retail product container 14 and face the products.
As shown in
The controller/data collector 12 may comprise a server having one or more processors, memory storage, a user interface, etc. and is configured to not only dictate what is displayed on the displays 16 of the retail product containers 14 and receive information and data from the retail product containers 14, but is also preferably configured to perform analytics based on information and data which has been received.
Preferably, the controller/data collector 12 is configured to control the display 16 of each retail product container 14 such that each display 16 provides planograms relating to retail products physically contained in the internal storage volume of the retail product container 14 (but not viewable through the display 16) based on inventory taken by the controller/data collector 12 using the inventory-taking hardware 20 of the retail product container 14. As such, there is no downside to the fact that, in reality, products stored in the retail product container 14 may not be neatly stored therein, or that products stored therein may be blocked from view by other products stored therein. The planograms which are displayed on the displays 16 of the retail product containers 14 effectively optimize what is presented to the customer.
Preferably, the controller/data collector 12 is configured to control the display 16 of each retail product container 14 to display current pricing information regarding products physically contained in the internal storage volume of the retail product container 14, wherein the pricing is dictated by the controller/data collector 12.
Preferably, the controller/data collector 12 is configured to detect customers using the customer-detecting hardware 18, and is configured to display promotions on the at least display 16 based on what is detected by the controller/data collector 12 using the customer-detecting hardware 18.
As discussed above, the controller/data collector 12 effectively takes an inventory of each retail product container 14 (i.e., using the inventory-taking hardware along with image recognition software). Preferably, the controller/data collector 12 is configured to transmit data to a distributor and/or store regarding overall stock relating to the inventory of the plurality of retail product containers 14. More specifically, the overall system can be configured to issue restock notifications, provide merchandising updates, pricing management, real time sales data and behavioral consumer analytics. As such, the system provides advertising, flawless merchandising, real time pricing and promotions, an inventory and direct store delivery solution and the power of analytics.
It should be noted that
Preferably, a plurality of retail product containers 14 and associated displays 16 are arranged side-by-side down an aisle of a retail store, such as a grocery store, and the controller/data collector 12 is configured to operate the displays 16 depending on what is detected by both the customer-detecting hardware 18 and inventory-taking hardware 20.
For example, preferably the controller/data collector 12 is configured to operate the displays 16 such that the displays 16 display full screen advertisements if no motion is detected (or motion is no longer detected) by the customer-detecting hardware 18. In other words, as a customer is approaching an aisle, the customer sees full screen advertisements on the displays 16.
Preferably, the controller/data collector 12 is configured such that if motion is detected within twenty feet of a retail product container 14 (i.e., by customer-detecting hardware 18) mounted on the door of that retail product container 14), the controller/data collector 12 operates that particular display 16 to stop displaying a full screen advertisement, and instead display a planogram of products which are located inside the retail product container 14. Preferably, banner advertisements and pricing tags are also loaded and displayed at the same time.
Preferably, the controller/data collector 12 is configured to manage merchandising advertising objects and layers. This can be facilitated and managed remotely, or locally at an individual retail location such as by using a content management application.
Preferably, the controller/data collector 12 is configured such that as a shopper approaches closer to a given retail product container 14, the display 16 associated with the retail product container 14 changes again. For example, the display 16 can change such that it then displays labels (e.g., organic, non-gmo, etc.) and tags (e.g., sale with local card, locally made, trending, etc.). These promotional labels and tags can be activated via the controller/data collector 12 and can be either selected from a pre-determined set of static animated icons, or they can be customized by the retailer and/or brands based on their business and marketing needs. The controller/data collector 12 can be configured to incorporate customer rating scores (e.g., 1-5 stars, trending, favorites, etc.) using one or more consumer review sources.
Preferably, the controller/data collector 12 is configured such that as a shopper stands or lingers in front of a given retail product container 14, the display 16 associated with the retail product container 14 changes yet again. At this point, preferably the controller/data collector 12 has been able to use the customer-detecting hardware 18 to effectively learn more about that particular customer, such as gender, age, mood, etc. The controller/data collector 12 is configured to take what has been detected about the customer to determine which advertisement and other information to present to that particular customer on the display 16 associated with the retail product container 14 in front of which the customer is standing. By tracking shopper data in parallel with which advertising content is being served on all displays 16 within the viewing range of the shopper, the retailer and the brands are better served, providing new influence opportunities at the moment of purchasing decision, optimizing marketing spend and generating new revenue streams.
Preferably, the controller/data collector 12 and displays 16 are configured such that, while a shopper is standing in front of a given display 16 (as detected by a camera mounted on a door of that particular retail product container 14) relevant conditional ads are displayed on that display 16 such as real-time data relating to news events, weather, sporting events, etc. (i.e., in addition to information and advertisements relating to the contents of the retail product container 14, such as banner advertisements, horizontal advertisements, full screen advertisements, tags, labels, spot advertisements, etc.).
As discussed above, preferably the controller/data collector 12 is configured to keep track of the inventory of a given retail product container 14 by using inventory-taking hardware 20, such as one or more cameras and/or sensors on the inside of the retail product container 14. Preferably, when the controller/data collector 12 has determined that a given product is out of stock, the controller/data collector 12 takes this into consideration when controlling what is displayed on the display 16 of that particular retail product container 14. For example, out of stock artwork can be displayed and/or advertisements can be swapped, such that out of stock advertising opportunities can be sold to different brands, wherein certain advertisements are displayed given certain out of stock scenarios.
Additionally, preferably all inputs collected by the IOT devices will be analyzed locally as well as remotely (via cloud) to provide the feedback inputs for the system to push more relevant/targeted content, tailored for the consumer. The analytics are preferably conducted anonymously, images captured by cameras are preferably processed to collect statistics on consumer demographic characteristics: (such as age and gender). This data is preferably subsequently analyzed for additional statistics for the retailors that are valuable for in-store merchandise layout design and smart merchandizing, including the ability to track the shoppers “traffic” areas, known as “heat maps”, areas were customers would concentrate more and spend more time exploring, etc.
Preferably, the interior facing cameras are utilized to ensure that products on the shelves are constantly monitored for out-of-stock warning. When a product is sold out, the signal preferably alerts the system to update the layout on the display as well as the store operator.
Preferably, by default, the platform is preferably programmed to run full screen advertisements. Customer-facing cameras are programmed to have motion detection capability, analyze an approaching customer, and switch the system to display planograms of products placed on the shelves. Further analytics based on the images captured by these cameras preferably triggers tailored promotional ads.
Preferably, at least a portion of the display is configured to have interactive touch capability, enabling the shoppers to explore further detailed information about various products, while minimizing the interference with other shoppers who would be able to see all the products displayed on the planogram and be able to proceed and pick up the product of their choice.
The system may be configured to track eye movements of the consumer (such as by using powerful software tools coupled with specific hardware that uses infrared sources to track eye movements) and correlate gaze spots with advertisements. This amounts to a very powerful demonstration tool to measure the success of the advertisement and its ability to impact the decision of the shopping consumer.
As discussed above, preferably each retail storage container 14 has a door and a display 16 is mounted on that door along with customer-detecting hardware 18, such as one or more proximity sensors (such as heat maps), cameras, facial sensors or scanners, and eye-sensors (i.e., iris-tracking sensors).
As shown in
The intelligent marketing and advertising platform provides an innovative merchandising solution for retailers by effectively transforming the glass surface of retail product containers (such as cooler doors) into a non-transparent display of planograms. The merchandising solution disclosed herein provides for digital planograms and pricing management, real time promotional updates and sales data, etc. The advertising method innovation for in-store retail signage disclosed herein is effectively accomplished by converting/transforming the simple glass surface of a retail product container (such as cooler/freezer doors) into digital “smart” screens that provide for innovative advertising solutions.
As a business model, the cooler doors of a retail store can be retrofit with “smart” cooler doors having displays, cameras and sensors that collectively measure, react, learn and communicate in order to optimize the metrics.
The intelligent marketing and advertising platform disclosed herein effectively prides for at least the following: the convergence of a brick and mortar retail establishment with e-commerce; in-store promotions as an electronic marketplace; real time and algorithmic-driven pricing and promotions; self-learning/machine-learning artificial intelligence algorithm-driven advertising which is personalized to a given consumer; behavior response and external data (i.e. weather, events, competition, etc.); and smart-sensor and digital merchandising for planogram compliance, automatic-restock, in-store audits, etc.
The advertising method for in-store retail signage disclosed herein effectively employs two primary technologies—large scale brand/product advertising rotations and planogram screens (preferably activated when a customer gets within three feet) with non-obtrusive ad banners or hot-spots.
As discussed above, each retail product container 14 preferably comprises inventory-taking hardware 20, such as cameras, that are disposed inside the retail product container 14 and face the products. Specifically, the cameras may be provided on the back 33 of the door 30 which the controller/data collector 12 (shown in
As such, an embodiment of the present invention effectively provides an automated ability to detect the inventory of products placed inside each retail product container 14 and update the planograms that are displaying the products on the display 16 (such as an LCD screen). If any item has gone “out of stock,” preferably the system is configured to detect and update the information on the display 16 to guide the consumers and store operators.
Preferably, the automatic detection system utilizes cameras, pointed to take pictures of the interior shelves of the retail product container 14 in conjunction with software that utilizes smart algorithms to conduct image analysis and subsequent analytics to extract the required information.
While one embodiment could provide that numerous cameras are placed in arrays to image and assemble the entire shelf area by adding individual pictures together using software, an array of cameras, however, poses challenges as all of them must be plugged into the operating PC which typically has limited number of inputs. Therefore, a more preferred embodiment is shown in
While an embodiment of the present invention could provide that wide field of view cameras (e.g. 180° FOV (“fisheye”)) are utilized because they provide for large area coverage, typically the image quality from such cameras is inadequate as objects get severely distorted on the edges, limiting the software's capability to “recognize” the images with regard to the image processing step. As such, a more preferred embodiment provides that miniature cameras with 120° Field of View (Diagonal) are mounted on 45° angled housings along the edge on the handle side of the door 30. As shown in
Preferably, the system is configured such that during door operation (i.e., when the door 30 is opened), the cameras 40 get triggered and take pictures at various intervals when the door 30 is open. Preferably, the system is configured to process the images and effectively reconstruct the entire shelf. The system may be configured such that the cameras get triggered in any number of ways. For example, the system could be configured such that the cameras get triggered by:
Regardless of exactly what triggers the cameras, preferably the system is configured such that images captured by the cameras are effectively “stitched” together to reconstruct the entire shelf space. Preferably, the system is configured to utilize image processing compare the images and conclude whether any of the products on shelves are missing. If any of the particular products are missing, preferably the system is configured to send a signal to content management software to update the planogram appropriately (i.e., on the display 16), preferably displaying that product as being “out of stock”. Additionally, preferably the system is configured to notify a store inventory management team, such that re-stocking of that particular item on that particular shelf can take place quickly.
In preceding embodiment involving the stitching together of multiple captured images, the disclosed system may capture a plurality of images using a first, second, and third camera as the door rotates. For example, the first photo may be taken by the first camera, the second photo may be taken by the second camera, and the third photo taken by the third camera. A controller on or near the retail product container may combine the pertinent portion of the three photos into a single, composite image. Alternatively, instead of a single photo from each of three cameras, the first camera may take multiple photos and combine them with one or more photos taken by the second camera and third camera. In other words, the controller may instruct the plurality of cameras to capture a number of images as the door rotates, thus capturing images from different positions/angles with respect to the internal contents of the retail product container. In one example, the controller may change the frequency of capture of photos—i.e., the controller may increase the number of images captured— if the controller receives measurements that the door is swinging at a higher rotational speed. In another example, the controller may change the frequency and/or other parameters of the cameras based on other situational or environmental factors. For example, in low light situations, the operating parameters of the one or more cameras may be adjusted to accommodate. In other examples, moisture or condensation on a camera lens may be detected and result in the controller modifying the operational parameters of affected cameras.
Furthermore, in other examples, the plurality of cameras installed on the swingable door of a retail product container may be one or more motion video cameras. As such, these video cameras may capture multiple video frames from a start position to a predefined end position of the rotation of the door. The end position may be a set predefined position, a predefined range, or a predefined relative rotation angle from the start position (e.g., once the door has rotated through 60 degrees from a start position.)
In one example, the plurality of images, which are captured by the system using a first, second, and third camera as the door rotates, may be sent to an electronic controller at or near the retail product container. The controller may execute computer-executable instructions to digitally combine the plurality of images into a single panoramic, composite image. The composite image may be large in size, thus consume a sizable amount of computer memory/storage at the controller. In some embodiments, the composite image may undergo post-processing to reduce the storage consumed by the image. In one example, artificial intelligence using a neural network, such as a convolutional neural network, may be used to identify a boundary edge of the pertinent portions of the contents of the retail product container. As a result, those areas of the image outside of the boundary edge may be affirmatively discarded by the post-processing module executing on the controller. The out-of-bound areas may be those portions of the image that are outside of the edge frame of the retail product container. At least one benefit of the preceding example of post-processing is a saving in memory storage capacity at the edge location. Moreover, a smaller image size also results in less network bandwidth consumption to transmit the image from the controller to one or more remote servers in the cloud.
In one example, a controller may be coupled with the one or more cameras installed in or on the door to automatically adjust parameters. The controller may adjust parameters of the camera, such as, but not limited to, exposure, focus position, sensor gain/ISO speed, aperture size, etc. Values for exposure may be in units of time, and other parameters may be in others units in different operating scenarios. Moreover, a person of skill in the art after review of the entirety disclosed herein will appreciate that one or more of the parameters may be interrelated or dependent. For example, an exposure of 1/25 sec at f/11, ISO 100 is equivalent to an exposure of 1/400 sec at f/2.8, ISO 100. In other words, because the shutter speed has been reduced by four stops, this means less light is being captured by the image sensor in the camera. As a result, the aperture is increased in size by four stops to allow more light into the camera assembly. A person having ordinary skill in the art would appreciate that there are benefits and disadvantages to adjusting the parameters in one way versus another. For example, to maximize exposure, camera settings/parameters might be set to a large aperture, 6400 ISO, and a slow shutter speed. In contrast, to minimize exposure, camera settings/parameters would be set to a small aperture, 100 ISO, and a fast shutter speed. Of course, the sharpness of the captured image might be effected by depth of field, aperture, and shutter speed parameters. With many examples disclosed herein involving a capture at a relatively close distance of an image of a stationary retail product stored in a retail product container, the ability to capture an image without introducing blurriness or planar warp is a consideration. Moreover, in some examples, the image sensor of the camera may be in motion at the time of image capture, thus introducing a relative motion between the subject being captured and the camera.
Regarding blurriness, the controller may execute computer-executable instructions (e.g., firmware, software, or application specific integrated circuits) to perform post-processing on the captured images to mitigate blurriness. In one example, a blurriness mitigation module in the controller may identify portions of captured images that are duplicative across the plurality of images captured. And, then select the appropriate portions from those duplicate images where the blurriness (and other undesirable image characteristics) are reduced. As explained herein, during the process of stitching together the multiple, captured images, the portions identified by the blurriness mitigation module may be incorporated into the final, composite image.
Referring to
In one illustrative embodiment, the camera to shelf distance may be positioned at four inches, and the retail products on the shelves may be generally centered relative to the door frame. Meanwhile, the camera may have an about 120° diagonal field of view (FOV) and oriented so that the view is wider in the direction of the camera's skinny dimension. That is, the camera is positioned near the handle side of the door with a 40° tilt towards the hinge side without having its view blocked by the display screen inside the door. While one embodiment identifies a 4″ positioning, the disclosure is not so limited. The camera-to-shelf distance may be set to a value other than four inches, in some examples.
Referring to
In those examples where the camera is built with the sensor oriented in the perpendicular direction, two cameras may be enough. Two cameras, positioned 28″ or more from the hinge, rotated about 50° towards the hinge edge of the door, can view the entire scene of the contents of a retail product container if the camera locations are 21″ from the top and bottom edges and 30″ from each other. Of course, a person of skill in the art after review of the entirety disclosed herein will appreciate that the disclosure is not limited to the 21″, 30″, and 50-degree specifications indicated in the preceding example. Rather, the specifications are understood to be adjusted based on the fundamental teachings of this disclosure.
Referring to
Referring to
While a specific embodiment of the invention has been shown and described, it is envisioned that those skilled in the art may devise various modifications without departing from the spirit and scope of the present invention.
This application is a continuation of and claims priority to co-pending U.S. application Ser. No. 16/763,448 (filed May 12, 2020), which is a 35 USC 371 National Phase of International Patent Application No. PCT/US2019/066869 (filed Dec. 17, 2019), which is a continuation-in-part claiming the priority filing date of U.S. patent application Ser. No. 16/222,643 (filed Dec. 17, 2018 and issued Sep. 8, 2020 as U.S. Pat. No. 10,769,666), which is a continuation-in-part of U.S. patent application Ser. No. 15/674,198 (filed Aug. 10, 2017 and issued Jun. 2, 2020 as U.S. Pat. No. 10,672,032) and a continuation-in-part of International Patent Application No. PCT/US18/46103 filed Aug. 9, 2018, and all of the aforementioned are hereby incorporated herein by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
2002322 | Kraemer | May 1935 | A |
4371870 | Biferno | Feb 1983 | A |
4671582 | Stromquist et al. | Jun 1987 | A |
4893902 | Baughman et al. | Jan 1990 | A |
4998382 | Kostos et al. | Mar 1991 | A |
5024023 | Kostos et al. | Jun 1991 | A |
5111618 | Kaspar et al. | May 1992 | A |
5116274 | Artwohl et al. | May 1992 | A |
5255473 | Kaspar et al. | Oct 1993 | A |
5270843 | Wang | Dec 1993 | A |
5645330 | Artwohl et al. | Jul 1997 | A |
6005652 | Matsuhira | Dec 1999 | A |
6148563 | Roche et al. | Nov 2000 | A |
6377228 | Jenkin et al. | Apr 2002 | B1 |
6427772 | Oden et al. | Aug 2002 | B1 |
6475087 | Cole | Nov 2002 | B1 |
6476885 | Murray et al. | Nov 2002 | B1 |
6606832 | Richardson et al. | Aug 2003 | B2 |
6606833 | Richardson et al. | Aug 2003 | B2 |
6874903 | Yang et al. | Apr 2005 | B2 |
7072096 | Holman et al. | Jul 2006 | B2 |
7121675 | Ter-Hovhannisian | Oct 2006 | B2 |
7259359 | Davey et al. | Aug 2007 | B2 |
7290707 | Sawasaki | Nov 2007 | B2 |
7319407 | Jang et al. | Jan 2008 | B2 |
7377125 | Seiden et al. | May 2008 | B2 |
7413233 | Jung | Aug 2008 | B1 |
7513637 | Kelly et al. | Apr 2009 | B2 |
7593867 | Deakin et al. | Sep 2009 | B2 |
7824056 | Madireddi et al. | Nov 2010 | B2 |
7870686 | Hines | Jan 2011 | B2 |
7934384 | Tuskiewicz et al. | May 2011 | B2 |
7978184 | Morrison | Jul 2011 | B2 |
8009864 | Linaker et al. | Aug 2011 | B2 |
8189855 | Opalach et al. | May 2012 | B2 |
8219438 | Moon et al. | Jul 2012 | B1 |
8335254 | Fernandez et al. | Dec 2012 | B1 |
8468059 | Enqvist | Jun 2013 | B2 |
8630924 | Groenevelt et al. | Jan 2014 | B2 |
8672427 | Hammonds | Mar 2014 | B2 |
8683745 | Artwohl et al. | Apr 2014 | B2 |
8955261 | Kobe et al. | Feb 2015 | B2 |
8972291 | Rimnac et al. | Mar 2015 | B2 |
9046294 | Lee et al. | Jun 2015 | B2 |
9052536 | Artwohl et al. | Jun 2015 | B2 |
9098825 | Bashkin | Aug 2015 | B2 |
9144328 | Seeley | Sep 2015 | B2 |
9155405 | Artwohl et al. | Oct 2015 | B2 |
9230386 | Roh et al. | Jan 2016 | B2 |
9264151 | Emigh et al. | Feb 2016 | B1 |
9280757 | Parpia et al. | Mar 2016 | B2 |
9336508 | Soon-Shiong | May 2016 | B2 |
9349238 | Tkachenko et al. | May 2016 | B2 |
9361628 | Stark | Jun 2016 | B2 |
9504338 | Artwohl et al. | Nov 2016 | B2 |
9514661 | Riegel | Dec 2016 | B2 |
9524419 | Chang | Dec 2016 | B2 |
9560777 | Krietzman et al. | Jan 2017 | B2 |
9575558 | Almen et al. | Feb 2017 | B2 |
9689603 | Roh et al. | Jun 2017 | B2 |
9710992 | Borra | Jul 2017 | B2 |
9860491 | Park et al. | Jan 2018 | B2 |
9881528 | Dunn | Jan 2018 | B2 |
9911377 | Howard et al. | Mar 2018 | B2 |
9915471 | Kim et al. | Mar 2018 | B2 |
9959568 | Klearman | May 2018 | B2 |
9967109 | Nishimura et al. | May 2018 | B2 |
9972284 | Lee et al. | May 2018 | B2 |
10062257 | Chau | Aug 2018 | B2 |
10085571 | Schiffman et al. | Oct 2018 | B2 |
10089520 | Motukuri et al. | Oct 2018 | B2 |
10089952 | Nam et al. | Oct 2018 | B2 |
10122915 | Maayan et al. | Nov 2018 | B2 |
10126849 | Lee et al. | Nov 2018 | B2 |
10169677 | Ren et al. | Jan 2019 | B1 |
10258170 | Dunn et al. | Apr 2019 | B2 |
10282695 | McNamara et al. | May 2019 | B1 |
10304417 | Park et al. | May 2019 | B2 |
10360571 | Garel et al. | Jul 2019 | B2 |
10419647 | Park et al. | Sep 2019 | B2 |
10477162 | Jain et al. | Nov 2019 | B2 |
10514722 | Artwohl et al. | Dec 2019 | B1 |
10575660 | Artwohl et al. | Mar 2020 | B2 |
10580333 | Artwohl et al. | Mar 2020 | B2 |
10674569 | Luckhardt et al. | Jun 2020 | B2 |
RE48115 | Artwohl et al. | Jul 2020 | E |
10838453 | Artwohl et al. | Nov 2020 | B2 |
10937344 | Artwohl et al. | Mar 2021 | B2 |
11047614 | Kim et al. | Jun 2021 | B2 |
11107023 | Glasgow et al. | Aug 2021 | B2 |
11141004 | Rolek | Oct 2021 | B1 |
RE49103 | Artwohl et al. | Jun 2022 | E |
11419434 | Artwohl et al. | Aug 2022 | B1 |
11435777 | Artwohl et al. | Sep 2022 | B2 |
11450247 | Artwohl et al. | Sep 2022 | B2 |
20010010516 | Roh et al. | Aug 2001 | A1 |
20010052741 | Yun | Dec 2001 | A1 |
20020007486 | Yun | Jan 2002 | A1 |
20030038912 | Broer et al. | Feb 2003 | A1 |
20030117790 | Lee et al. | Jun 2003 | A1 |
20030154141 | Capazario et al. | Aug 2003 | A1 |
20030205059 | Roche et al. | Nov 2003 | A1 |
20030207090 | Arora | Nov 2003 | A1 |
20030214619 | Masuda et al. | Nov 2003 | A1 |
20040073334 | Terranova | Apr 2004 | A1 |
20040093379 | Roh et al. | May 2004 | A1 |
20040144328 | Bonner et al. | Jul 2004 | A1 |
20040160388 | O'Keeffe | Aug 2004 | A1 |
20040194388 | Roche et al. | Oct 2004 | A1 |
20050068629 | Fernando et al. | Mar 2005 | A1 |
20050172654 | Rohrer et al. | Aug 2005 | A1 |
20050202178 | Roche et al. | Sep 2005 | A1 |
20050265019 | Sommers et al. | Dec 2005 | A1 |
20060103269 | Artwohl et al. | May 2006 | A1 |
20060127586 | Roche et al. | Jun 2006 | A1 |
20060145576 | Lee et al. | Jul 2006 | A1 |
20060158579 | Hasegawa | Jul 2006 | A1 |
20060174641 | Liu et al. | Aug 2006 | A1 |
20060192767 | Murakami | Aug 2006 | A1 |
20070003700 | Roche et al. | Jan 2007 | A1 |
20070016478 | Hill | Jan 2007 | A1 |
20070024822 | Cortenraad et al. | Feb 2007 | A1 |
20070058114 | Niiyama et al. | Mar 2007 | A1 |
20070076431 | Atarashi et al. | Apr 2007 | A1 |
20070138268 | Tuchman | Jun 2007 | A1 |
20070151274 | Roche et al. | Jul 2007 | A1 |
20070162182 | Marti et al. | Jul 2007 | A1 |
20070171647 | Artwohl et al. | Jul 2007 | A1 |
20070193280 | Tuskiewicz et al. | Aug 2007 | A1 |
20070195535 | Artwohl et al. | Aug 2007 | A1 |
20070216657 | Konicek | Sep 2007 | A1 |
20080004950 | Huang et al. | Jan 2008 | A1 |
20080024047 | Juo et al. | Jan 2008 | A1 |
20080158858 | Madireddi et al. | Jul 2008 | A1 |
20080172261 | Albertson et al. | Jul 2008 | A1 |
20080249857 | Angell et al. | Oct 2008 | A1 |
20090002990 | Becker et al. | Jan 2009 | A1 |
20090036208 | Pennington et al. | Feb 2009 | A1 |
20090052206 | Matsui et al. | Feb 2009 | A1 |
20090121970 | Ozbek | May 2009 | A1 |
20090146945 | Cho | Jun 2009 | A1 |
20090157472 | Burazin et al. | Jun 2009 | A1 |
20090276319 | Lungu et al. | Nov 2009 | A1 |
20090295731 | Kim et al. | Dec 2009 | A1 |
20090306820 | Simmons et al. | Dec 2009 | A1 |
20090313125 | Roh et al. | Dec 2009 | A1 |
20100013925 | Fowler et al. | Jan 2010 | A1 |
20100026912 | Ho | Feb 2010 | A1 |
20100043293 | Nicholson et al. | Feb 2010 | A1 |
20100062152 | Roche et al. | Mar 2010 | A1 |
20100068398 | Roche et al. | Mar 2010 | A1 |
20100083672 | Yoon et al. | Apr 2010 | A1 |
20100119705 | Roche et al. | May 2010 | A1 |
20100138037 | Adelberg et al. | Jun 2010 | A1 |
20100152892 | Gavra et al. | Jun 2010 | A1 |
20100180615 | Linder et al. | Jul 2010 | A1 |
20100214786 | Nichol | Aug 2010 | A1 |
20100268792 | Butler et al. | Oct 2010 | A1 |
20100275477 | Kim | Nov 2010 | A1 |
20100293827 | Suss et al. | Nov 2010 | A1 |
20110022980 | Segal et al. | Jan 2011 | A1 |
20110083460 | Thomas et al. | Apr 2011 | A1 |
20110098849 | Hudis et al. | Apr 2011 | A1 |
20110106624 | Bonner et al. | May 2011 | A1 |
20110116231 | Dunn et al. | May 2011 | A1 |
20110141011 | Lashina et al. | Jun 2011 | A1 |
20110150276 | Eckhoff et al. | Jun 2011 | A1 |
20110161160 | Carlson et al. | Jun 2011 | A1 |
20110173082 | Breitenbach et al. | Jul 2011 | A1 |
20110181792 | Hammonds | Jul 2011 | A1 |
20110191154 | Johnson et al. | Aug 2011 | A1 |
20110208569 | Fano et al. | Aug 2011 | A1 |
20120004769 | Hallenbeck et al. | Jan 2012 | A1 |
20120105424 | Lee et al. | May 2012 | A1 |
20120217254 | Cho et al. | Aug 2012 | A1 |
20120285089 | Artwohl et al. | Nov 2012 | A1 |
20120323620 | Hofman et al. | Dec 2012 | A1 |
20130027561 | Lee et al. | Jan 2013 | A1 |
20130063326 | Riegel | Mar 2013 | A1 |
20130110666 | Aubrey | May 2013 | A1 |
20130135455 | Hjelm et al. | May 2013 | A1 |
20130218721 | Borhan et al. | Aug 2013 | A1 |
20130271696 | Dunn | Oct 2013 | A1 |
20130325638 | Auclair et al. | Dec 2013 | A1 |
20140078407 | Green et al. | Mar 2014 | A1 |
20140126829 | Seeley et al. | May 2014 | A1 |
20140129354 | Soon-Shiong | May 2014 | A1 |
20140129393 | Soon-Shiong | May 2014 | A1 |
20140214547 | Signorelli et al. | Jul 2014 | A1 |
20140232958 | Venturas et al. | Aug 2014 | A1 |
20140316561 | Tkachenko et al. | Oct 2014 | A1 |
20140344118 | Parpia et al. | Nov 2014 | A1 |
20140365272 | Hurewitz | Dec 2014 | A1 |
20150002660 | Lee | Jan 2015 | A1 |
20150073590 | Garcia Manchado et al. | Mar 2015 | A1 |
20150178654 | Glasgow et al. | Jun 2015 | A1 |
20150269593 | Le | Sep 2015 | A1 |
20160027231 | Guzzone et al. | Jan 2016 | A1 |
20160040469 | Lietz et al. | Feb 2016 | A1 |
20160047587 | Sasaki et al. | Feb 2016 | A1 |
20160138860 | Kang | May 2016 | A1 |
20160143459 | Clein | May 2016 | A1 |
20160220039 | Chang et al. | Aug 2016 | A1 |
20160350715 | Minvielle | Dec 2016 | A1 |
20170027339 | Chang et al. | Feb 2017 | A1 |
20170046991 | Riegel | Feb 2017 | A1 |
20170124603 | Olson | May 2017 | A1 |
20170147971 | Morse et al. | May 2017 | A1 |
20170213184 | Lee et al. | Jul 2017 | A1 |
20170228776 | Walden | Aug 2017 | A1 |
20170329078 | Dunn et al. | Nov 2017 | A1 |
20180020847 | Dunn et al. | Jan 2018 | A1 |
20180053226 | Hutton et al. | Feb 2018 | A1 |
20180061283 | Kim et al. | Mar 2018 | A1 |
20180103778 | Olovsson | Apr 2018 | A1 |
20180125404 | Bott et al. | May 2018 | A1 |
20180226056 | Chan | Aug 2018 | A1 |
20180268367 | Bryan et al. | Sep 2018 | A1 |
20180335252 | Oh | Nov 2018 | A1 |
20180365630 | Seals et al. | Dec 2018 | A1 |
20190050900 | Avakian | Feb 2019 | A1 |
20190122263 | Avakian | Apr 2019 | A1 |
20190149725 | Adato et al. | May 2019 | A1 |
20190156273 | Fisher et al. | May 2019 | A1 |
20190213545 | Adato et al. | Jul 2019 | A1 |
20200286032 | Bogolea et al. | Sep 2020 | A1 |
Number | Date | Country |
---|---|---|
102113033 | Mar 2014 | CN |
105513052 | Apr 2016 | CN |
10205405 | Aug 2003 | DE |
102009003127 | Nov 2010 | DE |
102015203150 | Aug 2016 | DE |
2194222 | Jun 2010 | EP |
2843336 | Mar 2015 | EP |
3379184 | Sep 2018 | EP |
3740103 | Nov 2020 | EP |
2506575 | Apr 2014 | GB |
2001294308 | Oct 2001 | JP |
2003125904 | May 2003 | JP |
3450907 | Sep 2003 | JP |
2004183987 | Jul 2004 | JP |
2006209550 | Aug 2006 | JP |
2012161413 | Aug 2012 | JP |
2014206320 | Oct 2014 | JP |
2015169412 | Sep 2015 | JP |
2015222138 | Dec 2015 | JP |
2016533544 | Oct 2016 | JP |
2018151923 | Sep 2018 | JP |
20120044874 | May 2012 | KR |
10-2015-0093289 | Aug 2015 | KR |
20150093289 | Aug 2015 | KR |
20170033083 | Mar 2017 | KR |
I373735 | Oct 2012 | TW |
I443606 | Jul 2014 | TW |
9838547 | Sep 1998 | WO |
03025805 | Mar 2003 | WO |
2009152078 | Dec 2009 | WO |
2014165171 | Oct 2014 | WO |
2014175643 | Oct 2014 | WO |
2015095493 | Jun 2015 | WO |
2016183302 | Nov 2016 | WO |
2017127035 | Jul 2017 | WO |
2019032893 | Feb 2019 | WO |
Entry |
---|
Sep. 7, 2023—(MX) Office Action—App. No. MX/a/2020/001634. |
Nov. 8, 2023—(CN) Office Action—App. No. 201980090632.2. |
Aug. 19, 2023—(CN) Office Action—App. No. 201880064016.5 w/ translation. |
Jan. 25, 2024—(JP) Decision to Grant—App. No. 2020-520519. |
Apr. 22, 2024—(TW) Office Action and Search Report—App. No. 112137447 w/machine translation. |
Feb. 29, 2024—(EP) Office Action—App 19900402.9. |
May 27, 2024—(CN) 2nd Office Action—App. No. 201980090632.2 w/ machine translation. |
Jul. 6, 2023—(JP) Office Action—App. No. 2020-520519. |
Advertising with Transparent LCD Displays; Jul. 28, 2016, URL: http://www.displays2go.com/Article/Advertising-Transparent-LCD-Displays-78>. |
http://emotivev.com/products; at least as early as Jun. 5, 2017. |
https://www.youtube.com/watch?v=2tlM9lereLc; at least as early as Jun. 5, 2017, “Digital Signage Transparent LCD Beverage Cooler Demo”. |
Oct. 24, 2018—(WO) International Search Report and Written Opinion—App PCT/US2018/046103. |
Extended EP Search Report for EP 12782036, Dtaed Jun. 9, 2015, 7 Pages. |
(US) Notice of Allowance on U.S. Appl. No. 14/170,378, Dated Mar. 31, 2015, 10 Pages. |
File History of U.S. Appl. No. 16/547,288, filed Aug. 21, 2019. |
File History of U.S. Appl. No. 15/888,210, Filed Feb. 5, 2018. |
File History of U.S. Appl. No. 14/819,257, filed Aug. 5, 2015, (U.S. Pat. No. 9,504,338). |
File History of U.S. Appl. No. 14/686,958, filed Apr. 15, 2015, (U.S. Pat. No. 9,155,405). |
File History of U.S. Appl. No. 14/170,378, filed Jan. 31, 2004, (U.S. Pat. No. 9,052,536). |
File History of U.S. Appl. No. 13/286,053, filed Oct. 31, 2011, (U.S. Pat. No. 8,683,745). |
File History of U.S. Appl. No. 61/484,616, filed May 10, 2011. |
International Search Report of PCT/US19/66869 Dated Mar. 16, 2020. |
Written Opinion of PCT/US19/66869 Dated Feb. 11, 2020. |
Neff, J., “New System Puts Video Ads on Store Cooler Doors”, (2015), AdAge, http://adage.com/article/cmo-strategy/system-puts-video-ads-store-cooler-doors/301395, pp. 1-3. |
Jul. 1, 2020—(US) Notice of Allowance—U.S. Appl. No. 16/222,643. |
Jul. 27, 2020—(TW) Search Report App 108146270. |
Jul. 29, 2020—(TW) Office Action—App 108146270—Eng Tran. |
Nov. 19, 2020—(EP) Search Report—U.S. Appl. No. 18/843,780. |
Jun. 1, 2021—(TW) Notice of Allowance App 108146270. |
Jun. 3, 2021—(TW) Search Report App 108146270. |
Jul. 28, 2021—(EP) Office Action—EP19900402.9. |
Jul. 22, 2021—(KR) Office Action—10-2020-7006998. |
Jun. 3, 2021—(WO) International Search Report and Written Opinion—App PCT/US2021/013247. |
May 4, 2022—(WO) International Preliminary Examination Report (Chapter II)—App PCT/US2021/013247. |
May 26, 2022—(KR) Notice of Allowance App 10-2020-7006998. |
Jul. 18, 2022—(EP) Search Report—App. No. EP19900402.9. |
Test for report. |
Jun. 17, 2022—(TW) Search Report App 110132225 w/ Eng. Trans. |
Jun. 17, 2022—(TW) Search Report App 110132225 with English Translation. |
Kamoda, Rena. “Monitoring Grocery Inventory in Refrigerators using Appearance and Load Balance Features.” (2013). (Year: 2013). |
Dec. 20, 2022—(US) Non-Final Office Action—U.S. Appl. No. 16/741,581. |
Jan. 9, 2023—(US) Notice of Allownace—U.S. Appl. No. 15/931,971. |
Jan. 9, 2023—(US) Notice of Allowance—U.S. Appl. No. 15/931,971. |
Oct. 28, 2022—(US) Notice of Allowance—U.S. Appl. No. 16/763,448. |
Feb. 22, 2023—(US) Notice of Allowance—U.S. Appl. No. 16/763,448. |
Oct. 12, 2022—(CA) Office Action—App 3,123,879. |
Nov. 18, 2022—(EP) Office Action—App 18843780.0. |
Jan. 13, 2023—(CN) Office Action—App 201880064016.5. |
Jan. 31, 2023—(TW) Search Report—App 110132225. |
Apr. 4, 2023—(US) Notice of Allowance—U.S. Appl. No. 16/741,581. |
Mar. 24, 2023—(US) Notice of Allowance—U.S. Appl. No. 17/082,270. |
May 1, 2023—(US) Notice of Allowance—U.S. Appl. No. 16/763,448. |
May 2, 2023—(US) Notice of Allowance—U.S. Appl. No. 15/931,971. |
Jul. 5, 2023—(CN) First Office Action—App 202180011384.5. |
Jul. 25, 2023—(CA) Office Action—App. No. 3,123,879. |
May 18, 2024—(MX) 1st Office Action—App. No. MX/a/2021/007315 w/ translation. |
Jun. 25, 2024—(TW) Notice of Allowance—App. No. 112137447 w/translation. |
Number | Date | Country | |
---|---|---|---|
20240005271 A1 | Jan 2024 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16763448 | US | |
Child | 18368674 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16222643 | Dec 2018 | US |
Child | 16763448 | US | |
Parent | 15674198 | Aug 2017 | US |
Child | 16222643 | US |