Billboards and other large displays are an important component of most advertising portfolios. As with other real estate, location is key. Most displays are configured to catch the eyes of viewers, which can enhance the brands shown on the displays. In crowded display areas, it is important to make a display distinctive to optimize the impact of the display.
In accordance with certain aspects of the present disclosure, a display device includes a base structure, a plurality of modules coupled to the base structure, where each of the modules include a plurality of actuator assemblies. Each of the actuator assemblies is individually controllable to move the actuator assemblies between a retracted state and a plurality of extended states. A controller is coupled to each of the modules and is programmed to control the actuator assemblies to move the actuator assemblies between the retracted state and the plurality of extended states.
In accordance with further aspects, a system for displaying content includes a display device that has a base structure, a plurality of modules coupled to the base structure, with each of the modules including a plurality of actuator assemblies. Each of the actuator assemblies are individually controllable to move the actuator assemblies between a retracted state and a plurality of extended states. At least one light module is coupled to each of the modules, and a controller is coupled to each of the modules. The controller is programmed to control the actuator assemblies to move the actuator assemblies between the retracted state and the plurality of extended states. A computing device is configured to generate content for the display device and includes a memory and a processing unit encoding instructions that, when executed by the processing unit, cause the processing unit to control the actuator assemblies and the light modules.
In accordance with still further aspects of the disclosure, a method for controlling a display includes sending a first signal for controlling a plurality of actuator assemblies, with each of the actuator assemblies being individually controllable to move the actuator assemblies between a retracted state and a plurality of extended states. A second signal for controlling a plurality of light modules is also sent, with at least one of the light modules being coupled to each of the actuator assemblies. The first and second signals are synchronized to generate a desired effect on the display device.
The examples described herein are related to display devices used for advertising.
In some examples, the display devices incorporate lighting and movement. The lighting and movement are configured to catch a viewer's attention. This can enhance the impact of the brand shown on the display device.
Referring now to
The display device 100 includes a main display 102 and a side display 108. The main display 102 includes a stationary top channel logo area 104 (e.g., the stylized “Coca-Cola”) and a dynamic area 106.
The main display 102 and the side display 108 can include lighting to enhance the impact of the display device 100. For example, as described further below, the main display 102 and the side display 108 can include a plurality of elements that are lit. In addition, the dynamic area 106 includes a plurality of actuator assemblies that move.
For example, as shown in
Each of the modules 222 includes a plurality of actuator assemblies 232 positioned therein. The use of individual modules 222 allows the display device 100 to be installed in an efficient manner, since the modules 222 can be moved and manipulated more easily than the entire display device 100. In this example, the modules 222 extend from the base structure 210 a distance 218 of approximately four feet, although modules of different sizes can be used.
The example depicted shows that each of the modules 222 includes a matrix of five rows of five actuator assemblies 232, for a total of 25 actuator assemblies 232 in each module 222. This configuration allows the actuator assemblies 232 in each of the modules 222 to function as a unit, thereby addressing changes in the environment, including stresses caused by the elements, such as wind, temperature, etc. Because each of the modules 222 functions as a unit, such stresses are accommodated across the actuator assemblies 232.
In alternative designs, the modules 222 can include more or fewer actuator assemblies 232. In the depicted embodiment, some of the actuator assemblies 232 are nonmovable (i.e., stationary), in that the actuator assemblies 232 remain in place and do not move like other actuator assemblies, as described further below. For example, in some embodiments the actuator assemblies 232 that are nonmovable may be placed in the area surrounding the top channel logo area 104. In some embodiments, the top two rows of the modules 222 may be stationary or otherwise have actuator assemblies that are nonmovable.
Referring now to
In this example, the main display 102 has a 12.5 millimeter LED pitch size (see
For example,
Referring now to
The moving cube 314 is moved by a linear actuator assembly 318 in the directions 330, 332 in a plurality of extended positions. As depicted, the actuator assembly 232a is fully extended in the direction 330, the actuator assembly 232b is partially extended in the direction 330, and the actuator assembly 232c is fully retracted. In this example, the moving cube 314 moves approximately 20 inches when in the fully extended position, as depicted by the actuator assembly 232a.
For example, in the fully retracted state shown in
In this example, the linear actuator assembly 318 includes a driver with a servo motor. The servo motor is electrically controlled and moves the moving cube 314 in the directions 330, 332 to any of a plurality of extended positions. The movement can be precisely controlled, so that the position of the moving cube 314 is known. For example, in one embodiment, control is as precise as 0.0079 inches, with a position range of 1 to 29,000.
In some examples, the linear actuator assembly 318 is a F12-BC made by W-Robit of Taiwan. Such a linear actuator assembly 318 can drive up to 44 pounds, with a maximum drive speed of 40 inches per second. In another example, a PAC-UGT040D actuator made by PBC Linear of Roscoe, Ill., is used. The motor of the linear actuator assembly 318 is a BCH U04 Motor manufactured by Schneider Electric of Palatine, Ill. The motor includes a LXM23A servo driver system and Modicon M258 logic controller, both also manufactured by Schneider Electric. In still other examples, the motor is a SM23165DT motor made by Moog Animatics, of Santa Clara, Calif.
A Light-Emitting Diode (LED) module 316 is mounted to each of the moving cubes 314. The LED module 316 includes a plurality of LEDs, such as an NSSM032T LED module made by Nichia Corporation of Japan. Such an LED module is a 3-in-1 SMD LED, although other types can be used. In this example, the LED module 316 is 346 mm in height by 346 mm in width. The LED module 316 is configured to provide a plurality of colors, and each LED module 316 on each of the actuator assemblies 232 is individually controllable, as described below.
In example embodiments, the LED modules 316 can be configured to display one or a plurality of colors. For example, the LED modules 316 can be configured to display text, pictures, or other effects. By grouping the LED module 316, a larger effect, such as a larger picture or text, can be created on the main display 102.
In other embodiments, LED modules 316 could be placed on the sides (e.g., mounted on panels 314b and 314d), top (314c) and bottom (314a) of the moving cubes 314 near the end of the moving cubes 314. In such an embodiment, the light emitted from the LED modules 316 placed on the side, top and bottom of the moving cubes 314 could be seen when looking at the display device 100 from various angles. This embodiment, for example, would provide more continuous light when two adjacent moving cubes 314 are positioned at different distances and the display 100 is viewed from various angles.
In yet another alternative, the brightness of the lights in the LED modules 316 is configurable to create different appearances. For example, the lights can be dimmed or otherwise dulled to form depth and other visual effects, particularly around the edges of the display device 100.
A gap 317 (see
In these examples, the construction of the actuator assemblies 232 allows for ease in access and maintenance. Specifically, the way the actuator assemblies 232 are coupled to the modules 222 allows individual actuator assemblies 232 to be removed individually from a rear of the display device 100. For example,
Referring now to
Referring again to
The rod latches 250 are movable via the actuators 254 to move the rod latches 250 to selectively engage the latch hooks 256 into and out of engagement with the corresponding actuator assemblies 232. In the illustrated example, the rod latches 250 are threadably received by the base structure 210, such that rotating the actuators 254 in one direction moves the rod latches 250 up, and rotating them in the other direction moves the rod latches 250 down. In the example shown in
The display device 100 may include various other features to obtain and hold the attention of individuals capable of viewing the display device. These features could be facilitated by an interactive module 245 located on the display device 100 or be operably connected to it. For example, the interactive module 245 may include or be operably connected to sensors such as, microphones, cameras, motion detectors, moisture sensors, light sensors, etc. Additional features may include speakers, lasers, or other devices capable of producing light shows, which could also be used to attract and hold the attention of individuals. Such features could be operated separately or may be integrated with other sensors such that the display device 100 choreographs its display depending on input from the various sensors of the interactive module 245. For example, the interactive module 245 can include speakers and lasers controlled by a computing device that can choreograph the various sensors, speakers, or lasers to make the display device more attractive and entertaining.
In other examples, sensors, such as moisture, wind, temperature, etc. sensors, can be used to detect certain weather patterns. For example, the sensors can be used to detect certain weather conditions in which operation of the actuator assemblies is not advisable (icing conditions or extreme wind conditions, specifically). In such scenarios, the sensors detect the adverse weather condition and stop movement of the actuator assemblies until such time as the detected weather state passes. In some implementations, a delay period is included such that movement of the actuator assemblies is not restarted until some predetermined time passes in which the adverse weather condition is not detected. This prevents the actuator assemblies from repeatedly starting and stopping, for example, during periods of variable wind gusts.
In other examples, the interactive module 245 is programmed to transmit sound (e.g., music, voice, advertisements) so that passers can tune to a particular radio frequency to listen on their radios. In yet other examples, the interactive module 245 can be programmed to communicate wirelessly (e.g., through Bluetooth or via the Internet) with viewers' smartphones. In this example, the viewer can access content, such as a web site, that allows the viewer to upload or otherwise stream content that can be displayed on the display device 100. Such content could be pictures, etc.
In another example, the interactive module 245 could include microphones such that it could pay “live” sound. Such microphones could be directionally focused such that they could focus in on a particular sound source. In this regard the display device could include software such that the display device could be interactive with those viewing the display device's advertisement based on any number of factors such as movement, sound, recognizing elements in its surroundings, etc.
Referring now to
In this example, the computing device 502 is a local or remote computing device, such as a desktop, laptop, or tablet computer. The computing device 502 can use a standard communication protocol, such as DMX, CANOPEN, Ethernet or RS485 interface, to control the display device 100.
The control by the computing device 502 can include programming the movement of each of the actuator assemblies 232. In one example, an application programming interface (API) is provided that assists in the programming of the movement of the actuator assemblies 232.
In one example, the linear actuator assembly 318 is controlled by the computing device 502 according to a percentage of extension for the moving cube 314. For example, the computing device 502 defines a percentage, such as 0 percent, 10 percent, 25 percent, 50 percent, 75 percent, and/or 100 percent for the moving cube 314 at a given point in time. The percentage is translated to instructions transmitted to the appropriate linear actuator 318 to extend or retract the moving cube 314 the desired amount. By defining a changing percentage over time, the movement of the moving cube 314 can be choreographed, as desired.
In addition, the computing device 502 can define colors to be displayed by the LED module carried by the moving cube 314. The colors of the LEDs on the LED modules 316 can be changed to create the desired effect.
Since each of the actuator assemblies 232 can be individually controlled separately, the movement and color of each of the actuator assemblies 232 can be controlled to create patterns or other visual effects for the display device 100.
For example, the actuator assemblies 232 in a certain area of the display can be extended and retracted in coordination to give the appearance of movement of the display device 100. In one such example, the actuator assemblies 232 are controlled to provide a wave-like effect across the display device 100. In another example, the control is randomized, so that the actuator assemblies 232 move in random patterns. Other configurations and patterns are possible.
By controlling the display device 100 in this manner, the overall visual impact of the display device 100 is increased. Specified patterns can be used to further enhance the visual effect of the display device 100, thereby catching the eye of a viewer.
In some examples, the patterns are configured to make certain shapes and depictions. For example, as shown in
For example, the computing device 502 can be programmed to create various shapes on the display device 100 depending on the time of day, as well as control the sequence of those shapes. The sequence can be choreographed or randomized, as desired. For example, in one embodiment, the computing device 502 can control the sign to depict fluid flowing out of a bottle. Many other examples are possible.
In addition, the computing device 502 may also control other aspects of the display device 100, such as microphones, speakers, cameras or other sensors, such as motion detectors, light sensors, and moisture sensors. For example, the computing device 502 could be configured such that it controls cameras located on or near the display device 100 such that it could display images or video a camera captures. The computing device 502 may also control speakers located on or near the sign such that it can play music or other desired sounds, such as sounds obtained by a microphone. Thus the display device 100 would be able to play stored sound, stream sound from the internet, or also play “live” sound detected by the microphone.
In addition, the computing device 502 may be equipped with video recognition software such that, for example, the camera could recognize a passerby and focus on and display that individual on the display 100. In addition, the computing device may be configured or programmed to play stored sounds at relevant times to communicate with, for example, the individual being displayed on the display device 100 by also controlling the speakers to project relevant sounds, slogans or speech. In this aspect of one example embodiment the computing device 502 allows the sign to be interactive with those in its surroundings and drawing more attention to the display device 100.
The computing device 502 may also control other sensors placed on the display, such as for example, motion detectors, light sensors, and moisture sensors. The display device may also include lasers that can be controlled by the computing device 502 such that the display device can produce a laser light show. The computing device may also integrate one or more of the sensors, or the information obtained therefrom, speakers, lasers, etc. to work in combination to enhance the display device 100. In yet another example embodiment of the display 100, the computing device 502 could be connected to the Internet and obtain a variety of information and display it on the display 100, such as weather, news, etc. Such sensors, such as a light sensor, could be used to adjust the brightness of the display device 100, for example, to adjust the brightness of the LEDs depending on the weather or time of day or night. Likewise the microphones could also be used to adjust the volume of the speakers to the appropriate level given the particular surroundings.
Referring now to
At operation 610, a video creation software application such as a three-dimensional visualization software is used to author content for the display device 100. The software, which is executed by the computing device 502 (or any other computing device, not necessarily connected to the display device 100), allows for the creation and/or manipulation of video content that will be used to control the display device 100. The software optionally includes an emulator that depicts the display device 100 to allow a user to author different content for the display device. One example of such content is an advertisement featuring a bottle. The advertisement can define the shape, motion, and color of the bottle to be depicted on the display device 100.
Next, at operation 620, the content is edited into video (i.e., color) and motion components. This is accomplished by extracting the video and motion components so that dual synchronized video files are formed. The first video file is for controlling the light display (the LED modules), and the second video file is for controlling the motion (i.e., the moving cubes).
The first video file is transferred to operations 630, 640, whereat the LED modules of the display device 100 are controlled. This includes controlling which of the LED modules are active and any content displayed on the LED modules. In this example, the LED modules are controlled using the GigE protocol.
The second video file is transferred to operation 650, whereat the motion file is interpreted and translated into the DMX protocol. This protocol is, in turn, used at operation 660 to control movement of the moving cubes of the actuator assemblies by the servo motor.
By synchronizing the first and second video files, the visual and motion components of the display device 100 are synchronized to create the desired effects as defined by the author.
In these examples, the computing device 502 includes one or more processing units and computer readable media. Computer readable media includes physical memory such as volatile memory (such as RAM), non-volatile memory (such as ROM, flash memory, etc.) or some combination thereof. Additionally, the computing device can include mass storage (removable and/or non-removable) such as a magnetic or optical disks or tape. An operating system, such as Linux or Windows, and one or more application programs can be stored on the mass storage device. The computing device can include input devices (such as a keyboard and mouse) and output devices (such as a monitor and printer).
The computing device also includes network connections to other devices, computers, networks, servers, etc., such as through the network 504. In example embodiments, the computing device communicates with other components through one or more networks, such as a local area network (LAN), a wide area network (WAN), the Internet, or a combination thereof. Communications can be implemented using wired and/or wireless technologies.
The display device 100 is configured to be resistant to the forces of nature. For example, the display device (including the base structure 210 and the modules 222) is configured to withstand rain and wind as the display device 100 is used outside during the different seasons. As noted herein, in certain weather conditions, certain functions of the display device 100 can be suspended temporarily.
Embodiments of the APT 700 receive as inputs a display video intended to be shown on the LED modules 316, and a movement video which is an encoded representation of the LED module movement. As shown in
Examples of the APT 700 further provide the ability to export the content once the user has completed creating and previewing it. With the system illustrated in
Valid export data 724 include, for example, a file with video correctly formatted for display by the LED controllers 370, and a file with video correctly formatted for interpretation for movement by the actuator assemblies 232.
As noted above, some embodiments of the APT 700 provide the interface for including actuator assembly 232 movement along with the displayed video content. End users may either create movement to go along with their display videos using a video editing application 702 of choice, or they may select default movement files provided within the APT. For example, the APT 700 may include a library of pre-generated movement videos that define predetermined movement patterns available for users of the APT 700.
Embodiments of the APT 700 are configured to verify that the video and motion files are the same length. If the files are not the same length, various solutions may be employed. For example, if the content video is longer than motion video, an error message is presented to the user informing them if they continue the motion content will be looped. If the motion video is longer than the content video, an error message is presented to the user informing them if they continue the motion content will be truncated.
To combine the content and movement video files to simulate the video and motion together, both a content video file and corresponding movement file are loaded to the APT 700 from the editing application 702. For the content video file, the APT 700 checks for the appropriate file type, length, etc. in the validation process 722. Each video frame is read in sequence and converted to an image for manipulation by a three dimensional simulator. As noted above, the disclosed example display device 100 includes a grid having movable LED modules 316. The content video file is thus split into a corresponding grid for display on the individual LED modules 316 of each module 222. The movement file is the same size as the content video file, and is also split into a corresponding grid.
As noted above in conjunction with
In certain implementations, the MSC 800 is installed at the location of the display device 100 and provides operational functionality for the movement of the actuator assemblies 232. In some embodiments, the DMX protocol (DM512) is used for communicating to the actuator assemblies 232. The signals output by the MSC are thus converted to DMX instructions suitable for controlling the actuator assemblies 232. In some embodiments, the LightFactory control system from dreamsolutions of Auckland, New Zealand is used to convert the greyscale video signal data into DMX512 instructions.
A conversion process 808 converts the motion data to visual data, and the MSC 800 displays the motion data as a visual output (the greyscale data is displayed to the MSC monitor 810). Each frame of movement data is converted to a greyscale red, green and blue value. This greyscale value is drawn to the screen 810 as 28 pixel wide by 28 pixel high squares arranged in a grid (exactly like the movement video file exported from the APT 700). The visual motion data is converted to an internal representation of motion. The greyscale value for video for each module is converted into a numeric value between 0 and 255 (0 being completely black and 255 being fully white). The greyscale numeric value is then converted to DMX512 instructions such that the numeric values correspond to the extension of the actuator assemblies 232 as described above.
As illustrated in the example of
In the some examples, the environmental sensors include a weather server 822 that provides data regarding weather conditions such as wind speed, temperature, humidity, etc. During normal operation, the MSC 800 regularly requests updates (for example, each second) from the motion panel and environmental servers 820,822. In some implementations, each of these services has a separate timeout period (e.g., 60 seconds for the maintenance panel 820, 30 seconds for the weather sensors 822,824). If the service returns a negative status response during the entire timeout period, the MSC 800 will disable movement of the actuator assemblies 232. The MSC 800 will enable movement once the web service has again returned a positive status response for the entire timeout period. Additionally, if the web service is completely unresponsive during this timeout period, the MSC 800 will disable movement. The MSC 800 will enable movement once the web service has again been responsive for the entire timeout period.
In some implementations, the MCS 800 further includes a power usage detection process that monitors power consumption of the display device 100. For example, a power consumption threshold parameter may be determined and used as an input to the MSC 800. Power usage is monitored for module movement, LEDs, and other ancillary components. If power usage exceeds the threshold parameter, a warning or message is sent to an event log 826.
The various embodiments described above are provided by way of illustration only and should not be construed to limiting. Those skilled in the art will readily recognize various modifications and changes that may be made to the embodiments described above without departing from the true spirit and scope of the disclosure or the following claims.
This application claims the benefit of provisional application Ser. No. 61/800,611, filed on Mar. 15, 2013, which is incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4654989 | Fleming | Apr 1987 | A |
4757626 | Weinreich | Jul 1988 | A |
5077608 | Dubner | Dec 1991 | A |
5086287 | Nützel | Feb 1992 | A |
5525000 | Belobraydich et al. | Jun 1996 | A |
5717423 | Parker | Feb 1998 | A |
5793918 | Hogan | Aug 1998 | A |
5907312 | Sato et al. | May 1999 | A |
6189246 | Gorthala | Feb 2001 | B1 |
6433761 | Remitz | Aug 2002 | B1 |
6462840 | Kravtsov | Oct 2002 | B1 |
6546655 | Hillstrom | Apr 2003 | B1 |
6606809 | Hillstrom et al. | Aug 2003 | B2 |
7055271 | Lutz et al. | Jun 2006 | B2 |
7277080 | Goulthorpe | Oct 2007 | B2 |
7352339 | Morgan et al. | Apr 2008 | B2 |
7356357 | DeCost et al. | Apr 2008 | B2 |
7436388 | Hillis et al. | Oct 2008 | B2 |
7439950 | Carlberg | Oct 2008 | B2 |
7525510 | Beland et al. | Apr 2009 | B2 |
7551771 | England, III | Jun 2009 | B2 |
7552553 | Kelly | Jun 2009 | B2 |
7605772 | Syrstad | Oct 2009 | B2 |
7620026 | Anschutz et al. | Nov 2009 | B2 |
7653569 | Zbib | Jan 2010 | B1 |
7852333 | Nishikawa et al. | Dec 2010 | B2 |
7866075 | Meeker et al. | Jan 2011 | B2 |
7905413 | Knowles et al. | Mar 2011 | B2 |
7928968 | Shon et al. | Apr 2011 | B2 |
7948450 | Kay et al. | May 2011 | B2 |
8040361 | Bachelder et al. | Oct 2011 | B2 |
8081158 | Harris | Dec 2011 | B2 |
8152062 | Perrier et al. | Apr 2012 | B2 |
8232976 | Yun et al. | Jul 2012 | B2 |
8254338 | Anschutz et al. | Aug 2012 | B2 |
8289274 | Sliwa et al. | Oct 2012 | B2 |
8413073 | Lee | Apr 2013 | B2 |
8552883 | Su | Oct 2013 | B1 |
8576212 | Lee et al. | Nov 2013 | B2 |
8588517 | Lee et al. | Nov 2013 | B2 |
8606043 | Kwon et al. | Dec 2013 | B2 |
9257061 | Jurewicz et al. | Feb 2016 | B2 |
9269283 | Jurewicz et al. | Feb 2016 | B2 |
9640118 | Jurewicz et al. | May 2017 | B2 |
9885466 | Jurewicz et al. | Feb 2018 | B2 |
10208934 | Jurewicz et al. | Feb 2019 | B2 |
20030080923 | Suyama et al. | May 2003 | A1 |
20040077285 | Bonilla et al. | Apr 2004 | A1 |
20040165006 | Kirby et al. | Aug 2004 | A1 |
20050017977 | Simpson et al. | Jan 2005 | A1 |
20050150147 | Berryman | Jul 2005 | A1 |
20050264430 | Zhang et al. | Dec 2005 | A1 |
20060055641 | Robertus et al. | Mar 2006 | A1 |
20060107567 | Liao | May 2006 | A1 |
20060285832 | Huang | Dec 2006 | A1 |
20070171674 | Deutsch | Jul 2007 | A1 |
20070244417 | Escriba Nogues | Oct 2007 | A1 |
20080010041 | McDaniel | Jan 2008 | A1 |
20080115187 | Decost et al. | May 2008 | A1 |
20080201208 | Tie et al. | Aug 2008 | A1 |
20080238889 | Thorne | Oct 2008 | A1 |
20080246757 | Ito | Oct 2008 | A1 |
20080266204 | Bartels et al. | Oct 2008 | A1 |
20090084010 | Dykstra | Apr 2009 | A1 |
20090177528 | Wu et al. | Jul 2009 | A1 |
20090184892 | Eberle et al. | Jul 2009 | A1 |
20100182340 | Bachelder et al. | Jul 2010 | A1 |
20100219973 | Griffin et al. | Sep 2010 | A1 |
20110128283 | Lee et al. | Jun 2011 | A1 |
20110175992 | Lee et al. | Jul 2011 | A1 |
20110193277 | Christenson | Aug 2011 | A1 |
20110225860 | Troiano et al. | Sep 2011 | A1 |
20110228058 | Hatasawa | Sep 2011 | A1 |
20110231231 | Cruz | Sep 2011 | A1 |
20110235332 | Cheung | Sep 2011 | A1 |
20120092337 | Tsao | Apr 2012 | A1 |
20120139919 | Shintani | Jun 2012 | A1 |
20120154438 | Cohen | Jun 2012 | A1 |
20120159820 | Van Saanen | Jun 2012 | A1 |
20120188235 | Wu et al. | Jul 2012 | A1 |
20120195463 | Shinkai | Aug 2012 | A1 |
20120224311 | Sutherland et al. | Sep 2012 | A1 |
20120242958 | Zuloff | Sep 2012 | A1 |
20130009951 | Kwon et al. | Jan 2013 | A1 |
20130041730 | LoBianco | Feb 2013 | A1 |
20130063561 | Stephan | Mar 2013 | A1 |
20130117121 | Raman et al. | May 2013 | A1 |
20130194059 | Parr | Aug 2013 | A1 |
20130265213 | Chen | Oct 2013 | A1 |
20130312300 | Lee | Nov 2013 | A1 |
20130321394 | Fisher et al. | Dec 2013 | A1 |
20140104389 | Dharmatilleke | Apr 2014 | A1 |
20140114708 | Campbell | Apr 2014 | A1 |
20140259824 | Jurewicz et al. | Sep 2014 | A1 |
20140267457 | Jurewicz et al. | Sep 2014 | A1 |
Number | Date | Country |
---|---|---|
101546503 | Sep 2009 | CN |
10 2009 019 400 | Nov 2010 | DE |
2 363 506 | Dec 2001 | GB |
2 393 560 | Mar 2004 | GB |
2 428 433 | Jan 2007 | GB |
2 435 540 | Aug 2007 | GB |
4-14086 | Jan 1992 | JP |
10-268796 | Oct 1998 | JP |
3127447 | Jan 2001 | JP |
2001-333438 | Nov 2001 | JP |
2007-225638 | Sep 2007 | JP |
2005025377 | Mar 2005 | WO |
Entry |
---|
International Search Report and Written Opinion for PCT/US2017/053290 dated Jan. 4, 2018. |
International Search Report and Written Opinion for PCT/US2014/029573 dated Sep. 18, 2014. |
Number | Date | Country | |
---|---|---|---|
20190257506 A1 | Aug 2019 | US |
Number | Date | Country | |
---|---|---|---|
61800611 | Mar 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15854513 | Dec 2017 | US |
Child | 16271395 | US | |
Parent | 15469863 | Mar 2017 | US |
Child | 15854513 | US | |
Parent | 14996049 | Jan 2016 | US |
Child | 15469863 | US | |
Parent | 14213358 | Mar 2014 | US |
Child | 14996049 | US |