This section is intended to introduce the reader to various aspects of art that may be related to various aspects of the present techniques, which are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present disclosure. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
Amusement parks and other entertainment venues contain experiences (e.g., ride vehicle experiences, animated figures (e.g., robotic characters), scenes, attractions and so on) to entertain park guests. As the experiences become more technologically advanced and complex, the components within the experience may benefit from more robust monitoring and maintenance to ensure the safety of the park guests and optimal or improved operation of the experience. As such, it is now recognized that it is advantageous to utilize a variety of sensors, machine learning, and multi-dimensional modeling techniques to enable faster anomaly detection, automated alerting, predictive maintenance, and dynamic correction with regard to the experience.
Certain embodiments commensurate in scope with the originally claimed subject matter are summarized below. These embodiments are not intended to limit the scope of the disclosure, but rather these embodiments are intended only to provide a brief summary of certain disclosed embodiments. Indeed, the present disclosure may encompass a variety of forms that may be similar to or different from the embodiments set forth below.
In an embodiment, a method of monitoring an amusement park experience may include receiving, via multiple sensors, multiple layers of first sensor data indicative of characteristics of the experience. The method may generate a profile of the experience based on the first sensor data, wherein the profile includes a baseline and a threshold indicating an acceptable range of the characteristics. The method may receive second sensor data and third sensor data via the multiple sensors. The method may determine, in response to identifying characteristics of the second sensor data that deviate from the baseline but do not exceed the threshold, that the experience is operating properly. The method may also, in response to identifying characteristics of the third sensor data that deviate from the baseline and exceed the threshold, perform a corrective action.
In an embodiment, a system may include a network and one or more communication hubs communicatively coupled with one another via the network. The system may also include a ride vehicle including one or more ride vehicle sensors, where the one or more ride vehicle sensors are communicatively coupled to at least one of the one or more communication hubs. Further, the system may include an animated figure, the animated figure including an actuator and an actuator sensor, wherein the actuator sensor is communicatively coupled to at least one of the one or more communication hubs. The system may also include a controller that may receive data from the one or more ride vehicle sensors and data from the actuator sensor via the one or more communication hubs. The controller may adjust the actuator and control the ride vehicle based on a discrepancy between a baseline relative value and a currently detected relative value of the data from the one or more ride vehicle sensors and the data from the actuator sensor.
In an embodiment, one or more tangible, non-transitory, computer-readable media is provided in accordance with the present disclosure. The one or more tangible, non-transitory, computer-readable media comprises instructions that, when executed by at least one processor, cause the at least one processor to receive, via multiple sensors, first sensor data indicative of characteristics of an amusement park experience. The processor may generate a profile of the amusement park experience based on the first sensor data, and the profile may include a baseline and a threshold indicating an acceptable range of the characteristics. The processor may receive a second sensor data and third sensor data via the multiple sensors and determine, in response to identifying characteristics of the second sensor data that deviate from the baseline but do not exceed the threshold, that the experience is operating properly. However, in response to identifying characteristics of the third sensor data that deviate from the baseline and exceed the threshold, the processor may perform a corrective action.
These and other features, aspects, and advantages of the present disclosure will become better understood when the following detailed description is read with reference to the accompanying drawings in which like characters represent like parts throughout the drawings, wherein:
One or more specific embodiments of the present disclosure will be described below. In an effort to provide a concise description of these embodiments, all features of an actual implementation may not be described in the specification. It should be appreciated that in the development of any such actual implementation, as in any engineering or design project, numerous implementation-specific decisions must be made to achieve the developers' specific goals, such as compliance with system-related and business-related constraints, which may vary from one implementation to another. Moreover, it should be appreciated that such a development effort might be complex and time consuming, but would nevertheless be a routine undertaking of design, fabrication, and manufacture for those of ordinary skill having the benefit of this disclosure.
When introducing elements of various embodiments of the present disclosure, the articles “a,” “an,” and “the” are intended to mean that there are one or more of the elements. The terms “comprising,” “including,” and “having” are intended to be inclusive and mean that there may be additional elements other than the listed elements. Additionally, it should be understood that references to “one embodiment” or “an embodiment” of the present disclosure are not intended to be interpreted as excluding the existence of additional embodiments that also incorporate the recited features.
Theme parks and other such entertainment venues are becoming increasingly popular. Further, immersive experiences within such entertainment venues are in high demand. In order to provide new and exciting experiences, attractions, such as ride experiences and scenes (e.g., visual shows including live action, animated figures, computer-generated imagery, and so on) have become increasingly complex, involving integration of lighting, sound, movement, interactive elements, visual media, and so on. Accordingly, the components within the attractions may benefit from more exhaustive and holistic monitoring and maintenance. Conventional monitoring and maintenance systems include disparate clusters or networks of sensors and a manual regimen of prescribed inspections by technicians. Such conventional systems may fail to detect certain maintenance and/or user experience issues due to human error, or may fail to detect maintenance and/or user experience issues due to a lack of automated monitoring (e.g., via sensors, cameras, and so on). As may be appreciated, a system utilizing a multifaceted, layered, globally-connected network of sensors, multidimensional models, and machine learning engines may effectuate a more robust, accurate, responsive, and intelligent monitoring and maintenance system for amusement park attractions.
In view of the foregoing, present embodiments are generally directed to a method and system for automated monitoring, anomaly detection, correction, mitigation, and/or predictive maintenance in theme park experiences. Specifically, present embodiments are directed to collecting, via a network of communicatively coupled sensors, data on a variety of aspects of and components within an experience, using machine learning to detect and/or predict quantitative and qualitative issues, dynamically addressing the anomaly and/or triggering an alert regarding the anomaly. Visual aspects, control features, audio, aspects of the animated figures, and monitoring and maintenance components and/or aspects are prepared in accordance with present techniques to provide multifaceted and layered monitoring and maintenance of the experience. Multifaceted and layered monitoring and maintenance provides robust and intelligent monitoring of the experience. Layering is achieved at least in part by utilizing a number of data sources or streams (e.g., audiovisual data, infrared data, social media data and so on) to detect, predict, and/or correct for subtle anomalies that may nevertheless impact the quality of a guest's enjoyment of the experience. Due to the complex nature of interactions between the guest and the experience, in some embodiments, control instructions from an artificial intelligence (AI) or machine learning engine may coordinate sensor data, equipment performance and the like to detect and address (e.g., correct or mitigate the effects of) an anomaly. Further, numerous and variable iterative routines are included to gradually improve aspects of the monitoring and maintenance system.
Procedures, in accordance with the present disclosure (applicable to procedures illustrated in
The one or more memory devices 108 may store information such as control software, look up tables, configuration data, and so on. In some embodiments, the one or more processors 106 and/or the one or more memory devices 108 may be external to the controller 104 and/or the electronic device 102. The one or more memory devices 108 may include a tangible, non-transitory, machine-readable-medium, such as a volatile memory (e.g., a random access memory (RAM)) and/or a nonvolatile memory (e.g., a read-only memory (ROM)). The one or more memory devices 108 may store a variety of information and may be used for various purposes. For example, the one or more memory devices 108 may store machine-readable and/or processor-executable instructions (e.g., firmware or software) for the one or more processors 106 to execute, such as instructions for determining a likelihood that an entertainment experience (e.g., automated positioning of an animated figure) will trend past a threshold or envelope of acceptable operation (e.g., due to wear, component failure, control degradation or the like) and adjusting operation or control features accordingly. The one or more memory devices 108 may include one or more storage devices (e.g., nonvolatile storage devices) that may include read-only memory (ROM), flash memory, a hard drive, or any other suitable optical, magnetic, or solid-state storage medium, or a combination thereof.
The electronic device 102 may also include an electronic display 110 that enables graphical and/or visual output to be displayed to a user. The electronic display 110 may use any suitable display technology, and may include an electroluminescent (ELD) display, liquid crystal (LCD) display, light-emitting diode (LED) display, organic LED (OLED) display, active-matrix OLED display, plasma display panel (PDP), quantum dot LED (QLED) display, and so on.
As illustrated, the electronic device 102 may include a data acquisition system (DAQ) 112 operable to send and receive information to and from a sensor network 120. For example, the electronic device 102 may include a communication interface that enables the controller 104 to communicate with servers and/or other computing resources of the sensor network 120 via a communication network (e.g., a mobile communication network, a WiFi network, local area network (LAN), wide area network (WAN), the Internet, and the like). In some cases, at least a portion of the information received from the sensor network 120 may be downloaded and stored on the memory or storage devices 108 of the electronic device 102. The sensor network 120 may include ultraviolet sensors 122, infrared sensors 124, vibration sensors 126, imaging sensors 128, audio sensors 130, biometric sensors 132, and so on. While the preceding list of sensors is illustrated in
The sensor network 120 may enable the system 100 to collect an expansive amount of data regarding an amusement park attraction, such that the system 100 may obtain a fine-grain view of the various aspects of and components within the system 100. For example, the infrared sensor 124 may detect if the wheels or motor of a ride vehicle are emitting heat beyond a heat threshold, and if so may trigger a warning (e.g., via the controller 104) to technical operators, disable the ride vehicle, or take any other appropriate action. Combinations of input may also be assessed using algorithms, lookup tables, or artificial intelligence. For example, data from the infrared sensor 124 may be analyzed in combination with the vibration sensors 126 to detect a cause of heat related to misalignment of wheels of a ride vehicle, or the like. The sensor network 120 may enable the system 100 to address quantitative and qualitative issues in an amusement park attraction. As used herein, quantitative issues may be defined as issues that can be quantified, counted, or measured. For example, a temperature sensor detecting that the temperature in a room is 85 degrees Fahrenheit when the expected temperature is 75 degrees Fahrenheit. In contrast, qualitative issues may be defined as issues regarding the appearance, sound, or feel of the experience or components within the experience, particularly as they relate to a guest's enjoyment of the experience.
For example, the sensor network 120 may enable the system 100 to address qualitative issues in the case that a lightbulb of a set of lightbulbs burns out during a show or scene, negatively impacting the quality of a user's experience. However, if a camera or other imaging sensor (e.g., 130) detects a difference, such as a deviation from a predetermined profile, in the lighting of a scene, the electronic device 102 may cause the brightness or direction of other lightbulbs in the set of lightbulbs to adjust to compensate for the malfunctioning lightbulb. The predetermined profile may be established by utilizing a machine learning engine (e.g., 114), as will be discussed in greater detail below. Further, the sensor network 120 may enable the system 100 to detect qualitative deviations such as a guest being in an area that the guest is not expected to be in. Upon making this determination, the system 100 may alert a technical operator or take other corrective action (e.g., stopping or pausing operation of the experience) or automatically perform such actions.
In addition to the sensor network 120, the DAQ 112 may collect web data 134 using software applications such as web crawlers to obtain information regarding anomalies or other issues within an attraction. For example, if a scene experiences an issue that may not be easily detected by a sensor (e.g., an animated character's wig falls off or the animated character experiences a costuming malfunction), a guest may notice the issue and post about the issue on social media. The DAQ 112 may detect, via the web crawler, the social media post and may trigger an alert.
Through utilization of one or more centralized processing servers (e.g., the electronic device 102), machine learning (ML) algorithms (e.g., executed via the ML engine 114) and anomaly detection algorithms, the system 100 may realize improvements in computer-related technologies by increasing the accuracy and efficiency of anomaly detection, predictive analysis, and intelligent maintenance in amusement park anomaly detection and maintenance systems. The sensor network 120 along with other data resources such as the web data 134 may provide a layered system of anomaly detection and maintenance. The sensors in the sensor network 120 may provide redundancy to ensure detection of anomalies that may go unnoticed in a conventional monitoring and maintenance system. For example, a spotlight in an attraction may be equipped with a current or voltage sensor allowing the system 100 to determine if an electrical issue is causing the spotlight to dim or turn off when considered in conjunction with other detected values (e.g., lighting level values in a staging area) provided to the system 100. An imaging sensor 128 may also be fixed upon an area illuminated by the light of the spotlight and the system 100 may be trained (e.g., via the ML engine 114) to detect luminance or brightness variations based on the imaging sensor 128. In this way, a lighting issue with the spotlight may be detected by the system 100 based on the current or voltage sensor, the imaging sensor 128, or combinations thereof.
In some embodiments, the electronic device 102 may include the (ML) engine 114. While the ML engine 114 may be implemented outside of the controller 104 as illustrated in
The electronic device 102, in particular the controller 104 of the electronic device 102, may receive data from various components (e.g., the ride vehicle 310, the animated character 312, the set lights 302, the environmental components 314, and so on) and/or may receive data from sensors of the sensor network 120 communicatively coupled to or otherwise monitoring the various components and/or conditions. The controller 104 may receive the data and adjust the actuators of the animated character 312, adjust lighting, adjust climate control, control effects, adjust airflow, control the ride vehicle 310, and the like (including combinations thereof). For example, the controller 104 may, based on the data received regarding the animated character 312, cause the actuators to adjust to make the animated character 312 move in a certain direction, look towards a particular guest or group of guests, or interact with elements of the experience 300. The controller 104 may, based on the data received regarding the ride vehicle 310, cause the ride vehicle 310 to speed up, slow down, stop, change direction, engage a hydraulic system of the ride vehicle 310, and so on. The controller 104 may adjust the set lights 302 (e.g., adjust position, brightness, color, and so on), control the environmental components 314, and so on.
In the process block 202 of the method 200, the experience 300 may undergo several cycles under normal operating conditions (e.g., as verified by a human technical operator) in order to obtain this data. For example, the cycling may include movement of a ride vehicle experience to determine proper triggering and response in order to verify that not only is the ride vehicle experience operating as per the original creative intent, but that the ride vehicle experience is responsive to the guest presence and/or precise location as intended. In this phase, the ML engine 114 may collect and train upon the data gathered through cycling to determine desired operating parameters of the experience. Upon collecting a sufficient amount of training data, the ML engine 114 may begin processing and labelling the data. In the processing and labelling phase, supervised and/or unsupervised machine learning will be performed by the ML engine 114. Based on the results of the processing and labelling phase, the ML engine 114 may determine and encode the parameters of a normally functioning experience.
The experience may be cycled repeatedly to enable the ML engine 114 to perform multiple layered processing passes, gradually learning additional aspects and parameters of the observed experience. For example, there may be a processing pass in which lighting color and intensity (e.g., detected via the imaging sensors 128 or features of the set lights 302) is learned and stored across the system 100. By sampling the experience 300 frame-by-frame over known time units, the system 100 may ascertain whether the set lights 302 are functioning properly to specification and are aimed as intended. The system 100 may detect anomalies such as flickering, outages, degradation or timing and triggering errors. Another processing pass may determine whether the animated
Based on the information obtained in process block 202, the method 200 may, in process block 204, generate profiles of the experience 300 based on the received sensor data. The profiles may include a baseline and a threshold indicating an expected range of the characteristics or aspects of the experience 300. For example, a profile where the equipment in the experience 300 (e.g., the set lights 302, audio equipment 306, environmental components 314, and animated
In query block 206, the method 200 may determine whether the experience 300 is operating properly according to specification and expectation. That is, data 207 from a current experience (e.g., essentially real time data from the experience 300) may be compared with established profiles established as part of the process block 204. Through the cycling and processing passes of the experience 300, profile thresholds may be determined outside of which the profile may receive a different designation, but within which a designation may remain even if deviation in the aspects and characteristics of the experience 300 are detected. For example, if during operation of the experience 300 the system 100 detects that the set lights 302 are dimmer than expected due to a technical error, the system 100 may determine this to be a deviation from the A profile, but not a deviation that exceeds the A profile threshold. Under these conditions, it may be determined in the query block 206 that the experience 300 is operating properly, and thus the method 200 may continue to receive sensor data indicative of characteristics of the experience 300 (e.g., in the process block 202). However, if the animated
Depending on the parameters and boundaries designated by a user of the system 100, a B profile or C profile may be determined to be an improper operation of the experience 300, and thus, in process block 208, a corrective action may be determined and performed. The corrective action may include performing automated maintenance on malfunctioning equipment, such as having the controller 104, based on information received from the ML engine 114, send a command causing the experience 300 to adjust current to a malfunctioning set light 302 outputting less brightness than expected, or having the controller 104 send a command causing the experience 300 to adjust other equipment to account for malfunctioning equipment (e.g., adjusting the brightness of another set light 302 to compensate for the malfunctioning set light 302).
As previously discussed, the controller 104 may receive data from various components such as the ride vehicle 310 and/or the animated character 312 and adjust the actuators of the animated character 312 and/or control the ride vehicle 310 accordingly. Additionally or alternatively, the controller 104 may adjust the actuators of the animated character 312 and/or control the ride vehicle 310 based on the profile designation of the experience 300. For example, if the method 200 designates the profile of the experience 300 as an A profile, the controller 104 may cause the actuators to adjust to make the animated character 312 move in a certain direction, look towards a particular guest or group of guests, or interact with elements of the experience 300, and the controller may cause the ride vehicle 310 to accelerate at a particular location of the experience 300. However, if the method 200 designates the profile of the experience 300 as an X profile, the controller 104 may prevent the actuators from adjusting to move the animated character 312 and may prevent the ride vehicle 310 from accelerating—or may stop the ride vehicle 310—at the particular location of the experience 300.
The corrective action may also include sending an alert (e.g., to an alert panel 308) informing the technical operators of the malfunction. If the malfunction requires urgent action, the ML engine 114 may send an urgent alert to the technical operators and cause the experience 300 to stop or pause operation.
Through the method 200, the ML engine 114 may learn to identify known and new anomalies as well as anticipate anomalies that may occur in the future. For example, if the ML engine 114 determines that the set lights 302 gradually experience a reduction in brightness prior to a bulb burning out, the ML engine 114 may trigger an alert to the technical operators if the ML engine 114 detects that a set light 302 has gradually experienced a reduction in brightness, and may include in the alert an estimation of how long it may take for the bulb of the set light 302 to burn out. In this way, the system 100 may enable predictive maintenance within the experience 300 and may permit the technical operators to take corrective action prior to a malfunction, preserving the quality of the experience 300.
The system 100, through the ML engine 114, may also be able to apply the machine learning performed on the experience 300 to a different experience. The learned data from the experience 300 may be extrapolated and retargeted to another sufficiently similar experience by implementing a deep learning process known as transfer learning. Transfer learning allows for a sufficiently trained ML model to be repurposed and reused to make observations or predictions about a different but related problem set. For example, using the data learned about the operating characteristics of the set lights 302 in the experience 300, the ML engine 114 may be able to identify and anticipate issues that may be experienced by set lights in a different experience.
In certain embodiments, the system 100 and the experience 300 may utilize multiple localized sensor networks that are not linked together via a single network.
In contrast to
Utilizing the network 500, an issue detected in one sensor (e.g., 504B) in the network 500 may be communicated to other sensors (e.g., 504A, 504C, 510) in the network 500 as well as to other equipment (e.g., the device 508) such as controllers in the animated
With respect to
Additionally, using the 3D modeling engine 116, the system 100 may be able to detect and correct certain positional errors of the equipment or sensors. For example, if the system 100 (e.g., via the ML engine 114) detects unexpected readings from an image sensor 128, one or more other image sensors 128 may observe the position of the problematic image sensor 128, compare the position to an expected position based on the global 3D model, determine that the problematic image sensor 128 is misaligned, and send feedback to the system 100, thereby enabling the system 100 to correct the position of the problematic image sensor 128. Indeed, the system 100 can be trained based on 3D modeling and/or operate using 3D modeling as a baseline template for comparison to ongoing data (e.g., essentially real-time data from an attraction or experience).
While only certain features of the disclosure have been illustrated and described herein, many modifications and changes will occur to those skilled in the art. It is, therefore, to be understood that the appended claims are intended to cover all such modifications and changes as fall within the true spirit of the disclosure. It should be appreciated that any of the features illustrated or described with respect to the figures discussed above may be combined in any suitable manner.
The techniques presented and claimed herein are referenced and applied to material objects and concrete examples of a practical nature that demonstrably improve the present technical field and, as such, are not abstract, intangible or purely theoretical. Further, if any claims appended to the end of this specification contain one or more elements designated as “means for (perform)ing (a function) . . . ” or “step for (perform)ing (a function) . . . ”, it is intended that such elements are to be interpreted under 35 U.S.C. 112(f). However, for any claims containing elements designated in any other manner, it is intended that such elements are not to be interpreted under 35 U.S.C. 112(f).
This application claims priority to and the benefit of U.S. Provisional Application No. 63/296,733, entitled “SYSTEMS AND METHODS FOR VISUAL SCENE MONITORING,” filed Jan. 5, 2022, which is hereby incorporated by reference in its entirety for all purposes.
Number | Date | Country | |
---|---|---|---|
63296733 | Jan 2022 | US |