The invention pertains to digital data and, more particularly, the collection and systemic anticipatory intelligence of extremely large data sets a/k/a “big data” from industrial assets and systems. The invention has application in manufacturing, energy, utilities, aerospace, marine, defense and other enterprises that generate vast sums of asset data. The invention also has application to the collection and anticipatory analysis of asset-related data in other fields such as, by way of non-limiting example, financial services and health care.
With the rise in computing power, growth of digital networks and fall in sensor prices, equipment of all sorts are becoming increasingly instrumented. Nowhere is this trend more obvious, for example, than in industry, where virtually every piece of physical equipment, from the most complex electromechanical devices to the most mundane materials mixing vessels, have hundreds of sensors and supporting diagnostic systems. It is no small feat to provide the distributed infrastructure necessary to carry that information to factory control rooms, where technicians and automated workstations can monitor it to ensure optimum and safe plant operation. The trend in health care and other enterprises has equally been toward instrumentation. This has resulted in the inclusion of sensors and diagnostics in equipment residing everywhere from computed tomography (CT) scanners used in hospitals to color copiers in corporate offices.
Turning our attention again to industrial enterprises, the technologies behind the so-called Industry 4.0 hold promise in allowing that same data to be collected and analyzed at still higher levels in the enterprise. Also, Industry 4.0 holds great promise for industrials to connect design, manufacturing, operations and service—through horizontal and vertical integration with suppliers and customers—to enable digital ecosystems to be created. A narrow view of Industry 4.0, focused purely on IoT (internet of things) as sensory networks connected to interact with external systems and the environment, fails to address business process automation across partner networks driven by the complementary technologies that will fuel Industry 4.0
An object of the invention is to provide improved methods and apparatus for digital industrial data and, more particularly, for example, for the collection and automated analysis of extremely large data generated by industrial assets, health, enterprise and other assets.
A related object of the invention is to provide such methods and apparatus as an integrated suite of predictive self-service applications in health care manufacturing, industrials (such as Power, Oil & Gas, Mining, Chemicals and defense) and other enterprises that generate vast sums of industrial asset data. A further related object is to provide such methods and apparatus as find application in financial industries, e.g., in support of real time physical asset risk assessment, valuation and financing of equipment- and other equipment-intensive businesses.
Still another object of the invention is to provide such methods and apparatus as capitalize on existing Industry 4.0 technologies and other that are yet to overcome their shortcomings.
The foregoing are among the objects attained by the invention which provides distributed, hierarchical systems and methods for the ingestion of data generated by a instrumented assets in manufacturing and/or industrial plants, hospitals and other health care facilities and other enterprises. The systems and methods employ an architecture that is capable of (1) collecting and standardizing industrial and other protocols, (2) preliminarily autonomous processing data and analytics, as well as (3) executing predictive diagnostic (and, potentially, remedial) applications, at the plant or facility level for detection of error and other conditions (and, potentially, correcting same), and (4) forwarding those data for more in-depth fleet/enterprise processing in a private or public cloud or a combination—ensuring the architecture is Cloud Neutral (i.e. operate on any cloud provider and cloud instance). Those systems and methods include edge services to process and intelligently identify the nearest, most readily available and/or highest throughput/cost effective cloud services provider to which to transmit data for further analysis or applications. The architecture takes into account the varied data throughput as well as storage and processing needs at each level of the hierarchy.
Related aspects of the invention provide such systems and methods having an architecture as shown in
Related aspects of the invention provide such systems and methods, e.g., as described shown in
Further related aspects of the invention provide such systems and methods, e.g., as described above, in which the aforesaid computing apparatus include control nodes, a command unit and network, security services, encryption and/or threat protection. They can further include a physical firewall and cloud operating system. See
Still further related aspects of the invention provide such systems and methods, e.g., as described above, in which the aforesaid computing apparatus translate protocols, aggregate, filter, standardize, learn, store and forward, data received from plant sensors, devices, in the plant or other facility.
Yet still further related aspects of the invention provide such systems and methods, e.g., as described above, in which the aforesaid computing apparatus execute microservices (
Still yet further related aspects of the invention provide such systems and methods, e.g., as described above, in which software executing in the aforesaid computing apparatus also runs in the main cloud (public or private) platform to which those (local) computing apparatus are coupled for communication. The public and/or private cloud instance of that software samples data in sub seconds time intervals and edge cloud services can handle data generated in frequencies of MHz or GHz; and has ‘store and forward’ capabilities of data to the public and/or private cloud on an as needed basis.
Further related aspects of the invention provide such systems and methods, e.g., as described above, including an self-learning optimization model (referred to herein as PARCS™) as described, e.g., in
Further aspects of the invention provide a hierarchical system for data ingestion that includes one or more computing apparatus coupled for communication via a first network to one or more local data sources. The computing apparatuses preliminarily process data from the data sources, including executing predictive diagnostics to detect error and other conditions, and forward one or more of those data over a second network for processing by a selected remote computing platform, which performs in-depth processing on the forwarded data.
Related aspects of the invention provide systems, e.g., as described above, wherein the first network includes a private network and the second network includes a public network, and wherein the local computing apparatus select, as the computing platform, one that is nearest, most readily available and/or has the best cost performance.
Further related aspects of the invention provide systems, e.g., as described above, wherein one or more of the local computing apparatus process data from the data sources sampled down to a first time interval (e.g., milliseconds—MHz or GHz), and wherein the remote computing platform processes data sampled down to a second time interval. The remote computing platform aggregates data from multiple local computing apparatus, to aggregate, consolidate and provide enterprise view of system and ecosystem performance across multiple facilities and Assets.
Still further related aspects of the invention provide systems, e.g., as described above, wherein the data sources comprise instrumented manufacturing, industrial, health care or vehicular or other equipment. The latter can include, by way of example, equipment on autonomous vehicles to determine real time PARCS™ score per vehicle. In the case of manufacturing and/or industrial equipment, aspects of the invention provide systems in which such equipment is coupled to one or more of the computing apparatus via digital data processing apparatus that can include, for example, programmable logic controllers.
Still further related aspects of the invention provide systems, e.g., as described above, wherein one or more of the local computing apparatus execute the same software applications for purposes of preliminarily processing data from the data sources as the remote computing platform executes for purposes of in-depth processing that data.
Yet still further related aspects of the invention provide systems, e.g., as described above, wherein one or more of the computing apparatus aggregate, filter and/or standardize data for forwarding to the remote computing platform. In other related aspects, the invention provides such systems wherein one or more of the computing apparatus forward data for more in-depth processing by the selected remote computing platform via any of (i) a shared folder or (ii) posting time series datapoints to that platform via a representational state transfer (REST) applications program interface. In such systems, the remote computing platform can perform in-depth processing on the time series datapoints to predict outcomes and identify insights that can be integrated in incumbent IT (Information Technology) and OT (Operational Technology) systems.
The invention provides, in other aspects, a hierarchical system for data ingestion that includes a computing platform executing an engine (“cloud edge engine”) providing a plurality of software services to effect processing on data. One or more computing apparatus that are local to data sources but remote from the computing platform execute services of the cloud edge engine to (i) collect, process and aggregate data from sensors associated with the data sources, (ii) forward data from those data sources for processing by the computing platform and (iii) execute in-memory advanced data analytics. The edge computing apparatuses process data from the data sources sampled down to millisecond time intervals (MHz or GHz), while the remote computing platform processes forwarded data. According to aspects of the invention, services of the cloud edge engine executing on the computing apparatus support continuity of operations of the instrumented equipment even in the absence of connectivity between those the edge computing apparatus and the computing platform.
Related aspects of the invention provide systems, e.g., as described above, wherein the services of the cloud edge engine executing on the computing apparatus are registered, managed and scaled through the use of platform as a service (PaaS) functionality.
Other aspects of the invention provide systems, e.g., as described above, wherein the computing apparatuses forward data to the computing platform using a push protocol. Related aspects of the invention provide such systems wherein the computing apparatuses forward data to the platform by making that data available in a common area for access via polling.
Still other aspects of the invention provide systems, e.g., as described above, wherein the cloud edge engine comprises an applications program interface (API) that exposes a configuration service to configure any of a type of data source, a protocol used for connection, security information required to connect to that data source, and metadata that is used to understand data from the data source. The cloud edge engine can, according to further aspects of the invention, comprise a connection endpoint to connect a data source as per the configuration service, wherein the endpoint is a logical abstraction of integration interfaces for the cloud edge engine. Such an endpoint can support, according to further aspects of the invention, connecting any of (i) relational and other storage systems, (ii) social data sources, and (ii) physical equipment generating data.
Yet still other related aspects of the invention provide systems, e.g., as described above, wherein the cloud edge engine includes a messaging system to support ingestion of streams of data in MHz or GHz speeds directly from industrial assets, process in-memory predictive analytics and forward data to remote private or public cloud systems.
Still other related aspects provide systems, e.g., as described above, wherein the cloud edge engine comprises an edge gateway service comprising an endpoint to which sensors connect to create a network. The Edge Cloud can have multiple Gateways connected to the Edge Cloud, and data ingestion and lightweight applications can be installed on Gateways to reduce latency and improve processing.
Still yet other related aspects of the invention provide systems, e.g., as described above, in which the cloud edge engine comprises an edge data routing service that time-stamps and routes data collected from the data sources to a persistent data store. The edge data routing service can, according to other related aspects of the invention, analyze data for a possibility of generating insights based on self-learning algorithms.
Further aspects of the invention provide systemic asset intelligence systems constructed and operated in the manner of the systems described above that additionally include an self-learning optimization engine executing on one or more of the computing apparatus and computing platform to identify and predict failure of one or more data sources that comprise smart devices. That self learning optimization engine (as shown, by way of example, for systems according to some practices of the invention, in
In further related aspects of the invention, the self learning optimization engine of systems, e.g., as described above, executes a model that performs a device performance measurement step to calculate any of asset performance, availability, asset reliability, asset capacity and asset serviceability. In still further related aspects of the invention, that model can perform a real time PARCS™ score to generate asset health indices and/or to predict asset maintenance and optimization.
Other aspects of the invention provide methods for data ingestion and for systemic asset intelligence paralleling operation of the systems described above.
The foregoing and other aspects of the invention are evident in the drawings and in the detailed description that follows.
A more complete understanding of the invention may be attained by reference to the drawings, in which:
For the sake of simplicity and without the loss of generality, the discussion below focuses largely on practices of the invention in connection with predictive enterprise-level plant and industrial monitoring and control. The invention has application, as well, in health care, financial services and other enterprises that benefit from the collection and systemic anticipatory analysis of large data sets generated by hospitals, office buildings and other facilities, as will be evident to those skilled in the art from the discussion below and elsewhere herein. In these regards, it will be appreciated that whereas industrial “plants” are often referenced in regard to the embodiments discussed below, in other embodiments, the other embodiments, the term “facility” may apply.
Industry 4.0 holds great promise, yet, is hugely overhyped. A narrow view of Industry 4.0 as sensory networks connected to interact with external systems and the environment, fails to address the complementary technologies that will enable Industry 4.0.
Systems according to the invention embrace those technologies. They feature architectures to meet the strategic Industry 4.0 needs of enterprises into the future; functionality that ingests data from different industrial protocols and systems at the edge cloud, with each data connection defined as microservices to facilitate the delivery of predictive analytics and application functionality. Such cloud systems, moreover, can support multi-tenancy by client and asset, allowing data for multiple customers (e.g., enterprises) to be transmitted to, stored on, and/or processed within a single, cloud-based data processing system without risk of data commingling or risk to data security. Multi-tenancy further facilitates the delivery of Industrial SaaS (software as a service) application functionality by taking advantage of economies of scale, pay on usage, lower cost and re-use.
One such system, suitable for supporting industrial and enterprise data from a manufacturing, industrial or other enterprise, is shown in
The items identified (explicitly or implicitly) in
The items identified (explicitly or implicitly) in
The items identified (explicitly or implicitly) in
The items identified (explicitly or implicitly) in
These may be implemented in micro-servers or other computing apparatus of the type available in the marketplace as adapted in accord with the teachings hereof—see
The items identified (explicitly or implicitly) in
The items identified (explicitly or implicitly) in
The items identified (explicitly or implicitly) in
Micro-services provide the ability to distribute data logic, API's, algorithms and application features between edge cloud services and public/private cloud hosted applications and analytics Micro-services are registered, managed and scaled through the use of a PaaS (Platform as a Service) components within the NAUTILIAN™ platform. In systems according to the invention that employ it, the micro-services architecture provides the following advantages over the traditional service oriented architecture:
The benefits of the micro services architecture for Industry 4.0 approach include:
The same version of the NAUTILIAN™ software running in the main cloud platform (e.g., Amazon's AWS service or Microsoft Azure) also executes local to the plant in a microserver-based Cloud in a Box (or in other computing apparatus local to the plant). The cloud instance of Edge Cloud samples data in sub-seconds time intervals and can handle data generated in frequencies of MHz or GHz. The local Cloud in a Box instance samples in milliseconds, and has ‘store and forward’ capabilities if connectivity is lost to the main cloud instance, hereinafter occasionally referred to as “Edge Cloud” or the like. Edge Cloud Services in AWS or MS Azure public or private cloud aggregates, filters and standardizes data from local Edge Cloud instances, e.g., at different locations in plant and/or in different plants. Edge cloud services hosted on the cloud-in-a-box can ingest data at Giga hertz speeds (streaming) from industrial assets such as a turbine in test mode, and provide local analytics to identify and predict potential performance issues.
Edge Cloud services provide for standardization, aggregation, learning through the PARCS™ engine and filtering of data from industrial devices. There exists the ability to store and forward data from the Edge Cloud to public or private cloud instances based on availability of network connectivity, bandwidth, latency and application/analytical needs. Equally the ability to deploy analytical models and applications developed in the main cloud (public or private) to the cloud-in-a-box (bi-directional) is also possible.
Public or private cloud (main) hosted Edge Cloud software services can manage thousands or more of industrial assets, plant and manufacturing site instances—standardization, aggregation, learning and filtering of site data, as suggested in
As above, same version of the SaaS (software as a service) Industrial Performance Applications and analytics running on the public or private cloud as local Cloud in a Box instance and augmented with data from SAP ERP or other business systems or social media networks to supplement production information. Site level industrial performance applications for real time analytics (milliseconds) and aggregated site manufacturing line analysis (standalone or connected modes).
Industrial protocol translator from proprietary industrial equipment and PLC manufacturers to OPC (Open Protocol Communications) via the installation of OPC UA client and server software hosted both in the Cloud in a Box and the public/private cloud configurations provides the ability to connect to proprietary vendor specific protocols, ingest data and apply standards and learning machine (via PARCS™) to proprietary data formats. The OPC UA client is configured with the Edge Cloud services to determine the frequency of data collection from industrial assets and PLC systems and provide edge to main cloud connectivity.
Each cloud-a-box instance running OPC UA and edge cloud services connects back to the public and or private NAUTILIAN™ in the same way and all data is keyed by Site Identifier (tenant)
Provides consolidated view across all industrial plants and manufacturing sites, including integration with business systems such as SAP, Oracle ERP or IBM Maximo. Can be configured to group sites by tenant, asset, asset type, region, product lines, and/or manufacturing lines.
SaaS Industrial Performance applications and analytics are shown in
The use of open source software technologies (predominately Apache Kafka, Apache Spark and Cassandra) to consolidate data from multiple sites—either in real-time or on a batch basis.
To aggregate data from multiple sites within one database schema for sites, assets and customers through the use of Tenant-ID's per asset allows for segmentation and isolation of tenant data, ability to add Blockchain keys to tenant data to uniquely identify source data and location. Information on tenant and asset utilization is integrated in the billing engine service (see
The illustrated system (a/k/a the QiO NAUTILIAN Platform) uses the Edge Cloud Engine for data ingestion. Data ingestion is the process of obtaining, importing, learning and processing data for later use or storage in a database. This process often involves connectivity, loading and application of standards and aggregation rules. Data is then presented via API's to application services. In built learning engine (PARCS™) automates the time to map data and apply intelligence to the underlying data structures.
Edge Cloud Engine data ingestion methodology systematically to validate the individual files; transform them into required data models; analyze the models against rules; serve the analysis to applications requesting it.
A UML diagram for an Edge Cloud implementation according to one practice of the invention is shown in
Building blocks for such a system include open source and other big data technologies, all adapted in accord with the teachings hereof. For example, data was loaded onto secure ftp folders within the public or Private cloud. Edge cloud services according to the invention were written to pre-process the data, sequence Apache Spark jobs to load the data into Big Data stores such as Cassandra and Hadoop (HDFS).
More generally, Edge Cloud Services are the ingestion endpoint of QiO's NAUTILIAN™ Platform. In some embodiments, uses HDFS and or Cassandra to store data in distributed fashion; Apache Spark for high speed data transformation and analysis; Cassandra for efficient storage and retrieval of Time Series Data. Cassandra also allows data storage for complex lookup structures; and/or Apache Kafka is used for defining routing rules and weaves all technologies together to allow interoperability, synchronicity and order.
Billing Engine
The billing engine serves as the general purpose metrics calculator for the entire platform with principal responsibility of providing feedback to the NAUTILIAN platform architecture for optimising resource utilisation and also provide a framework for charging the tenants based on usage of platform services. For such an optimisation it computes and reports the overall utilisation of resources consumed, referred to as Asset Use Model. The integration of the Billing Engine with Syniverse (a leading mobile roaming telecom services provider) provides the ability to leverage Syniverse's software services to generate usage based pricing (akin to data plans on a cell phone) per client, per asset on a global basis. The above billing service and integration with Syniverse can occur at the edge or on a remote cloud.
Referring to
Log Aggregator: This component reads ingestion, API and cloud billing logs and converts them into statistics that can be used readily to generate the Utilisation Report.
Invoice Generator: This component reads a billing configuration (which is very simple that says total cost of processing+storing per KB data is $xxx—broken into several sections—for a specific subscription) and creates an invoice based on attached excel template below:
Predictive Analysis (PARCS™) engine: This component is responsible for forecasting the subsequent month usage by a particular tenant and asset to ensure capacity, service and quality are maintain proactively. In the table, the estimation is same as the current month's utilisation, although, that is not necessarily the case in most circumstances.
The cost incurring components are placed to the right of the following mind map whereas the chargeable components are placed on the left in mind-map of
Representative source code for an embodiment of the billing engine follows:
Cloud Edge Engine is a set of services that can be deployed rapidly on any cloud compute infrastructure to enable collection, processing, learning and aggregation of data collected from various types of equipment and data sources. Cloud Edge Engine pushes the frontier of QiO Platform-based applications, data, analytics and services away from centralized nodes to the logical extremes of a network. The CEE enables analytics and knowledge generation to occur at the source of the data.
The REST interface of Cloud Edge Engine exposes a configuration service to configure the usage. Configuration includes the type of data source, the protocol used for connection, and security information required to connect to that data source. Configuration also includes metadata that is used to understand data from the data source.
Connection Endpoint is used for connecting to the data source as per configuration set. The endpoint is a logical abstraction of Integration interfaces for the Cloud Edge Engine and it supports connecting to relational, NoSQL and Batch Storage systems. It can also connect to social data sources like Twitter and Facebook. It can also connect to physical equipment generating data over a variety of protocols including, but not limited to, SNMP and MQTT.
Apache Kafka is a fast, scalable, durable and distributed publish subscribe messaging system. It is used in Cloud Edge Engine to handle ingestion of huge streams of data. This component receives live feeds from equipment or other data generating applications.
Cassandra and/or HDFS provides high throughput access to application data and are used for storage of raw datasets that are required to be processed by the Edge Engine. Cassandra is highly fault-tolerant and designed to be deployed on low-cost hardware. Using Cassandra a large file is split and distributed across various machines in Cassandra cluster to run distributed operations on the large datasets. Synchronization of Cassandra data nodes at the edge and with public/private cloud nodes guarantees no data loss.
Edge Cloud Engine uses Apache Spark for high speed parallel computing on the distributed datasets or data streams—enabling the implementation of the LAMBDA architecture (in memory and batch data processing and analytics). Apache Spark is used for defining series of transformations on raw datasets and converting them into datasets representing meaningful analysis. Moreover Edge Cloud uses Apache Spark to cache frequently needed data.
Edge cloud uses Cassandra to store the Master Datasets, time series datasets and analysis results for faster access from applications needing this data. Being master less, Cassandra has no single point of failure and once the Edge Cloud Engine stores data into Cassandra, it remains highly available for the applications.
Interfacing Edge Cloud Engine with Other Services
Discussed below are techniques for interfacing the Edge Cloud Engine with other services.
Apache Kafka is used for defining routing rules and weaves all technologies together to allow interoperability, synchronicity and order.
During the data standardization phase of the ingestion process, each raw data record is published to the Kafka Topic “INGESTION_RAW_DATA” with the following format:
The raw data record is then mapped and transformed into a standardized record.
A JSON message is then formed with the foregoing plus missing parameters and send it to a “Batch Streaming” process step, after all the raw data lines for all parameters of an asset for a specific timestamp have been processed and standardized. This is a pivoted standardized message.
It is possible that the asset data points for a specific timestamp are spread across two or more .dat files within a customer file—a .zip file. This process step ensures that the data from all the files is obtained before forming the pivoted standardized message for the asset/timestamp combination Batch Streaming
The Batch Streaming process step publishes all pivoted standardized messages to a single Kafka Topic called INGESTION_PIVOTED_DATA as Keyed Messages, where the Key is the asset ID string.
The Storage microservice as well as the Analytics service are consumers of that Kafka topic.
When it is done with all the data from the file, it logs step status and completion date under the file log via the Ingestion Logs service—status “Data ingested to Kafka”.
Pivoted Standardized Messages can include the following fields
The edge cloud machine is set of services that can be deployed on any cloud compute infrastructure to enable collection, processing and aggregation of data collected from various types of sensors. The sensor data can be actively pushed using RESTFul service/AMQP (Advanced Message Queueing Protocol)/MQTT (MQ Telemetry Transport protocol) to the edge cloud machine. In scenarios where active push is not practical the services can be configured to poll sensor data using SNMP/MODBUS protocols. The collected data is saved to a common access Cassandra data store.
Edge cloud machine primarily consists of three interdependent services viz.,
Referring to
To support active data push using Apache Kafka, AMQP or MQTT or REST interface, Apache ActiveMQ is used. It the most popular and powerful open source messaging and Integration Patterns server. Apache ActiveMQ was chosen for implementing the data push considering the requirement of supporting lightweight clients as the sensor data adaptors would be.
The Edge Gateway Services exposes a queue with name “SensorDataQueue”. For supporting AMQP a broker needs to be configured as
For enabling communication over MQTT following configuration is needed in the broker configuration file
For communicating over REST simply use the http POST method like
To enable data polling the Edge Gateway Service can be configured using a configuration message. This message is sent to the Edge Cloud Machine from the Data Access API.
Edge Data Routing service routes the data collected by the data gateway service to a persistent datastore and timestamps it by tenant and asset. The service also tests the possibility of generating event based on preconfigured rules or learnt rules from the PARCS™ engine. If the rule is satisfied the event is generated. This event is further enriched with the information available in rule configuration and time series data available in datastore.
The datastore is implemented using a Cassandra cluster. Cassandra is chosen for its features such as high availability, high scalability and high performance.
For routing Apache Camel is used in this example, but Apache Kafka can also be used. Apache Camel is used to define routing and mediation rules. Leveraging Java based route definitions to route messages internally in the Edge Cloud Machine. These routing rules enable the Edge Cloud Machine functional and operative. The rules dictate when to collect data, where to collect data from, how this data is transformed, aggregated processed and finally stored.
Referring to
Systemic Asset Intelligence' across products, product systems and ecosystem. In other words, the ability to seamlessly connect, integrate, secure and drive business outcomes in real time using both human generated (ERP, SCM, CRM, Social Networks etc.) and machine generated data (engines, turbines, compressors etc.). Creating outcomes that cut across horizontal and vertical value chains as well as time horizons (past, present and future). Developing cloud-native, data science-driven, collaborative applications that enable the improvement of safety, optimization of operations and inventories, the guaranteeing of customer service times, and create dynamic pricing models based on product usage patterns.
Described below the systemic asset intelligent model framework based on the automated collection and processing of data in a system according to the invention. The sources of information, proprietary or not, are accessible through connected assets and systems. The processing of this information is done through cloud-based ‘Big Data’ approaches and data science services. The SAI model framework tracks different variables of assets related to performance, availability, reliability, capacity and serviceability (PARCS™)—attributes any industrial asset will either generate or create within a product system. These variables correlate with each other and can predict the health and behavior of an Asset. Based on the prognostic information, a predictive model can be constructed to decide assets optimal performance, maintenance and warranty management cycles and performance. The model outputs can be integrated into application services to enable devices to achieve near-zero downtime.
System components suffer wear with usage and age as a deterioration process, which causes low reliability, poor performance and—potentially—huge losses to their owners, especially if they are part of large and complex industrial systems. Therefore, risk assessment, maintenance and warranty management are important factors in keeping devices in good operation, both to decrease failure rates and increase performance.
Asset manufacturers often face the problem of being responsible for provision of products with service level agreements. Failure eradication is then a problem for the manufacturer—not a trivial task if the product or service is being provided as part of a large system with complex interactions. The common protocol to deal with Asset breakdown is to investigate notifications from the customer and give recommendations to carry out typical and easy checks. If the fault is not rectified then onsite diagnosis and fixing of devices is carried out by maintenance experts. This asset repair supply chain process is typically reactive, slow, tedious and costly. The most important aspect is cost associated with device down-time. Failure-based maintenance, scheduled maintenance and preventive maintenance models are positive and efficient but how to decide any maintenance interval is crucial task where these traditional models are not effective.
The optimal performance of any depends on several dimensions such as Performance, Availability, Reliability, Capacity and Serviceability aka PARCS™—which are highly correlated. Individual and system asset health and behavior are governed by these dimensions. Traditional models and approaches are not capable of measuring and correlating these dimensions accurately and usually ignore them—due to the cost and infrastructure required to calculate all the permutations—with the use cloud technologies and big data technologies, these limitations are now removed.
Much to the contrary, an systemic asset intelligence model attempts to learn in advance—through connected assets, systems and ecosystems and cloud-based information systems—the prognosis for assets, predicting the likelihood of faults and preventing them through collaborative applications. The prevention of asset failure can dramatically reduce the serving cost of the repair, improve safety and increase operational performance from reduced down time.
The SAI model relies on its ability to collect all relevant information about connected asset, system, sub systems, ecosystem and then process and analyze that information, giving any recommendations/alerts/anomalies in real time. This ability to process the massive amount of asset data (Big Data) in real time using data science tools—and delivering customer feedback in real time—is innovative and game-changing. The formulation of the SAI model framework is likely to be expressed mathematically and statistically to comprehend different objectives and constraints. The SAI model is predictive, self-learning, agile and more cost-effective than traditional alternatives based on legacy software architectures such as Microsoft SQL or Oracle databases.
What can be Achieved with an SAI Model?
The aim of System Anticipatory Intelligence (SAI) is optimal performance whilst ensuring zero-downtime. This means the model attempts to predict the likelihood of any type of industrial asset downtime or asset performance anomaly.
SAI is to be achieved through a self-learning optimization process, i.e. one intended to obtain the maximum effectiveness of an Asset. This involves data being parsed (possibly at different frequencies) and then certain patterns being detected: an incident becomes known to the system. Then the system provides a response/recommendation and predicts the future occurrence of a certain event. SAI using the PARCS™ engine can occur at individual component level within an Asset (compressor), the Asset (Turbine), system level (two aircraft turbines or MRO facility) or ecosystem (all airlines with the similar turbine or suppliers of compressor parts), and over time horizons—past, present and future.
The SAI process is carried out by means of a self-learning optimization engine. The engine gathers the device data at their source, possibly from Assets in motion (e.g. airlines), through edge cloud services. The typically enormous size of the collected data justifies the use of the expression Big Data to refer to them. Both the detection and response are done through application services, which mean they are running at (external) service provider premises. Lastly the prediction is often presented in a graphical manner, also referred as visualization.
The platform of the SAI optimization engine can be rapidly deployed in a Model-View-Presenter (MVP), i.e. is a user's graphical interface showing the outcomes of the statistical models. Moreover, the SAI optimization engines are economically designed using appropriate technologies and adapted to the specific needs of the customers. The edge cloud potentially allows the collection of high frequency data which could be exploited in economically disruptive ways. The SAI optimization model is designed to help determine the condition of in-service assets in order to predict when maintenance should be performed. This predictive maintenance will be more cost effective compared with routine or time-based preventive maintenance (often seen in Annual Maintenance Contracts) because maintenance tasks are performed only when required. Also a convenient scheduling of corrective actions is enabled, and one would usually see a reduction in unexpected device failure.
This is possible by performing periodic or continuous equipment condition monitoring. The accurate prediction of future device condition trends uses principles of data science to determine what type and at what point in the future maintenance activities will be appropriate. This is part of reliability-centered maintenance (RCM) which emphasizes the use of predictive maintenance techniques. In addition to traditional preventive measures, RCM seeks to provide companies with a tool for achieving lowest asset net present costs (NPC) for a given level of performance and risk.
Thus, in the development of SAI optimization models we will end up looking at computerized maintenance management systems (CMMS), distributed control systems (DCS) and certain protocols like Highway addressable remote transducer protocol (HART), IEC61850 and OLE for process control (OPC).
Sources of data can include non-destructive testing technologies (infrared, acoustic/ultrasound, corona detection, vibration analysis, wireless sensor network and other specific tests or sources). As well as data sourced from IT/Enterprise systems such as SAP, Maximo, Oracle ERP and industrial systems such as SCADA and/or Historians.
The self learning optimization model discussed takes SAI to the next level by putting the service requirement prediction of the device under consideration in the context of the service environment in which it is operating.
SAI delivers the following:
The SAI self-learning optimization model attempts to identify and predict the likelihood of any potential reason for failure of a device. Consider the well-known bathtub curve (Smith, et al, “The bathtub curve: an alternative explanation,” Reliability and Maintainability Symposium, 1994. Proceedings, Annual, pp. 241-247) in
A major part in the normal function of an asset is regular maintenance to ensure the safe and reliable operation of equipment. Effective maintenance can be achieved ensuring a balance between the predicted needs and the PARCS™ parameters. The optimization model framework model—PARCS™—is shown graphically in
Salient features of PARCS™ model that enable SAI:
This is a demonstration of the SAI optimization model using failure data of a device given in the table below. This is a simple data set used to illustrate how some “−abilities” are calculated. Events are put into categories of up-time and down-time for a device. Because the data lacks specific failure details, the up-time intervals are often considered as generic age-to-failure data. Likewise, the specific maintenance details are often consider as generic repair times.
To calculate the optimization model parameters for this Asset:
Availability deals with the duration of up-time for operations and is a measure of how often the system is alive and well. Availability is defined as
where
Using the data set provided in the table above, the availability of device is 98.6% based on MTBM=8205.3 hours and MTTR=112.5 hours.
Reliability deals with reducing the frequency of failures over a time interval and is a measure of the probability for failure-free operation during a given interval, i.e., it is a measure of success for a failure free operation. It is often expressed as
where X is constant failure rate and MTBF is mean time between failure (same as MTBM). MTBF measures the time between system failures.
The data in the table above shows the mean time between maintenance is 683.8 hours. If we want to calculate the device reliability for a period of one year (8760 hours). The device has a reliability of exp(−8760/683.8)=0.00027%. The reliability value is the probability of completing the one year operation without failure. In short, the system is highly unreliable (for a one year time) and maintenance requirement is high as the device is expected to have 8760/683.8=12.8 maintenance actions per year.
The above calculations for reliability were done by available historical data given in the table above. The more accurate predictions will be found by building a probability plot from the data in the table above. This probability plot shows the mean time between maintenance events is 730 hours.
Serviceability deals with duration of service outages or how long it takes to achieve (ease and speed) the service actions. The mathematical formulae is expressed as
where S is constant service rate and MTTR is mean time to repair.
Data in the table above shows mean down time due to service is 9.4 hours. If we want to calculate the device serviceability with an allowed repair time of 10 hours. The device has a Serviceability of 1-exp(−10/9.4)=65.5%. The serviceability value is the probability of completing the repairs in the allowed interval of 10 hours. Therefore, the device has a modest serviceability value (for the allowed repair interval of 10 hours).
The above calculations for serviceability were done using available historical data given in the table above. The more accurate predictions will be found by building a probability plot from the data in the table above. This probability plot shows the mean time to repair is 10 hours.
Thus, the SAI Optimization Model allows:
The SAI Optimization Model Is a holistic model which gives solutions for predicting and resolving failures/anomalies and/or performance issues.
Described below is the architecture of a smart device integration a key piece of capability for assets with smart sensors—sensors that are self discoverable, automatically connect to Wifi, Bluetooth, and ZigBee. These sensors will connect to IOT gateways and/or directly to Cloud in a Box appliances and communicate through Edge Cloud Services defined earlier.
The intention behind building this device, and a system according to the invention in which it is embodied, is to measure different gas levels in the atmosphere at different parts of geography & send all these measured variables & locations to Edge Cloud to be get transmitted over Internet where it can be analyzed and accessed through one URI. The weather at each location depends mostly on presence of these gases. Excess of these gases can cause pollution to environment & very serious harms to human being.
Here we decided to measure CO, CO2, NO, NO2, O3, PM10 & PM2.5 contents in PPM. A sensor for CO & LPG measuring was connected (MQ7—CO Sensor, MQ5—LPG sensor)—individual sensor modules were used had their own supply & analog output circuitry. Sensors were connected to a Raspberry Pi1 & Raspberry Pi2 module as gateway.
The sensors used in the illustrated embodiment include those described below.
MQ7 Sensor: CO Sensor (for Example from Sparkfun)
The illustrated smart device incorporates, as a microconverter module, an EVAL ADuC832 evaluation board available from Analog Devices.
The microcomputer utilized in the embodiment of
All of the sensors give an analog output proportional to amount of gas sensed in PPM. This analog signal can't be directly connected to edge cloud services or to the RPi board, since PC or RPi doesn't have their inbuilt analog to digital convertors (ADC). The interface required either external serial ADC or another convertor which will/can directly read these analog signals of multiple sensors & can give direct digital data in our required format to edge cloud services.
Here we have used a simple convertor micro-controller board of “Analog Devices” 7—the “EVAL ADuC832”. It is 89X52 core 8 bit microcontroller having inbuilt 12 bit ADC with 12 channels. I.e. allows the ability to connect at max 12 sensors to this board. This micro-convertor board with some program burnt in it will then select one sensor Chanel one by one sequentially & will read its output & will give direct digital read out at its serial terminal which then can be directly connected to the edge cloud services to display on via a visualization tool
The system of
The next part is a RS232 bridge, which acts as an interface between micro-convertor & microcomputer. The micro-convertor is sending data at baud rate of 9600 to external interface. This data is then given to RS232 pins (RXD & TXD pins) Raspberry Pi board (Pins 8 & 10).
In Raspberry Pi the operating system used was an Raspbian Wheezy as well as Snappy OS. Development via edge cloud services to connect and ingest data.
Possible modifications, features and other characteristics of the illustrated embodiment follow:
Components selected for the illustrated embodiment are all by way of example. E.g., any microcontroller with inbuilt ADC & UART can be used (e.g., LPC2148 micro-controller, which is a power full ARM-7 series of micro-controller). However, system integration and cost of device in customized equipment with compared to ADuC832 can be higher and programming more complex.
For Sensor assembly care should be taken of fixture design such that local air (The environment where sensor & unit is installed) should get flown on every sensor. Also Sensors should not get directly exposed to open environment such as direct rain, storm, flame or other hazardous conditions like electrical sparks etc.
Source of power is important—depending on whether power is from a battery or mains supply. It is suggested to keep power of system mains operated in normal operation mode & keep battery operation in case of mains power failure. Battery operation requires rechargeable batteries with its charging circuit. Also our main system of micro-converter & microcomputer requires very less energy (3.3V/300 mA=0.99 W==1 W approx.). But if you consider sensor part, each of sensors requires about 5V DC & about 50 mA to 100 mA of current. There for while designing a compact system a special care is required to be taken while designing its power supply section. A separate 3.3V & 5V power supply section is required with their different current requirements.
UART Bridge is preferred between micro-converter & microcomputer, since it gives a facility of debugging & checking output of micro-convertor unit.
This section outlines communication protocol between SmartDevice and the Edge Cloud. SmartDevice communicates with an Edge Cloud for archiving and analysing data. This data exchange can be of various types.
Terms Used with Description:
<packet SmartDeviceid=“SmartDevice 4” datetime=“27022015-12:40 PM” type=“ACTIVATE” passkey=“HASHED_KEY”></packet>
This packet will make a handshake between Edge Cloud & SmartDevice. Only after a handshake has happened Edge Cloud will start accepting data.
Format for sending sensor data as notifications to Edge Cloud for SmartDevice. Request id is optional in this format. If SmartDevice is posting data to Edge Cloud on request then it will contain request id. If its posting data on time intervals then it will be blank.
Format for sending sensor data as notifications to Edge Cloud for SmartDevice. Request id is optional in this format. If SmartDevice is posting data to Edge Cloud on request then it will contain request id. If its posting data on time intervals then it will be blank.
This format of packet is used to set function values of the SmartDevice. Here, the type of packet is “CONFIG”. Function elements will contain function id's & values to set. When this process of resetting will be completed SmartDevice will send empty packet with same id & type as “RESPONSE”.
Format sent from SmartDevice to Edge Cloud for asking Edge Cloud if there is anything for SmartDevice or not.
This format is sent from Edge Cloud to SmartDevice for querying sensors given in the packet. In response to this SmartDevice will send the following packet format with same id & type as “RESPONSE”.
Format sent from Edge Cloud to SmartDevice for querying all sensors present in SmartDevice. In response to this SmartDevice will send the following packet format with same id & type as “RESPONSE” with current data of all sensors.
The RESTful web service on Edge Cloud would exposes the following functions used for communication.
This function call will activate SmartDevice to start accepting data further. Unless the SmartDevice is in activated mode data will not be accepted. But before this SmartDevice should be registered into the system. The XML format needed for this is as below:
After invoking web service to activate SmartDevice, if successfully activated, it will return “OK” status code. Otherwise it will return “BadRequest”.
This function is used by the SmartDevice to post data to the system and generate notifications for respective SmartDevices features. Note that this data will be accepted to system if & only if SmartDevice is registered & SmartDevice is activated. All the packets with type “RESPONSE” should be posted to this function
This function is used by SmartDevice to fetch request or command xml from the Edge Cloud & process further according to that. Note that this function will return a xml string which be either to set configuration of the SmartDevice or to query sensor values.
If there is any error at edge cloud then edge cloud will reply as “edge cloud error”
Some embodiments of the invention provide a Sustainability Index feature, building on the PARCS™ model discussed above, to collect data across the supply chain, e.g., from the farmer to the retailer, and create a sustainability index that can then be shown on each consumer product to drive smarter buying habits. The analogy is the Energy Index shown on electrical products such as washing machines, to illustrate the cost of energy consumption per annum.
A benefit of the foregoing is to provide Industrial Engineers with a workbench for developing, collaborating and deploying reusable Systemic Asset Intelligence analytics and applications. Embodiments of the invention constructed and operated as discussed above and adapted for systemic asset intelligence (referred to below, as the “NAUTILIAN Foresight Engine”) comprise cloud-based software that supersedes legacy modelling tools such as Matlab and OSlsoft PI for Industrial Engineers to collaborate on data ingestion, asset models (pumps, compressors, valves etc.), analytical models (vibration, oil temperature, EWMA) using standard software libraries in R, Python, Scala etc. and a user interface where engineering communities can share, critique and deploy code to rapidly develop cloud native predictive applications. NAUTILIAN Foresight Engine is a toolkit, with open interfaces and a SDK (software development kit) for Engineers (physical sciences and computer science) to collaborate, and has the following key features:
Ingestion Manager: to connect, extract, filter, standardize and load data from any source (machine or human generated), at any frequency (streaming, snapshot, or batch);
Asset Discovery: to provide a default set of visualizations, parameters, manufacturer configurations and allow the user to define reusable mathematical functions, relationships and metadata;
User Profiler: ability to create user personas (roles and responsibilities) tied to organizational structure and relationships. Allowing the ability to control users and group access rights to view, modify and delete;
Analytical/Machine Learning Framework (PARCS): for industrial and software engineers to write code in Java, R, Scala, Python etc. creating analytics that monitor & predict the behaviour of an asset, group of assets or system over time periods, and generate confidence indices and diagnostic networks to validate the accuracy of the analytical models;
Insight Manager: to visualize, share and distribute charts to review and get feedback. Analytics generated as anomalies can be reviewed, commented on and tracked across engineering teams. Workflows can be configured to route specific anomalies to engineering teams and feedback captured.
At the core of Foresight Engine is PARCS, providing a multi-dimensional view of any industrial system and the interconnections to systems. Providing a Digital Twin of the physical asset, through logical data definitions and parameter configurations.
The NAUTILIAN™ Platform provides manufacturing and industrial customers with a software framework of open services to create industrial agility, where engineers can experiment, rapidly test mathematical models and develop smart applications. NAUTILIAN™ is a horizontal platform based on open-source technologies and is cloud neutral.
Foresight Engine is deployed on NAUTILIAN Platform as set of microservices.
An overview of the NAUTIAN Platform architecture is shown in
Kubernetes is used to provide cloud neutrality and deploy NAUTILIAN Templates and applications anywhere. Docker images are used to deliver stateless and stateful microservices as containers.
Kubernetes Helm is used to provide installation scripts (Helm Charts) and offer a catalog of all components and application templates. The catalog is stored on Artifactory together with all Docker images used by the charts.
httpslidocker.qiotec.com:5555 is QiO's official Docker repository protected by secure layer.
Provide the following functionality:
Consists of:
Support the Oauth2 Standard, and JWT standard implementations.
Provides integration to physical devices and sensors to extract, load and transform (ELT) time series data at speed and low cost, apply standards, and aggregate data at the edge. Edge Services support communication to various protocols such as BacNet, Modbus, Hart, etc., and convert proprietary protocols into standards such as OPC UA (Unified Architecture).
Integration with Blockchain (Guardtime KSI) provides digital asset identity services and validation of asset integration.
Microservices architecture and the associated application development refers to building software as a number of small independent processes which communicate with each other through language-agnostic APIs. The key is to have modular blocks which focus on a specific task and are highly decoupled so they can be easily swapped in and out rapidly with no detrimental effect.
The independent application features and functions, and APIs are self-contained, can be re-used and monitored across applications, and enable functionality to be scaled at a granular level.
The implementation of microservices follow these principles:
All microservices must be highly available and elastic so that they can scale up and down. For instance, Kubernetes uses the concept of replica sets to maintain a specified number of instances of a particular service to maintain availability and resiliency and Nautilian services leverage this functionality.
Kubernetes provides this capability with liveness (indicate when to restart a container) and health (readiness to start accepting requests) checks. When liveness or health checks run, and they find that a particular service is not in a healthy state, the service will be killed and restarted. Combined with replica sets, Kubernetes will restore the service to maintain the desired number of replicas of a particular service. Nautilian provides the tooling for enabling liveness and health checks by default when services are deployed.
When dependent services, e.g. other microservices, databases, message queues, caches, etc., start to experience faults, the impact of the failure needs to be limited in scope to avoid potential cascading failures. At the application level, tools, such as Netflix Hystrix, provide bulkheading to compartmentalize functionality in order to:
From the domain perspective, the service must be able to degrade gracefully when downstream components are faulting. This provides the benefit of limiting the blast radius of a faulting component, but how does a particular service maintain its service level? The use of
Hystrix enables providing fallback methods and workflows to allow a service to provide some level of service, possibly at a degraded level, in the event of dependent service failures.
Prove the System has been Designed for Failure
When a system is designed with failure in mind and able to withstand faults, a useful technique is to continuously prove whether or not this is true. Nautilian provides a tool that can access Kubernetes namespaces in environment, up to and including production, and randomly kill pods with running services. If a particular service was not designed to be able to withstand these types of faults, the Chaos Monkey tool will quickly provide that feedback.
Services are implemented to define a logical set of one or more pods to provide resiliency and elasticity for a particular microservice. Due to scaling requirements, resource utilization balancing, or hardware failures, pods related to a microservice can come and go. Service discovery enables the dynamic discovery of pods to be added, or removed, from the logical set of pods that are supporting the implemented service.
The default way to discover the pods for a Kubernetes services is via DNS names.
For a service named foo-bar, the host name foo-bar might be hard coded in the application code.
For example, to access an HTTP URL use http://foo-bar/ or for HTTPS use https://foo-bar/(assuming the service is using the port 80 or 443 respectively). Or, if a non-standard port number is used, e.g. 1234, then that port number is appended to the URL such as http://foo-bar:1234/.
DNS works in Kubernetes by resolving to the service named foo-bar in the particular Kubernetes namespace being accessed where the application services are running. This provides the added benefit of not have having to configure applications with environment specific configuration and protects from inadvertently accessing a production service when working in a test environment. This also allows the application to be moved, i.e. its Docker images and Kubernetes metadata, into another environment and work without any changes.
When there is more than one pod implementing a particular service, Kubernetes service discovery automatically enables load balancing of requests across the related pods. To expose these services, such as APIs and UIs, the Rancher Kubernetes ingress load balancer provider will be used.
To properly capture logs, when microservices are written, developers should:
Capturing historical metrics is essential to diagnose issues involving microservices. These metrics are also useful for auto scaling of services based on load.
Nautilian uses Prometheus as the back end storage service and REST API to capture metrics, and then Graphana is used as the console to view, query, and analyse the metrics.
Each microservice will implement metrics capture, and reporting.
For microservice names and locations, Kubernetes service discovery will be used.
With respect to sensitive information, such as passwords, ssh keys, and OAuth tokens, Kubernetes secrets will be used rather than storing this type of information in a pod definition or in a docker image.
Used to create reusable APIs to access source and target systems and applications without direct point to point interfaces. Includes ability to monitor the performance and usage of APIs per application and system usage.
Consists of:
Ability to publish standard integration message, route to subscribers, process contributions by subscribers, integrate with workflow services and complete business event/transactions.
Consists of:
Provides the ability to create, test and deploy workflow rules and agents to simplify business processes, data validation and automate user actions based on business rules and configurations. And, to monitor performance of workflow rules and configurations.
Consists of:
Provides an integration toolkit for accessing batch, real time and near real time data—cleaning the data, reformatting, and integration with other applications.
Consists of:
Referring to
Menu of catalog services with service levels, pricing and default configurations that allows a PaaS admin to select standard services and deploy these for a customer tenant with minimal manual intervention and direction.
Catalog: List of PaaS services per customer tenant(s) to provision Data, LAMBDA, Asset and Analytical services. Each service has a service owner, price and SLA
Billing: The ability to monitor consumption by tenant and asset on a real time basis for all services consumed, and the ability to then automatically generate an invoice for payment. Tracking of payment against services, ability to accept payment by PayPal, Credit Card or Purchase Order.
Consists of:
Provide the ability to connect to different data sources with multi-tenancy at the asset and tenant level, with varying time horizons (milliseconds, seconds to snapshots), and to extract, transform and load the data into structured and non-structured databases. The consumption of data loaded into big data technologies, such as Cassandra and Hadoop, are provided via direct access tools, such as Hive, BI tools, and via RESTful API's.
The following provides an overview of the data services technologies employed in the Nautilian™ Platform.
Hadoop Filesystem used for fault tolerant distributed storage of large volumes of all types of data.
Used for metadata and transactional data storage. Hive is used in conjunction with HDFS and provides a SQL-like query interface to Hadoop filesystems. MariaDB is a relational database that is used by the Hive metastore repository to maintain the metadata for Hive tables and partitions.
Additionally, MariaDB provides a relational SQL repository for transactional data.
Distributed document database storage using JSON-like documents that can allow the data structure to change over time. MongoDB is used predominantly for APIs.
Distributed DB for time series and large volume storage. Apache Cassandra is an open-source distributed NoSQL database platform that provides high availability without a single point of failure. Cassandra's data model is an excellent fit for handling data in Time Series, regardless of data type or size.
In-memory database for key value storage used for caching and fast access.
Distributed RESTful search engine for dealing with unstructured and semi structured data.
AWS S3 (Simple Storage Service) is an object based storage system with high durability that is used for archiving the incoming data ingestion feeds for reference.
Real-Time and Batch: LAMBDA Architecture
Provides the ability to simultaneously ingest real time streaming data and batch data, and to perform calculations and analysis in memory to provide outputs from one model to another in parallel while leveraging data in motion (in memory) and data at rest (data stores).
The Lambda Architecture aims to satisfy the needs for a robust system that is fault-tolerant, both against hardware failures and human mistakes, being able to serve a wide range of workloads and use cases, and in which low-latency reads and updates are required.
Consists of:
Assignment of cryptographic keys (via Guardtime Blockchain Keyless Signature Infrastructure—KSI) to create digital identities and ensure any device connection is provided with a KSI key to ensure trust of the device.
Assignment of KSI key to customer tenant data to ensure data resides in only approved and authorized cloud environments, any unauthorized access or movement of data outside of approved cloud environments is immediately known.
Allocation of KSI identity to PARCS™ score to allow the creation of Digital Register per asset and ensure complete traceability and governance of all asset data across Cloud instances and changes.
Rich UI interface allowing users to interact with visual charts, maps, videos, chat, presence, notifications etc. Visualize complex analytical charts and ability to change configuration/settings of charts provided.
Framework for Runtime Configurations and Customizations of all User Interfaces delivered via application templates.
Consists of:
Referring to
UI to load data sources (Realtime, Batch—Any data type, and any frequency) and normalize and standardize.
Utilizes NAUTILIAN Platform Edge Services and Integration services.
Consists of:
Provides a default set of visualizations, parameters, manufacturer configurations and allow the user to define reusable mathematical functions, relationships and metadata.
Consists of:
Visualize, share and distribute charts to review and get feedback. Analytics generated as anomalies can be reviewed, commented on and tracked across engineering teams. Workflows can be configured to route specific anomalies to engineering teams and feedback captured.
Consists of:
Used by industrial and software engineers to write code in Java, R, Scala, Python etc. creating analytics that monitor & predict the behaviour of an asset, group of assets or system over time periods, and generate confidence indices and diagnostic networks to validate the accuracy of the analytical models.
Utilizes NAUTILIAN Platform services for real time streaming and batch execution of machine learning (ML) algorithms, such as Spark ML, H2O, TensorFlow, etc.
Consists of:
The QiO solution provides re-usable application templates to accelerate the development of bespoke applications with all the scaffolding and best-practices of mobile-responsive web applications already baked in.
This allows rapid development of business ready applications, in production with low cost and good quality.
An example of an application template would be the Predictive Maintenance template which would be installed on Foresight Engine. Configuring the organizational structure and adding users through user management would provide the basic application framework to develop a Predictive Maintenance application that can be enhanced over time.
Consists of:
The PARCS™ scores are based on asset specific data including asset type, asset characteristics, sensor data, and historical log data. In principle, the goal is to have the PARCS™ architecture auto detect the asset type, read asset type characteristics from a database, and automatically identify and clean sensor data and log data. The functionality requires a significant amount of data for each asset, which is not always the case. Therefore, we will require user approval for some calculations. Furthermore, the use of ontology that relate asset types to one another so that we can map new data to related historical data used to train our models.
The asset type ontology is used to group together similar assets based on their features. Leveraging existing data to define reference states, i.e. statistical description of historical performance, reliability, etc. Then, the reference states can be used to normalize new data into a Z-score metric. The PARCS™ Z-score metrics can be applied even in cases when there are minimal amounts of data available. To build the asset type ontology, we leverage content from third party providers, such as Asset Performance Technologies (APT), which has over 600 assets described in terms of device function, preventative maintenance, failure causes, failure modes, and failure effects.
The PARCS™ score are complemented by further calculations that provide predictions and recommendations. First, there are data specific to assets that can provide further indication of a change in a PARCS™ score. For example, vibrational data can indicate if a motor has a greater chance of failure in the future. Therefore, the PARCS™ framework allows peripheral models to indicate future trends in performance, availability, etc. A recommendation engine will also be built to aid serviceability. By leveraging available data, we can indicate expected costs and time needed to perform corrective maintenance. Optimization algorithms will be used to minimize cost and time and optimize the maintenance of an asset by recommending optimized maintenance plans. The maintenance plans will be dynamically updated based on the data continuously collected from the assets as well as the factory environment.
In
Asset Value Calculator: This service(s) is used to apply the PARCS™ scores to additional contexts such as risk prediction, insurance/warranty models, and financial planning. These services are outside of the scope of PARCS™, although they are closely connected. The asset value calculators depend on external data sources that provide insight into additional contexts above.
An architecture illustrating the use of Foresight Engine, PARCS™ and the NAUTILIAN™ platform is covered below.
This system, depicted in
The process adopted is summarized below:
Control variables are defined as all variables that can be adjusted by the operator of an asset. Telematics data for the Asset per minute from sensors on the Asset and by aggregating the time series data over events and time.
Uncontrolled variables are defined as variables, such as environmental data such as outside temperature direction, that cannot be altered by the Operator of the Asset.
Involves the transformation and aggregation of controlled and uncontrolled variables. For example uncontrolled variable such as Wind direction (in degrees) in converted into unit vectors, to reduce data errors in analysis. In addition controlled and uncontrolled variables are aggregated per Asset Event (for example a shutdown or at start-up), using Apache SparkSQL interface and partitioning each unique event. Normalization of events and clustering is via the use of data science algorithms such as KDTree and KMeans. After aggregating the variables scatter plot diagrams are produced to validate results for the aggregation process.
The PARCS™ engine (as shown in
A code snippet of the clustering logic as an input into the PARCS™ model is provided below:
Although the discussion above focuses largely on practices of the invention in connection with enterprise-level plant and industrial monitoring and control (as well as autonomous vehicle operation and maintenance), it will be appreciated that the invention has application, as well, in health care, financial services and other enterprises that benefit from the collection and systemic anticipatory analysis of large data sets. In regard to health care, for example, it will be appreciate that the teachings hereof can be applied to the monitoring, maintenance and control of networked, instrumented (i.e., “sensor-ized”) health care equipment in a hospital or other health-care facility, as well as in the monitoring of care of patients to which that equipment is coupled. In regard to financial services, it will be appreciated that the teachings hereof can be applied to the monitoring, value estimation and PARCS-based expected life predication of networked, instrumented equipment of all sorts (e.g., consumer product, construction, office/commercial, to name a few) in a plant, office building or other facility, thereby, enabling insurers, equity funds and other financial services providers (and consumers) to estimate actual depreciation, current and future value of such assets.
Described above are systems and methods meeting the objects set forth previously, among many others. It will be appreciated that the embodiments shown in the drawings and discussed here are merely examples of embodiments of the invention, and that other embodiments incorporating changes to those shown here fall within the scope of the invention. It will be appreciated, further, that the specific selections of hardware and software components discussed herein to construct embodiments of the invention are merely by way of example and that alternates thereto may be utilized in other embodiments.
In view of the foregoing, what we claim is:
This application is a continuation of U.S. patent application Ser. No. 15/631,685, filed Jun. 23, 2017, which claims the benefit of filing of commonly-owned, United States Provisional Patent Application Serial Nos. 62/354,540, filed Jun. 24, 2016, and 62/356,171, filed Jun. 29, 2016, the contents of all which are incorporated herein by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
62356171 | Jun 2016 | US | |
62354540 | Jun 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15631685 | Jun 2017 | US |
Child | 16561940 | US |