Integration platforms allow organizations to design, implement, and deploy software systems that harness heterogeneous resources (e.g., applications, services, and data sources) from across an organization's technical landscape. A data record may traverse a data pipeline, for example from a source to a target/destination of the integration platform may undergo various transformations and exchanges between complex and disparate systems/resources. Lineage information (e.g., data lineage, etc.) for the data record may include processes/executions affecting a data record, for example, such as a source/origin of the data record, what happens to the data record (e.g., extraction of the data record from a source, transformation of the data record, loading of the data record to a target/destination, etc.), and/or where the data record moves throughout the integration platform over time. Determining lineage information for a data record within an integration platform is a hard and manual task. For example, during replication of a data record through complex and disparate systems of the integration platform, it may be possible to determine where the data record came from, but since a data pipeline may be defined by various complex and disparate systems/resources information describing how the data record has been transformed may be non-existent. Open-source data analysis tools do not provide data lineage capabilities. To understand the data lineage of a data record, a user must write code based on the data available. However, since information describing how a data record has been transformed may be non-existent, it may be impossible to determine where which systems/resources processed the data record and/or how it has been transformed. Thus, troubleshooting issues related to the data record and/or determining diagnostics/metrics for the data record is challenging.
The accompanying drawings, which are incorporated herein and form a part of the specification, illustrate embodiments of the present disclosure and, together with the description, further serve to explain the principles of the disclosure and to enable a person skilled in the arts to make and use the embodiments.
The present disclosure will be described with reference to the accompanying drawings. In the drawings, like reference numbers indicate identical or functionally similar elements. Additionally, the left-most digit of a reference number identifies the drawing in which the reference number first appears.
Provided herein are system, apparatus, device, method, computer program product embodiments, and/or combinations and sub-combinations thereof, for determining metrics for data records. For example, the system, apparatus, device, method, computer program product embodiments, and/or combinations and sub-combinations thereof, may be used to determine, for a given data record, data indicative of what upstream sources and/or downstream assets are affected as the data record traverses a pipeline configured within the technical landscape and/or infrastructure of an organization, business, and/or operating entity, who/what are is generating the data, and who/what is relying on the data for decision making.
The technical landscape and/or infrastructure of an organization, business, and/or operating entity may incorporate a wide array of applications, services, data sources, servers, resources, and/or the like. Applications in the landscape/infrastructure may include custom-built applications, legacy applications, database applications, cloud-based applications, enterprise-resource-planning applications, and/or the like. The applications in the landscape and/or associated data may be configured with/on different devices (e.g., servers, etc.) at different locations (e.g., data centers, etc.), and/or may be accessed via a network (e.g., cloud, Internet, wide-area network, etc.). Additionally, the organization, the business, and/or the operating entity may be in communication with and/or connect to a plurality of third-party systems, applications, services, and/or APIs to access data and incorporate additional functions into their technical landscape/infrastructure.
An integration platform may allow users to create useful business processes, applications, and other software tools that will be referred to herein as integration applications, integration scenarios, or integration flows. An integration flow may leverage and incorporate data from the organization's disparate systems, services, and applications and from third-party systems. An integration platform may bridge divides between these disparate technical resources by centralizing communications, using connectors that allow integration flows to authenticate and connect to external resources, databases, Software-as-a-service (SaaS) applications, and incorporate data and functionality from these external resources into an integration flow.
In some instances and/or use cases, an organization, business, and/or operating entity may use a data pipeline, such as an extract, transform, and load (ETL) pipeline, that aggregates data from disparate sources, transforms the aggregate data, and stores the data in a data warehouse, relational data store, and/or other destination for reporting, analysis, or other client-facing applications. As the data item travels it may be replicated and/or transformed to standardize it, or used in calculations to generate other data records that enrich an overall data environment. As data is replicated within the integration platform, it may be not possible to determine where the data came from or how the data has been transformed. If an organization, business, and/or operating entity does not know where their data comes from or goes, they have uncontrolled environments within which it is very difficult to extract value from data. If an organization, business, and/or operating entity cannot extract value from data, troubleshooting issues related to the data and/or producing useful diagnostic information (e.g., quality control and/or metric information, versioning information, etc.) for a data record is challenging. For example, to understand how a particular data record is processed (e.g., replicated, transformed, etc.) by the data pipeline and/or to determine the cause of a fault in the processing, a developer must create custom code to request data/information describing the processes (e.g. replication, aggregation, filtering, etc.) executed on the data record over time, which may become arduous.
In some instances and/or use cases, an organization, business, and/or operating entity may use data visualization tools (e.g., software, etc.) that allows developers to analyze separate and/or discrete components of data. However, these visualization tools do not enable developers to navigate the data lineage to determine the changes applied to data and the relationships of the data at the same time. For example, no visualization software enables the developer to see, for a particular entity, all the fields that are transferred to a target and how each transformation is applied over it, such as merge operations, custom operations, and/or the like. Again, a developer must create custom code to see, for a particular entity, all the fields that are transferred to a target and how each transformation is applied over it—which, if even possible, may become arduous.
In some instances and/or use cases, an organization, business, and/or operating entity may use data pipelines, such as extract, transform, and load (ETL) pipelines, that aggregate data from disparate sources, transforms the aggregate data, and stores the data in a data warehouse, relational data store, and/or other destination for reporting, analysis, or other client-facing applications. As data is exposed and/or processed through the pipelines, it may be subjected to disparate data quality qualifiers and/or metrics output by disparate APIs, systems, and/or the like. An organization, business, and/or operating entity is unable to effectively measure data quality based on disparate forms and/or values of metrics exposed from pipelines (e.g., ETL pipelines, etc.).
Accordingly, a need exists for systems and methods that provide a non-coding solution for determining detailed lineage information and metric information for data records within an integration platform. The methods and systems described herein define a complete data pipeline process, including each transformation and data schema involved during the process. Data and/or metrics may be determined/collected at each step of a data pipeline process, and the data and/or metrics may be used to determine, for each entity and field, the source/origin and how it was transformed based on which transformation was defined by each. Metrics collected may also be transformed to customer-facing metrics that allow an end-user to analyze the maturity and quality of data.
Data sources 102 (e.g., data source 102A, data source 102B, data source 102C, etc.) may include an application programming interface (API) and/or any other technical resource. Although only three data sources 102 (e.g., data source 102A, data source 102B, data source 102C, etc.) are shown in
According to some embodiments, one or more of the data sources 102 may be and/or include data storage mediums (e.g., data lakes, data silos, data buckets, virtual storage, remote storage, physical storage devices, relational databases, etc.) of any type/form and configured to store data in any form and/or representation, such as raw data, transformed data, replicated data, semi-structured data (CSV, logs, XML, etc.), unstructured data, binary data (images, audio, video, etc.), and/or the like.
According to some embodiments, one or more of the data sources 102 may be resources that are not APIs or storage mediums. For example, according to some embodiments, data sources 102 may include any appropriate data source that may be modeled using a dialect (e.g., a set of keywords and semantics that can be used to evaluate a schema, etc.).
Data targets 104 (e.g., data target 104A, data target 104B, data target 104C, etc.) may be any type of API, technical resource, and/or system to be included in an integration flow. Although only three data targets 104 (e.g., data target 104A, data target 104B, data target 104C, etc.) are shown in
Although the data sources 102 are shown in
The integration platform 110 may be and/or include a system and/or software platform configured to access a plurality of software applications, services, and/or data sources. The integration platform 110 may be configured to design, maintain, and deploy integration flows based on the disparate software applications, services, and/or data sources. For example, the integration platform 110 may include/incorporate an enterprise service bus (ESB) architecture, a micro-service architecture, a service-oriented architecture (SOA), and/or the like. According to some embodiments, the integration platform 110 may allow a user to build and deploy integrations that communicate with and/or connect to third-party systems and provide additional functionalities that may be used to further integrate data from a plurality of organizational and/or cloud-based data sources. The integration platform 110 may allow users to build integration flows and/or APIs, and to design integration applications that access data, manipulate data, store data, and leverage data from disparate technical resources.
The integration platform 110 may include a design module 112, a runtime services module 114, connectors 116, a data bridge module 118, a versioning module 120, and data visualization module 122.
The interface module 112 may allow users to design and/or manage integration applications and integration flows that access disparate data sources 102 and data targets 104. The interface module 112 may standardize access to various data sources, provide connections to third-party systems and data, and provide additional functionalities to further integrate data from a plurality of organizational and/or cloud-based sources. The interface module 112 may include a graphical design environment and/or generate a graphical user interface (GUI) that enables a user to build, edit, deploy, monitor, and/or maintain integration applications. For example, the interface module 112 may include a GUI that may be used to define a complete data pipeline process, including each transformation and data schema involved during the process. During the definition of the pipeline, a user/developer may customize/personalize how data from a source (e.g., the data sources 102, etc.) will arrive and/or end up at a target (the data targets 104, etc.). Customizations and/or user preferences may be stored, for example, as a script for an expression language and/or programming language designed for transforming data, such as DataWeave and/or the like, and the entities and fields affected by the mutations for a given schema may be tracked.
The interface module 112 may communicate with a data visualization tool to cause display of visual lineage information for a data record detailing each transformation and/or the like of the data record from a data source 102 to a data target 104. As described later herein, once a pipeline has been defined, lineage information about the data (e.g., a genealogical tree, a data graph, etc.), version, and transformation may be stored, for example as metadata (and/or the like). The metadata may be associated with records/information produced/output when the pipeline runs. The lineage information may be stored and accessed at any time to determine how changes to a data record are related and how the lineage for the data record evolved.
The runtime services module 114 may include runtime components for building, assembling, compiling, and/or creating executable object code for specific integration scenarios at runtime. According to some embodiments, runtime components may create interpreted code to be parsed and applied upon execution. In some embodiments, runtime components may include a variety of intermediary hardware and/or software that runs and processes the output of integration flows. The runtime services module 114 may provide a point of contact between the data sources 102, the data targets 104, and the data bridge module 118. The runtime services module 114 may also include various system APIs.
The connectors 116 may provide connections between the integration platform and external resources, such as databases, APIs for software as a service (SaaS) applications, and many other endpoints. The connectors 116 may be APIs that are pre-built and selectable within the interface module 112, for example, using a drag-and-drop interface. The connectors 116 may provide reliable connectivity solutions to connect to a wide range of applications integrating with any other type of asset (e.g., Salesforce, Amazon S3, Mongo Db, Slack, JIRA, SAP, Workday, Kafka, etc.). The connectors 116 may enable connection to any type of API, for example, APIs such as SOAP APIs, REST APIs, Bulk APIs, Streaming APIs, and/or the like. The connectors 116 may facilitate the transfer of data from a source (e.g., the data sources 102, etc.) and a target (e.g., the data targets 104, etc.) by modeling the data into a file and/or the like, such as separated value files (CSV*, TSV, etc.), JavaScript Object Notation (JSON) text files delimited by new lines, JSON Arrays, and/or any other type of file. The connectors 116 may be responsible for and/or facilitate connecting to the data sources 102 and the data targets 104, authenticating, and performing raw operations to receive and insert data. The connectors 116 may support OAuth, Non-Blocking operations, stateless connection, low-level error handling, and reconnection.
The data bridge module 118 may be configured to receive/take data from a data source (e.g., the data sources 102, etc.) and replicate to a data target (e.g., the data targets 104, etc.), normalizing the data and schema (e.g., schema determined based on custom pipeline definitions, etc.). For example, the data bridge module 118 may support modeling of any API (and/or source otherwise capable of being modeled using a dialect) as an entity-relationship model. The data bridge module 118 may create and store a relational model based on raw data retrieved from a data source (e.g., the data sources 102, etc.) and translate received raw data to the relational data model. The data bridge module 118 may include software that translates data from a source (e.g., the data sources 102, etc.) into an entity-relationship model representation of the source model. The data bridge module 118 may also facilitate the use of data virtualization concepts to more readily interact with analytics and business intelligence applications, such as that described below with reference to data visualization tool 122. In this regard, the data bridge module 118 may create a relational model in a format that allows analytics and business intelligence tools to ingest/view the data.
The data bridge module 118, for example, during a replication process, may apply deduplication, normalization, and validation to the derived relational model before sending the results to a target destination, which may be a data visualization tool or another data storage location. The data bridge module 118 may employ the connectors 116 to authenticate with and connect to a data source 102. The data bridge module 118 may then retrieve a model or unstructured data in response to an appropriate request. According to some embodiments, the data bridge module 118 may include a data bridge adapter 200 to move data from a data source in data sources 102 to a data target in data targets 104 while applying data and schema normalizations.
As illustrated in
The data bridge adapter 200 may perform data virtualization by using definitions in a dialect file (described below as dialects 202) and an expression script (described below as expression scripts 204). With an appropriate dialect (e.g., a set of keywords and semantics that can be used to evaluate a schema, etc.) and an appropriate script selected based on the type of the data source, the data bridge adapter 200 may programmatically build an entity-relationship model of data received from a source (e.g., the data sources 102, etc.).
Returning to
According to some embodiments, dialects of the dialects 202 may specify a format/model of a particular data pipeline configuration. As described, the dialects 202 may be written using AML definitions, and an AML processor tool can parse and validate the instances of the metadata document. For example, an example pipeline configuration dialect is as follows:
The expression scripts 204 may be written in an expression language for accessing and transforming data. For example, the expression scripts 204 may be written in DataWeave expression language and/or the like. According to some embodiments, the expression scripts 204 may be written in any programming, expression, and/or scripting languages. The expression scripts 204 may parse and validate data received from a source according to a dialect in dialects 202. The outcome of this parsing may be, for example, a JSON document and/or the like that encodes a graph of information described in the dialect. As with dialects 202, a unique script may be created in expression scripts 204 for each API design methodology. Thus, an expression script may exist for WebAPI, one for Salesforce, etc. The expression scripts 204 may serve to transform the source model received from the API into the adapter model as defined by the associated dialect in dialects 202. The expression script may move the data from the responses received from the API to the entity-relationship model. According to some embodiments, the expression scripts 204 may be provided by integration platform 110 as stock functionality for a finite list of APIs and thus operate behind the scenes to perform the needed transformations. According to some embodiments, the expression scripts 204 may be extensible and/or customizable by particular customers to meet particular needs.
The connectivity configuration 206 may provide the services for handling the connections to various sources and targets (e.g., the data sources 102, the data targets 104, etc.) The connectivity configuration 206 may store login information, addresses, URLs, and other credentials for accessing the data sources 102 and/or the data targets 104. the connectivity configuration 206 may be employed by the data bridge adapter 200 to establish a connection and to maintain the connection itself, e.g., through connectors-as-a-service (CaaS) and/or the like.
The job processor 208 may perform additional transformations on an entity-relationship model derived from a data source (e.g., the data sources 102, etc.). The job processor 208 may perform configurations specified by a user in the interface when creating the entity-relationship model and/or standardized jobs needed based on the selected data target. The job processor 208 may transform and/or replicate a particular data field based on the unique requirements of the user or the target system. For example, the job processor 208 may modify the data as required for a particular data visualization tool's requirements. According to some embodiments, when a pipeline has been defined and is running, the job processor 208 may modify data (e.g., metadata, etc.) collected for a data record (e.g., each record with its input source(s) and which version of code processed it, etc.) to be used by the visualization module 122 to generate visual lineage information and/or tracing. The lineage information may be stored and/or used to determine how data mutations/changes are related and how the lineage evolved.
The adapters 210 may include information required to connect to various types of API and other data sources. For example, the adapters 210 may include components for connecting to APIs, via JDBC, stream adapters, file adapters, etc. Components needed for the adapters 210 may vary based on the type of adapter used.
Returning to
The visualization module 122 may be an analytics platform that allows data analysts to use advanced visualization techniques to explore and analyze data. For example, a user may use TABLEAU and/or a similar visualization tool to generate advanced visualizations, graphs, tables, charts, and/or the like. The visualization module 122 can output, for example, for an entity all the fields that are transferred to a target (e.g., the data targets 104, etc.) and how each transformation is applied, including merge operation and/or custom operations (e.g., determined by the associated dialect, etc.).
The visualization module 122 may be deployed locally (e.g., on a premises device, etc.), remotely (e.g., cloud-based, etc.), and/or within the integration platform 110. The visualization module 122 may have unique requirements for ingesting data. For example, according to some embodiments, the visualization module 122 may receive a JSON file and/or other representation of a relational model. According to some embodiments, the visualization module 122 may receive CSV data, PDF data, textual input, and/or any other type of input of data. The visualization module 122 may employ connectors (e.g., the connectors 116, etc.) specific to various data sources to ingest data.
A replication pipeline defines a job of copying data from a data source instance into a data destination instance. Each replication pipeline may have a replication source schema that specifies the data subset from the data source instance to be copied. A replication pipeline may include configurations such as scheduling frequency, filtering, and validation. The replication pipeline is in charge of moving the data from data source instances to the data destination instance.
A replication source schema may be a selection of a subset of data source relational schema for a replication pipeline. Based on the replication source schema, the data bridge platform 400 may generate a corresponding data destination schema for a replication pipeline. The data destination schema may represent the relational model schema for the data destination type.
A destination instance may represent an instance of a data destination type, for example, such as a RedShift database with a particular access credential. Each data destination instance may be used as the destination for many replication pipelines. A data destination type may be a type of destination for data (e.g., Tableau Hyper Type, RDS Type, etc.). Each data destination type may have multiple data destination instances.
According to some embodiments, the interface module 112 in communication with a data bridge pipeline experience API (XAPI) 402 may define a data pipeline process, including each transformation and data schema involved during the process. The data bridge pipeline XAPI 402 may be a single point of accessing the data bridge platform 400 (e.g., the data bridge module 118, etc.) capabilities serving as both a proxy to redirect requests to the internal services of the data bridge platform and/or apply some level of orchestration when multiple requests are needed. During the definition of the data pipeline, the interface module 112 may receive and submit to the data bridge pipeline XAPI 402 user preference information that may be used to customize to personalize how the data will arrive and/or end up at a target (e.g., the data targets 104 of
A data bridge pipeline service (DBPS) 404 may handle the create, read, update, and delete (CRUD) storage operations for pipelines and pipelines configurations. Pipeline configurations 410, such as user preference information that specify source and/or target/destination information, may be stored by the DBPS 404 in a data bridge database 408. The data bridge database 408 may be a relational database and/or suitable storage medium. The DBPS 404 may store lineage configurations 414, such as pre-defined and/or user preference information (e.g., dialects, schemas, etc.) that customizes/personalizes how data will arrive and/or end up at a target/destination (e.g., the data targets 104 of
A data bridge pipeline job (DBJS) 406 may be responsible for triggering a pipeline and keeping track of the progress of a pipeline (e.g., each pipeline of the integration platform 110 of
As previously described herein, once a pipeline has been defined, for example, via the interface module 112 (e.g., a data bridge UI, etc.), the data bridge module 118 may collect information, such as Run job pipeline metadata 412, about the running pipeline and the result for each record processed. The run job pipeline metadata 412 may be linked to the pipeline configurations 410 (e.g., the pipeline defined, etc.) and the lineage information 414 (e.g., defined ERD and transformation, etc.). The combined information (e.g., metadata, etc.) may be used to generate a lineage traceability route from source to target over time and based on each schema version provided at the source and target.
The data bridge platform 400 supports multiple ways to retrieve the lineage information. For example, to retrieve lineage information for a data record, a user may interact with the interface module 112 (e.g., a GUI, etc.) to request the current lineage traceability for the latest pipeline from the DBPS 404. As another example, to retrieve lineage information for a data record, a user may submit a request, using an ANG Query, to determine the traceability of a field/entity based on their lineage and version history to compare lineage changes over time.
The data bridge platform 400, for example, based on lineage information/metadata and data generated by the DBJS 406, can trace/determine why a pipeline fails. The data bridge platform 400 can determine if a pipeline failure is related to some transformation, and if so, over which entity and field.
Returning to
As shown in
Returning to
For example, as the data bridge module 118 applies deduplication, normalization, and/or validation to data (e.g., relational models) the data bridge module 118 may apply one or more algorithms to the results to produces information about the quality and freshness of data. For example, the data bridge module 118 may collect data duplication (nro records), new data (nro records), updated data (new records), errors (record not allowed to insert or update), data freshness measures, and/or the like. The data bridge module 118 may collect any type of metric information, and the metric information collected may be agnostic of a data source and/or source technology. The data bridge module 118 may scale its metric information collection and analysis according to anything that may be modeled using a dialect (e.g., the dialects 202, etc.).
The data bridge module 118, for example, when applying deduplication, normalization, and/or validation to data (e.g., relational models), may apply one or more of the following algorithms to the results to produce custom qualifying information regarding data freshness, data duplication, new data, updated data, errors, and/or the like:
The data bridge module 118 may use relational models (ERD models) based on dialects (e.g., AML Dialects, the dialects 202, etc.) transformed to JSON-LD and/or the like to determine data quality. For example, using an AML Dialect (Sematic) and the linked data representation of the data, the data bridge module 118 may determine/match which entities/fields are changed, updated, and/or deleted. The data bridge module 118 may use the custom qualifying information to define data integrity rules (error detection). For example, the data bridge module 118 may be configured with a model checker such as AML Model checker and/or the like to identify data accuracy and completeness. Based on a defined relational model (ERD model), the data bridge module 118 can identify the primary key and duplicate records. With an understanding of the primary key and duplicate records, the data bridge module 118 may generate KPI information for the data being processed. The KPI information may be stored in a data quality repository to produce insight and/or generate alarms regarding data quality, for example, for each entity. KPI information stored in the data quality repository can be used, for example, by the data bridge module 118 to determine, for each entity, the quality of a data warehouse and produce changes in an associated model, rules, data ingestion, and/or the like to facilitate any improvements needed.
The data bridge module 118 may communicate with the visualization module 122 to output a visual display of data quality. Data quality metrics can be displayed in any way to show different insights for the same metrics. For example, a histogram can be used to show how data quality evolves, where an x-axis represents time and a y-axis represents a data quality metric. According to some embodiments, the data bridge module 118 may communicate with the visualization module 122 to display data quality and freshness metrics in a radar chart 600 as shown in
Method 700 may be performed by processing logic that can comprise hardware (e.g., circuitry, dedicated logic, programmable logic, microcode, etc.), software (e.g., instructions executing on a processing device), or a combination thereof. It is to be appreciated that not all steps may be needed to perform the disclosure provided herein. Further, some of the steps may be performed simultaneously, or in a different order than shown in
In 710, a computer-based system (e.g., the integration platform 110 comprising the data bridge module 118, etc. may receive a request for metric information for a data record traversing a data pipeline. The request may be, for example, from a user device, a user application, a user interface, and/or the like. For example, the request may be received via a graphical user interface (GUI), an API, and/or any other interface and/or the like. At least the portion of the data record may include at least one of an entity of the data record or a field of the data record.
In 720, the computer-based system may determine each data flow component of the data pipeline that executes a respective task on at least a portion of the data record. For example, the computer-based system may determine each data flow component of the data pipeline that executes the respective task based on formatting information that defines data flow components of the data pipeline. For example, the computer-based system may determine each data flow component of the data pipeline that executes the respective task by determining, based on the formatting information, a relational model for a dataflow component that executes the respective task. The computer-based system may determine, based on a primary key affected by the respective task of the data flow component, the respective task of each other data flow component of the data pipeline. For example, the respective task for each of the other data flow components may affect a version of the primary key.
In 730, the computer-based system may determine, for each data flow component of the data pipeline that executes the respective task, respective metric information associated with the respective task in a first format. The first format may include a format associated with an application programming interface (API) and/or the like associated with the data component. The respective metric information for each data flow component of the data pipeline may include a measure of duplication of the data record, an update to the data record, an error associated with the data record, an amount of the data record, and/or the like.
In 740, the computer-based system may transform, for each data flow component of the data pipeline that executes the respective task, the respective metric information in the first format into a second format. For example, the second format may be configured for a user device, a user application, a user interface, and/or the like. Transforming the respective metric information in the first format into the second format may be based on a transformation template that maps disparate metric values to a common metric value representation. The second format may include and/or be based on a histogram, a radar chart, a gauge, a line graph, and/or the like.
In 750, the computer-based system may send the respective metric information for each data flow component of the data pipeline that executes the respective task in the second format. For example, the computer-based system may send the respective metric information for each data flow component of the data pipeline that executes the respective task in the second format to a user device, a user application, a user interface, and/or the like associated with the second format. Sending the respective metric information for each data flow component of the data pipeline that executes the respective task to the user device may cause the user device to display the respective metric information for each data flow component in the second format.
In the above description, numerous specific details such as resource partition
Computer system 800 may include one or more processors (also called central processing units, or CPUs), such as a processor 804. Processor 804 may be connected to a communication infrastructure or bus 806.
Computer system 800 may also include user input/output device(s) 802, such as monitors, keyboards, pointing devices, etc., which may communicate with communication infrastructure or bus 806 through user input/output device(s) 802.
One or more of processors 804 may be a graphics processing unit (GPU). In an embodiment, a GPU may be a processor that is a specialized electronic circuit designed to process mathematically intensive applications. The GPU may have a parallel structure that is efficient for parallel processing of large blocks of data, such as mathematically intensive data common to computer graphics applications, images, videos, etc.
Computer system 800 may also include a main or primary memory 808, such as random access memory (RAM). Main memory 808 may include one or more levels of cache. Main memory 808 may have stored therein control logic (i.e., computer software) and/or data.
Computer system 800 may also include one or more secondary storage devices or memory 810. Secondary memory 810 may include, for example, a hard disk drive 812 and/or a removable storage device or drive 814. Removable storage drive 814 may be a floppy disk drive, a magnetic tape drive, a compact disk drive, an optical storage device, a tape backup device, and/or any other storage device/drive.
Removable storage drive 814 may interact with a removable storage unit 818. The removable storage unit 818 may include a computer-usable or readable storage device having stored thereon computer software (control logic) and/or data. Removable storage unit 818 may be a floppy disk, magnetic tape, compact disk, DVD, optical storage disk, and/any other computer data storage device. Removable storage drive 814 may read from and/or write to the removable storage unit 818.
Secondary memory 810 may include other means, devices, components, instrumentalities, and/or other approaches for allowing computer programs and/or other instructions and/or data to be accessed by computer system 800. Such means, devices, components, instrumentalities, and/or other approaches may include, for example, a removable storage unit 822 and an interface 820. Examples of the removable storage unit 822 and the interface 820 may include a program cartridge and cartridge interface (such as that found in video game devices), a removable memory chip (such as an EPROM or PROM) and associated socket, a memory stick and USB port, a memory card and associated memory card slot, and/or any other removable storage unit and associated interface.
Computer system 800 may further include a communication or network interface 824. Communication interface 824 may enable computer system 800 to communicate and interact with any combination of external devices, external networks, external entities, etc. (individually and collectively referenced by reference number 828). For example, communication interface 824 may allow computer system 800 to communicate with external or remote devices 828 over communications path 826, which may be wired and/or wireless (or a combination thereof), and which may include any combination of LANs, WANs, the Internet, etc. Control logic and/or data may be transmitted to and from computer system 800 via communication path 826.
Computer system 800 may also be any of a personal digital assistant (PDA), desktop workstation, laptop or notebook computer, netbook, tablet, smartphone, smartwatch or other wearables, appliance, part of the Internet-of-Things, and/or embedded system, to name a few non-limiting examples, or any combination thereof.
Computer system 800 may be a client or server, accessing or hosting any applications and/or data through any delivery paradigm, including but not limited to remote or distributed cloud computing solutions; local or on-premises software (“on-premise” cloud-based solutions); “as a service” models (e.g., content as a service (CaaS), digital content as a service (DCaaS), software as a service (SaaS), managed software as a service (MSaaS), platform as a service (PaaS), desktop as a service (DaaS), framework as a service (FaaS), backend as a service (BaaS), mobile backend as a service (MBaaS), infrastructure as a service (IaaS), etc.); and/or a hybrid model including any combination of the foregoing examples or other services or delivery paradigms.
Any applicable data structures, file formats, and schemas in computer system 800 may be derived from standards including but not limited to JavaScript Object Notation (JSON), Extensible Markup Language (XML), Yet Another Markup Language (YAML), Extensible Hypertext Markup Language (XHTML), Wireless Markup Language (WML), MessagePack, XML User Interface Language (XUL), or any other functionally similar representations alone or in combination. Alternatively, proprietary data structures, formats, and/or schemas may be used, either exclusively or in combination with known or open standards.
In some embodiments, a tangible, non-transitory apparatus or article of manufacture comprising a tangible, non-transitory computer useable or readable medium having control logic (software) stored thereon may also be referred to herein as a computer program product or program storage device. This includes, but is not limited to, computer system 800, main memory 808, secondary memory 810, and removable storage units 818 and 822, as well as tangible articles of manufacture embodying any combination of the foregoing. Such control logic, when executed by one or more data processing devices (such as computer system 800), may cause such data processing devices to operate as described herein.
Based on the teachings contained in this disclosure, it will be apparent to persons skilled in the relevant art(s) how to make and use embodiments of this disclosure using data processing devices, computer systems, and/or computer architectures other than that shown in
It is to be appreciated that the Detailed Description section, and not any other section, is intended to be used to interpret the claims. Other sections can set forth one or more but not all exemplary embodiments as contemplated by the inventor(s), and thus, are not intended to limit this disclosure or the appended claims in any way.
Additionally and/or alternatively, while this disclosure describes exemplary embodiments for exemplary fields and applications, it should be understood that the disclosure is not limited thereto. Other embodiments and modifications thereto are possible and are within the scope and spirit of this disclosure. For example, and without limiting the generality of this paragraph, embodiments are not limited to the software, hardware, firmware, and/or entities illustrated in the figures and/or described herein. Further, embodiments (whether or not explicitly described herein) have significant utility to fields and applications beyond the examples described herein.
implementation One or more parts of the above implementations may include software. Software is a general term whose meaning of specified functions and relationships thereof. The boundaries of these functional building blocks have been arbitrarily defined herein for the convenience of the description. Alternate boundaries can be defined as long as the specified functions and relationships (or equivalents thereof) are appropriately performed. Also, alternative embodiments can perform functional blocks, steps, operations, methods, etc. using orderings different than those described herein.
References herein to “one embodiment,” “an embodiment,” “an example embodiment,” or similar phrases, indicate that the embodiment described can include a particular feature, structure, or characteristic, but every embodiment can not necessarily include the particular feature, structure, or characteristic. Moreover, such phrases are not necessarily referring to the same embodiment. Further, when a particular feature, structure, or characteristic is described in connection with an embodiment, it would be within the knowledge of persons skilled in the relevant art(s) to incorporate such feature, structure, or characteristic into other embodiments whether or not explicitly mentioned or described herein. Additionally, some embodiments can be described using the expression “coupled” and “connected” along with their derivatives. These terms are not necessarily intended as synonyms for each other. For example, some embodiments can be described using the terms “connected” and/or “coupled” to indicate that two or more elements are in direct physical or electrical contact with each other. The term “coupled,” however, can also mean that two or more elements are not in direct contact with each other, but yet still co-operate or interact with each other.
The breadth and scope of this disclosure should not be limited by any of the above-described exemplary embodiments, but should be defined only in accordance with the following claims and their equivalents.