Computational methods exist to predict interactions between ligands and proteins. They are generally classified as ‘ligand-based’ or ‘structure-based’ in accordance with the type of information used to make the prediction.
Protein-based predictions have the potential to learn biophysical compatibility and therefore may be more universally generalizable but are highly data-bound. Specifically, protein-based predictions use 3D molecular structures of ligands co-crystallized with proteins to evaluate or predict interaction. These methods are computationally demanding, and they tend to be trained on only 100's to 1000's of different proteins. Neural networks trained on these tend to have a very high feature space to data ratio. As a result, the method may produce a significant number of false negatives and/or false positives in docking when applied to previously unseen protein systems or drug scaffolds.
Ligand-based predictions may be performed using Drug Target Interaction (DTI) databases that have millions of records. Examples of publicly-available DTI databases include ChEMBL, NCBI's Bioassay, and STITCH. However, these records tend to only represent ˜2,000 of 20,000 human proteins. With high ligand-to-protein data ratios, one standard approach is to derive many different models for each of the 2,000 proteins. These tend to be successful, even beating high-throughput experimental results in many cases, but (1) only represent ˜10% of human proteins, (2) may be weaker when there is not much chemical diversity among examples for an individual protein, and (3) individual models do not learn the physical attributes of drug-protein compatibility, missing to benefit from data used to generate other models.
In general, in one aspect, one or more embodiments relate to a method for predicting drug-target binding using synthetically-augmented data, the method comprising: generating a plurality of ghost ligands for a plurality of proteins in a protein structure database; generating a plurality of drug-target interaction (DTI) features for proteins and ligands in a DTI database, using the plurality of ghost ligands; generating a machine learning model using the plurality of DTI features; and predicting a likelihood of interaction for a combination of a query protein and a query ligand using the machine learning model.
In general, in one aspect, one or more embodiments relate to a non-transitory computer readable medium comprising computer readable program code for predicting drug-target binding using synthetically-augmented data, the computer readable program code causing a computer system to: generate a plurality of ghost ligands for a plurality of proteins in a protein structure database; generate a plurality of drug-target interaction (DTI) features for proteins and ligands in a DTI database, using the plurality of ghost ligands; generate a machine learning model using the DTI features; and predict a likelihood of interaction for a combination of a query protein and a query ligand using the machine learning model.
In general, in one aspect, one or more embodiments relate to a system for differential drug discovery, the system comprising: a protein structure database; a ghost ligand identification engine configured to generate a plurality of ghost ligands for a plurality of proteins in the protein structure database; a ghost ligand database storing the plurality of ghost ligands; a drug-target interaction (DTI) database storing proteins and ligand; a feature generation engine configured to generate a plurality of DTI features for the proteins and the ligands in the DTI database, using the plurality of ghost ligands in the ghost ligand database; a machine learning model training engine configured to generate a machine learning model using the DTI features; and a DTI prediction engine configured to predict a likelihood of interaction for a combination of a query protein and a query ligand using the machine learning model.
The present embodiments are illustrated by way of example and are not intended to be limited by the figures of the accompanying drawings.
Specific embodiments disclosed herein will now be described in detail with reference to the accompanying figures. Like elements in the various figures may be denoted by like reference numerals and/or like names for consistency.
The following detailed description is merely exemplary in nature, and is not intended to limit the embodiments disclosed herein or the application and uses of embodiments disclosed herein. Furthermore, there is no intention to be bound by any expressed or implied theory presented in the preceding technical field, background, brief summary or the following detailed description.
In the following detailed description of some embodiments disclosed herein, numerous specific details are set forth in order to provide a more thorough understanding of the various embodiments disclosed herein. However, it will be apparent to one of ordinary skill in the art that the embodiments may be practiced without these specific details. In other instances, well-known features have not been described in detail to avoid unnecessarily complicating the description.
Throughout the application, ordinal numbers (e.g., first, second, third, etc.) may be used as an adjective for an element (i.e., any noun in the application). The use of ordinal numbers is not to imply or create any particular ordering of the elements nor to limit any element to being only a single element unless expressly disclosed, such as by the use of the terms “before”, “after”, “single”, and other such terminology. Rather, the use of ordinal numbers is to distinguish between the elements. By way of an example, a first element is distinct from a second element, and the first element may encompass more than one element and succeed (or precede) the second element in an ordering of elements.
In one or more embodiments of the invention, the elements of the protein-based prediction methods and ligand-based prediction methods may be combined to obtain superior predictions of drug-target interactions. In one or more embodiments of the invention, a machine learning model is used to predict drug-target interactions (DTIs).
An abundance of data points exists for protein-ligand interactions in DTI databases. For example, the ChEMBL database contains ˜15,000,000 records describing pairs of proteins and ligands as binding or non-binding (and frequently providing a measure of affinity or confidence). However, using the content of these DTI databases to predict interactions of new pairs of proteins and ligands has certain limitations, such as limited coverage of the human proteome and weak predictions when the diversity among examples for an individual protein is limited. Accordingly, the quality of predictions by machine learning models operating purely on DTI database records may be limited. In the alternative, a machine learning model may operate on 3D molecular structures of ligands co-crystallized with proteins to capture the biophysics of interactions between proteins and ligands. A few databases that capture protein-ligand interactions exist (e.g., sc-PDB), but these databases contain relatively few data points and tend to be highly redundant. Further, these databases tend to lack protein structure diversity and chemotype diversity. Training a machine learning model on these data points may, thus, be challenging due to an insufficient amount of data. In particular, a machine learning model operating on 3D molecular structures of ligands co-crystallized with proteins has a high-dimensional feature space which, in combination with the limited availability of suitable training samples, makes the machine learning model vulnerable to overfitting.
In one or more embodiments of the disclosure, local 3D features and DTI records are used in combination by a machine learning model to enable superior predictions over the above-described conventional protein-based and ligand-based predictions. More specifically, synthetic data is generated by projecting DTI records onto 3D structure models of known protein-ligand complexes, in order to generate otherwise unavailable local protein features. The synthetic data generated in this manner may be used to train a machine learning model. The machine learning model may then be used to make predictions for pairs of a query protein and a query ligand.
Turning to
The ghost ligand identification engine (110), in accordance with one or more embodiments, includes instructions in the form of computer readable program code to perform at least one of the steps described in
The feature generation engine (120), in accordance with one or more embodiments, includes instructions in the form of computer readable program code to perform at least one of the steps described in
Continuing with
The drug-target interaction (DTI) prediction engine (150), in accordance with one or more embodiments, includes instructions in the form of computer readable program code to perform at least one of the steps described in
The protein structure database (160), in accordance with one or more embodiments, may be any type of storage unit and/or device (e.g., a file system, database, collection of tables, or any other storage mechanism) for storing data. The protein structure database (160) is described below with reference to
The ghost ligand database (170), in accordance with one or more embodiments, may be any type of storage unit and/or device (e.g., a file system, database, collection of tables, or any other storage mechanism) for storing data. The ghost ligand database (170) is described below with reference to
The drug-target interaction database (180), in accordance with one or more embodiments, may be any type of storage unit and/or device (e.g., a file system, database, collection of tables, or any other storage mechanism) for storing data. The drug-target interaction database (180) is described below with reference to
The protein annotation database (190), in accordance with one or more embodiments, may be any type of storage unit and/or device (e.g., a file system, database, collection of tables, or any other storage mechanism) for storing data. The protein annotation database (190) is described below with reference to
Turning to
Turning to
Turning to
Turning to
Turning to the flowchart of
In Step 200, a ghost ligand database is generated from proteins obtained from a protein structure database. A detailed description of Step 200 is provided in
In Step 202, drug-target interaction (DTI) features are generated, including ligand features and protein features. A detailed description of Step 202 is provided in
In Step 204, the machine learning model for DTI prediction is generated. A detailed description of Step 204 is provided in
Turning to the flowchart of
In Step 300, proteins are obtained from a protein structure database. For each of the proteins, a 3D model may be retrieved.
In Step 302, the obtained proteins are clustered by sequence or domain.
In one or more embodiments, protein clusters are defined as any set of two or more proteins sharing similarity in primary sequence or three-dimensional topology, commonly known as a fold. Protein clusterings may be obtained directly from publicly-available databases such as the PDB, SCOP, CATH, PFAM, or Uniprot for example. Protein clusters may be manually created on the basis of sequence similarity using protein sequence alignment tools and clustering tools such as BLAST, CD-HIT, or UCLUST for example. Alternatively, proteins with three-dimensional structure models may be clustered by grouping unrelated proteins that share a common topology or fold.
In Step 304, one of the clusters is selected for further processing.
In Step 306, a pairwise structural alignment is performed for all proteins in the selected cluster.
In Step 308, ghost ligands are obtained by projecting each ligand to the cluster peers. A confidence score may be obtained for each of the projections, comprised of individual scores representing confidence in different components of the ligand projection. Examples for projecting ligands to cluster peers are provided in
In Step 310, the ghost ligands and associated confidence scores are stored in the ghost ligand database.
In Step 312, a determination is made about whether additional clusters are remaining for processing. If additional clusters are remaining, the execution of the method may return to Step 304 to select another cluster for processing as described in Steps 306-310. If no additional clusters are remaining, the execution of the method may terminate. Once the method of
Turning to the flowchart of
In Step 400, a drug-target interaction for a combination of a ligand and a protein is selected from the DTI database. The subsequent steps are performed for this considered combination of the ligand and the protein. The steps may later be repeated for other combinations of a ligand and a protein.
In Step 402, features are generated for the selected ligand. These features may include a ligand fingerprint and ligand descriptors. The fingerprint may capture the structure of the ligand in a descriptor format and may be based on a SMILES representation of the underlying molecules using a fixed length vector. Molecular fingerprinting methods may include atom pair, extended connectivity fingerprint, graph-based fingerprints, torsion fingerprints, or pharmacophore fingerprints, for example. Molecular weight, number of rotatable bonds, number of hydrogen bond donors, number of hydrogen bond receptors, hydrophobicity, aromaticity, and functional group compositions may be used as ligand descriptors, for example. Molecular shape descriptors such as ellipticity, geometric descriptors, branching descriptors, or chirality descriptors, for example, may also be used.
In Step 404, ghost ligands are retrieved from the ghost ligand database for the selected protein.
In Step 406, each of the ghost ligands is scored, based on its difference to the drug, or more specifically, to the selected ligand. A higher similarity results in a higher score. Similarity between ligands may be scored using a distance metric for comparing molecular fingerprints, such as a Tanimoto distance, D ice distance, or cosine distance for example.
Steps 404 and 406 may be performed for all protein models (e.g., homology models or experimental derived models) that are available for the selected protein.
In Step 408, the ghost ligand that is most similar to the selected ligand is selected for further processing.
In Step 410, a confidence vector is generated for the DTI features comprised of individual scores representing confidence in different components of the DTI features and its representative ghost ligand. The confidence vector may include confidence scores representing ghost ligand projections from Step 308. The confidence vector may further include a confidence score for the fingerprint similarity between the selected ligand and the most similar ghost ligand selected in Step 408. In addition, the confidence vector may include a score for the confidence of the selected DTI. The confidence of the selected DTI may be scored based on the source of the DTI data (e.g., a different score may be assigned depending on whether the DTI data were obtained using high or low throughput screening, etc.).
In Step 412, local features are obtained for the selected protein surrounding the most similar ghost ligand. The local features may include binding site features existing in concentric shells of increasing radii, as illustrated in
In Step 414, global features for the structure and/or sequence of the selected protein are obtained by expanding to outer shells with large radii as described in step 412. Shell radii corresponding to local features may include thresholds of 5Å, 10Å, or 15Å for example. Shell radii corresponding to domain-level descriptions or global protein descriptions may have larger distance thresholds of 20Å, 25Å, or 30Å for example, or may have no distance threshold. Global features may further include descriptions of domains or folds and may be derived from publicly-available databases such as SCOP, CATH, or PFAM, for example. Global features may further include features derived from protein sequence and may include the presence of common sequence motifs, for example. Global features may include descriptions of the protein folding state, such as the presence and biophysical properties of intrinsically disordered regions, hinges, loops, ordered regions, or modulatory domains, for example. Global features may also be described with relation to distance from a ligand binding site.
In Step 416, functional annotations of the selected protein are obtained. The functional annotations may be obtained from a protein annotation database. Functional annotations may include Enzyme Commission (EC) numbers, Gene Ontology (GO) annotations, or Uniprot Keywords, for example. Functional annotations may also include the presence or absence of documented positional-specific properties of a protein, such as catalytic sites, post translational modifications, disease associations, or genetic variations, for example.
In Step 418, features are generated for the selected protein. These features may include the local features, the global features and/or the functional annotations.
In Step 420, a determination is made about whether additional DTIs are remaining for processing. If additional DTIs are remaining, the execution of the method may return to Step 400 to select another DTI for processing as described in Steps 402-418. If no additional clusters are remaining, the execution of the method may terminate. Once the method of
Turning to the flowchart of
In Step 500, ligand features and protein features are obtained. The obtaining of ligand and protein features may be performed as described in Steps 402 and 418 of
In Step 502, the ligands and proteins are filtered by a function of the confidence vectors established in Step 408 of
In Step 504, ligand and protein features are concatenated to generate positive training samples.
In Step 506, ligand and protein features are shuffled. The shuffled ligand and protein features are concatenated to generate negative training samples. This step may be repeated multiple times to evaluate different ratios of positive-to-negative training samples, 1:1, 1:5, 1:10, 1:19, or 1:20, for example
In Step 508, the machine learning model for DTI prediction is trained using the positive and negative training samples. A learning algorithm based on backpropagation may be used, for example. In one or more embodiments, training samples may be weighted, based on their associated confidence vectors. In one or more embodiments, transfer learning is used to more effectively train the machine learning model. Initially, the machine learning model may be trained by applying an initial confidence threshold in step 502. In subsequent re-training phases, a confidence threshold may be increased to reduce the number and increase the quality of training examples. In addition or alternatively, subsequent re-training phases may limit training examples to select classes of drugs or targets. The machine learning model may be a supervised discriminative classification or regression model such as a random forest, support vector machine, single layer perceptron, or multiple layer artificial neural network, for example. Artificial neural networks are especially well-suited for this task, given the magnitude of training data points (100,000 s to 10,000,000 s) and the dimensionality of training data features (1000 s to 10,000 s). In one embodiment, an artificial neural network representation takes the form of fully-connected network with a feature input layer, two hidden layers with, for example, 512 and 256 nodes, respectively, and two output nodes corresponding to interacting and non-interacting pairs. In one embodiment, an artificial neural network with multiple hidden layers omits connections between input types, for the creation of separate latent spaces representing ligand fingerprints, global protein features, local protein features, and protein functional features.
Turning to the flowchart of
In Step 600, the query protein and the query ligand are obtained. The query protein and the query ligand may be obtained from a user wanting to obtain a prediction of the interaction between the query protein and the query ligand.
In Step 602, possible binding sites and associated local features are obtained for the query protein, as previously described in Steps 404-412 of
In Step 604, global features and protein annotations are obtained for the query protein. The global features and protein annotations may be obtained as previously described in Steps 414 and 416 of
In Step 606, features are generated for the query protein. These features may include the local features, the global features and/or the functional annotations.
In Step 608, the ligand fingerprint and ligand descriptors are obtained, as previously described in Step 402 of
In Step 610, features are generated for the query ligand. These features may include the ligand fingerprint and the ligand descriptors.
In Step 612, the machine learning model for DTI prediction is applied to the features of the query ligand and the features of the query protein to obtain a numerical score for a likelihood of interaction between the query ligand and the query protein.
The following paragraphs further illustrate embodiments of the disclosure based on various examples. Those skilled in the art will appreciate that the disclosure is not limited to these examples.
(i) Sample Ghost Ligands:
Turning to
(ii) Sample Confidence Vector:
Embodiments of the disclosure rely on a heuristic process to augment drug-target interaction (DTI) data with assumed three-dimensional structure representations. These assumed DTI representations may provide information-rich features for machine learning, improving models aimed at predicting protein-ligand interactions. Obtaining these approximated DTI representations for any given DTI data point requires several assumptions outlined throughout Steps 200 and 202 of
Confidence Vectors are comprised of multiple metrics that describe measurable uncertainty in the approximated DTI representations. These metrics are accrued in the creation of the ghost ligand database (Step 200) and projecting known DTI data onto the ghost ligand database (Step 202). In one example, the Confidence Vector contains four elements, including: (1) Percentage of sequence identity between the homology model representing the DTI protein and its source template, (2) RMSD from the alignment between the source structure of the ghost ligand and the homology model representing the DTI protein, (3) Tanimoto similarity between morgan3 fingerprints of the DTI ligand and the ghost ligand template (4) confidence in the DTI data point.
In the example, a Drug-Target Interaction (DTI) database indicates that ligand gefitinib, interacts with protein Aurora Kinase A. The DTI database assigned the interaction a probability of 85%, based accuracy of the source biophysical experiment. The specific interaction between gefitinib and Aurora Kinase A is not present in the source three-dimensional structure database. In creating the ghost ligand database, a homology model for the Aurora Kinase A protein was created from close homolog Aurora Kinase B, which shares a sequence identity of 72.5%. The closest molecule to gefitinib, that was successfully mapped onto the homology model was erlotinib which shares a Morgan3 fingerprint Tanimoto similarity of 0.372. The erlotinib ghost ligand location was approximated based on a structural alignment between the Aurora Kinase A homology model and erlotinib-EGFR co-complex crystal structure, with a ligand binding site RMSD of 2.345 Å. Accordingly, the corresponding confidence vector would be: [85%, 72.5%, 0.372, 2.345 Å].
(iii) Sample Training Data and Negative Shuffling:
The described methodology focuses on augmenting drug-target interacting (DTI) pairs from a DTI database with a mixture of relevant features obtained through a series of deterministic mappable relationships and heuristically modeled features (local structure features). Each row in a DTI database may be transformed into a feature vector as exemplified in
When a neural network is trained only by true drug-target interactions positive examples, extracted from drug-target interaction datasets, the model might learn to ignore the core and obvious patterns of interactions since they will not provide any signal to the model. Further, it may be necessary to control a potentially significant bias toward highly represented drugs and targets in drug-target interaction datasets. Hence, it may be beneficial to sample negative examples with the relative proportion for each drug and target. As a result, the model may learn patterns based on positive examples and negative examples. In
Embodiments of the disclosure use ghost ligands to create local protein features for proteochemometrics (PCM) from protein-ligand datasets. More specifically, drug-target interactions (DTI) data are threaded onto 3D atomistic models of protein-ligand complexes to derive local protein features. Mixed feature datasets of the PCMs may include local (pocket), regional (domain) and global (whole protein) annotations and/or functional annotations.
Conventionally, training data for machine learning should be high-confidence ‘model quality data’. The use of a prediction (ghost ligands+threading), in accordance with one or more embodiments, to produce training data for a machine learning algorithm appears to be counterintuitive. Specifically, if the heuristic approach is not sufficiently accurate, then, conventional wisdom suggests that the introduction of local features derived from the combination of ghost ligand and threading has the potential to introduce additional noise, thereby reducing the performance over conventional DTI PMCs. However, as the performance comparison (900) of
Specifically,
Various embodiments of the disclosure have one or more of the following advantages. Embodiments of the invention enable the prediction of drug-target interactions (DTIs) using machine learning models that reflect compatibility between a protein environment and ligand attributes. Localized 3D features are created to represent binding sites, even when there is no 3D information available for the interaction under consideration.
Mapping known drug-target interactions onto homology models synthetically augments abundant DTI training data with highly-dimensional biophysical information to train a deep neural network. The approach, thus, enables the use of comprehensive DTI databases, even though it is not necessarily known where a drug binds to a protein, for the entries of the DTI databases.
Methods in accordance with one or more embodiments do not require in-detail knowledge of the biophysics of protein-ligand interactions. Accordingly, no precise 3D coordinates of atoms are required, thereby enabling the use of DTI databases and homology models to map drug-target interactions onto pockets of proteins.
Embodiments of the disclosure require a reduced feature space and permit vastly larger training data in comparison to structure-based deep learning approaches relying on 3D atomic coordinates. Further, embodiments of the disclosure were found to generalize well. An initial performance evaluation suggests that the method as described performs about 1,000,000 times faster, in comparison to docking simulations. The methods in accordance with one or more embodiment require no human intervention. Specifically, the most likely protein representation and binding site is automatically identified. As discussed in Exhibits A and B, methods in accordance with one or more embodiments may be used as accurate in-silico alternatives or additions to other in- silico and/or experimental methods for predicting drug-target interactions.
Embodiments of the disclosure may have various applications. For example, embodiments may be used for proteome screenings (e.g., to perform toxicity predictions or phenotypic deconvolution predictions), virtual screenings, and for drug discovery and development in general.
Embodiments of the disclosure may be implemented on a computing system. Any combination of mobile, desktop, server, router, switch, embedded device, or other types of hardware may be used. For example, as shown in
The computer processor(s) (1002) may be an integrated circuit for processing instructions. For example, the computer processor(s) may be one or more cores or micro-cores of a processor. The computing system (1000) may also include one or more input devices (1010), such as a touchscreen, keyboard, mouse, microphone, touchpad, electronic pen, or any other type of input device.
The communication interface (1012) may include an integrated circuit for connecting the computing system (1000) to a network (not shown) (e.g., a local area network (LAN), a wide area network (WAN) such as the Internet, mobile network, or any other type of network) and/or to another device, such as another computing device.
Further, the computing system (1000) may include one or more output devices (1008), such as a screen (e.g., a liquid crystal display (LCD), a plasma display, touchscreen, cathode ray tube (CRT) monitor, projector, or other display device), a printer, external storage, or any other output device. One or more of the output devices may be the same or different from the input device(s). The input and output device(s) may be locally or remotely connected to the computer processor(s) (1002), non-persistent storage (1004), and persistent storage (1006). Many different types of computing systems exist, and the aforementioned input and output device(s) may take other forms.
Software instructions in the form of computer readable program code to perform embodiments of the disclosure may be stored, in whole or in part, temporarily or permanently, on a non-transitory computer readable medium such as a CD, DVD, storage device, a diskette, a tape, flash memory, physical memory, or any other computer readable storage medium. Specifically, the software instructions may correspond to computer readable program code that, when executed by a processor(s), is configured to perform one or more embodiments of the disclosure.
The computing system (1000) in
Although not shown in
The nodes (e.g., node X (1022), node Y (1024)) in the network (1020) may be configured to provide services for a client device (1026). For example, the nodes may be part of a cloud computing system. The nodes may include functionality to receive requests from the client device (1026) and transmit responses to the client device (1026). The client device (1026) may be a computing system, such as the computing system shown in
The computing system or group of computing systems described in FIG.
10A and 10B may include functionality to perform a variety of operations disclosed herein. For example, the computing system(s) may perform communication between processes on the same or different system. A variety of mechanisms, employing some form of active or passive communication, may facilitate the exchange of data between processes on the same device. Examples representative of these inter-process communications include, but are not limited to, the implementation of a file, a signal, a socket, a message queue, a pipeline, a semaphore, shared memory, message passing, and a memory-mapped file. Further details pertaining to a couple of these non-limiting examples are provided below.
Based on the client-server networking model, sockets may serve as interfaces or communication channel end-points enabling bidirectional data transfer between processes on the same device. Foremost, following the client-server networking model, a server process (e.g., a process that provides data) may create a first socket object. Next, the server process binds the first socket object, thereby associating the first socket object with a unique name and/or address. After creating and binding the first socket object, the server process then waits and listens for incoming connection requests from one or more client processes (e.g., processes that seek data). At this point, when a client process wishes to obtain data from a server process, the client process starts by creating a second socket object. The client process then proceeds to generate a connection request that includes at least the second socket object and the unique name and/or address associated with the first socket object. The client process then transmits the connection request to the server process. Depending on availability, the server process may accept the connection request, establishing a communication channel with the client process, or the server process, busy in handling other operations, may queue the connection request in a buffer until server process is ready. An established connection informs the client process that communications may commence. In response, the client process may generate a data request specifying the data that the client process wishes to obtain. The data request is subsequently transmitted to the server process. Upon receiving the data request, the server process analyzes the request and gathers the requested data. Finally, the server process then generates a reply including at least the requested data and transmits the reply to the client process. The data may be transferred, more commonly, as datagrams or a stream of characters (e.g., bytes).
Shared memory refers to the allocation of virtual memory space in order to substantiate a mechanism for which data may be communicated and/or accessed by multiple processes. In implementing shared memory, an initializing process first creates a shareable segment in persistent or non-persistent storage. Post creation, the initializing process then mounts the shareable segment, subsequently mapping the shareable segment into the address space associated with the initializing process. Following the mounting, the initializing process proceeds to identify and grant access permission to one or more authorized processes that may also write and read data to and from the shareable segment. Changes made to the data in the shareable segment by one process may immediately affect other processes, which are also linked to the shareable segment. Further, when one of the authorized processes accesses the shareable segment, the shareable segment maps to the address space of that authorized process. Often, only one authorized process may mount the shareable segment, other than the initializing process, at any given time.
Other techniques may be used to share data, such as the various data described in the present application, between processes without departing from the scope of the disclosure. The processes may be part of the same or different application and may execute on the same or different computing system.
Rather than or in addition to sharing data between processes, the computing system performing one or more embodiments of the disclosure may include functionality to receive data from a user. For example, in one or more embodiments, a user may submit data via a graphical user interface (GUI) on the user device. Data may be submitted via the graphical user interface by a user selecting one or more graphical user interface widgets or inserting text and other data into graphical user interface widgets using a touchpad, a keyboard, a mouse, or any other input device. In response to selecting a particular item, information regarding the particular item may be obtained from persistent or non-persistent storage by the computer processor. Upon selection of the item by the user, the contents of the obtained data regarding the particular item may be displayed on the user device in response to the user's selection.
By way of another example, a request to obtain data regarding the particular item may be sent to a server operatively connected to the user device through a network. For example, the user may select a uniform resource locator (URL) link within a web client of the user device, thereby initiating a Hypertext Transfer Protocol (HTTP) or other protocol request being sent to the network host associated with the URL. In response to the request, the server may extract the data regarding the particular selected item and send the data to the device that initiated the request. Once the user device has received the data regarding the particular item, the contents of the received data regarding the particular item may be displayed on the user device in response to the user's selection. Further to the above example, the data received from the server after selecting the URL link may provide a web page in Hyper Text Markup Language (HTML) that may be rendered by the web client and displayed on the user device.
Once data is obtained, such as by using techniques described above or from storage, the computing system, in performing one or more embodiments of the disclosure, may extract one or more data items from the obtained data. For example, the extraction may be performed as follows by the computing system in
Next, extraction criteria are used to extract one or more data items from the token stream or structure, where the extraction criteria are processed according to the organizing pattern to extract one or more tokens (or nodes from a layered structure). For position-based data, the token(s) at the position(s) identified by the extraction criteria are extracted. For attribute/value-based data, the token(s) and/or node(s) associated with the attribute(s) satisfying the extraction criteria are extracted. For hierarchical/layered data, the token(s) associated with the node(s) matching the extraction criteria are extracted. The extraction criteria may be as simple as an identifier string or may be a query provided to a structured data repository (where the data repository may be organized according to a database schema or data format, such as XML).
The extracted data may be used for further processing by the computing system. For example, the computing system of
The computing system in
The user, or software application, may submit a statement or query into the DBMS. Then the DBMS interprets the statement. The statement may be a select statement to request information, update statement, create statement, delete statement, etc. Moreover, the statement may include parameters that specify data, or data container (database, table, record, column, view, etc.), identifier(s), conditions (comparison operators), functions (e.g. join, full join, count, average, etc.), sort (e.g. ascending, descending), or others. The DBMS may execute the statement. For example, the DBMS may access a memory buffer, a reference or index a file for read, write, deletion, or any combination thereof, for responding to the statement. The DBMS may load the data from persistent or non-persistent storage and perform computations to respond to the query. The DBMS may return the result(s) to the user or software application.
The computing system of
For example, a GUI may first obtain a notification from a software application requesting that a particular data object be provided within the GUI. Next, the GUI may determine a data object type associated with the particular data object, e.g., by obtaining data from a data attribute within the data object that identifies the data object type. Then, the GUI may determine any rules designated for displaying that data object type, e.g., rules specified by a software framework for a data object class or according to any local parameters defined by the GUI for presenting that data object type. Finally, the GUI may obtain data values from the particular data object and render a visual representation of the data values within a display device according to the designated rules for that data object type.
Data may also be provided through various audio methods. In particular, data may be rendered into an audio format and provided as sound through one or more speakers operably connected to a computing device.
Data may also be provided to a user through haptic methods. For example, haptic methods may include vibrations or other physical signals generated by the computing system. For example, data may be provided to a user using a vibration generated by a handheld computer device with a predefined duration and intensity of the vibration to communicate the data.
The above description of functions presents only a few examples of functions performed by the computing system of
While the disclosure has been described with respect to a limited number of embodiments, those skilled in the art, having benefit of this disclosure, will appreciate that other embodiments can be devised which do not depart from the scope of the disclosure as disclosed herein. Accordingly, the scope of the disclosure should be limited only by the attached claims.
The embodiments and examples set forth herein were presented in order to best explain the present invention and its particular application and to thereby enable those skilled in the art to make and use the invention. However, those skilled in the art will recognize that the foregoing description and examples have been presented for the purposes of illustration and example only. The description as set forth is not intended to be exhaustive or to limit the invention to the precise form disclosed.
While the invention has been described with respect to a limited number of embodiments, those skilled in the art, having benefit of this disclosure, will appreciate that other embodiments can be devised which do not depart from the scope of the invention as disclosed herein. Accordingly, the scope of the invention should be limited only by the attached claims.
The present application claims priority under 35 U.S.C. § 119(e) to U.S. Provisional Application No. 62/788,682, filed on Jan. 4, 2019, having at least one of the same inventors as the present application, and entitled, “METHOD AND SYSTEM FOR PREDICTING DRUG BINDING USING SYNTHETIC DATA”. U.S. Provisional Application No. 62/788,682 is incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CA2020/050005 | 1/2/2020 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62788682 | Jan 2019 | US |