Trust is an essential component to many social and business interactions, but trust can be both hard to measure and difficult to quantify. People typically looks towards a variety of different factors, experiences, and influences to determine how much to trust another party or entity in a transaction. For example, a potential customer deciding whether to dine at a particular restaurant may take into account how many times he or she has eaten at the restaurant, word of mouth from friends and family, and any ratings from online feedback sites. As another example, a bank may look up the credit score of a potential borrower as a measure of their financial responsibility when determining whether to issue a loan. Often, people can have wildly different preferences as to which factors are the most important in determining trust levels, and these preferences may change depending on the type and details of the transaction. Trust can also change over time, reflecting the cumulative experiences, transaction history, and recent trends between entities. A single negative event can destroy trust, and trust can also be rebuilt over time. All of the above considerations make “trust” an elusive measure to capture.
Systems, devices, and methods are described herein for calculating a trust score. The trust score may be calculated between entities including, but not limited to, human users, groups of users, organizations, businesses/corporations, products/product lines, and/or locations. The trust score may reflect the trustworthiness, reputation, membership, status, and/or influence of the entity in a particular community or in relation to another entity. The trust score may take into account data from any suitable data sources, including, but not limited to, network connectivity information, social network information, credit score, available court data, transaction history, ratings/feedback data, group/demographics data, search engine data, or any publically available information. The trust score may also include certain non-publically available information provided by the entities themselves (e.g., non-public transaction history, targeted ratings, etc.).
As used herein, a “system trust score” refers to a trust score calculated for an entity based on information available for the entity, without specific reference to another entity or activity/transaction. The system trust score may represent a base level of trustworthiness for the entity that does not take into account information about a specific activity/transaction. In some embodiments, the system trust score may be calculated based on publicly available information, such as verification data, a network connectivity score, and/or ratings data. As defined herein, a “network community” may include any collection or group of entities connected through a network, including, but not limited to a computer network or a social network. In some embodiments, a user may set an initial trust score as a minimum trust level. In these embodiments, the initial trust score may be retrieved and updated based on publicly available information in order to determine the system trust score. In some embodiments, the system trust score may be provided to an end user upon request without the end user having to identify themselves. For example, an end user may query the system trust scores of other entities, for example through a website or a mobile application, without having to sign into the website or mobile application or otherwise having to identify themselves.
As used herein, a “peer trust score” refers to a trust score calculated for a first entity in relation to a second entity. The peer trust score may take into account certain information that is specific to the first and second entity, such as specific transaction history between the first and second entity, number of common contacts/friends, etc. In some embodiments, the peer trust score may be derived from the system trust score and represent an update of the system trust score. For example, in some embodiments, the peer trust score may be calculated based on substantially the same data sources as the system trust score, where some components may be updated in order to further weight or take into account additional information that is specific to the first and second entity. In other embodiments, the peer trust score may be calculated independently from the system trust score and may be based on a different set of data sources than the system trust score.
As used herein, a “contextual trust score” refers to a trust score calculated for a first entity in relation to a specific activity or transaction. The contextual trust score may take into account certain information that is particular to the specific activity or transaction. In some embodiments, the contextual trust score may be derived from the system trust score or the peer trust score and represent an update of the system trust score or the peer trust score. For example, in some embodiments, the contextual trust score may be calculated based on substantially the same data sources as the system trust score, where some components may be updated in order to take into account information that is particular to the activity/transaction. In other embodiments, the contextual trust score may be calculated based on a different set of data sources than the system trust score and the peer trust score. In some embodiments, the contextual trust score may be calculated by weighting data from different data sources based on the type of activity/transaction. For example, the trust score of a potential borrower who is seeking a mortgage from a bank may heavily weight the borrower's credit score and financial history rather than their level of connectivity in a social network. In this manner, the contextual trust score may be based on the same or similar data sources as the system trust score and/or the peer trust score, but with a different weighting to combine the data from the data sources. In some embodiments, specific details of the transactions may also affect the calculation of the contextual trust score. For instance, the contextual trust score for a friend borrowing $10 may focus more on social network connectivity (e.g., the number of friends they have in common, etc.), while the contextual trust score for a borrower seeking a $100K loan from the bank may focus more on financial factors. In some embodiments, the details of the transaction may affect the weighting of the combination of data from the data sources.
According to one aspect, a method for updating a trust score may comprise identifying paths from a first entity to a second entity, calculating a network connectivity score based on the identified paths, receiving data about the second entity from a remote source, and calculating a ratings score based on the received data from the remote source. A trust score for the second entity may be determined by combining the network connectivity score and the ratings score. An indication of an activity to be performed by the first entity and the second entity may be received, and the trust score may be updated based on the indication of the activity. In some embodiments, the first and second entity may be connected by a social network. In such embodiments, identifying paths from the first entity to the second entity may comprise identifying an intermediate entity in the social network that connects the first entity to the second entity. For example, the intermediate entity may be a common friend between a first user and a second user. Calculating the network connectivity score may comprise determining a number of mutual friends between the first entity and the second entity. For example, the network connectivity score may be assigned according to a graduated scale based on the number of mutual friends between the first entity and the second entity. The network connectivity score may also be calculated based on the number of identified paths between the first and the second entity and whether the number of identified paths exceeds a certain threshold.
In some embodiments, the ratings data may be one of a credit score, criminal history data, financial transaction history data, and/or business reviews data. The ratings data may be combined with the network connectivity score according to a weighted sum in order to determine the trust score for the second entity. The weighted sum may be based on a default set of weights or based on user-assigned weights. The trust score for the second entity may then be updated based on the indication of the activity. For example, the indication of the activity may adjust the weighted sum such that a different weighted sum is used to calculate the trust score for the second entity.
In some embodiments, at least one of the first entity and the second entity is a human user. For instance, the trust score may be calculated between two users who are participating in a certain activity. In another embodiment, at least one of the first entity and the second entity may be a business. For example, the trust score between a user and a restaurant may be calculated in order to aid the user in determining whether to eat at the restaurant. In yet other embodiments, at least one of the first entity and the second entity may be a group of users or an organization. As an illustrative example, the second entity may be the Boy Scouts of America, and the trust score may be calculated between a first user and the Boy Scouts of America. In some embodiments, at least one of the first and second entity may be a product or an object. For instance, the first entity may be a first user, and the second entity may be a chainsaw, and a trust score may be calculated between the chainsaw and the first user. In this example, the trust score may take into account any user reviews of the chainsaw received from a third-party ratings source. In some embodiments, at least one of the first and second entity may be a location, city, region, nation, or any other geographic place. For instance, a trust score between a first user and a city, such as New York City, may be calculated. In this example, the trust score may take into account number of contacts that the first user has in New York City, traveler reviews received from third-party ratings sources, and/or and activities, transactions, or interactions that the first user has had with New York City.
In some embodiments, a decision related to the activity may be automatically resolved based, at least in part, on a calculated trust score. For instance, a bank may request the trust score of a potential borrower in order to evaluate the suitability of the borrower for a loan. Based on the updated trust score, the bank may automatically issue the loan, for example, if the trust score exceeds a certain threshold. In this manner, the system trust score, peer trust score, and/or the contextual trust score can, either alone or in combination, form the basis for automatic decision making.
In some embodiments, at least one of the system, peer, and/or contextual trust score may include a confidence range. For example, each of the components from the data sources may comprise a confidence range (such as a variance or a standard deviation) indicating a level of uncertainty in the data, and the component scores may be combined to form one of the system, peer, and/or contextual trust score. Thus, the resulting trust score may be represented by a mean score and a confidence range, and in some embodiments, the confidence range may be represented by a mean and standard deviation.
The foregoing and other features and advantages will be apparent upon consideration of the following detailed description, taken in conjunction with the accompanying drawings, and in which:
To provide an overall understanding of the systems, devices, and methods described herein, certain illustrative embodiments will be described. It will be understood that the systems, devices, and methods described herein may be adapted and modified for any suitable application and that such other additions or modifications will not depart from the scope hereof.
Communication network 104 may include any wired or wireless network, such as the Internet, WiMax, wide area cellular, or local area wireless network. Communication network 104 may also include personal area networks, such as Bluetooth and infrared networks. Communications on communications network 104 may be encrypted or otherwise secured using any suitable security or encryption protocol.
Application server 106, which may include any network server or virtual server, such as a file or web server, may access data sources 108 locally or over any suitable network connection. Application server 106 may also include processing circuitry (e.g., one or more computer processors or microprocessors), memory (e.g., RAM, ROM, and/or hybrid types of memory), and one or more storage devices (e.g., hard drives, optical drives, flash drives, tape drives). The processing circuitry included in application server 106 may execute a server process for calculating trust scores, while access application 102 executes a corresponding client process. The access application 102 may be executed by processing circuitry on a user's equipment, such as a computer or a mobile device (e.g., a cell phone, a wearable mobile device such as a smartwatch, etc.). The processing circuitry included in application server 106 and/or the processing circuitry that executes access application 102 may also perform any of the calculations and computations described herein in connection with calculating a trust score. In some embodiments, a computer-readable medium with computer program logic recorded thereon is included within application server 106. The computer program logic may calculate trust scores and may generate such trust scores for display on a display device. In some embodiments, application 102 and/or application server 106 may store a calculation date of a trust score and may generate for display the trust score together with a date of calculation.
Application server 106 may access data sources 108 over the Internet, a secured private LAN, or any other communications network. Data sources 108 may include one or more third-party data sources, such as data from third-party social networking services and third-party ratings bureaus. For example, data sources 108 may include user and relationship data (e.g., “friend” or “follower” data) from one or more of Facebook, MySpace, openSocial, Friendster, Bebo, hi5, Orkut. PerfSpot, Yahoo! 360, LinkedIn, Twitter, Google Buzz, Really Simple Syndication readers or any other social networking website or information service. Data sources 108 may also include data stores and databases local to application server 106 containing relationship information about users accessing application server 106 via access application 102 (e.g., databases of addresses, legal records, transportation passenger lists, gambling patterns, political and/or charity donations, political affiliations, vehicle license plate or identification numbers, universal product codes, news articles, business listings, and hospital or university affiliations).
Application server 106 may be in communication with one or more of data store 110, key-value store 112, and parallel computational framework 114. Data store 110, which may include any relational database management system (RDBMS), file server, or storage system, may store information relating to one or more network communities. For example, one or more of data tables 1100 (
Parallel computational framework 114, which may include any parallel or distributed computational framework or cluster, may be configured to divide computational jobs into smaller jobs to be performed simultaneously, in a distributed fashion, or both. For example, parallel computational framework 114 may support data-intensive distributed applications by implementing a map/reduce computational paradigm where the applications may be divided into a plurality of small fragments of work, each of which may be executed or re-executed on any core processor in a cluster of cores. A suitable example of parallel computational framework 114 includes an Apache Hadoop cluster.
Parallel computational framework 114 may interface with key-value store 112, which also may take the form of a cluster of cores. Key-value store 112 may hold sets of key-value pairs for use with the map/reduce computational paradigm implemented by parallel computational framework 114. For example, parallel computational framework 114 may express a large distributed computation as a sequence of distributed operations on data sets of key-value pairs. User-defined map/reduce jobs may be executed across a plurality of nodes in the cluster. The processing and computations described herein may be performed, at least in part, by any type of processor or combination of processors. For example, various types of quantum processors (e.g., solid-state quantum processors and light-based quantum processors), artificial neural networks, and the like may be used to perform massively parallel computing and processing.
In some embodiments, parallel computational framework 114 may support two distinct phases, a “map” phase and a “reduce” phase. The input to the computation may include a data set of key-value pairs stored at key-value store 112. In the map phase, parallel computational framework 114 may split, or divide, the input data set into a large number of fragments and assign each fragment to a map task. Parallel computational framework 114 may also distribute the map tasks across the cluster of nodes on which it operates. Each map task may consume key-value pairs from its assigned fragment and produce a set of intermediate key-value pairs. For each input key-value pair, the map task may invoke a user-defined map function that transmutes the input into a different key-value pair. Following the map phase, parallel computational framework 114 may sort the intermediate data set by key and produce a collection of tuples so that all the values associated with a particular key appear together. Parallel computational framework 114 may also partition the collection of tuples into a number of fragments equal to the number of reduce tasks.
In the reduce phase, each reduce task may consume the fragment of tuples assigned to it. For each such tuple, the reduce task may invoke a user-defined reduce function that transmutes the tuple into an output key-value pair. Parallel computational framework 114 may then distribute the many reduce tasks across the cluster of nodes and provide the appropriate fragment of intermediate data to each reduce task.
Tasks in each phase may be executed in a fault-tolerant manner, so that if one or more nodes fail during a computation the tasks assigned to such failed nodes may be redistributed across the remaining nodes. This behavior may allow for load balancing and for failed tasks to be re-executed with low runtime overhead.
Key-value store 112 may implement any distributed file system capable of storing large files reliably. For example, key-value store 112 may implement Hadoop's own distributed file system (DFS) or a more scalable column-oriented distributed database, such as HBase. Such file systems or databases may include BigTable-like capabilities, such as support for an arbitrary number of table columns.
Although
Cluster of mobile devices 202 may include one or more mobile devices, such as PDAs, cellular telephones, mobile computers, or any other mobile computing device. Cluster of mobile devices 202 may also include any appliance (e.g., audio/video systems, microwaves, refrigerators, food processors) containing a microprocessor (e.g., with spare processing time), storage, or both. Application server 106 may instruct devices within cluster of mobile devices 202 to perform computation, storage, or both in a similar fashion as would have been distributed to multiple fixed cores by parallel computational framework 114 and the map/reduce computational paradigm. Each device in cluster of mobile devices 202 may perform a discrete computational job, storage job, or both. Application server 106 may combine the results of each distributed job and return a final result of the computation.
The system trust score 302, peer trust score 304, and contextual trust score 306 may be represented in any suitable fashion. As an illustrative example, the system trust score 302, peer trust score 304, and contextual trust score 306 may each be represented as a percentage out of 100 or as a numerical score out of 1000. In other embodiments, the system trust score 302, peer trust score 304, and contextual trust score 306 may be represented by different categories of trustworthiness (e.g., “reliable,” “flaky,” “honest,” “fraudulent,” etc.) or by a graphical scheme (e.g., a color spectrum representing level of trustworthiness). For ease of illustration, the trust score and component scores that comprise the trust scores will be discussed herein as numerical values. However, other methods of portraying a calculated trust score will be contemplated by those of ordinary skill in the art and will not depart from the scope hereof.
Each type of trust score may combine data from data sources according to a specific weighting. For instance, a weighting for a system trust score may be set as:
The following is an example that illustrates one application of a system trust score 302, peer trust score 304, and contextual trust score 306. It will be understood that the following is provided for illustrative purposes only and that the systems, devices, and methods described herein may be further adapted or modified.
John sees an ad at ABC Restaurant for a short order cook and is trying to decide if he should apply. John opens an app on his mobile device and searches for ABC Restaurant. The app shows there are multiple matches to this search, but the nearest one is sorted to the top. After tapping on the correct restaurant, the app shows the ABC Restaurant profile page. The ABC Restaurant profile page includes a system trust score for ABC Restaurant, which is calculated based in part on the ratings from three blogs. John taps to see more details and sees a list of most recent blogs from bloggers. By tapping on individual blogs, he can read the actual article. He can also tap on the bloggers to see their profile page in the app.
The system trust score for ABC Restaurant is also calculated based on previous transactions where ABC Restaurant was the employer. John taps to show a list of previous transactions, ratings of those transactions, and comments.
John taps on the social graph to see how he is connected to the restaurant through one or more networks (e.g., Facebook, MySpace, Twitter, LinkedIn, etc.). From the social graph he sees that Bob, the manager, is a friend of a friend. Based on the social graph data, the app updates the system trust score to calculate a peer trust score between John and ABC Restaurant. The peer trust score is higher than the system trust score to indicate the incremental increase in trustworthiness based on the connections between John and Bob the manager. The app also displays Bob's system trust score, calculated based on publicly available information and a default weighting, and Bob's peer trust score with respect to John, which also takes into account the social graph data.
John decides to apply for the job. After an interview, Bob the manager is deciding whether or not to hire John as a short order cook. Bob uses the app to search for John. There are multiple results for John, but Bob eventually finds him and taps on his entry. John's profile page displays his system trust score, calculated based on publicly available information (e.g., credit score, verification data, search engine mining, employment history, etc.) and a default weighting. Bob taps on the social graph to see how he is connected to John. He discovers that they are connected through a friend of a friend. The app updates John's system trust score based on the social network data to calculate a peer trust score between John and Bob, which is higher than John's system trust score to indicate the incremental increase in trustworthiness due to the connections between John and Bob. The app also shows average ratings from previous transactions where John was the employee. Bob taps to show a list of transactions, which can be ordered into chronological order and filtered by type of job. Bob also indicates to the app that he wishes to hire John as an employee. The app adjusts the weightings of the trust score to give a higher weight to the employee history rather than other components (such as credit score). The app uses the adjusted weightings to update the peer trust score to calculate the contextual trust score, which represents John's trustworthiness as a potential employee.
After reviewing the information in the app, Bob has decided to hire John. From John's profile page, he taps on the Action icon and chooses “Hire”. The app prompts Bob to fill in relevant information such as position, start date, annual salary, and vacation days per year. After confirming the data, the transaction appears in Bob's Notification list, with the status of “Waiting for John . . . ” John receives a notification on his phone. He opens the app and sees a new transaction in his Notifications list. The app prompts John to confirm the details of his new job. John chooses to confirm, and Bob receives a notification that John has confirmed the transaction.
As illustrated in the above example, a user may request a system trust score for another entity, which may then be subsequently refined into a peer trust score based on information specific to the parties involved and into a contextual trust score based on the details of an activity/transaction to be performed by the parties.
Data verification component 404 may include data that verifies information associated with the target entity. In some embodiments, the data verification component 404 may include verification of contact information, including, but not limited to, email address, phone number, and/or mailing address. The data verification component may also comprise email, IM, and other messaging factors, such as frequency of messages, time of day of messages, depth of thread, or a review of threads for key transaction/activity types (e.g., loan, rent, buy, etc.). Data verification component 404 may take into account data from passport and/or other government IDs, tax return factors (e.g., a summary of a tax return to prove income), educational data (e.g., certificates of degree/diploma), group affiliation factors (e.g., invoices that prove membership to a group), achievements (e.g., proof of awards, medals, honorary citations, etc.), employment data (e.g., paystub data). The data verification component 404 may also incorporate facial recognition software to verify certain documents, such as IDs. In some embodiments, this facial recognition software may be used for subsequent verification of the user's identity. As an illustrative example, the data verification component 404 may be used as a part of an airport scanning system to verify the user's identity. The data verification component 404 may comprise subcomponents such as data corresponding to the above illustrative examples, and as more subcomponents are verified, the higher the data verification component 404. The subcomponents may be combined to determine the data verification component 404 in any suitable manner, such as a weighted sum or the method discussed further below in relation to
Network connectivity component 406 is discussed further below in relation to
Credit score component 408 may comprise any suitable financial information associated with the target entity, including income, checking/savings account information (number of accounts, value), and credit score information from one or more institutions. The credit score information may be received from any typical credit score agency, including, but not limited to, Transunion, Equifax, and Experian. Credit score factors may also be taken into account, such as number of credit accounts, credit utilization, length of credit history, number of late payments, etc. Other financial information taken into account may include prior loan and payment data, data on net worth or assets/liabilities, and information on any prior infractions. The various financial data may be combined using any suitable approach, including, but not limited to, the methods discussed below in relation to
Court data component 410 may include any data on activity associated with the target entity in a criminal or civil court. For example, court data component 410 may comprise data on how many cases involve the entity suing someone else and the type of suit, how many cases involve the target entity as the defendant, any criminal cases that may have a negative impact on trustworthiness, and the final holding/disposition of any concluded cases (e.g., acquitted, convicted, settled, etc.). Court data may be derived from any publicly available sources and from any available municipal, state, federal, or international court.
A ratings/feedback data component 412 may include any data that reflects a rating or feedback associated with the target entity. For instance, online rating sites such as Yelp may provide ratings information on various businesses. Any ratings of the target entity, information on volume, number of ratings, average rating, who rates the target entity, and whether the target entity responds to comments may be taken into account. In some embodiments, ratings data may be received from ratings institutions, such as the Better Business Bureau. Feedback data may include any positive or negative comments associated with the target entity. In some embodiments, feedback data may include comments made by peers in a social network. In some embodiments, the number and timing of ratings by other users or entities may be used to affect the ratings/feedback data component 412. For instance, a lack of negative feedback for a specified period of time may result in an increase (or decrease) in the ratings/feedback data component 412. Similarly, a lack of positive feedback for a specified period of time may result in a decrease (or increase) in the ratings/feedback data component 412.
Group/demographics component 414 may include information on group membership of the target entity or demographic information such as age, sex, race, location, etc. The group data may suggest an activity performed by the target entity. For instance, membership to a national sailing club may indicate an interest in sailing and boats. In some embodiments, a peer trust score may be adjusted to take into account the group/demographic component. For instance, the peer trust score for a target entity may be increased if a first entity and the target entity are both members of the same national sailing club. As another example, similarities in demographic information (age, sex, race, location, etc.) may indicate an incremental increase in trustworthiness between a first and the target entity, and the peer trust score for the target entity may be adjusted accordingly.
The search engine mining component 416 may include analytics performed on suitable search engines, such as Google or Yahoo. Websites/blogs/articles may be searched and scanned for entries about the target entry and a positive or negative sentiment may be detected and stored for such entries. Number of articles, sentiment, timing of the articles, may indicate a positive or negative adjustment to the search engine mining component 416. In some embodiments, online shopping or auction websites such as eBay may be scanned for information associated with the target entity, such as rating and volume of transactions, feedback comments, number of bought/sold items, average value of items, and category of items (e.g., hardware, software, furniture, etc.).
Transaction history component 418 may comprise any information on past transactions associated with the target entity. Successful transactions or activities may be identified and positively impact the transaction history component score. For example, if I loan John $100 and he promptly pays me back, I may be more inclined to loan him money in the future. Transaction history data may be locally tracked and stored (e.g., by application 102 in
In some embodiments, the transactions history component 418 may comprise interactions between previous transactions in the transaction history between a first entity and a second entity. In this manner, processing circuitry may take into account elements of regret and forgiveness in determining a trust score. For example, a first transaction may correspond to an increase or decrease in a trust score, while a second, subsequent transaction related to the first transaction may result in an adjustment to the peer trust score in the opposite direction. The adjustment may be either a decrease in the trust score (e.g., regret or suspicion) or an increase in the trust score (e.g., forgiveness or redemption). As an illustrative example, a subject may have stolen a car in the past and be subsequently convicted of the theft and sentenced to serve 3 years in prison for the crime. The initial theft may serve to decrease the subject's trust score, reflecting the increased suspicion associated with a known delinquent, while the subsequent conviction and sentence might serve to increase the subject's trust score, reflecting a level of redemption in the trustworthiness of the subject.
In some embodiments, the transactions that comprise the transactions history component 418 may be associated with an increase or decrease in a trust score over time. For example, a transaction may contribute to an initial increase in a trust score, and over time, the initial increase may decay until the trust score returns to an initial value. Similarly, a transaction may cause an initial decrease in a trust score, and over time, the initial decrease may decay until the trust score returns to an initial value.
In some embodiments, any one of the system, peer, or contextual trust score may also include a location component that takes into account a geographic location of an entity. For example, the location of an end user as determined by GPS coordinates or an address of a business may be incorporated into the calculation of a trust score. In some embodiments, a peer trust score may take into account the location of a first entity and a second entity and adjust the trust score accordingly. For instance, if a first user and a second user happen to be from the same hometown, then the peer trust scores may be increase to reflect this common information. In some embodiments, the location of the entity may provide an automatic increase/decrease in the trust score. For instance, a particular location may be known as a dangerous neighborhood, city, or region, and the trust scores of all entities located or associated with the dangerous location may be automatically decreased to reflect this danger. As an illustrative example, a user who travels to a country close to a known warzone may not be as comfortable trusting strangers in the country. The trust levels of others located in the same location as the user may be automatically decreased to reflect the increased suspicion. In some embodiments, the user may be traveling with his friends, as indicated by the high level of peer trust scores the user has with the plurality of people located around the user. Processing circuitry may determine that the user is surrounded by friends in any suitable manner, including explicit indications of friendship, common hometown, place of work, or any other common information. If the user is traveling to a dangerous location, but is traveling with friends, then the trust scores of other entities associated with the dangerous location may still be decreased, but they may be decreased by a smaller amount than if the user was not traveling with friends.
In some embodiments, any of the system, peer, and/or contextual trust scores may take into account biological responses of an end user. For instance, mobile devices may include cell phones, smart watches, heart rate monitors, and other wearable mobile devices that can monitor one or more biological responses of an end user (e.g., heart rate, breathing rate, brain waves, sweat response, etc.). These detected biological responses of an end user, in conjunction with location information, may be used, in part, to determine a trust score. For example, an increase in heart rate may be an indication of anxiety, and may result in a decrease in trust score. The increase in heart rate may be caused by the user moving to a new location, in which case the trust score associated with that location may be decreased. The increase in heart rate may have been caused by a first user moving into close proximity with a second user, in which case the peer trust score with respect to the second user may be decreased, to reflect the increased anxiety that the first user feels around the second user.
wherein wi is the weighting as given by the default weighting above, and c, is the component score.
In some embodiments, the default weightings may be adjusted according to user-specified values. For example, as discussed above, users who care more about network connectivity may increase the weighting for the network connectivity component 506, and users who care less about financial responsibility may choose to decrease credit score component 508. In some embodiments, the default weightings above may be automatically adjusted, for example by application 102, to reflect a peer trust score or contextual trust score. For example, application 102 may detect that a first and second entity are entering into a financial transaction and may automatically adjust the weight for the credit score component 508 to reflect the importance of this component to the type of activity. Thus, the users may be provided with an contextual trust score that weights factors in a more relevant manner than the default weightings.
In some embodiments, at least one of the system trust score, peer trust score, and contextual trust score may be represented by a mean value and confidence band. The confidence band may represent a statistical variance in the calculated trust score. For example, each of the component scores may be associated with a mean score μ and a standard deviation σ based on how trustworthy the data source is. The mean and standard deviation for each of the component scores may be combined accordingly. As will be understood by those of ordinary skill in the art, the mean value of the total component scores may be represented by a sum of the mean value of each component score. The variance of two component scores together may be combined using the following equation:
V(A+B)=V(A)+V(B)+2*Covar(A,B)
where V(A) is the variance (i.e., the square of the standard deviation) of component A, V(B) is the variance of component B, and Covar(A,B) is the covariance of components A and B.
Icon 602 and initial score 604 may graphically represent a first trust score of a target entity. Although icon 602 is depicted as a smiley face, it will be understood that any suitable graphical representation may be utilized to represent a relative trust level of the target entity. In some embodiments, the initial score 604 may be a system trust score for the target entity calculated using a default set of weights. In other embodiments, the initial score 604 may be a peer trust score calculated in relation to the user of the mobile app. For instance, the initial score 604 may represent a trust level that takes into account mutual friends of the requesting user and the target user.
The requesting user may use transaction selector 606 to indicate an activity/transaction to be performed with the target user. In some embodiments, transaction selector 606 may be optional, and no transaction is needed to calculate a revised score. Although transaction selector 606 is depicted as a dropdown box, any suitable input method (e.g., text input box, radio buttons, etc.) may be utilized to receive an indication of an activity/transaction from the requesting user. After an activity/transaction is selected, transaction details field 608 may provide further details or options. For example, if the requesting user indicates that the target entity wishes to request a loan, then the transaction details field 608 may include a field for indicating the amount of the loan. In this manner, a different weighting of components may be used for a $10 loan as opposed to a $100,000 loan. The requesting user may add an additional transaction using additional transaction button 610. In cases where multiple transactions are indicated, weightings for the multiple transactions may be averaged.
Revised score icon 612 may indicate a revised trust score calculated based on the information entered into transaction selector 606 and transaction details field 608. In some embodiments, the revised score icon 612 may reflect a peer trust score, for example, when a transaction is not selected in transaction selector 606. In other embodiments, the revised score icon 612 may reflect a contextual trust score calculated based on the activity/transaction and transaction details indicated in transaction selector 606 and transaction details field 608. The revised score icon 612 may include a graphical representation of the revised trust score, similar to icon 602. In the illustrative example depicted in
The first profile score 614 and the second profile score 616 may indicate one or more of a system trust score, peer trust score, and/or contextual trust score for the requesting user. As with icon 602 and icon 612, the first profile score 614 and second profile score 616 may include a graphical representation, such as a smiley face, of the respective trust score.
As discussed above in relation to
In some embodiments, the weighting profiles may be stored, for example in data store 110 depicted in
In some embodiments, the user may indicate an initial or base trust score factor that may be applied to every other user. At least one of the system trust score, peer trust score, and contextual trust score may then be calculated as updates to the initial or base trust score that the user has indicated. For example, each of the components discussed in relation with
First profile score 708 and second profile score 710 may be substantially similar to first profile score 614 and second profile score 616 depicted in
According to the graded scale depicted in
The metric 802 and the steps of the graded scale may be determined by a server, such as application server 106 depicted in
The metric for the target user may fall into one of regions 906, 908, 910, 914, 916, and 918. As will be understood by those of ordinary skill in the art, regions 906 and 918 represent about 2.5% each of distribution 900, regions 908 and 916 represent about 13.5% each of distribution 900, and regions 910 and 914 represent about 34% each of distribution 900. The application or application server may assign a component score depending on which of regions 906, 908, 910, 914, 916, and 918 the metric of the target user falls into. For instance, the component score for the target user may be relatively low if the metric falls within regions 906 or 918 and may be relatively high if the metric falls within regions 910 or 914. A graded scale, similar to table 800 depicted in
The network graph 1000 may represent a visualization of a network that connects a requesting entity, depicted by source node 1002, and a target entity, depicted by target node 1004. One or more intermediate nodes, such as intermediate node 1006, may also be displayed, as well as paths 1008 that connect nodes 1002, 1004, and 1006. In some embodiments, a dominant path 1010 may be displayed and visually distinguished from other paths 1008. The dominant path 1010 may be determined using any suitable algorithm. For example, the dominant path 1010 may represent the shortest-length path from source node 1002 to source node 1004. In other embodiments, the dominant path 1010 may represent a path through specific intermediate nodes, such as nodes with relatively high trust values. For example, a longer path from node 1002 through node 1006 to node 1004 may have higher trust at each link of the path than the shorter path 1010.
In some embodiments, each of the nodes 1002, 1004, and 1006 may include images, text, or both, such as a profile picture associated with the entity depicted by the nodes. In some embodiments, the network graph 1000 may be generated for display in a scrollable display, wherein a user may scroll and zoom the network graph 1000 to see more and less nodes as desired.
In an embodiment, a path counting approach may be used where processing circuitry is configured to count the number of paths between a first node n1 and a second node n2 within a network community. A connectivity rating Rn1n2 may then be assigned to the nodes. The assigned connectivity rating may be proportional to the number of subpaths, or relationships, connecting the two nodes, among other possible measures. Using the number of subpaths as a measure, a path with one or more intermediate nodes between the first node n1 and the second node n2 may be scaled by an appropriate number (e.g., the number of intermediate nodes) and this scaled number may be used to calculate the connectivity rating.
In some embodiments, weighted links are used in addition to or as an alternative to the subpath counting approach. Processing circuitry may be configured to assign a relative user weight to each path connecting a first node n1 and a second node n2 within a network community. A user connectivity value may be assigned to each link. For example, a user or entity associated with node n may assign user connectivity values for all outgoing paths from node n1. In some embodiments, the connectivity values assigned by the user or entity may be indicative of that user or entity's trust in the user or entity associated with node n2. The link values assigned by a particular user or entity may then be compared to each other to determine a relative user weight for each link.
The relative user weight for each link may be determined by first computing the average of all the user connectivity values assigned by that user (i.e., the out-link values). If ti is the user connectivity value assigned to link i, then the relative user weight, wi, assigned to that link may be given in accordance with:
wi=1+(ti−
To determine the overall weight of a path, in some embodiments, the weights of all the links along the path may be multiplied together. The overall path weight may then be given in accordance with:
wpath=Π(w1) (2)
The connectivity value for the path may then be defined as the minimum user connectivity value of all the links in the path multiplied by the overall path weight in accordance with:
tpath=wpath×tmin (3)
To determine path connectivity values, in some embodiments, a parallel computational framework or distributed computational framework (or both) may be used. For example, in one embodiment, a number of core processors implement an Apache Hadoop or Google MapReduce cluster. This cluster may perform some or all of the distributed computations in connection with determining new path link values and path weights.
The processing circuitry may identify a changed node within a network community. For example, a new outgoing link may be added, a link may be removed, or a user connectivity value may have been changed. In response to identifying a changed node, in some embodiments, the processing circuitry may re-compute link, path, and weight values associated with some or all nodes in the implicated network community or communities.
In some embodiments, only values associated with affected nodes in the network community are recomputed after a changed node is identified. If there exists at least one changed node in the network community, the changed node or nodes may first undergo a prepare process. The prepare process may include a “map” phase and “reduce” phase. In the map phase of the prepare process, the prepare process may be divided into smaller sub-processes which are then distributed to a core in the parallel computational framework cluster. For example, each node or link change (e.g., tail to out-link change and head to in-link change) may be mapped to a different core for parallel computation. In the reduce phase of the prepare process, each out-link's weight may be determined in accordance with equation (1). Each of the out-link weights may then be normalized by the sum of the out-link weights (or any other suitable value). The node table may then be updated for each changed node, its in-links, and its out-links.
After the changed nodes have been prepared, the paths originating from each changed node may be calculated. Once again, a “map” and “reduce” phase of this process may be defined. During this process, in some embodiments, a depth-first search may be performed of the node digraph or node tree. All affected ancestor nodes may then be identified and their paths recalculated.
In some embodiments, to improve performance, paths may be grouped by the last node in the path. For example, all paths ending with node in, may be grouped together, all paths ending with node n2 may be grouped together, and so on. These path groups may then be stored separately (e.g., in different columns of a single database table). In some embodiments, the path groups may be stored in columns of a key-value store implementing an HBase cluster (or any other compressed, high performance database system, such as BigTable).
In some embodiments, one or more threshold functions may be defined. The threshold function or functions may be used to determine the maximum number of links in a path that will be analyzed in a connectivity determination or connectivity computation. Threshold factors may also be defined for minimum link weights, path weights, or both. Weights falling below a user-defined or system-defined threshold may be ignored in a connectivity determination or connectivity computation, while only weights of sufficient magnitude may be considered.
In some embodiments, a user connectivity value may represent the degree of trust between a first node and a second node. In one embodiment, node n1 may assign a user connectivity value of l1 to a link between it and node n2. Node n2 may also assign a user connectivity value of l2 to a reverse link between it and node n1. The values of l1 and l2 may be at least partially subjective indications of the trustworthiness of the individual or entity associated with the node connected by the link. A user (or other individual authorized by the node) may then assign this value to an outgoing link connecting the node to the individual or entity. Objective measures (e.g., data from third-party ratings agencies or credit bureaus) may also be used, in some embodiments, to form composite user connectivity values indicative of trust. The subjective, objective, or both types of measures may be automatically harvested or manually inputted for analysis.
Table 1104 may store user connectivity values. In some embodiments, user connectivity values may be assigned automatically by the system (e.g., by application server 106 (
In some embodiments, user connectivity values may be manually assigned by members of the network community. These values may represent, for example, the degree or level of trust between two users or nodes or one node's assessment of another node's competence in some endeavor. User connectivity values may include a subjective component and an objective component in some embodiments. The subjective component may include a trustworthiness “score” indicative of how trustworthy a first user or node finds a second user, node, community, or subcommunity. This score or value may be entirely subjective and based on interactions between the two users, nodes, or communities. This manual user connectivity score may “override” one or more of the system trust score, peer trust score, or contextual trust score. When a user “overrides” one of the above trust scores with a manual trust score, the user-specified trust score may be provided concurrently with, or instead of, the overridden trust score.
In some embodiments, a system administrator may override one or more of the system trust score, peer trust score, or contextual trust score. For example, a system administrator may override a system trust score of an entity to take into account recent trends or events. When a trust score is overridden by the system administrator, the administrator's trust score may be provided concurrently with, or instead of, the overridden trust score. When the overridden trust score reaches a specified range or threshold of the administrator's trust score, the system may automatically revert back to the overridden trust score. As an illustrative example, the system administrator may decrease a system trust score of an entity that has taken negative public attention in the news. The overridden trust score will continue to be calculated by the system and will gradually reflect the negative public attention of the entity. When the overridden trust score reaches within a certain range of the administrator's trust level (e.g., within 10%), then the system will automatically revert back to the calculated score. In some embodiments, the administrator's trust score will be provided to a user with a notification that the score was overridden and/or a reason why the trust score was overridden.
Table 1104 may store an identification of a link head, link tail, and user connectivity value for the link. Links may or may not be bidirectional. For example, a user connectivity value from node n1 to node n2 may be different (and completely separate) than a link from node n2 to node n1. Especially in the trust context described above, each user can assign his or her own user connectivity value to a link (i.e., two users need not trust each other an equal amount in some embodiments).
Table 1106 may store an audit log of table 1104. Table 1106 may be analyzed to determine which nodes or links have changed in the network community. In some embodiments, a database trigger is used to automatically insert an audit record into table 1106 whenever a change of the data in table 1104 is detected. For example, a new link may be created, a link may be removed, or a user connectivity value may be changed. This audit log may allow for decisions related to connectivity values to be made prospectively (i.e., before an anticipated event). Such decisions may be made at the request of a user, or as part of an automated process. This prospective analysis may allow for the initiation of a transaction (or taking of some particular action) in a fluid and/or dynamic manner. After such a change is detected, the trigger may automatically create a new row in table 1106. Table 1106 may store an identification of the changed node, and identification of the changed link head, changed link tail, and the user connectivity value to be assigned to the changed link. Table 1106 may also store a timestamp indicative of the time of the change and an operation code. In some embodiments, operation codes may include “insert,” “update,” or “delete” operations, corresponding to whether a link was inserted, a user connectivity value was changed, or a link was deleted, respectively. Other operation codes may be used in other embodiments.
Data structure 1210 may include node table 1212. In the example shown in
In some embodiments, the processes described with respect to
At step 1302, a determination is made whether at least one node has changed in the network community. As described above, an audit record may be inserted into table 1106 (
If a node change is not detected at step 1304, then process 1300 enters a sleep mode at step 1306. For example, in some embodiments, an application thread or process may continuously check to determine if at least one node or link has changed in the network community. In other embodiments, the application thread or process may periodically check for changed links and nodes every in seconds, where n is any positive number. After the paths are calculated that go through a changed node at step 1316 or after a period of sleep at step 1306, process 1300 may determine whether or not to loop at step 1308. For example, if all changed nodes have been updated, then process 1300 may stop at step 1318. If, however, there are more changed nodes or links to process, then process 1300 may loop at step 1308 and return to step 1304.
In practice, one or more steps shown in process 1300 may be combined with other steps, performed in any suitable order, performed in parallel (e.g., simultaneously or substantially simultaneously), or removed.
If there are no more link changes at step 1322, then, in reduce phase 1324, a determination may be made at step 1326 that there are more nodes and link changes to process. If so, then the next node and its link changes may be retrieved at step 1328. The most recent link changes may be preserved at step 1330 while any intermediate link changes are replaced by more recent changes. For example, the timestamp stored in table 1106 (
As shown in
If there are no more changed nodes at step 1350, then, in reduce phase 1352, a determination may be made at step 1354 that there are more nodes and paths to process. If so, then the next node and its paths may be retrieved at step 1356. At step 1358, buckets may be created by grouping paths by their head. If a bucket contains only the NULL path at step 1360, then the corresponding cell in the node table may be deleted at step 1362. If the bucket contains more than the NULL path, then at step 1364 the bucket is saved to the corresponding cell in the node table. If there are no more nodes and paths to process at step 1356, the process may stop at step 1374.
As shown in
If there are no more changed nodes at step 1378, then, in reduce phase 1380, a determination may be made at step 1384 that there are more node and deletion pairs to process. If so, then the next node and its deletion pairs may be retrieved at step 1384. At step 1386, for each deletion pair, any paths that go through the changed node in the old bucket may be deleted. If there are no more nodes and deletion pairs to process at step 1382, the process may stop at step 1394.
As shown in
If there are no more changed nodes at step 1398, then, in reduce phase 1400, a determination may be made at step 1402 that there are more node and paths to process. If so, then the next node and its paths may be retrieved at step 1404. Each path may then be added to the appropriate node bucket at step 1406. If there are no more nodes and paths to process at step 1402, the process may stop at step 1416.
At step 1422, processing circuitry, such as processing circuitry of access application 102 or application server 106, may verify at least one entry in an entity's profile. The entry may be one or more pieces of verification data, such as verification data described in connection with data verification component 404 depicted in
At step 1434, the processing circuitry may determine whether user-specified weightings have been received. For example, a user may have specified custom weightings through a user interface such as interface 700 depicted in
At step 1502, processing circuitry, such as processing circuitry of access application 102 or application server 106, may receive a system trust score. The system trust score may have been calculated previously, such as by a method similar to process 1420 depicted in
At step 1508, the processing circuitry may update component scores based on the information from steps 1502-1506. In some embodiments, updating component scores comprises updating less than all of the component scores that comprise the system trust score. For example, the processing circuitry may only update the network connectivity component to take into account the mutual contacts of the first entity and the second entity. Other component scores that were calculated with respect to the second entity's system trust score, such as credit score or court data, may not be affected by the additional social graph information. At step 1510, the processing circuitry may calculate the peer trust score based on the updated components by, for instance, combining the component scores using a weighted average. In some embodiments, the processing circuitry may calculate the peer trust score in response to a user request for the peer trust score. For example, the user may press calculate button 618 depicted in
At step 1602, processing circuitry, such as processing circuitry of access application 102 or application server 106, may receive a peer trust score. The system trust score may have been calculated previously, such as by a method similar to process 1500 depicted in
The foregoing is merely illustrative of the principles of the disclosure, and the systems, devices, and methods described herein are presented for purposes of illustration, and not of limitation. Variations and modifications will occur to those of skill in the art after reviewing this disclosure. The disclosed features may be implemented, in any combination and subcombination (including multiple dependent combinations and subcombinations), with one or more other features described herein. The various features described or illustrated above, including any components thereof, may be combined or integrated in other systems. Moreover, certain features may be omitted or not implemented. Examples, changes, substitutions, and alterations ascertainable by one skilled in the art can be made without departing from the scope of the information disclosed herein.
This application is a continuation of U.S. patent application Ser. No. 15/400,471, filed Jan. 6, 2017, entitled “CALCULATING A TRUST SCORE”, which is a continuation of U.S. patent application Ser. No. 14/664,285, filed Mar. 20, 2015, issued as U.S. Pat. No. 9,578,043 and entitled “CALCULATING A TRUST SCORE,” the entirety of which is incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5555542 | Ogura et al. | Sep 1996 | A |
5860605 | Van Der Zanden | Jan 1999 | A |
6108308 | Flavin et al. | Aug 2000 | A |
6286007 | Miller et al. | Sep 2001 | B1 |
6356902 | Tan et al. | Mar 2002 | B1 |
6446048 | Wells et al. | Sep 2002 | B1 |
6509898 | Chi et al. | Jan 2003 | B2 |
6633886 | Chong | Oct 2003 | B1 |
6708308 | De Souza et al. | Mar 2004 | B2 |
6738777 | Bliss et al. | May 2004 | B2 |
6751729 | Giniger et al. | Jun 2004 | B1 |
6823299 | Contreras | Nov 2004 | B1 |
7069259 | Horvitz et al. | Jun 2006 | B2 |
7086085 | Brown et al. | Aug 2006 | B1 |
7130262 | Cortez et al. | Oct 2006 | B1 |
7130908 | Pecus et al. | Oct 2006 | B1 |
7139837 | Parekh et al. | Nov 2006 | B1 |
7266649 | Yoshida et al. | Sep 2007 | B2 |
7272719 | Bleckmann et al. | Sep 2007 | B2 |
7451365 | Wang et al. | Nov 2008 | B2 |
7458049 | Tuncer et al. | Nov 2008 | B1 |
7512612 | Akella et al. | Mar 2009 | B1 |
7539697 | Akella et al. | May 2009 | B1 |
7664802 | Aaltonen et al. | Feb 2010 | B2 |
7668665 | Kim | Feb 2010 | B2 |
7685192 | Scofield et al. | Mar 2010 | B1 |
7743208 | Yoshida et al. | Jun 2010 | B2 |
7805407 | Verbeke et al. | Sep 2010 | B1 |
7822631 | Vander Mey | Oct 2010 | B1 |
7856449 | Martino et al. | Dec 2010 | B1 |
7865551 | Mcculler | Jan 2011 | B2 |
7886334 | Walsh | Feb 2011 | B1 |
7930255 | Choi et al. | Apr 2011 | B2 |
8010458 | Galbreath et al. | Aug 2011 | B2 |
8010460 | Work et al. | Aug 2011 | B2 |
8010602 | Shen et al. | Aug 2011 | B2 |
8108536 | Hernacki et al. | Jan 2012 | B1 |
8156558 | Goldfeder et al. | Apr 2012 | B2 |
8170958 | Gremett et al. | May 2012 | B1 |
8180804 | Narayanan et al. | May 2012 | B1 |
8214883 | Obasanjo et al. | Jul 2012 | B2 |
8234688 | Grandison et al. | Jul 2012 | B2 |
8237714 | Burke | Aug 2012 | B1 |
8244848 | Narayanan | Aug 2012 | B1 |
8261078 | Barriga et al. | Sep 2012 | B2 |
8301617 | Muntz et al. | Oct 2012 | B2 |
8306973 | Ohazama et al. | Nov 2012 | B2 |
8316056 | Wable | Nov 2012 | B2 |
8392590 | Bouchard et al. | Mar 2013 | B2 |
8443366 | Yancey | May 2013 | B1 |
8468103 | Galbreath et al. | Jun 2013 | B2 |
8489641 | Seefeld et al. | Jul 2013 | B1 |
8516196 | Jain et al. | Aug 2013 | B2 |
8572129 | Lee et al. | Oct 2013 | B1 |
8601025 | Shajenko et al. | Dec 2013 | B1 |
8606721 | Dicker | Dec 2013 | B1 |
8621215 | Iyer | Dec 2013 | B1 |
8682837 | Skelton | Mar 2014 | B2 |
8683423 | Amaral et al. | Mar 2014 | B2 |
8688701 | Ghosh et al. | Apr 2014 | B2 |
8725673 | Kast et al. | May 2014 | B2 |
8832093 | Redstone et al. | Sep 2014 | B2 |
8832790 | Villa et al. | Sep 2014 | B1 |
8874572 | Broyles | Oct 2014 | B1 |
8949250 | Garg et al. | Feb 2015 | B1 |
9143503 | Lo et al. | Sep 2015 | B2 |
9147273 | Allen et al. | Sep 2015 | B1 |
9154491 | Leske | Oct 2015 | B1 |
9171336 | Englar et al. | Oct 2015 | B2 |
9171338 | Chrapko et al. | Oct 2015 | B2 |
9223978 | Kraemer | Dec 2015 | B2 |
9319419 | Sprague et al. | Apr 2016 | B2 |
9390243 | Dhillon et al. | Jul 2016 | B2 |
9438619 | Chan et al. | Sep 2016 | B1 |
9443004 | Chan | Sep 2016 | B2 |
9443044 | Gou et al. | Sep 2016 | B2 |
9460475 | Chrapko et al. | Oct 2016 | B2 |
9584540 | Chan et al. | Feb 2017 | B1 |
9613341 | Shivakumar | Apr 2017 | B2 |
9721296 | Chrapko | Aug 2017 | B1 |
9747650 | Chrapko et al. | Aug 2017 | B2 |
9846896 | Shah | Dec 2017 | B2 |
10007895 | Vanasco | Jun 2018 | B2 |
10585893 | Pantressangle et al. | Mar 2020 | B2 |
20030046280 | Rotter et al. | Mar 2003 | A1 |
20030076825 | Guruprasad | Apr 2003 | A1 |
20030133411 | Ise et al. | Jul 2003 | A1 |
20030227924 | Kodialam et al. | Dec 2003 | A1 |
20040018518 | Krieb et al. | Jan 2004 | A1 |
20040088147 | Wang et al. | May 2004 | A1 |
20040122803 | Dom et al. | Jun 2004 | A1 |
20040181461 | Raiyani et al. | Sep 2004 | A1 |
20040181518 | Mayo et al. | Sep 2004 | A1 |
20040239674 | Ewald | Dec 2004 | A1 |
20050083936 | Ma | Apr 2005 | A1 |
20050096987 | Miyauchi | May 2005 | A1 |
20050243736 | Faloutsos et al. | Nov 2005 | A1 |
20050256949 | Gruhl et al. | Nov 2005 | A1 |
20060212931 | Shull et al. | Sep 2006 | A1 |
20060248573 | Pannu et al. | Nov 2006 | A1 |
20060259957 | Tam et al. | Nov 2006 | A1 |
20060271564 | Meng Muntz et al. | Nov 2006 | A1 |
20060287842 | Kim | Dec 2006 | A1 |
20060290697 | Madden et al. | Dec 2006 | A1 |
20060294134 | Berkhim et al. | Dec 2006 | A1 |
20070087819 | Van Luchene et al. | Apr 2007 | A1 |
20070109302 | Tsuboshita et al. | May 2007 | A1 |
20070124291 | Hassan et al. | May 2007 | A1 |
20070136086 | Luerssen | Jun 2007 | A1 |
20070143629 | Hardjono et al. | Jun 2007 | A1 |
20070162761 | Davis et al. | Jul 2007 | A1 |
20070180495 | Hardjono et al. | Aug 2007 | A1 |
20070214259 | Ahmed et al. | Sep 2007 | A1 |
20070220146 | Suzuki | Sep 2007 | A1 |
20070263012 | Panditharadhya et al. | Nov 2007 | A1 |
20070282886 | Purang et al. | Dec 2007 | A1 |
20080005096 | Moore | Jan 2008 | A1 |
20080015916 | Cossey et al. | Jan 2008 | A1 |
20080059576 | Liu et al. | Mar 2008 | A1 |
20080086442 | Dasdan et al. | Apr 2008 | A1 |
20080101343 | Monette et al. | May 2008 | A1 |
20080104225 | Zhang et al. | May 2008 | A1 |
20080133391 | Kurian et al. | Jun 2008 | A1 |
20080183378 | Weidner | Jul 2008 | A1 |
20080281694 | Kretz et al. | Nov 2008 | A1 |
20080288457 | Aaltonen et al. | Nov 2008 | A1 |
20080288612 | Kwon | Nov 2008 | A1 |
20090024629 | Miyauchi | Jan 2009 | A1 |
20090027392 | Jadhav et al. | Jan 2009 | A1 |
20090043489 | Weidner | Feb 2009 | A1 |
20090049517 | Somiotti et al. | Feb 2009 | A1 |
20090063157 | Seo | Mar 2009 | A1 |
20090064293 | Li et al. | Mar 2009 | A1 |
20090094134 | Toomer | Apr 2009 | A1 |
20090106822 | Obasanjo et al. | Apr 2009 | A1 |
20090198562 | Wiesinger et al. | Aug 2009 | A1 |
20090276233 | Brimhall et al. | Nov 2009 | A1 |
20090296568 | Kitada | Dec 2009 | A1 |
20090327054 | Yao | Dec 2009 | A1 |
20100004940 | Choi et al. | Jan 2010 | A1 |
20100010826 | Rosenthal et al. | Jan 2010 | A1 |
20100043055 | Baumgart | Feb 2010 | A1 |
20100076987 | Schreiner | Mar 2010 | A1 |
20100100963 | Mahaffey | Apr 2010 | A1 |
20100106557 | Buss | Apr 2010 | A1 |
20100121707 | Goeldi | May 2010 | A1 |
20100161662 | Jonas et al. | Jun 2010 | A1 |
20100169137 | Jastrebski et al. | Jul 2010 | A1 |
20100180048 | Guo et al. | Jul 2010 | A1 |
20100205541 | Rapaport et al. | Aug 2010 | A1 |
20100217525 | King et al. | Aug 2010 | A1 |
20100250605 | Pamu et al. | Sep 2010 | A1 |
20100262610 | Acosta et al. | Oct 2010 | A1 |
20100274815 | Vanasco | Oct 2010 | A1 |
20100309915 | Pirbhai et al. | Dec 2010 | A1 |
20110029467 | Spehr et al. | Feb 2011 | A1 |
20110055897 | Arasaratnam | Mar 2011 | A1 |
20110113098 | Walsh et al. | May 2011 | A1 |
20110173344 | Mihaly et al. | Jul 2011 | A1 |
20110208866 | Marmolejo-Meillon et al. | Aug 2011 | A1 |
20110219034 | Dekker et al. | Sep 2011 | A1 |
20110246237 | Vdovjak | Oct 2011 | A1 |
20110246412 | Skelton | Oct 2011 | A1 |
20110265011 | Taylor et al. | Oct 2011 | A1 |
20110283205 | Nie et al. | Nov 2011 | A1 |
20110295626 | Chen et al. | Dec 2011 | A1 |
20110314557 | Marshall | Dec 2011 | A1 |
20120109714 | Azar | May 2012 | A1 |
20120110005 | Kuo et al. | May 2012 | A1 |
20120182822 | Hayashi | Jun 2012 | A1 |
20120182882 | Chrapko et al. | Jul 2012 | A1 |
20120197758 | Zhong et al. | Aug 2012 | A1 |
20120204265 | Judge | Aug 2012 | A1 |
20120278767 | Stibel et al. | Nov 2012 | A1 |
20120282884 | Sun | Nov 2012 | A1 |
20120290427 | Reed et al. | Nov 2012 | A1 |
20120317149 | Jagota et al. | Dec 2012 | A1 |
20120317200 | Chan | Dec 2012 | A1 |
20120324027 | Vaynblat et al. | Dec 2012 | A1 |
20130013807 | Chrapko | Jan 2013 | A1 |
20130054598 | Caceres | Feb 2013 | A1 |
20130073387 | Heath | Mar 2013 | A1 |
20130097180 | Tseng | Apr 2013 | A1 |
20130097184 | Berkhim et al. | Apr 2013 | A1 |
20130110732 | Uppal | May 2013 | A1 |
20130124542 | Lee et al. | May 2013 | A1 |
20130138741 | Redstone et al. | May 2013 | A1 |
20130166601 | Chrapko et al. | Jun 2013 | A1 |
20130173457 | Chrapko | Jul 2013 | A1 |
20130198811 | Yu et al. | Aug 2013 | A1 |
20130254305 | Cheng et al. | Sep 2013 | A1 |
20130282884 | Chandrasekaran et al. | Oct 2013 | A1 |
20130290226 | Dokken | Oct 2013 | A1 |
20130291098 | Chung et al. | Oct 2013 | A1 |
20130332740 | Sauve et al. | Dec 2013 | A1 |
20140081652 | Klindworth | Mar 2014 | A1 |
20140089189 | Vasireddy | Mar 2014 | A1 |
20140114962 | Rosenburg et al. | Apr 2014 | A1 |
20140156274 | You et al. | Jun 2014 | A1 |
20140172708 | Chrapko et al. | Jun 2014 | A1 |
20140173723 | Singla et al. | Jun 2014 | A1 |
20140258160 | Chrapko et al. | Sep 2014 | A1 |
20140278730 | Muhart et al. | Sep 2014 | A1 |
20140279352 | Schaefer et al. | Sep 2014 | A1 |
20140280151 | Micaelian | Sep 2014 | A1 |
20140287725 | Lee | Sep 2014 | A1 |
20140304339 | Hamilton | Oct 2014 | A1 |
20140317003 | Shah | Oct 2014 | A1 |
20140317107 | Gharpure et al. | Oct 2014 | A1 |
20150026120 | Chrapko et al. | Jan 2015 | A1 |
20150089568 | Sprague et al. | Mar 2015 | A1 |
20150121456 | Milman et al. | Apr 2015 | A1 |
20150142595 | Acuña-Rohter | May 2015 | A1 |
20150163217 | Lo et al. | Jun 2015 | A1 |
20150169142 | Ongo et al. | Jun 2015 | A1 |
20150213407 | Cabler et al. | Jul 2015 | A1 |
20150220835 | Wilson et al. | Aug 2015 | A1 |
20150242856 | Dhurandhar et al. | Aug 2015 | A1 |
20150271206 | Schultz et al. | Sep 2015 | A1 |
20150347591 | Bax et al. | Dec 2015 | A1 |
20150359039 | Haque et al. | Dec 2015 | A1 |
20150370801 | Shah | Dec 2015 | A1 |
20160035046 | Gupta et al. | Feb 2016 | A1 |
20160073271 | Schultz et al. | Mar 2016 | A1 |
20160171011 | Drogobetski et al. | Jun 2016 | A1 |
20160171113 | Fanous et al. | Jun 2016 | A1 |
20160197788 | Chrapko et al. | Jul 2016 | A1 |
20160277424 | Mawji et al. | Sep 2016 | A1 |
20170024749 | Barathy et al. | Jan 2017 | A1 |
20170236078 | Rasumov | Aug 2017 | A1 |
20180068010 | Paterson et al. | Mar 2018 | A1 |
20190385130 | Mossoba et al. | Dec 2019 | A1 |
Number | Date | Country |
---|---|---|
2600344 | Sep 2006 | CA |
2775899 | Apr 2011 | CA |
1619567 | May 2005 | CN |
101167093 | Apr 2008 | CN |
101383695 | Mar 2009 | CN |
101393566 | Mar 2009 | CN |
101403978 | Apr 2009 | CN |
101438279 | May 2009 | CN |
101443806 | May 2009 | CN |
101841539 | Sep 2010 | CN |
101902459 | Dec 2010 | CN |
102136114 | Jul 2011 | CN |
102668457 | Sep 2012 | CN |
102685661 | Sep 2012 | CN |
102855572 | Jan 2013 | CN |
103095728 | May 2013 | CN |
103456233 | Dec 2013 | CN |
103493049 | Jan 2014 | CN |
103593764 | Feb 2014 | CN |
104504043 | Apr 2015 | CN |
102823225 | Sep 2015 | CN |
104954492 | Sep 2015 | CN |
109690608 | Apr 2019 | CN |
1511232 | Mar 2005 | EP |
2001-298453 | Oct 2001 | JP |
2002123649 | Apr 2002 | JP |
2003-259070 | Sep 2003 | JP |
2005149202 | Jun 2005 | JP |
2005339281 | Dec 2005 | JP |
2006-113900 | Apr 2006 | JP |
2006-260099 | Sep 2006 | JP |
2007-004411 | Jan 2007 | JP |
2007-249413 | Sep 2007 | JP |
2008-129990 | Jun 2008 | JP |
2009-025871 | Feb 2009 | JP |
2009-064433 | Mar 2009 | JP |
2009-146253 | Jul 2009 | JP |
2013506204 | Feb 2013 | JP |
2019860 | Sep 2019 | KR |
201250611 | Dec 2012 | TW |
2006019752 | Feb 2006 | WO |
2006115919 | Nov 2006 | WO |
WO-2007085903 | Aug 2007 | WO |
2009002193 | Dec 2008 | WO |
2009020964 | Feb 2009 | WO |
2009109009 | Sep 2009 | WO |
2010048172 | Apr 2010 | WO |
2011106897 | Sep 2011 | WO |
2011134086 | Nov 2011 | WO |
2011143761 | Nov 2011 | WO |
2011127206 | Apr 2012 | WO |
2013026095 | Feb 2013 | WO |
2013173790 | Nov 2013 | WO |
2014144114 | Sep 2014 | WO |
2015047992 | Apr 2015 | WO |
2015106657 | Jul 2015 | WO |
2016011371 | Jan 2016 | WO |
2017019203 | Feb 2017 | WO |
Entry |
---|
Notice Before Examining Pursuant to Section 41 of Law and Rule 66 of the Regulations received for Israel patent Application No. 261207 dated Oct. 27, 2019, 2 pages. |
Notice Before Examining Pursuant to Section 41 of Law and Rule 66 of the Regulations received for Israel patent Application No. 261921 dated Nov. 4, 2019, 2 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/474,785 dated Nov. 12, 2019, 127 pages. |
Extended European Search Report for European Patent Application No. 17840662.5 dated Jan. 2, 2020, 9 pages. |
Second Office Action received for Chinese Patent Application Serial No. 201610392929.2 dated Nov. 20, 2019, 8 pages.(Including English Translation). |
Final Office Action received for U.S. Appl. No. 15/644,356 dated Dec. 30, 2019, 68 pages. |
Notice Before Examining Pursuant to Section 41 of Law and Rule 66 of the Regulations received for Israel patent Application No. 261464 dated Nov. 4, 2019, 2 pages. |
Notice Before Examining Pursuant to Section 41 of Law and Rule 66 of the Regulations received for Israel patent Application No. 261465 dated Nov. 4, 2019, 2 pages. |
Notification under Section 18 of the Law received for Israel patent Application No. 264827 dated Jan. 7, 2020, 2 pages. |
Communication pursuant to Rules 70(2) and 70a(2) EPC received for EP Patent Application Serial No. 17840662.5 dated Jan. 21, 2020, 1 page. |
Office Action received for Brazialian Patent Application Serial No. BR112012007316-8 dated Mar. 10, 2020, 5 pages (Including English Translation). |
Final Office Action received for U.S. Appl. No. 15/644,356 dated Apr. 30, 2020, 56 pages. |
Notice of allowance received for Israel patent Application No. 261465 dated Mar. 17, 2020, 3 pages. |
Notice of allowance received for Israel patent Application No. 261921 dated Mar. 19, 2020, 02 pages. |
Notice of allowance received for Israel patent Application No. 261207 dated Mar. 15, 2020, 02 pages. |
Angwin et al., “Scrapers' Dig Deep for Data on Web”, http://online.wsj.com/article/SB10001424052748703358504575544381288117888.html?, Oct. 12, 2010, printed on Nov. 6, 2010, 5 pages. |
Anthes, “The Search Is On-Computerworld”, http://www.computerworld.com/s/article/70041lThe_Search_Is_On, Apr. 15, 2002, printed Nov. 6, 2010, 8 pages. |
Baras et al., “Dynamic Self-Organization and Clustering in Distributed Networked Systems for Performance Improvement,” Proceedings of the 47th annual Allerton conference on Communication, Control, and Computing, Alierton'09, Illinois, USA, pp. 968-975, Sep. 30-Oct. 2, 2009. |
Chakraborty et al., “TrustBAC-Integrating Trust Relationships into the RBAC Model for Access Control in Open Systems,” Proceedings of the eleventh ACM symposium on Access Control Models and Technologies, SACMAT '06, 10 pages, Jun. 7-9, 2006. |
Ding et al., “Transitive Closure and Metric Inequality of Weighted Graphs: Detecting Protein Interaction Modules Using Cliques”, Int. J. Data Mining and Bioinformatics, vol. X, No. X, 200X, pp. 162-177, 2006. |
“Facebook announces groups, downloading”, http://content.usatoday.com/communities/technologylive/post/2010/10/live-facebook-announcesdownloading-other-features/1, Oct. 6, 2010, printed Nov. 6, 2010, 7 pages. |
Feldman et al., “Robust Incentive Techniques for Peer-to-Peer Networks”, Proceedings of the fifth ACM Conference on Electronic Commerce EC'04, New York, New York, USA, 10 pages, May 17-20, 2004. |
Final Office Action for U.S. Appl. No. 15/079,952, dated Dec. 16, 2016, 14 pages. |
Gan et al., “A Novel Reputation Computing Model for Mobile Agent-Based E-Commerce Systems”, Proceedings of the International Conference on Information Security and Assurance, ISA 2008, pp. 253-260, Apr. 24-26, 2008. |
Geisberger, et al., “Contraction Hierarchies: Faster and Simpler Hierarchical Routing in Road Networks”, LNCS 6038, 2008, pp. 319-333. |
Golbeck et al., “Inferring Trust Relationships in Web-based Social Networks”, Journal of ACM Transactions of Internet Technology (TOIT), vol. 6, issue 4, Nov. 2006, 41 pages. |
Gu et al., “Processing Massive Sized Graphs Using Sector/Sphere”, Proceedings of the 2010 IEEE Workshop on Many-Task Computing on Grids and Supercomputers (MTAGS), New Orleans, LA, USA, 10 pages, Nov. 15, 2010. |
Gupta et al., “A Reputation System for Peer-to-Peer Networks,” Proceedings of the 13th International Workshop on Network and operating systems support for digital audio and video NOSSDAV'03, Monterey, California, USA, Jun. 1-3, 2003. |
Gupta et al., “Reputation Management Framework and its use as Currency in large-Scale Peer to-Peer Networks”, Proceedings of the Fourth IEEE International Conference on Peer-to-Peer Computing P2P2004, Zurich, Switzerland, pp. 1-9, Aug. 2004. |
Hartley et al., “MSSG: A Framework for Massive-Scale Semantic Graphs,” Proceedings of 2006 IEEE International Conference on Cluster Computing, CLUSTER'2006, Barcelona, Spain, 10 pages, Sep. 25-28, 2006. |
Huynh et al., “An Integrated Trust and reputation model for open multi-Agent systems,” Journal of Autonomous Agents and Multi-Agent Systems, vol. 13, issue 2, pp. 119-154, Sep. 2006. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2016/050305 dated Jun. 3, 2016, 7 pages. |
Josang et al., “Simplification and Analysis of Transitive Trust Networks”, Journal of Web Intelligence and Agent Systems, vol. 4, issue 2, Apr. 2006, pp. 1-26. |
Kamola et al., “Reconstruction of a Social Network Graph from Incomplete Call Detail Records”, Conference Proceedings of the International Conference on Computational Aspects of SocialNetworks (CASoN), Oct. 19, 2011, pp. 136-140. |
Kang et al., “Pegasus: A Peta-Scale Graph Mining System—Implementation and Observations”, Proceedings of the Ninth IEEE International Conference on Data Mining, ICDM'09, Miami, FL, USA, pp. 229-238, Dec. 6-9, 2009. |
Kim et al., “Design and Implementation of the location-based Personalized Social Media Service,” Conference Proceedings of the International Conference on Internet and Web Applications and Services (ICIW), May 9, 2010, pp. 116-121. |
Lumsdaine et al., “Challenges in Parallel Graph Processing”, Parallel Processing letters, vol. 17, No. 1, Mar. 2007, 16 pages. |
Malewicz et al., “Pregel: a System for large-Scale Graph Processing”, Proceedings of the 2010 International Conference on Management Data, SIGMOD'10, Indianapolis, Indiana, USA, Jun. 6-11, 2010, pp. 135-145. |
Meyer “Outrage as Credit Agency Plans to Mine Facebook Data,” Gigaom.com, https://gigaom.com/2012/06/07/credit-agency-mines-facebook-data/ Jun. 7, 2012, 3 pages. |
Mining Social Networks, Untangling the social Web, http://www.economist.com/node/16910031?story_id=16910031&fsrc=rss, Sep. 2, 2010, printed Nov. 6, 2010, 5 pages. |
Mori et al., “Improving Deployability of Peer-Assisted CDN Platform with Incentive”, Proceedings of IEEE Global Telecommunications Conference GLOBECOM 2009, Honolulu, Hawaii, USA, 2009, 7 pages. |
Mui et al., “A Computational Model of Trust and Reputation”, Proceedings of the 35th Annual Hawaii International Conference on System Sciences, HICSS '02, vol. 7, 2002, 9 pages. |
Notice of Allowance for U.S. Appl. No. 15/224,172, dated Oct. 14, 2016, 35 pages. |
Notice of Allowance for U.S. Appl. No. 15/056,484, dated Jul. 13, 2016, 31 pages. |
Non- Final Office Action for U.S. Appl. No. 15/079,952, dated Jul. 14, 2016, 26 pages. |
Notice of Allowance for U.S. Appl. No. 14/664,285, dated Oct. 7, 2016, 27 pages. |
Notice of Allowance for U.S. Appl. No. 15/070,643, dated Apr. 7, 2017, 28 pages. |
Notice of Allowance for U.S. Appl. No. 15/406,405, dated Apr. 6, 2017, 27 pages. |
Non-Final Office Action for U.S. Appl. No. 15/070,643, dated Feb. 24, 2017, 34 pages. |
Safaei et al., “Social Graph Generation & Forecasting Using Social Network Mining”, Proceedings of the 33rd Annual IEEE International Computer Software and Applications Conference, COMPSAC '09, 2009, pp. 31-35. |
“Startup Wants to Base Your Credit Score on Social Media Profiles”, Mashable, Technology Review 7, Jun. 7, 2012, 18 pages. |
Taiwanese Office Action for Taiwanese Patent Application No. 105108584 dated Oct. 31, 2016, 7 pages. |
Wallentin et al., “A Cross-Layer Route Discovery Strategy for Virtual Currency Systems in Mobile Ad Hoc Networks”, Proceedings of the Seventh International Conference on Wireless On-demand Network Systems and Services IEEE/IFIP WONS 2010, Kranjska Gora, Slovenia, 2010, pp. 91-98. |
Zetter, “Tracking Terrorists the Las Vegas Way”, http://pcworld.about.com/news/Aug072002id103692.htm, printed Nov. 6, 2010, 3 pages. |
Zhang et al., “A Review of Incentive Mechanisms in Peer-to-Peer Systems”, Proceedings of the First International Conference on Advances in P2P Systems AP2PS'09, Sliema, Malta, 2009, pp. 45-50. |
Zhang et al., “MARCH: A Distributed Incentive Scheme for Peer-to-Peer Networks”, Proceedings of the 26th Annual IEEE Conference on Computer Communication INFOCOM 2007, Anchorage, Alaska, USA, 2007, pp. 1091-1099. |
Non-Final Office Action for U.S. Appl. No. 13/498,429, dated Apr. 17, 2014, 24 pages. |
Non-Final Office Action for U.S. Appl. No. 13/498,429, dated Jul. 25, 2014, 18 pages. |
Final Office Action for U.S. Appl. No. 13/498,429, dated Jan. 16, 2015, 21 pages. |
Non-Final Office Action for U.S. Appl. No. 14/282,935, dated Jul. 21, 2014, 26 pages. |
Final Office Action for U.S. Appl. No. 14/282,935, dated Jan. 30, 2015, 42 pages. |
Non-Final Office Action for U.S. Appl. No. 14/282,935, dated Aug. 5, 2015, 7 pages. |
Final Office Action for U.S. Appl. No. 14/282,935, dated Feb. 12, 2016, 13 pages. |
Non-Final Office Action for U.S. Appl. No. 15/254,642, dated Dec. 28, 2016, 29 pages. |
EPO. Mitteilung des Europischen Patentamts vom 1. Oct. 2007 ber Geschftsmethoden = Notice from the European Patent Office dated Oct. 1, 2007 concerning business methods = Communiqu de l'Office europen des brevets,en date du 1er Oct. 2007, concernant les mthodes dans le domaine des activites economiques. vol. 30,Nr:11,pp. 592-593. Journal Officiel de l'Office Europeen des Brevets.Official Journal of the European Patent Office.Amtsblattt des Europaeischen Patenta, Nov. 1, 2007 Oeb, Munchen, DE. |
Tang, et al. Social influence analysis in large-scale networks. p. 807. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining ; KDD '09: Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '09—Jun. 28, 2009 to Jul. 1, 2009—Paris), Jan. 1, 2009; Jun. 28, 2009-Jul. 1, 2009 ACM, 2 Penn Plaza, Suite 701 New York NY 10121-0701 USA. |
Non-Final Office Action for U.S. Appl. No. 13/521,216, dated Jun. 20, 2014. |
Final Office Action for U.S. Appl. No. 13/521,216, dated May 21, 2015. |
Non-Final Office Action for U.S. Appl. No. 13/521,216, dated Apr. 13, 2016. |
Final Office Action for U.S. Appl. No. 13/521,216, dated Oct. 21, 2016. |
Final Office Action for U.S. Appl. No. 13/503,352, dated Jan. 5, 2015, 28 pages. |
Non-Final Office Action for U.S. Appl. No. 13/503,352, dated Aug. 21, 2014, 24 pages. |
Non-Final Office Action for U.S. Appl. No. 13/503,352, dated Aug. 18, 2015, 26 pages. |
Non-Final Office Action for U.S. Appl. No. 15/224,063, dated Sep. 1, 2016, 47 pages. |
Non-Final Office Action for U.S. Appl. No. 13/695,419, dated Oct. 31, 2014, 20 pages. |
Final Office Action for U.S. Appl. No. 13/695,419, dated Jun. 2, 2015, 31 pages. |
Non-Final Office Action for U.S. Appl. No. 13/695,419, dated Dec. 17, 2015, 43 pages. |
Non-Final Office Action for U.S. Appl. No. 13/695,419, dated Apr. 5, 2016, 72 pages. |
Non-Final Office Action for U.S. Appl. No. 13/695,419, dated Oct. 12, 2016, 147 pages. |
Non-Final Office Action for U.S. Appl. No. 13/824,324, dated Mar. 9, 2017, 17 pages. |
Non-Final Office Action for U.S. Appl. No. 14/664,285, dated Jun. 4, 2015, 25 pages. |
Final Office Action for U.S. Appl. No. 14/664,285, dated Sep. 21, 2015, 33 pages. |
Non-Final Office Action for U.S. Appl. No. 14/664,285, dated Mar. 4, 2016, 24 pages. |
Final Office Action for U.S. Appl. No. 14/664,285, dated Jul. 29, 2016, 29 pages. |
Non-Final Office Action for U.S. Appl. No. 15/070,643, dated Aug. 26, 2016, 36 pages. |
Non-Final Office Action for U.S. Appl. No. 15/589,841, dated Jun. 5, 2017, 40 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2017/050257 dated May 19, 2017, pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2017/050197 dated May 30, 2017, 15 pages. |
Non- Final Office Action For U.S. Appl. No. 15/466,590, dated Jul. 13, 2017, 20 pages. |
Non- Final Office Action For U.S. Appl. No. 15/055,952, dated Aug. 12, 2016, 33 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2017/050255 dated Jun. 19, 2017, 9 pages. |
Non-Final Office Action for U.S. Appl. No. 13/695,419, dated Jun. 14, 2017, 44 pages. |
Japanese Office Action dated Jul. 28, 2017, mailed Aug. 2, 2017 for Japanese Patent Application No. 2016-131310, 4 pages. |
Non-Final Office Action for U.S. Appl. No. 15/224,063, dated Jul. 14, 2017, 36 pages. |
Non-Final Office Action for U.S. Appl. No. 15/400,471, dated Aug. 7, 2017, 24 pages. |
Non-Final Office Action for U.S. Appl. No. 15/630,299, dated Aug. 10, 2017, 17 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2017/050351 dated Jun. 19, 2017, 10 pages. |
Non-Final Office Action for U.S. Application Serial No. 15/671,102 dated Sep. 22, 2017, 21 pages. |
Non-Final Office Action for U.S. Appl. No. 15/675,041 dated Oct. 6, 2017, 72 pages. |
Canadian Office Action for Canadian Application Serial No. 2,775,899 dated Oct. 12, 2017, 6 pages. |
Notice of Allowance for U.S. Appl. No. 13/695,419 dated Nov. 8, 2017, 92 pages. |
Final Office Action for U.S. Appl. No. 13/824,324, dated Nov. 30, 2017, 33 pages. |
Decision to Grant a Patent and English Translation thereof dated Nov. 14, 2017 for Japanese Patent Application No. 2016-131310, 6 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2010/001531 dated Jan. 10, 2011, 9 pages. |
Notification of Reasons for Refusal and English Translation thereof dated Mar. 29, 2016 for Japanese Patent Application No. 2015-085003, 4 pages. |
Decision to Grant a Patent and English Translation thereof dated Jun. 2, 2016 for Japanese Patent Application No. 2015-085003, 6 pages. |
Notification of Reasons for Refusal and English Translation thereof dated Apr. 3, 2014 for Japanese Patent Application No. 2012-531192, 7 pages. |
Decision to Grant a Patent and English Translation thereof dated Feb. 16, 2015 for Japanese Patent Application No. 2012-531192, 6 pages. |
Extended European Search Report for European Patent Application No. 10819770.8 dated Oct. 9, 2014, 7 pages. |
Communication pursuant to Rules 70(2) for European Patent Application No. 10819770.8 dated Oct. 28, 2014, 1 page. |
Communication pursuant to Article 94(3) EPC for European Patent Application No. 10819770.8 dated Feb. 8, 2016, 7 pages. |
Decision to Refuse European Patent Application No. 10819770.8 dated Apr. 18, 2017, 6 pages. |
Chinese First Office Action and English Translation thereof dated Mar. 24, 2014 for Chinese Application Serial No. 201080051338.X, 31 pages. |
Chinese Second Office Action and English Translation thereof dated Jan. 16, 2015 for Chinese Application Serial No. 201080051338.X, 26 pages. |
Chinese Third Office Action and English Translation thereof dated Sep. 28, 2015 for Chinese Application Serial No. 201080051338.X, 12 pages. |
Notification to Grant Patent Right for Invention and English Translation thereof dated Mar. 22, 2016 for Chinese Application Serial No. 201080051338.X, 4 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2011/050017 dated Apr. 13, 2011, 9 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2010/001658 dated Jan. 26, 2011, 9 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2011/050260 dated Jul. 26, 2011, 7 pages. |
Final Office Action dated Dec. 15, 2017 for U.S. Appl. No. 15/224,063, 52 pages. |
Non-Final Office Action dated Dec. 15, 2017 for U.S. Appl. No. 15/589,841, 66 pages. |
Final Office Action dated Jan. 19, 2018 for U.S. Appl. No. 15/466,590, 56 pages. |
Final Office Action dated Jan. 26, 2018 for U.S. Appl. No. 15/400,471, 78 pages. |
Final Office Action dated Feb. 8, 2018, for U.S. Appl. No. 15/630,299, 40 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2017/050962 dated Nov. 20, 2017, 9 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/644,356, dated Apr. 5, 2018, 74 pages. |
Notice of Allowance received for U.S. Appl. No. 15/589,841, dated Mar. 20, 18, 44 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/046,041, dated Apr. 26, 2018, 70 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/907,164, dated May 24, 2018 29 pages. |
Final Office Action received for U.S. Appl. No. 15/675,041 dated May 17, 2018, 73 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/400,471, dated May 29, 2018, 34 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/630,299, dated May 17, 2018,19 pages. |
Notice of Allowance for U.S. Appl. No. 15/953,011 dated Aug. 28, 18, 72 pages. |
First Action Interview Pilot Program Pre-Interview Communication for U.S. Appl. No. 15/953,011 dated Jul. 10, 2018, 10 pages. |
First Action Interview Pilot Program Pre-Interview Communication for U.S. Appl. No. 15/907,166 dated Jul. 6, 2018, 6 pages. |
Final Office Action for U.S. Appl. No. 15/400,471 dated Oct. 29, 2018, 42 pages. |
Final Office Action for U.S. Appl. No. 15/644,356 dated Oct. 18, 2018, 41 pages. |
Non-Final Office Action for U.S. Appl. No. 16/014,032 dated Oct. 4, 2018, 74 pages. |
First Action Interview Pilot Program Communication for U.S. Appl. No. 15/907,166 dated Oct. 31, 2018, 46 pages. |
Final Office Action for U.S. Appl. No. 15/907,164 dated Nov. 14, 2018, 70 pages. |
Chinese First Office Action and English Translation thereof dated Sep. 4, 2018 for Chinese Application Serial No. 201610392926.9, 17 pages. |
Final Office Action for U.S. Appl. No. 15/046,041 dated Nov. 20, 2018, 32 pages. |
Non-Final Office Action for U.S. Appl. No. 16/141,615 dated Nov. 29, 2018, 15 pages. |
www.trustscience.com Inc., et al., “Defendants' Motion for Stay of Discovery Pending Decision on Motion to Dismiss,” Case 6:18-cv-01174-CEM-DCI, Document 32, Filed Sep. 21, 2018, 12 pages. |
www.trustscience.com Inc., et al., “Plaintiff's Response in Opposition to Defendants' Motion to Stay Discovery Pending Decision on Motion to Dismiss,” Case 6:18-cv-01174-CEM-DCI, Document 35, Filed Oct. 5, 2018, 20 pages. |
www.trustscience.com Inc., et al., “Motion to Dismiss the Complaint and Memorandum of Law in Support,” Case 6:18-cv-01174-CEM-DCI, Document 41, Filed Oct. 19, 2018, 36 pages. |
www.trustscience.com Inc., et al., “Plaintiff's Response to Defendants' Motion to Dismiss and Notice of Intent to File Amended Complaint,” Case 6:18-cv-01174-CEM-DCI, Document 44, Filed Nov. 2, 2018, 3 pages. |
www.trustscience.com Inc., et al., “First Amended Complaint and Demand for Injunctive Relief and Jury Trial,” Case 6:18-cv-01174-CEM-DCI, Document 46 Filed Nov. 6, 2018, 49 pages. |
www.trustscience.com Inc., et al., “Order and Permanent Injunction,” Case 6:18-cv-01174-CEM-DCI, Document 49 Filed Nov. 21, 2018, 4 pages. |
Examiner's Report for Canadian Application Serial No. 2,775,899 dated Oct. 2, 18, 8 pages. |
Non-Final Office Action recevied for U.S. Appl. No. 15/675,041 dated Jan. 11, 2019, 91 pages. |
Non-Final Office Action for U.S. Appl. No. 16/204,651 dated Jan. 29, 2019, 47 pages. |
Notice of Allowance for U.S. Appl. No. 16/195,946 dated Feb. 14, 2019, 69 pages. |
Notice of Allowance for U.S. Appl. No. 15/400,471 dated Mar. 21, 2019, 47 pages. |
Non-Final office Action recevied for U.S. Appl. No. 15/644,356 dated Mar. 21, 2019, 39 pages. |
First Office Action received for Chinese Patent Application Serial No. 201610392929.2 dated Mar. 14, 2019, 20 pages. (Including English Translation). |
Final Office Action received for U.S. Appl. No. 15/907,166 dated Apr. 24, 2019, 27 pages. |
Final Office Action received for U.S. Appl. No. 16/014,032 dated Apr. 17, 2019, 38 pages. |
International Search Report and Written Opinion for International Patent Application No. PCT/CA2011/050569 dated Dec. 5, 2011, 9 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/046,041 dated May 22, 2019, 37 pages. |
Final Office Action received for U.S. Appl. No. 16/204,651 dated Jul. 19, 2019, 34 pages. |
Final Office Action received for U.S. Appl. No. 16/141,615 dated Jul. 30, 2019, 78 pages. |
Final Office Action received for U.S. Appl. No. 15/675,041 dated Jul. 29, 2019, 61 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/907,164 dated Aug. 20, 2019, 52 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/166,581 dated Sep. 18, 2019, 92 pages. |
Final Office Action received for U.S. Appl. No. 15/046,041 dated Sep. 10, 2019, 23 pages. |
Final Office Action received for U.S. Appl. No. 16/014,032 dated Oct. 1, 2019, 35 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/410,272 dated Oct. 11, 2019, 65 pages. |
Canadian Office Action received for Patent Application Serial No. 2,775,899 dated Oct. 8, 2019, 6 pages. |
Indian First Office Action received for Indian Patent Application Serial No. 735/KOLNP/2012 dated Aug. 23, 2019, 7 pages. |
Notice of Allowance received for U.S. Appl. No. 16/410,272 dated Apr. 28, 2020, 54 pages. |
Notice of allowance received for Israel patent Application No. 261464 dated Jul. 1, 2020, 5 pages. |
Final Office Action received for U.S. Appl. No. 15/474,785 dated Oct. 16, 2020, 55 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/166,581 dated Jul. 28, 2020, 74 pages. |
Canadian Office Action for Canadian Application Serial No. 3,033,793 dated Sep. 2, 2020, 6 pages. |
Communication pursuant to Article 94(3) EPC for European Patent Application No. 17840662.5 dated Aug. 31, 2020, 3 pages. |
Canadian Office Action for Canadian Application Serial No. 2,775,899 dated Oct. 7, 2020, 3 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/835,433 dated Mar. 18, 2021, 94 pages. |
Notice of Reasons for Refusal received for Japanese Application Serial No. 2019-500705 dated Apr. 6, 2021, 09 pages. (Including English Translation). |
Non-Final Office Action received for U.S. Appl. No. 17/158,840 dated May 19, 2021, 47 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/801,295 dated Jun. 15, 2021, 111 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/801,295 dated Jun. 16, 2021, 88 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/774,744 dated Jun. 24, 2021, 126 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/745,498 dated Aug. 3, 2021, 105 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/773,382 dated Jul. 28, 2021, 105 pages. |
Liu et al., “Hybrid content filtering and reputation-based popularity for recommending blog articles”, Bradford vol. 38, No. 6, Jun. 23, 2014, pp. 788-805. |
Office Action received for Mexican Patent Application Serial No. MX/a/2018/010005 dated May 18, 2021, 8 pages. (Including English Translation). |
Non-Final Office Action received for U.S. Appl. No. 17/079,600 dated Dec. 3, 2021, 84 pages. |
Office Action received for Indian Patent Application Serial No. 201847032375 dated Oct. 29, 2021, 8 pages. |
Office Action received for taiwan Patent Application No. 106127464 dated Oct. 1, 2021, 17 pages. |
Notice of Allowance received for U.S. Appl. No. 17/158,840, dated Dec. 24, 2021, 97 pages. |
Office Action received for Indian Patent Application Serial No. 201847032355 dated Aug. 31, 2021, 6 pages. |
Office Action received for Canadian Patent Application Serial No. 3,015,926 dated Aug. 17, 2021, 3 pages. |
Notification to Grant Patent Right for Invention received for Chinese Patent Application Serial No. 201780025229.2 dated May 28, 2021, 5 pages. (Including English Translation). |
Office Action received for Israel Patent Application Serial No. 261464 dated Jul. 25, 2021, 3 pages. |
Office Action received for Canadian Patent Application Serial No. 3,014,995 dated Aug. 12, 2021, 4 pages. |
Office Action received for Canadian Patent Application Serial No. 3,014,361 dated Aug. 18, 2021, 5 pages. |
Office Action received for Canadian Patent Application Serial No. 3,016,091 dated Aug. 19, 2021, 7 pages. |
Final Office Action received for U.S. Appl. No. 16/774,744 dated Apr. 14, 2022, 86 pages. |
Final Office Action received for U.S. Appl. No. 16/745,498 dated Apr. 19, 2022, 54 pages. |
Office Action received for Canadian Patent Application Serial No. 3014361 dated Mar. 16, 2022, 4 pages. |
Final Office Action received for U.S. Appl. No. 16/773,382 dated May 24, 2022, 40 pages. |
Non Final Office Action received for U.S. Appl. No. 17/231,658 dated Jun. 16, 2022, 129 pages. |
Non Final Office Action received for U.S. Appl. No. 16/995,293 dated Jul. 6, 2022, 123 pages. |
Summons to attend oral proceedings received for European Patent Application Serial No. 17840662.5 dated May 23, 2022, 11 pages. |
Office Action received for Mexican Patent Application Serial No. MX/a/2018/010005 dated May 17, 2022, 12 pages. |
Mexican Notice of Allowance dated May 19, 2022 for Mexican Patent Application No. MX/a/2018/010423, 2 pages. |
Office Action received for Mexican Patent Application Serial No. MX/a/2018/010005 dated Nov. 30, 2021, 9 pages. |
Office Action received for Mexican Patent Application Serial No. MX/a/2018/010423 dated Nov. 25, 2021, 11 pages. |
Japanese Notice of Allowance dated Feb. 15, 2022 for Japanese Patent Application No. 2019-500705, 3 pages. |
First Office Action received for Mexican Patent Application Serial No. MX/a/2019/001858 dated Nov. 24, 2022, 8 pages.(Including English Translation). |
Notice of Allowance received for U.S. Appl. No. 16/773,382 dated Dec. 21, 2022, 43 pages. |
Notice of Allowance received for U.S. Appl. No. 16/995,293 dated Jan. 24, 2023, 73 pages. |
Non-Final Office Action received for U.S. Appl. No. 16/745,498 dated Nov. 29, 2022, 44 pages. |
Decision to refuse a European Patent Application received for European Patent Application Serial No. 17840662.5 dated Nov. 18, 2022, 14 pages. |
First Office Action received for Chinese Patent Application Serial No. 201780025677.2 dated Oct. 13, 2022, 12 pages.(Including English Translation). |
Shang Jiang et al., “Application of incidence matrix method in credit evaluation”, Journal of Shenyang Normal University (Natural Science), Jul. 15, 2007, pp. 308-310, 11 pages (including English Translation). |
Shen Limin et al., “Adaptive trust model based on time series analysis in opportunistic network”, Journal of Chinese Computer Systems No. 7, Jul. 15, 2015, pp. 1553-1558, 37 pages (including English Translation). |
First Office Action received for Chinese Patent Application Serial No. 201780024074.0 dated Oct. 10, 2022, 20 pages (Including English Translation). |
First Office Action received for Chinese Patent Application Serial No. 201780030761.3 dated Nov. 21, 2022, 22 pages (Including English Translation). |
Final Office Action received for U.S. Appl. No. 17/079,600 dated Aug. 25, 2022, 36 pages. |
Mexican Office Action received for Mexican Patent Application Serial No. MX/a/2018/010426 dated Jul. 12, 2022, 7 pages. |
Notice of Allowance received for Canadian Patent Application No. 3014995 dated Sep. 6, 2022, 1 page. |
Mexican Office Action dated May 27, 2022 for Mexican Patent Application No. MX/a/2018/011618, 3 pages. |
Non-Final Office Action received for U.S. Appl. No. 18/046,382 dated Mar. 16, 2023, 139 pages. |
Non-Final Office Action received for U.S. Appl. No. 17/659,292 dated Mar. 28, 2023, 124 pages. |
Non-Final Office Action received for U.S. Appl. No. 17/079,600 dated Apr. 7, 2023, 30 pages. |
Notice of Allowance received for U.S. Appl. No. 16/995,293 dated Mar. 1, 2023, 7 pages. |
Final Office Action received for U.S. Appl. No. 17/231,658 dated May 10, 2023, 97 pages. |
Non-Final Office Action received for U.S. Appl. No. 17/805,750 dated Jun. 2, 2023, 124 pages. |
Office Action received for Chinese Patent Application Serial No. 201780024074.0 dated Apr. 11, 2023, 22 pages. |
Notice of Allowance received for Canadian Patent Application Serial No. 3014361 dated Apr. 27, 2023, 1 page. |
Office Action received for Mexican Patent Application Serial No. MX/a/2018/011618 dated Jan. 27, 2023, 7 pages (Original Copy only). |
Notice of Allowance received for Canadian Patent Application No. 3016091 dated Mar. 2, 2023, 1 page. |
Office Action received for Chinese Patent Application Serial No. 201780025677.2 dated Mar. 10, 2023, 2 pages (Original Copy only). |
Office Action received for Mexican Patent Application Serial No. MX/a/2019/001858 dated Apr. 21, 2023, 12 pages (including machine translation). |
Notice of Allowance received for Taiwanese Patent Application Serial No. 106127464 dated May 25, 2023, 2 pages (including English translation). |
Notice of Allowance received for Chinese Patent Application Serial No. 201780030761.3 dated Jul. 10, 2023, 12 pages (including English machine translation). |
Li, Han “Research on Trust Recommendation and Friend Search Filtering Algorithm in Social Networks” Yanshan University, 2012, 160 pages (including English Translation). |
Non-Final Office Action received for U.S. Appl. No. 16/745,498 dated Aug. 25, 2023, 51 pages. |
Non-Final Office Action received for U.S. Appl. No. 17/660,167 dated Sep. 21, 2023, 131 pages. |
Non-Final Office Action received for U.S. Appl. No. 18/187,040 dated Sep. 29, 2023, 57 pages. |
Notice of Allowance received for Mexican Patent Application Serial No. MX/a/2018/011618, dated Aug. 8, 2023, 3 pages (English Translation). |
Non-Final Office Action received for U.S. Appl. No. 18/310,590 dated Oct. 11, 2023, 94 pages. |
Office Action received for Mexican Patent Application Serial No. MX/a/2019/001858 dated Oct. 9, 2023, 14 pages (including English translation). |
Final Office Action received for U.S. Appl. No. 17/079,600 dated Dec. 5, 2023, 47 pages. |
Notice of Allowance received for U.S. Appl. No. 17/659,292 dated Dec. 18, 2023, 95 pages. |
Final Office Action received for U.S. Appl. No. 18/046,382 dated Dec. 26, 2023, 127 pages. |
Notice of Allowance received for Chinese Patent Application Serial No. 201780024074.0 dated Dec. 8, 2023, 4 pages. |
Number | Date | Country | |
---|---|---|---|
20190378219 A1 | Dec 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15400471 | Jan 2017 | US |
Child | 16534474 | US | |
Parent | 14664285 | Mar 2015 | US |
Child | 15400471 | US |