Embodiments of the present invention are directed to computerized systems and methods for preparing an electronic tax return using a tax return preparation application; and more particularly, to systems and methods for generating a sub-graph of a tax calculation graph used by a tax calculation engine to perform tax calculation operations which can be used for creating a calculator for a tax topic and/or for modifying the granularity of questions requesting tax data in preparing a tax return.
The embodiments of the present invention may be implemented on and/or within a tax return preparation system comprising a tax preparation software application executing on a computing device. The tax return preparation system may operate on a new construct in which tax rules and the calculations based thereon are established in declarative data-structures, namely, completeness model(s) and tax calculation graph(s). The tax calculation graph(s) comprise a plurality of nodes including input nodes, functional nodes, and function nodes. The tax calculation graph(s) may be a single overall tax calculation graph which includes all calculations for all of the tax topics (e.g., gross income, which has sub-topics of interest income, investment income, employment income, self-employment income; affordable care act penalty; deductions, such as dependent deductions, itemized deductions, etc.) of the tax code, or a plurality of tax calculation graphs each covering particular tax topics and sub-topics which may be compiled to form the overall tax calculation graph.
The tax calculation graph(s) are configured with a plurality of calculation paths wherein each calculation path connects a plurality of nodes which are data dependent such that a node is connected to another node if the node depends on the other node. In addition, as used herein, a calculation path from a first node “leads” or is “leading” to a second node if the second node is directly or indirectly dependent on the first node along a connected path of nodes. Also, as used herein, a calculation path “leading” or “along” a calculation path from a start node to an end node on a tax calculation graph includes the start node and end node. Use of these data-structures permits the user interface to be loosely connected or even divorced from the tax calculation engine and the data used in the tax calculations. Tax calculations are dynamically calculated based on tax-related data that is input from a user, derived from sourced data, or estimated. A smart tax logic agent running on a set of rules can review current run time data and evaluate missing tax data necessary to prepare and complete a tax return. The tax logic agent proposes suggested questions to be asked to a user to fill in missing blanks. This process can be continued until completeness of all tax topics has occurred. A completed tax return (e.g., a printed tax return or an electronic tax return) can then be electronically prepared and filed (electronically and/or in paper form) with respect to the relevant taxing jurisdictions.
The completeness model(s) comprise a data structure that captures all the conditions necessary to obtain all of the tax data necessary to complete a tax return that can be filed with the pertinent tax agency. Similar to the calculation graph(s), the completeness model(s) (also referred to as “completion model(s)) can comprehensively cover all tax topics, or they can be a plurality of completeness models each covering particular tax topics and sub-topics which may combined form the overall completeness model. The completeness model may be embodied in various forms. As a couple examples, the completeness model(s) may be completeness graphs (also referred to as “completion graphs”) such as a decision tree, or the completeness model(s) may be in the form of decision tables representing tax questions for obtaining tax data for each tax topic and the logic relating the tax questions to other tax questions and/or completion of the tax topic. For instance, answers and/or entry of tax data in response to certain tax questions are logically related to other tax questions in the decision table and/or a completion goal for the tax topic indicating that the tax topic is completed.
In another aspect of the tax return preparation system, the system is configured to operate the computing device to establish a connection to a data store configured to store user-specific tax data therein. The computing device executes a tax calculation engine configured to read and write tax calculation data to and from the shared data store, the tax calculation engine using one or more of the tax calculation graphs specific to particular tax topics. The computing device executes a tax logic agent, the tax logic agent reading from the shared data store and a plurality of decision tables collectively representing a completeness model for computing tax liability or a portion thereof, the tax logic agent outputting one or more suggestions for missing tax data based on an entry in one of the plurality of decision tables. The computing device executes a user interface manager configured to receive the one or more suggestions and present to a user one or more questions based on the one or more suggestions via a user interface, wherein a user response to the one or more questions is input to the shared data store. The user interface manager is configured to generate and display a question screen to the user. The question screen includes a question for the user requesting tax data and is also configured to receive the tax data from the user in the form of input from the user. The user interface manager which receives the suggestion(s) selects one or more suggested questions to be presented to a user. Alternatively, the user interface manager may ignore the suggestion(s) and present a different question or prompt to the user.
In the event that all tax topics are covered, the tax logic agent, instead of outputting one or more suggestions for missing tax data may output a “done” instruction to the user interface manager. The computing device may then prepare a tax return based on the data in the shared data store. The tax return may be a conventional paper-based return or, alternatively, the tax return may be an electronic tax return which can then be e-filed.
In some situations, it may be desirable to modify the tax return preparation system to adjust the granularity of tax questions presented to a user. For example, at a high level of granularity, such as for basic users, the system may ask itemized tax questions requesting an itemized listing of tax data. As an example, for investment income, the tax questions may request tax data for each investment transaction, such as each purchase and sale of stock. At a lower level of granularity, such as for professional users, it may be preferable to only ask for the total investment income. In one manner of accomplishing this modification of the granularity of tax questions, the nodes on the tax calculation graphs corresponding to the desired level of granularity are re-configured to be user enterable nodes, and the higher level nodes of the tax calculation graph which lead to the re-configured nodes (which includes the higher level input nodes corresponding to tax data to be input by a user) are ignored (i.e., excluded from the tax calculation graph). In other words, a new sub-graph of the tax calculation graph is generated which excludes the nodes of the tax calculation graph having a calculation path leading to the user enterable nodes.
Accordingly, in one embodiment of the present invention, a system is provided for modifying a level of detail of tax questions for obtaining user enterable tax data. The system may comprise various components of the tax preparation system as described above, including a computing device and a data store in communication with the computing device and configured to store user-specific tax data therein. The system also includes a tax calculation graph as described above, having an additional feature in that one or more of the functional nodes and/or function nodes each include a user selectable tag for selectably configuring such node as a user enterable node, wherein the value for the user enterable node is entered by a user. As used herein, the term “user enterable node” means a non-input node which is re-configured to be a node in which a user enters a value for the node. In other words, a “user enterable node” is a node which a user enters a set value for the node, as opposed to the value of the node being calculated based on the value of input node(s) and function(s) defined for the node, as originally configured in the tax calculation graph. Thus, to modify the granularity of tax questions, such as for a tax calculation graph for an entire tax return or for a certain tax topic, a user selects one or more nodes of the tax calculation graph to selectably configure such node(s) as user enterable nodes. Typically, a target node is also defined, which may be user selected or a default value. For example, a target node may be an endpoint node for a tax topic, such as gross income, or adjusted gross income, etc. A default target node may be a final tax return result, such as tax refund or tax owed, or the endpoint node of a tax topic for the nodes selected as user enterable nodes. As described below, the target node is useful in generating a sub-graph of the tax calculation graph.
In another aspect, the system further comprises a sub-graph engine configured to generate a sub-graph of the tax calculation graph based upon the target node and the user enterable nodes selected by the user. The sub-graph engine generates a sub-graph which includes all nodes along each calculation path leading from each user enterable node to the target node; and excludes all nodes of the tax calculation graph which only have a calculation path leading to only one or more user enterable nodes, and all nodes of the tax calculation graph which do not have a calculation path leading to target node. The sub-graph engine can exclude the excluded nodes by actually deleting them from the tax calculation graph or by setting an excluded tag for the excluded nodes indicating the node is excluded.
In another feature, the sub-graph engine may generate the sub-graph by analyzing the calculation paths. The sub-graph engine determines all of the nodes along each calculation path of the tax calculation graph leading from each user enterable node to the target node. These nodes will be included in the sub-graph, and for convenience of nomenclature, they may be referred to as “first sub-graph nodes.”
In some cases, the selection of the user enterable nodes may not cover all of the calculation paths leading to the target node. For instance, there may still be a calculation path from an input node to the target node which does not pass through one of the user enterable nodes. This may indicate that the selection of user enterable nodes was incomplete, or that it is intended that the nodes along calculation paths not including a user enterable node should be part of the final sub-graph. Thus, the sub-graph engine determines any calculation paths leading from an input node to the target node which do not include a user enterable node. These calculation paths are referred to as “open paths.” When open paths are identified, the sub-graph engine may request the user to select additional user enterable nodes along the open paths, and receives a selection of one or more additional user enterable node(s). The sub-graph engine determines all of the nodes along each calculation path leading from each additional user enterable node to the target node. These nodes will also be a part of the final sub-graph and are referred to as “second sub-graph nodes.”
The sub-graph engine then generates the final sub-graph consisting essentially (or consisting only) of the first sub-graph nodes and the second sub-graph nodes, if any of each type of sub-graph node have been determined as described above. The term “consisting essentially of” in regards to the nodes of a sub-graph means that the sub-graph does not include any nodes which materially affect the compactness of the sub-graph, but may include a small number of nodes as artifacts (e.g. less than 2%, less than 5% or less than 10% of the total number of nodes in the sub-graph) which do not affect the calculation of the sub-graph.
Even after a selection of additional user enterable node(s), there may still be open paths. So, in still another aspect, the sub-graph engine may determine any remaining open paths which do not include a user enterable node or an additional user enterable node. The sub-graph engine determines all nodes along each of the remaining open paths, which will also be a included in the final sub-graph, and are referred to as “third sub-graph nodes.” In this case, the sub-graph engine generates the final sub-graph consisting essentially of the first sub-graph nodes, the second sub-graph nodes, and the third sub-graph nodes, if any of each type of sub-graph node have been determined as described above.
In yet another aspect, the selection of user enterable nodes may be invalid in that there is a calculation path between a first user enterable node (including additional user enterable node(s), as described above) and a second user enterable node (including additional user enterable node(s)), in other words the first user enterable node and second user enterable node are interdependent. This may be problematic because the values for the two interdependent user enterable nodes may be inconsistent with the proper tax calculations, and/or have other inconsistencies. Thus, the sub-graph engine may be further configured to determine whether there is a calculation path between a first user enterable node and a second user enterable node. When there are interdependent user enterable nodes, the sub-graph engine notifies a user that the selection of the first user enterable node and second user enterable node is invalid, and requests a revised selection of user enterable nodes. The sub-graph engine may then repeat the interdependency check for the revised selection of user enterable nodes. Once a valid selection is obtained, the sub-graph engine determines the final sub-graph based upon the revised and valid selection of user enterable nodes.
In another feature, the sub-graph engine may generate the sub-graph by first determining an initial construct sub-graph. The sub-graph engine is configured to generate a construct sub-graph comprising all of the nodes of the tax calculation graph along each calculation path leading to the target node. Then, for each user enterable node, the sub-graph engine determines the first sub-graph nodes, second sub-graph nodes and/or third sub-graph nodes, as described above, using the construct sub-graph rather than the entire tax calculation graph.
The sub-graph may be useful for a number of purposes. For example, the sub-graph can be used to modify the granularity of tax questions by modifying the completeness model(s) based on the sub-graph such as by configuring questions and logic represented in the completeness model(s) and corresponding to a node leading to a user enterable node as not needed to complete particular the tax topic. In addition, a sub-graph may be useful as a calculator for calculating the result of a tax topic of the overall tax calculation graph.
Another embodiment of the present invention is directed to computer-implemented methods for modifying a level of detail of tax questions for obtaining user enterable tax data. The methods may be implemented on a tax preparation system, same or similar to that described above. For example, the method may comprise a tax preparation system executing a sub-graph engine determining all of the nodes along each calculation path of the tax calculation graph leading from each user enterable node to the target node wherein such nodes are referred to as first sub-graph nodes. The sub-graph engine also determines any calculation paths leading from an input node to the target node which do not include a user enterable node, such calculation paths referred to as open paths. When one or more open paths are determined, the system prompts the user to select one or more nodes on one or more of the open paths as an additional user enterable node and determining all of the nodes along each calculation path leading from each additional user enterable node to the target node, such nodes referred to as second sub-graph nodes. The sub-graph engine determines any remaining open paths leading from an input node to the target node which do not include a user enterable node or an additional user enterable node, for any remaining open paths, determine all nodes along such remaining open paths, such nodes referred to as third sub-graph nodes. Finally, the sub-graph engine generates a sub-graph consisting essentially of the first sub-graph nodes, the second sub-graph nodes and the third sub-graph nodes.
In additional aspects of the computer implemented method may also include any of the additional aspects and features described herein for the system embodiments for modifying a level of detail of tax questions for obtaining user enterable tax data.
Still another embodiment of the present invention is directed to is directed to an article of manufacture comprising a non-transitory computer readable medium embodying instructions executable by a computer to execute a process according to any of the method embodiments of the present invention for modifying a level of detail of tax questions for obtaining user enterable tax data. For instance, the non-transitory computer readable medium embodying instructions executable by a computer may be configured to execute a process comprising: determining a construct sub-graph comprising all nodes of the tax calculation graph having a calculation path leading to the target node; for each user enterable node, determining all of the nodes along each calculation path of the construct sub-graph leading from each user enterable node to the target node wherein such nodes are referred to as first sub-graph nodes; determining any calculation paths leading from an input node of the construct sub-graph to the target node which do not include a user enterable node, such calculation paths referred to as open paths; when one or more open paths are determined, receiving a selection of a node on one or more of the open paths to be configured as an additional user enterable node, and determining all nodes along each calculation path leading from each additional user enterable node to the target node, such nodes referred to as second sub-graph nodes; determining any remaining open paths leading from an input node to the target node which do not include a user enterable node or an additional user enterable node, and for any remaining open paths, and determining all nodes along such remaining open paths, such nodes referred to as third sub-graph nodes; and generating a final sub-graph consisting essentially of the first sub-graph nodes, second sub-graph nodes third sub-graph nodes.
In additional aspects, the article of manufacture may be further configured according to the additional aspects described herein for the systems and methods for modifying a level of detail of tax questions for obtaining user enterable tax data.
In additional embodiments of the invention, the sub-graph may be utilized in various ways. For example, the system may be configured to modify the detail of questions across different tax preparation applications, such as between a retail version and a professional tax preparer version. The retail version would typically include more detailed questions about tax topics, whereas a profession tax prepare version would typically have less detail such as by excluding questions requesting itemized lists within a tax topic and only requesting a total or previously calculated result. In such case, the system may comprise a granularity control engine which is configured to generate a sub-graph of a tax calculation graph based upon a target node and one or more user enterable nodes, as described above for the sub-graph engine. Indeed, the granularity control engine may comprise a sub-graph engine same as that described above.
The granularity control engine is also configured to generate a modified completeness model of the full completeness model for a particular tax topic in which each question and logic of the full completeness model (also referred to as the “first completeness model” as opposed to the modified completeness model) which is directed to a node on the calculation graph leading to the user enterable nodes is configured as not needed to complete the particular tax topic. In this way, the tax logic agent of the tax preparation software utilizing the sub-graph and the modified completeness model will not recommend questions directed to the nodes leading to the user enterable nodes, thereby modifying the granularity of the questions requesting tax data for a taxpayer in preparing a tax return.
In an additional aspect, the system may further comprise a tax logic agent configured to analyze the modified completeness model to determine one or more suggested tax questions for obtaining missing tax data required to complete the tax calculation graph of the sub-graph. Of course, the tax logic agent will not suggest tax questions which are configured as not needed in the completeness model.
In another aspect, the system may further comprise a user interface manager configured to receive the one or more suggested tax questions from the tax logic agent, analyze the suggested tax questions, and determine one or more tax questions to present in preparing a tax return using a tax preparation system.
In further aspects, the granularity control engine may include one or more of the aspects and features of the system having a sub-graph engine, as described above.
Another embodiment of the present invention is directed to computer-implemented methods for modifying the level of detail of questions requesting tax data for a taxpayer in preparing a tax return in a tax preparation system. The methods may be implemented on the tax preparation system having a granularity control engine, same or similar to the system described above. For instance, the method may comprise a granularity control engine generating a sub-graph of the tax calculation graph based upon a target node and one or more user enterable nodes selected from the nodes of the tax calculation graph. The granularity control engine generates a modified completeness model of the first completeness model in which each question and logic of the first completeness model directed to a node on the calculation graph leading to the user enterable nodes is configured as not needed to complete the tax topic.
In another aspect, the method may also include any of the additional aspects and features described herein for the system embodiments for modifying a level of detail of tax questions for obtaining user enterable tax data. Still another embodiment of the present invention is directed to is directed to an article of manufacture comprising a non-transitory computer readable medium embodying instructions executable by a computer to execute a process according to any of the method embodiments of the present invention for modifying a level of detail of tax questions for obtaining user enterable tax data using a granularity control engine. For instance, the non-transitory computer readable medium embodying instructions executable by a computer may be configured to execute a process comprising: generating a sub-graph of the tax calculation graph based upon a target node and one or more user enterable nodes selected from the nodes of the tax calculation graph; and generating a modified completeness model of the first completeness model in which each question and logic of the first completeness model directed to a node on the calculation graph leading to the user enterable nodes is configured as not needed to complete the tax topic.
In additional aspects, the article of manufacture may be further configured according to the additional aspects described herein for the systems and methods for modifying a level of detail of tax questions for obtaining user enterable tax data utilizing a granularity control engine.
It is understood that the steps of the methods and processes of the present invention are not required to be performed in the order as shown in the figures or as described, but can be performed in any order that accomplishes the intended purpose of the methods and processes.
Embodiments of the present invention are directed to systems, methods and articles of manufacture for generating a sub-graph of a tax calculation graph for a tax preparation system. The sub-graph may be advantageously utilized in various ways, such as for modifying a level of detail of tax questions for obtaining user enterable tax data from a user of the tax preparation system in preparing a tax return. The system comprises a tax preparation system including a computing device, a data store in communication with the computing device and configured to store user-specific tax data. The system includes a tax calculation graph, as described herein, which is usable by a tax calculation engine of the tax preparation system to calculate a tax return. The system includes a sub-graph engine which is configured to generate a sub-graph of the tax calculation graph based upon a user selected target node of the tax calculation graph (e.g., a target calculated result of the sub-graph) and at least one user selected node to be defined as user-enterable node(s). For example, a particular function node of may normally be a calculated value based on several values of leaf nodes (input nodes), but it is being re-defined as a user-enterable node in which the value of the node is requested from a user and is input by the user, and received by the tax preparation system. The sub-graph engine is configured to generate the sub-graph, wherein the sub-graph excludes the nodes of the original tax calculation graph which are rendered unnecessary for the calculation of the sub-graph. Thus, the sub-graph excludes (a) all nodes of the original tax calculation graph which only have a calculation path leading to only one or more user enterable nodes and (b) all nodes which do not have a calculation path leading to the selected target node. In other words, the sub-graph includes only those nodes along calculation path(s) leading from a user enterable node or a remaining input node to the target node.
Tax preparation is a time-consuming and laborious process. It is estimated that individuals and businesses spend around 6.1 billion hours per year complying with the filing requirements of the United States federal Internal Revenue Code. Tax return preparation software has been commercially available to assist taxpayers in preparing their tax returns. Tax return preparation software is typically run on a computing device such as a computer, laptop, tablet, or mobile computing device such as a Smartphone. Traditionally, a user has walked through a set of rigidly defined user interface interview screens that selectively ask questions that are relevant to a particular tax topic or data field needed to calculate a taxpayer's tax liability.
In contrast to the rigidly defined user interface screens used in prior iterations of tax preparation software, the present design provides tax preparation software 100 that runs on computing devices 102, 103 (see
Use of these data-structures permits the user interface to be loosely connected or even detached from the tax calculation engine and the data used in the tax calculations. Tax calculations are dynamically calculated based on tax data derived from sourced data, estimates, user input, or even intermediate tax calculations that are then utilized for additional tax calculations. A smart tax logic agent running on a set of rules can review current run time data and evaluate missing data fields and propose suggested questions to be asked to a user to fill in missing blanks. This process can be continued until completeness of all tax topics has occurred. An electronic return can then be prepared and filed with respect to the relevant taxing jurisdictions.
Note that in
The completeness graph 12 and the tax calculation graph 14 represent data structures that can be constructed in the form of a tree.
As one can imagine given the complexities and nuances of the tax code, many tax topics may contain completeness graphs 12 that have many nodes with a large number of pathways to completion. However, many branches or lines within the completeness graph 12 can be ignored, for example, when certain questions internal to the completeness graph 12 are answered that eliminate other nodes 20 and arcs 22 within the completeness graph 12. The dependent logic expressed by the completeness graph 12 allows one to minimize subsequent questions based on answers given to prior questions. This allows a minimum question set that can be generated that can be presented to a user as explained herein.
As explained herein, the directed graph or completion graph 12 that is illustrated in
Referring to
After an initial question has been presented and rows are eliminated as a result of the selection, next, a collection of candidate questions from the remaining available rows 32a and 32b is determined. From this universe of candidate questions from the remaining rows, a candidate question is selected. In this case, the candidate questions are questions Qc and QG in columns 34c, 34g, respectively. One of these questions is selected and the process repeats until either the goal 34h is reached or there is an empty candidate list.
In
In still other embodiments, values for leaf nodes 24 may be derived or otherwise calculated. For example, while the number of dependents may be manually entered by a taxpayer, those dependents may not all be “qualifying” dependents for tax purposes. In such instances, the actual number of “qualified” dependents may be derived or calculated by the tax preparation software 100. In still other embodiments, values for leaf nodes 24 may be estimated as described herein.
Still other internal nodes, referred to as functional nodes 26, semantically represent a tax concept and may be calculated or otherwise determined using a function node 28 (also referred to as a “function 28”). The functional node 26 and the associated function 28 define a particular tax operation 29. For example, as seen in
Interconnected functional node 26 containing data dependent tax concepts or topics are associated with a discrete set of functions 28 that are used to capture domain specific patterns and semantic abstractions used in the tax calculation. The discrete set of functions 28 that are associated with any particular functional node may be commonly re-occurring operations for functions that are used throughout the process of calculating tax liability. For instance, examples of such commonly reoccurring functions 28 include copy, capping, thresholding, accumulation or adding, look-up operations, phase out calculations, comparison calculations, exemptions, exclusions, and the like.
In one embodiment, the entire set of functions 28 that is used to compute or calculate a tax liability is stored within a data store 30 which in some instances may be a database. The various functions 28 that are used to semantically describe data connections between functional nodes 26 can be called upon by the tax preparation software 100 for performing tax calculations. Utilizing these common functions 28 greatly improves the efficiency of the tax preparation software 100 and can be used by a programmer to more easily track and follow the complex nature of the ever-evolving tax code. The common functions 28 also enable easier updating of the tax preparation software 100 because as tax laws and regulations change, fewer changes need to be made to the software code as compared to prior hard-wired approaches.
Importantly, the tax calculation graph 14 and the associated functional node 26 and functions 28 can be tagged and later be used or called upon to intelligently explain to the user the reasoning behind why a particular result was calculated or determined by the tax preparation software 100 program, as explained in more detail below. The functions 28 can be de-coupled from a specific narrow definition and instead be associated with one or more explanations. Examples of common functions 28 found in tax legislation and tax rules include the concepts of “caps” or “exemptions” that are found in various portions of the tax code. One example of a “cap” is the portion of the U.S. tax code that limits the ability of a joint filer to deduct more than $3,000 of net capital losses in any single tax year. There are many other instances of such caps. An example of an “exemption” is one that relates to early distributions from retirement plans. For most retirement plans, early distributions from qualified retirement plans prior to reaching the age of fifty nine and one-half (59%) require a 10% penalty. This penalty can be avoided, however, if an exemption applies such as the total and permanent disability of the participant. Other exemptions also apply. Such exemptions are found throughout various aspects of the tax code and tax regulations.
In some embodiments, the function node 28 may include any number of mathematical or other operations. Examples of functions 28 include summation, subtraction, multiplication, division, and look-ups of tables or values from a database 30 or library as is illustrated in
Still referring to
As seen in
In the event there is a penalty, the ACA requires that the penalty be the greater of a percentage of income, net of specified deductions, or a specified penalty that is applied per individual or family. For example, for the 2015 year, the percentage is 2.0 percent and increases to 2.5 percent in subsequent years.
In order to determine the non-income or “fixed” penalty, an accumulator function 28i is used to determine the penalty. In this example, the calculation pertains to a family wherein the penalty includes a fixed amount for a child ($162.50 per child in 2015) and a fixed amount per adult ($325.00 per adult). Under the ACA, there is a maximum cap of this fixed penalty. For example, in 2015, the maximum family penalty is $975. As seen in
As seen in
The schema 44 may be a modified version of the MeF schema used by the IRS. For example, the schema 44 may be an extended or expanded version (designated MeF++) of the MeF model established by government authorities that utilizes additional fields. While the particular MeF schema 44 is discussed herein the invention is not so limited. There may be many different schemas 44 depending on the different tax jurisdiction. For example, Country A may have a tax schema 44 that varies from Country B. Different regions or states within a single country may even have different schemas 44. The systems and methods described herein are not limited to a particular schema 44 implementation. The schema 44 may contain all the data fields required to prepare and file a tax return with a government taxing authority. This may include, for example, all fields required for any tax forms, schedules, and the like. Data may include text, numbers, and a response to a Boolean expression (e.g., True/False or Yes/No). As explained in more detail, the shared data store 42 may, at any one time, have a particular instance 46 of the MeF schema 44 (for MeF++ schema) stored therein at any particular time. For example,
As seen in
For example, user input 48a is one type of data source 48. User input 48a may take a number of different forms. For example, user input 48a may be generated by a user using, for example, an input device such as keyboard, mouse, touchscreen display, voice input (e.g., voice to text feature). photograph or image, or the like to enter information manually into the tax preparation software 100. For example, as illustrated in
User input 48a may also include some form of automatic data gathering. For example, a user may scan or take a photographic image of a tax document (e.g., W-2 or 1099) that is then processed by the tax preparation software 100 to extract relevant data fields that are then automatically transferred and stored within the data store 42. OCR techniques along with pre-stored templates of tax reporting forms may be called upon to extract relevant data from the scanned or photographic images whereupon the data is then transferred to the shared data store 42.
Another example of a data source 48 is a prior year tax return 48b. A prior year tax return 48b that is stored electronically can be searched and data is copied and transferred to the shared data store 42. The prior year tax return 48b may be in a proprietary format (e.g., .txf, .pdf) or an open source format. The prior year tax return 48b may also be in a paper or hardcopy format that can be scanned or imaged whereby data is extracted and transferred to the shared data store 42. In another embodiment, a prior year tax return 48b may be obtained by accessing a government database (e.g., IRS records).
An additional example of a data source 48 is an online resource 48c. An online resource 48c may include, for example, websites for the taxpayer(s) that contain tax-related information. For example, financial service providers such as banks, credit unions, brokerages, investment advisors typically provide online access for their customers to view holdings, balances, and transactions. Financial service providers also typically provide year-end tax documents to their customers such as, for instance, 1099-INT (interest income), 1099-DIV (dividend income), 1099-B (brokerage proceeds), 1098 (mortgage interest) forms. The data contained on these tax forms may be captured and transferred electronically to the shared data store 42.
Of course, there are additional examples of online resources 48c beyond financial service providers. For example, many taxpayers may have social media or similar accounts. These include, by way of illustration and not limitation, Facebook, Linked-In, Twitter, and the like. User's may post or store personal information on these properties that may have tax implications. For example, a user's Linked-In account may indicate that a person changed jobs during a tax year. Likewise, a posting on Facebook about a new home may suggest that a person has purchased a home, moved to a new location, changed jobs; all of which may have possible tax ramifications. This information is then acquired and transferred to the shared data store 42, which can be used to drive or shape the interview process described herein. For instance, using the example above, a person may be asked a question whether or not she changed jobs during the year (e.g., “It looks like you changed jobs during the past year, is this correct?”. Additional follow-up questions can then be presented to the user.
Still referring to
Referring briefly to
Still referring to
As seen in
The following pseudo code generally expresses how a rule engine 64 functions utilizing a fact cache based on the runtime canonical data 62 or the instantiated representation of the canonical tax schema 46 at runtime and generating non-binding suggestions 66 provided as an input a UI control 80. As described in U.S. application Ser. No. 14/097,057 previously incorporated herein by reference, data such as required inputs can be stored to a fact cache so that the needed inputs can be recalled at a later time, and to determine what is already known about variables, factors or requirements of various rules:
The TLA 60 may also receive or otherwise incorporate information from a statistical/life knowledge module 70. The statistical/life knowledge module 70 contains statistical or probabilistic data related to the taxpayer. For example, statistical/life knowledge module 70 may indicate that taxpayers residing within a particular zip code are more likely to be homeowners than renters. More specifically, the statistical/life knowledge module may comprise tax correlation data regarding a plurality of tax matter correlations. Each of the tax matter correlations quantifies a correlation between a taxpayer attribute and a tax related aspect. For instance, a taxpayer attribute could be taxpayer age which may be correlated to a tax related aspect such as having dependents, or a taxpayer attribute might be taxpayer age which may be correlated to homeownership or other relevant tax related aspect. The tax correlation data also quantifies the correlations, such as by a probability of the correlation. For instance, the correlation between the taxpayer attribute and the tax related aspect may be a certain percentage probability, such as 10%, 20%, 30%, 40%, 50%, 60%, or any percentage from 0% to 100%. Alternatively, the quantification can be a binary value, such as relevant or not relevant. In other words, for a given taxpayer attribute, it may be determined that a tax related aspect is relevant or completely not relevant when a taxpayer has the given taxpayer attribute. As an example, if the taxpayer attribute is that the taxpayer is married, the correlation may indicate that spouse information is relevant and will be required.
The TLA 60 may use this knowledge to weight particular topics or questions related to these topics. For example, in the example given above, questions about home mortgage interest may be promoted or otherwise given a higher weight. The statistical knowledge may apply in other ways as well. For example, tax forms often require a taxpayer to list his or her profession. These professions may be associated with transactions that may affect tax liability. For instance, a taxpayer may list his or her occupation as “teacher.” The statistic/life knowledge module 70 may contain data that shows that a large percentage of teachers have retirement accounts and in particular 403(b) retirement accounts. This information may then be used by the TLA 60 when generating its suggestions 66. For example, rather than asking generically about retirement accounts, the suggestion 66 can be tailored directly to a question about 403(b) retirement accounts.
The data that is contained within the statistic/life knowledge module 70 may be obtained by analyzing aggregate tax data of a large body of taxpayers. For example, entities having access to tax filings may be able to mine their own proprietary data to establish connections and links between various taxpayer characteristics and tax topics. This information may be contained in a database or other repository that is accessed by the statistic/life knowledge module 70. This information may be periodically refreshed or updated to reflect the most up-to-date relationships. Generally, the data contained in the statistic/life knowledge module 70 is not specific to a particular tax payer but is rather generalized to characteristics shared across a number of tax payers although in other embodiments, the data may be more specific to an individual taxpayer.
Still referring to
The user interface manager 82, as explained previously, receives non-binding suggestions from the TLA 60. The non-binding suggestions may include a single question or multiple questions that are suggested to be displayed to the taxpayer via the user interface presentation 84. The user interface manager 82, in one aspect of the invention, contains a suggestion resolution element 88, which is responsible for resolving how to respond to the incoming non-binding suggestions 66. For this purpose, the suggestion resolution element 88 may be programmed or configured internally. Alternatively, the suggestion resolution element 88 may access external interaction configuration files. Additional details regarding configuration files and their use may be found in U.S. patent application Ser. No. 14/206,834, which is incorporated by reference herein.
Configuration files specify whether, when and/or how non-binding suggestions are processed. For example, a configuration file may specify a particular priority or sequence of processing non-binding suggestions 66 such as now or immediate, in the current user interface presentation 84 (e.g., interview screen), in the next user interface presentation 84, in a subsequent user interface presentation 84, in a random sequence (e.g., as determined by a random number or sequence generator). As another example, this may involve classifying non-binding suggestions as being ignored. A configuration file may also specify content (e.g., text) of the user interface presentation 84 that is to be generated based at least in part upon a non-binding suggestion 66.
A user interface presentation 84 may comprise pre-programmed interview screens that can be selected and provided to the generator element 85 for providing the resulting user interface presentation 84 or content or sequence of user interface presentations 84 to the user. User interface presentations 84 may also include interview screen templates, which are blank or partially completed interview screens that can be utilized by the generation element 85 to construct a final user interface presentation 84 on the fly during runtime.
As seen in
Still referring to
The TLA 60 also outputs a tax data that is used to generate the actual tax return (either electronic return or paper return). The return itself can be prepared by the TLA 60 or at the direction of the TLA 60 using, for example, the services engine 90 that is configured to perform a number of tasks or services for the taxpayer. The services engine 90 is operatively coupled to the TLA 60 and is configured to perform a number of tasks or services for the taxpayer. For example, the services engine 90 can include a printing option 92. The printing option 92 may be used to print a copy of a tax return, tax return data, summaries of tax data, reports, tax forms and schedules, and the like. The services engine 90 may also electronically file 94 or e-file a tax return with a tax authority (e.g., federal or state tax authority). Whether a paper or electronic return is filed, data from the shared data store 42 required for particular tax forms, schedules, and the like is transferred over into the desired format. With respect to e-filed tax returns, the tax return may be filed using the MeF web-based system that allows electronic filing of tax returns through the Internet. Of course, other e-filing systems may also be used other than those that rely on the MeF standard. The services engine 90 may also make one or more recommendations 96 based on the run-time data 62 contained in the TLA 60. For instance, the services engine 90 may identify that a taxpayer has incurred penalties for underpayment of estimates taxes and may recommend to the taxpayer to increase his or her withholdings or estimated tax payments for the following tax year. As another example, the services engine 90 may find that a person did not contribute to a retirement plan and may recommend 96 that a taxpayer open an Individual Retirement Account (IRA) or look into contributions in an employer-sponsored retirement plan. The services engine 90 may also include a calculator 98 that can be used to calculate various intermediate calculations used as part of the overall tax calculation algorithm. For example, the calculator 98 can isolate earned income, investment income, deductions, credits, and the like. The calculator 98 can also be used to estimate tax liability based on certain changed assumptions (e.g., how would my taxes change if I was married and filed a joint return?). The calculator 98 may also be used to compare analyze differences between tax years.
By using calculation graphs 14 to drive tax calculations and tax operations, it is possible to determine interdependencies of the nodes (including tax operations, functional nodes and function nodes) and the year-over-year calculation graphs 14 can be used to readily identify differences and report the same to a user. Differences can be found using commonly used graph isomorphism algorithms over the two respective calculation graphs 14.
Referring again to
These stored entries 112 can be recalled or extracted by the explanation engine 110 and then displayed to a user on a display 104 of a computing device 102, 103. For example, explanation engine 110 may interface with the UI control 80 in two-way communication such that a user may ask the tax preparation software 100 why a particular tax calculation, operation, or decision has been made by the system 40. For instance, the user may be presented with an on-screen link (
In some instances, a user is able to further “drill down” with additional questions to gain additional explanatory detail. This additional level of detailed explanations is possible by traversing the calculation graph(s) 14 to identify each of the preceding or upstream input node(s) 24, function node(s) 26 and/or function node(s) 28. In the context of the example listed above as shown in
With reference to
In one aspect of the invention, the choice of what particular explanation will be displayed to a user may vary. For example, different explanations associated with the same function node 26 and function 28 pair may be selected by the explanation engine 110 for display to a user based on the user's experience level. A basic user may be given a general or summary explanation while a user with more sophistication may be given a more detailed explanation. A professional user such as a CPA or other tax specialist may be given even more detailed explanations.
In some embodiments, the different levels of explanation may be tied to product types or codes. These may be associated with, for example, SKU product codes. For example, a free edition of the tax preparation software 100 may provide few or no explanations. In a more advanced edition (e.g., “Deluxe edition”), additional explanation is provided. Still more explanation may be provided in the more advanced editions of the tax preparation software 100 (e.g., “Premier edition”). Versions of the tax preparation software 100 that are developed for accountants and CPAs may provide even more explanation.
In still other embodiments a user may be able to “unlock” additional or more detailed explanations by upgrading to a higher edition of tax preparation software 100. Alternatively, a user may unlock additional or more detailed explanations in an a la carte manner for payment of an additional fee. Such a fee can be paid through the tax preparation software 100 itself using known methods of payment.
In one aspect of the invention, the natural language generator 114 may rely on artificial intelligence or machine learning such that results may be improved. For example, the explanation engine 110 may be triggered in response to a query that a user has typed into a free-form search box within the tax preparation software 100. The search that has been input within the search box can then be processed by the explanation engine 110 to determine what tax operation the user is inquiring about and then generate an explanatory response 115B.
As seen in
The narrative explanations 116 and their associated sub-explanations (e.g., 116′, 116a, 116b, 116d, 116e) are constructed as an explanation tree with the root of the tree representing a particular tax topic or tax operation. In the example of
Encapsulating the tax code and regulations within calculation graphs 14 results in much improved testability and maintainability of the tax preparation software 100. Software programming errors (“bugs”) can be identified more easily when the calculation graphs 14 are used because such bugs can be traced more easily. In addition, updates to the calculation graphs 14 can be readily performed when tax code or regulations change with less effort.
Further, the degree of granularity in the narrative explanations 116 that are presented to the user can be controlled. As explained in the context of
By capturing the tax code and tax regulations in one or more calculation graph(s) 14, the targeted calculations can be done on various tax topics or sub-topics within the overall tax calculation graph(s) 14 covering calculations for an entire tax return. In addition, the granularity or detail of tax questions presented to a user may also be adjusted utilizing the calculation graph(s) 14 and/or the completeness model(s) (such as completeness graphs, decision table, etc.). For example, sub-graphs for a particular tax topic can be generated from a tax calculation graph 14 which includes a subset of all of the nodes and calculation paths of the tax calculation graph which are needed to perform the desired calculation. In order to adjust the detail of questions, a sub-graph of a tax calculation graph 14 may be generated in which certain non-input nodes (e.g. function nodes 26 and 28) are re-configured as user enterable nodes such that the input nodes 24 leading to such re-configured nodes are no longer necessary. The sub-graph is generated which excludes certain nodes, such as excluding the re-configured input nodes 24 (and any function nodes 26 and 28 which are unnecessary to calculate the sub-graph). The completeness model associated with the respective tax topic is also modified to exclude tax questions directed to the excluded input nodes 24 (and any functions nodes 26 and 28 which are unnecessary to calculate the sub-graph). The sub-graph of the and modified completeness model can then be utilized by the tax preparation system 40 to prepare a tax return by presenting tax questions at a different level of detail than the original tax calculation graph(s) 14 and completeness mode(s).
In one aspect of enabling the functionality of modifying the level of detail of tax questions, the non-input nodes (e.g., the function nodes 24 and 26) on a calculation graph 14 include a user selectable tag for selectably configuring such node as a user enterable node, wherein the value for the user enterable node is entered by a user. As defined above, the term “user enterable node” means a non-input node which is re-configured to be a node in which a user enters a value for the node. Hence, a “user enterable node” is a node which a user enters a set value for the node, as opposed to the value of the node being calculated based on the value of input node(s) and function(s) defined for the node, as originally configured in the tax calculation graph 14. Therefore, to modify the granularity of tax questions, such as for a tax calculation graph for an entire tax return or for a certain tax topic, a user selects one or more nodes of the tax calculation graph 14 to selectably configure such node(s) as user enterable nodes. A target node is also defined, which may be user selected or a default value. A target node may be an endpoint node for a tax topic, such as gross income, or adjusted gross income, etc. A default target node may be a final tax return result, such as tax refund or tax owed, or the endpoint node of a tax topic for the nodes selected as user enterable nodes.
As an example, referring to the tax calculation graph 14 in
Referring back to
As an example, again referring to
The sub-graph engine 200 generates the sub-graph 204 by analyzing the calculation paths within the tax calculation graph 14 of interest. For instance, the sub-graph engine 200 traverses the calculation paths of the tax calculation graph 14 and determines whether each node along the calculation paths should be included or excluded from the sub-graph 200. In one determination, the sub-graph engine 200 traverses along each calculation path leading from each user enterable node to the target node, and determines all of the nodes along such calculation paths. Such nodes are included in the sub-graph (unless there are interdependent nodes, as described below), and are referred to a “first sub-graph nodes.”
The sub-graph engine 200 is also configured to determine any calculation paths leading to the target node which do not pass through a user enterable node. As described above in the example using
At this point, the sub-graph engine 200 may generate the final sub-graph 204 consisting essentially (or consisting only) of the first sub-graph nodes and the second sub-graph nodes, if any of each type of sub-graph node have been determined as described above. The term “consisting essentially of” in regards to the nodes of a sub-graph means that the sub-graph does not include any nodes which materially affect the compactness of the sub-graph, but may include a small number of nodes as artifacts (e.g. less than 2%, less than 5% or less than 10% of the total number of nodes in the sub-graph) which do not affect the calculation of the sub-graph.
Even after a selection of additional user enterable node(s), there may still be one or more open path(s). So, the sub-graph engine 200 may determine any remaining open paths which do not include a user enterable node or an additional user enterable node. The sub-graph engine 200 may be further configured to iteratively advise the user that there are still open paths and to select additional user enterable node(s) or to generate the sub-graph with the open path(s). Alternatively, the sub-graph engine 200 may not give the user the option, and may simply proceed. Either way, the sub-graph engine 200 determines all nodes along each of the remaining open paths, which will also be included in the final sub-graph, and are referred to as “third sub-graph nodes.” In this case, the sub-graph engine generates the final sub-graph consisting essentially of the first sub-graph nodes, the second sub-graph nodes, and the third sub-graph nodes, if any of each type of sub-graph node have been determined as described above.
It is possible that the selection of user enterable nodes may be invalid because there is a calculation path between a first user enterable node (including additional user enterable node(s), as described above) and a second user enterable node (including additional user enterable node(s)). In this case, first user enterable node and second user enterable node are interdependent. This may cause problems because the values for the two interdependent user enterable nodes may be inconsistent with the proper tax calculations, and/or have other inconsistencies, so it may be desirable to modify this situation. Referring again to
In order to reduce the possible calculation paths within the tax calculation graph 14 which the sub-graph engine 200 needs to analyze to generate a final sub-graph 204, the sub-graph engine 200 may first determine an initial construct sub-graph. Hence, before determining the first sub-graph nodes, second sub-graph nodes and/or third sub-graph nodes, the sub-graph engine generates a construct sub-graph comprising all of the nodes of the tax calculation graph 14 along each calculation path leading to the target node. Then, the sub-graph engine 204 determines the first sub-graph nodes, second sub-graph nodes and/or third sub-graph nodes, as described above, using the construct sub-graph rather than the entire tax calculation graph. Using the tax calculation graph 14 of
As explained above, a sub-graph may have a number of useful purposes. As an example, a sub-graph may be utilized as a calculator for calculating the result of a tax topic of the overall tax calculation graph. Referring to
In addition, a sub-graph can be used to modify the granularity of tax questions by modifying the completeness model(s) based on the sub-graph including configuring questions and logic represented in the completeness model(s) and corresponding to a node leading to a user enterable node as not needed to complete a particular tax topic. For instance, the tax preparation system 40 may be configured to modify the detail of questions across different tax preparation applications, such between a retail version and a professional tax preparer version. The retail version would typically include more detailed questions about tax topics, whereas a professional tax preparer version would typically have less detail such as by excluding questions requesting itemized lists within a tax topic and only requesting a total or previously calculated result.
Referring again to
Still referring to
The granularity control engine 200 may generate the modified completeness model 210 in any suitable manner. In one way. the granularity control engine modifies the applicable completion graph(s) 12 for the particular tax topic(s) being modified (e.g., in the example of
The granularity control engine 202 then uses the modified completion graph 12 to generate a modified decision table 30, similar to the process described above for converting a completion graph 12 into a decision table 30.
Once the modified decision table 30 is generated, the tax preparation system 40 utilizes the modified decision table 30 (i.e., the modified completeness model 210) in preparing a tax return, as described below, except that the system 40 utilizes the sub-graph 204 of tax calculation graph 12 instead of the original tax calculation graph 12, and the modified decision table 30 instead of the original decision table 30. In an additional aspect, the system may further comprise a tax logic agent configured to analyze the modified completeness model to determine one or more suggested tax questions for obtaining missing tax data required to complete the tax calculation graph of the sub-graph. Of course, the tax logic agent will not suggest tax questions which are configured as not needed in the completeness model.
In another aspect, the system may further comprise a user interface manager configured to receive the one or more suggested tax questions from the tax logic agent, analyze the suggested tax questions, and determine one or more tax questions to present in preparing a tax return using a tax preparation system.
Another embodiment of the present invention is directed to computer-implemented methods for modifying the level of detail of questions requesting tax data for a taxpayer in preparing a tax return in a tax preparation system. The methods may be implemented on the tax preparation system having a granularity control engine, same or similar to the system described above. For instance, the method may comprise a granularity control engine generating a sub-graph of the tax calculation graph based upon a target node and one or more user enterable nodes selected from the nodes of the tax calculation graph. The granularity control engine generates a modified completeness model of the first completeness model in which each question and logic of the first completeness model directed to a node on the calculation graph leading to the user enterable nodes is configured as not needed to complete the tax topic.
Turning to
Referring back to
In one embodiment, the gathering or importation of data sources such as prior tax returns 48b, online resources 48c, and third party information 48d is optional. For example, a taxpayer may want to start the process from scratch without pulling information from other sources. However, in order to streamline and more efficiently complete a tax return other users may desire to obtain tax related information automatically. This would reduce the number of interview or prompt screens that are presented to the user if such information were obtained automatically by the tax preparation software 100. A user may be given the opportunity to select which data sources 48 they want accessed and searched for relevant tax related data that will be imported into the shared data store 42. A user may be asked to submit his or her account and password information for some data sources 48 using the UI control 80. Other data sources 48 such as some third party data sources 48d may be accessed without such information.
Next, as seen in operation 1200, after the schema 44 is populated with the various imported or entered data fields from the data sources 48, the tax calculation engine 50, using the tax calculation graphs 14 (or sub-graphs), reads data from the shared data store 42, performs tax calculations, and writes back data to the shared data store 42.
In operation 1300, the tax logic agent 60 reads the run time data 62 which represents the instantiated representation of the canonical tax schema 44 at runtime. The tax logic agent 60 then utilizes the decision tables 30 (or modified completeness model(s) or modified decision table(s)) to generate and send non-binding suggestions 66 to the UI control 80 as seen in operation 1400. Alternatively, the tax logic agent 60 may determine that completeness has been achieved across the tax topics in which case a done instruction may be delivered to the UI control as seen in operation 1500. If not done, the process continues whereby the user interface manager 82 will then process the suggestion(s) 66 using the suggestion resolution element 88 for resolving of how to respond to the incoming non-binding suggestions 66 as seen in operation 1600. The user interface manager 82 then generates a user interface presentation 84 to the user as seen in operation 1700 whereby the user is presented with one or more prompts. The prompts may include questions, affirmations, confirmations, declaratory statements, and the like. The prompts are displayed on a screen 104 of the computing device 102 whereby the user can then respond to the same by using one or more input devices associated with the computing device 102 (e.g., keyboard, mouse, finger, stylus, voice recognition, etc.).
Still referring to
Accordingly, a tax preparation system 40 is provided which can prepare a tax return. In addition, the system 40 may be utilized to generate sub-graphs which can be used as tax topic calculators. The system 40 can also modify the granularity of tax questions presented to a user for obtaining tax data, which functionality can be utilized to create different versions of the tax preparation applications such as a basic version, expert version, professional tax preparer version, etc.
The described embodiments of the present invention, including the functions performed by the system 40 and its components, may also be embodied in, or readable from, a computer-readable medium or carrier, e.g., one or more of the fixed and/or removable data storage data devices and/or data communications devices connected to a computer. Carriers may be, for example, magnetic storage medium, optical storage medium and magneto-optical storage medium. Examples of carriers include, but are not limited to, a floppy diskette, a memory stick or a flash drive, CD-R, CD-RW, CD-ROM, DVD-R, DVD-RW, or other carrier now known or later developed capable of storing data. The processor 304 performs steps or executes program instructions 302 within memory 300 and/or embodied on the carrier to implement method embodiments.
Embodiments, however, are not so limited and implementation of embodiments may vary depending on the platform utilized. Accordingly, embodiments are intended to exemplify alternatives, modifications, and equivalents that may fall within the scope of the claims.
Number | Name | Date | Kind |
---|---|---|---|
4213251 | Foundos | Jul 1980 | A |
4809219 | Ashford et al. | Feb 1989 | A |
5006998 | Yasunobu | Apr 1991 | A |
5495607 | Pisello et al. | Feb 1996 | A |
5557761 | Chan et al. | Sep 1996 | A |
5673369 | Kim | Sep 1997 | A |
5742836 | Turpin | Apr 1998 | A |
5819249 | Dohanich | Oct 1998 | A |
6078898 | Davis | Jun 2000 | A |
6535883 | Lee et al. | Mar 2003 | B1 |
6601055 | Roberts | Jul 2003 | B1 |
6631361 | O'Flaherty et al. | Oct 2003 | B1 |
6670969 | Halstead et al. | Dec 2003 | B1 |
6690854 | Helbing | Feb 2004 | B2 |
6697787 | Miller | Feb 2004 | B1 |
6898573 | Piehl | May 2005 | B1 |
6912508 | McCalden | Jun 2005 | B1 |
7234103 | Regan | Jun 2007 | B1 |
7295998 | Kulkarni | Nov 2007 | B2 |
7331045 | Martin et al. | Feb 2008 | B2 |
7448022 | Ram et al. | Nov 2008 | B1 |
7539635 | Peak et al. | May 2009 | B1 |
7565312 | Shaw | Jul 2009 | B1 |
7603301 | Regan | Oct 2009 | B1 |
7668763 | Albrecht | Feb 2010 | B1 |
7680756 | Quinn | Mar 2010 | B2 |
7685082 | Coletta | Mar 2010 | B1 |
7693760 | Fiteni | Apr 2010 | B1 |
7693769 | Burlison | Apr 2010 | B1 |
7716094 | Sutter et al. | May 2010 | B1 |
7742958 | Leek | Jun 2010 | B1 |
7747484 | Stanley | Jun 2010 | B2 |
7761333 | Kapp | Jul 2010 | B2 |
7778895 | Baxter | Aug 2010 | B1 |
7818222 | Allanson | Oct 2010 | B2 |
7849405 | Coletta | Dec 2010 | B1 |
7860763 | Quinn et al. | Dec 2010 | B1 |
7865829 | Goldfield | Jan 2011 | B1 |
7895102 | Wilks et al. | Feb 2011 | B1 |
7899757 | Talan | Mar 2011 | B1 |
7900298 | Char et al. | Mar 2011 | B1 |
7908190 | Enenkiel | Mar 2011 | B2 |
7912767 | Cheatham et al. | Mar 2011 | B1 |
7912768 | Abeles | Mar 2011 | B2 |
7925553 | Banks | Apr 2011 | B2 |
8001006 | Yu et al. | Aug 2011 | B1 |
8019664 | Tifford et al. | Sep 2011 | B1 |
8082144 | Brown et al. | Dec 2011 | B1 |
8086970 | Achtermann et al. | Dec 2011 | B2 |
8108258 | Slattery | Jan 2012 | B1 |
8126820 | Talan | Feb 2012 | B1 |
8190499 | McVickar | May 2012 | B1 |
8204805 | Eftekhari | Jun 2012 | B2 |
8224726 | Murray | Jul 2012 | B2 |
8234562 | Evans | Jul 2012 | B1 |
8244607 | Quinn | Aug 2012 | B1 |
8346635 | Olim | Jan 2013 | B1 |
8346680 | Castleman | Jan 2013 | B2 |
8370795 | Sage | Feb 2013 | B1 |
8386344 | Christina | Feb 2013 | B2 |
8407113 | Eftekhari et al. | Mar 2013 | B1 |
8417596 | Dunbar et al. | Apr 2013 | B1 |
8417597 | McVickar | Apr 2013 | B1 |
8447667 | Dinamani et al. | May 2013 | B1 |
8452676 | Talan | May 2013 | B1 |
8473880 | Bennett et al. | Jun 2013 | B1 |
8478671 | Tifford | Jul 2013 | B1 |
8510187 | Dinamani | Aug 2013 | B1 |
8527375 | Olim | Sep 2013 | B1 |
8560409 | Abeles | Oct 2013 | B2 |
8583516 | Pitt | Nov 2013 | B1 |
8589262 | Gang | Nov 2013 | B1 |
8607353 | Rippert et al. | Dec 2013 | B2 |
8635127 | Shaw | Jan 2014 | B1 |
8639616 | Rolenaitis | Jan 2014 | B1 |
8682756 | Tifford et al. | Mar 2014 | B1 |
8682829 | Barthel | Mar 2014 | B2 |
8694395 | Houseworth | Apr 2014 | B2 |
8706580 | Houseworth | Apr 2014 | B2 |
8788412 | Hamm | Jul 2014 | B1 |
8812380 | Murray | Aug 2014 | B2 |
8813178 | Khanna | Aug 2014 | B1 |
8838492 | Baker | Sep 2014 | B1 |
8892467 | Ball | Nov 2014 | B1 |
8949270 | Newton et al. | Feb 2015 | B2 |
9372687 | Pai | Jun 2016 | B1 |
9690854 | Stent et al. | Jun 2017 | B2 |
9760953 | Wang et al. | Sep 2017 | B1 |
9916628 | Wang et al. | Mar 2018 | B1 |
9922376 | Wang et al. | Mar 2018 | B1 |
9990678 | Cabrera et al. | Jun 2018 | B1 |
20020065831 | DePaolo | May 2002 | A1 |
20020107698 | Brown et al. | Aug 2002 | A1 |
20020111888 | Stanley et al. | Aug 2002 | A1 |
20020174017 | Singh | Nov 2002 | A1 |
20020198832 | Agee | Dec 2002 | A1 |
20030101070 | Mahosky et al. | May 2003 | A1 |
20030126054 | Purcell | Jul 2003 | A1 |
20030139827 | Phelps | Jul 2003 | A1 |
20030174157 | Hellman | Sep 2003 | A1 |
20030182102 | Corston-Oliver et al. | Sep 2003 | A1 |
20040002906 | Von Drehnen et al. | Jan 2004 | A1 |
20040019540 | William | Jan 2004 | A1 |
20040019541 | William | Jan 2004 | A1 |
20040021678 | Ullah et al. | Feb 2004 | A1 |
20040078271 | Morano | Apr 2004 | A1 |
20040083164 | Schwartz et al. | Apr 2004 | A1 |
20040088233 | Brady | May 2004 | A1 |
20040117395 | Gong | Jun 2004 | A1 |
20040172347 | Barthel | Sep 2004 | A1 |
20040181543 | Wu et al. | Sep 2004 | A1 |
20040205008 | Haynie et al. | Oct 2004 | A1 |
20050171822 | Cagan | Aug 2005 | A1 |
20050216379 | Ozaki | Sep 2005 | A1 |
20050262191 | Mamou et al. | Nov 2005 | A1 |
20060112114 | Yu | May 2006 | A1 |
20060155618 | Wyle | Jul 2006 | A1 |
20060155632 | Cherkas et al. | Jul 2006 | A1 |
20060178961 | Stanley | Aug 2006 | A1 |
20060282354 | Varghese | Dec 2006 | A1 |
20060293990 | Schaub | Dec 2006 | A1 |
20070033116 | Murray | Feb 2007 | A1 |
20070033117 | Murray | Feb 2007 | A1 |
20070033130 | Murray | Feb 2007 | A1 |
20070055571 | Fox et al. | Mar 2007 | A1 |
20070094207 | Yu et al. | Apr 2007 | A1 |
20070136157 | Neher et al. | Jun 2007 | A1 |
20070150387 | Seubert et al. | Jun 2007 | A1 |
20070156564 | Humphrey et al. | Jul 2007 | A1 |
20070179841 | Agassi | Aug 2007 | A1 |
20070192166 | Van Luchene | Aug 2007 | A1 |
20070250418 | Banks et al. | Oct 2007 | A1 |
20080059900 | Murray | Mar 2008 | A1 |
20080097878 | Abeles | Apr 2008 | A1 |
20080126170 | Leck et al. | May 2008 | A1 |
20080147494 | Larson | Jun 2008 | A1 |
20080162310 | Quinn | Jul 2008 | A1 |
20080177631 | William | Jul 2008 | A1 |
20080215392 | Rajan | Sep 2008 | A1 |
20080243531 | Hyder et al. | Oct 2008 | A1 |
20090024694 | Fong | Jan 2009 | A1 |
20090037305 | Sander | Feb 2009 | A1 |
20090037847 | Achtermann et al. | Feb 2009 | A1 |
20090048957 | Celano | Feb 2009 | A1 |
20090064851 | Morris et al. | Mar 2009 | A1 |
20090117529 | Goldstein | May 2009 | A1 |
20090125618 | Huff | May 2009 | A1 |
20090138389 | Barthel | May 2009 | A1 |
20090150169 | Kirkwood | Jun 2009 | A1 |
20090157572 | Chidlovskii | Jun 2009 | A1 |
20090193389 | Miller | Jul 2009 | A1 |
20090204881 | Murthy | Aug 2009 | A1 |
20090239650 | Alderucci et al. | Sep 2009 | A1 |
20090248594 | Castleman | Oct 2009 | A1 |
20090248603 | Kiersky | Oct 2009 | A1 |
20100036760 | Abeles | Feb 2010 | A1 |
20100088124 | Diefendorf et al. | Apr 2010 | A1 |
20100131394 | Rutsch et al. | May 2010 | A1 |
20100153138 | Evans | Jun 2010 | A1 |
20110004537 | Allanson et al. | Jan 2011 | A1 |
20110078062 | Kleyman | Mar 2011 | A1 |
20110145112 | Abeles | Jun 2011 | A1 |
20110173222 | Sayal et al. | Jul 2011 | A1 |
20110225220 | Huang et al. | Sep 2011 | A1 |
20110258195 | Welling | Oct 2011 | A1 |
20110258610 | Aaraj et al. | Oct 2011 | A1 |
20110264569 | Houseworth | Oct 2011 | A1 |
20120016817 | Smith et al. | Jan 2012 | A1 |
20120027246 | Tifford | Feb 2012 | A1 |
20120030076 | Checco et al. | Feb 2012 | A1 |
20120030577 | Akolkar et al. | Feb 2012 | A1 |
20120072321 | Christian et al. | Mar 2012 | A1 |
20120109792 | Eftekhari | May 2012 | A1 |
20120109793 | Abeles | May 2012 | A1 |
20120136764 | Miller | May 2012 | A1 |
20120278365 | Labat et al. | Nov 2012 | A1 |
20130036347 | Eftekhari | Feb 2013 | A1 |
20130080302 | Allanson et al. | Mar 2013 | A1 |
20130097262 | Dandison | Apr 2013 | A1 |
20130111032 | Alapati et al. | May 2013 | A1 |
20130138586 | Jung et al. | May 2013 | A1 |
20130185347 | Romano | Jul 2013 | A1 |
20130187926 | Silverstein et al. | Jul 2013 | A1 |
20130198047 | Houseworth | Aug 2013 | A1 |
20130218735 | Murray | Aug 2013 | A1 |
20130262279 | Finley et al. | Oct 2013 | A1 |
20130282539 | Murray | Oct 2013 | A1 |
20130290169 | Bathula | Oct 2013 | A1 |
20140108213 | Houseworth | Apr 2014 | A1 |
20140172656 | Shaw | Jun 2014 | A1 |
20140201045 | Pai et al. | Jul 2014 | A1 |
20140207633 | Aldrich et al. | Jul 2014 | A1 |
20140241631 | Huang | Aug 2014 | A1 |
20140244455 | Huang | Aug 2014 | A1 |
20140244457 | Howell et al. | Aug 2014 | A1 |
20140337189 | Barsade | Nov 2014 | A1 |
20150142703 | Rajesh | May 2015 | A1 |
20150237205 | Waller et al. | Aug 2015 | A1 |
20150254623 | Velez et al. | Sep 2015 | A1 |
20150269491 | Tripathi et al. | Sep 2015 | A1 |
20160027127 | Chavarria et al. | Jan 2016 | A1 |
20160063645 | Houseworth et al. | Mar 2016 | A1 |
20160071112 | Unser | Mar 2016 | A1 |
20160078567 | Goldman et al. | Mar 2016 | A1 |
20160092993 | Ciaramitaro | Mar 2016 | A1 |
20160092994 | Roebuck et al. | Mar 2016 | A1 |
20160098804 | Mascaro et al. | Apr 2016 | A1 |
20160148321 | Ciaramitaro et al. | May 2016 | A1 |
20160275627 | Wang | Sep 2016 | A1 |
20170004583 | Wang | Jan 2017 | A1 |
20170004584 | Wang | Jan 2017 | A1 |
20170032468 | Wang et al. | Feb 2017 | A1 |
20180032855 | Wang et al. | Feb 2018 | A1 |
Number | Date | Country |
---|---|---|
2002-117121 | Apr 2002 | JP |
2005-190425 | Jul 2005 | JP |
2014-206960 | Oct 2014 | JP |
10-2012-0011987 | Feb 2012 | KR |
2017004094 | Jan 2017 | WO |
2017004095 | Jan 2017 | WO |
2017019233 | Feb 2017 | WO |
2017116496 | Jul 2017 | WO |
2017116497 | Jul 2017 | WO |
2018022023 | Feb 2018 | WO |
2018022128 | Feb 2018 | WO |
2018080562 | May 2018 | WO |
2018080563 | May 2018 | WO |
Entry |
---|
Final Office Action dated Jun. 6, 2017 in U.S. Appl. No. 14/462,411, (20pges). |
Amendment After Final Office Action dated Jun. 6, 2017 in U.S. Appl. No. 14/448,922, (8pages). |
Interview Summary dated Jun. 7, 2017 in U.S. Appl. No. 14/448,922, (2pages). |
Advisory Action dated Jun. 14, 2017 in U.S. Appl. No. 14/448,922, (4pages). |
Amendment After Final Office Action dated Jun. 20, 2017 in U.S. Appl. No. 14/448,922, (14pages). |
Office Action dated May 26, 2017 in U.S. Appl. No. 14/553,347, (43pages). |
Office Action dated Jun. 2, 2017 in U.S. Appl. No. 14/673,261, (65pages). |
Office Action dated May 25, 2017 in U.S. Appl. No. 14/529,736, (42pages). |
Office Action dated Jun. 6, 2017 in U.S. Appl. No. 14/462,315, (54pages). |
Amendment and Response dated Jun. 2, 2017 in U.S. Appl. No. 14/448,986, (12pages). |
Interview Summary dated Jun. 2, 2017 in U.S. Appl. No. 14/448,986, (3pages). |
Office Action dated Jun. 7, 2017 in U.S. Appl. No. 14/555,334, (54pages). |
Office Action dated Jun. 7, 2017 in U.S. Appl. No. 14/555,296, (7pages). |
Response dated Jun. 7, 2017 in U.S. Appl. No. 14/555,543, (21pages). |
Amendment dated Jun. 9, 2017 in U.S. Appl. No. 14/097,057, (26pages). |
Office Action dated Jun. 22, 2017 in U.S. Appl. No. 14/698,746, (50pages). |
Response to Restriction Requirement dated Jul. 5, 2017 in U.S. Appl. No. 14/555,902, (12pages). |
PCT International Search Report for PCT/US2016/067866 Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Jul. 26, 2017 (5pages). |
PCT Written Opinion of the International Search Authority for PCT/US2016/067866, Applicant: Intuit Inc., Form PCT/ISA/237, dated Jul. 26, 2017 (4pages). |
PCT International Search Report for PCT/US2016/067867 Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Jul. 26, 2017 (5pages). |
PCT Written Opinion of the International Search Authority for PCT/US2016/067867, Applicant: Intuit Inc., Form PCT/ISA/237, dated Jul. 26, 2017 (9pages). |
Response to Office Action dated Jul. 17, 2017 in U.S. Appl. No. 14/462,345, (17pages). |
Advisory Action dated Jul. 31, 2017 in U.S. Appl. No. 14/462,345, (3pages). |
Request for Continued Examination and Response dated Aug. 14, 2017 in U.S. Appl. No. 14/462,345, (17pages). |
Office Action dated Aug. 9, 2017 in U.S. Appl. No. 14/097,057, (47pages). |
Interview Summary dated Sep. 6, 2017 in U.S. Appl. No. 14/553,347, (2pages). |
Office Action dated Sep. 14, 2017 in U.S. Appl. No. 14/530,159, (41pages). |
Response dated Aug. 10, 2017 in U.S. Appl. No. 141448,678, (41pages). |
Office Action dated Jul. 28, 2017 in U.S. Appl. No. 14/555,553, (52pages). |
Office Action dated Aug. 21, 2017 in U.S. Appl. No. 14/755,684, (43pages). |
Response dated Jul. 5, 2017 in U.S. Appl. No. 141555,902, (12pages). |
Office Action dated Sep. 8, 2017 in U.S. Appl. No. 14/555,939, (92pages). |
Office Action dated Jun. 28, 2017 in U.S. Appl. No. 14/207,121, (29pages). |
Office Action dated Sep. 14, 2017 in U.S. Appl. No. 14/557,335, (57pages). |
Response dated Aug. 7, 2017 in U.S. Appl. No. 141462,315, (10pages). |
Advisory Action dated Aug. 24, 2017 in U.S. Appl. No. 14/462,315, (3pages). |
Request for Examination and Response dated Sep. 6, 2017 in U.S. Appl. No. 14/462,315, (43pages). |
Office Action dated Jun. 27, 2017 in U.S. Appl. No. 14/755,859, (174pages). |
Advisory Action dated Jul. 5, 2017 in U.S. Appl. No. 14/448,922, (4pages). |
Request for Continued Examination and Amendment dated Aug. 21, 2017 in U.S. Appl. No. 14/448,922, (37pages). |
Request for Continued Examination and Amendment dated Sep. 6, 2017 in U.S. Appl. No. 14/448,922, (36pages). |
Request for Continued Examination and Amendment dated Sep. 6, 2017 in U.S. Appl. No. 141462,411, (24pages). |
Office Action dated Aug. 25, 2017 in U.S. Appl. No. 14/673,646, (65pages). |
Office Action dated Jun. 27, 2017 in U.S. Appl. No. 14/675,166, (46pages). |
Response dated Jun. 23, 2017 in U.S. Appl. No. 14/555,293, (7pages). |
Office Action dated Jul. 10, 2017 in U.S. Appl. No. 14/555,222, (63pages). |
Office Action dated Aug. 18, 2017 in U.S. Appl. No. 14/555,543, (42pages). |
Office Action dated Aug. 25, 2017 in U.S. Appl. No. 14/673,555, (71pages). |
Response dated Sep. 21, 2017 in U.S. Appl. No. 14/448,481, (44pages). |
Office Communication dated Apr. 4, 2018 in Canadian Patent Application No. 2,959,230, (6pages). |
Supplementary Search Report dated Mar. 26, 2018 in European Patent Application No. 16843282.1-1217, (6pages). |
Amendment and Response to Office Action for U.S. Appl. No. 14/462,345 dated Apr. 12, 2018, (15pages). |
Response to Office Action for U.S. Appl. No. 14/553,347 dated Mar. 30, 2018, (26pages). |
Advisory Action for U.S. Appl. No. 14/553,347 dated Apr. 13, 2018, (7pages). |
Response and Request for Continued Examination for U.S. Appl. No. 14/553,347 dated Mar. 30, 2018, (41pages). |
Amendment and Response to Office Action for U.S. Appl. No. 14/673,261 dated Apr. 23, 2018, (39pages). |
Advisory Action for U.S. Appl. No. 14/673,261 dated May 14, 2018, (9pages). |
Communication pursuant to Rules 70(2) and 70a(2) EPC dated Apr. 25, 2018 in European Patent Application No. 16843282.1-1217, (1page). |
Office Action for U.S. Appl. No. 14/462,058 dated Apr. 27, 2018, (47pages). |
Amendment and Response to Final and Advisory Actions and Request for Continued Examination for U.S. Appl. No. 14/448,678 dated Mar. 5, 2018, (25pages). |
Amendment and Response for U.S. Appl. No. 14/555,553 dated Apr. 12, 2018, (24pages). |
Advisory Action for U.S. Appl. No. 14/555,553 dated Apr. 24, 2018, (3pages). |
Amendment and Response to Final Office Action and Request for Continued Examination for U.S. Appl. No. 14/555,553 dated May 11, 2018, (25pages). |
Amendment and Response for U.S. Appl. No. 14/529,798 dated Mar. 28, 2018, (23pages). |
Response for U.S. Appl. No. 14/755,684 dated Mar. 12, 2018, (23pages). |
Advisory Action for U.S. Appl. No. 14/755,684 dated Mar. 30, 2018, (2pages). |
Response for U.S. Appl. No. 14/755,684 dated Apr. 4, 2018, (23pages). |
Office Action for U.S. Appl. No. 14/555,902 dated May 17, 2018, (23pages). |
Response for U.S. Appl. No. 14/207,121 dated Mar. 19, 2018, (34pages). |
Advisory Action for U.S. Appl. No. 14/207,121 dated Apr. 6, 2018 (3pages). |
Response for U.S. Appl. No. 14/462,315 dated May 9, 2018, (33pages). |
Office Action for U.S. Appl. No. 14/698,746 dated Feb. 28, 2018, (14pages). |
Response for U.S. Appl. No. 14/698,746 dated Apr. 30, 2018, (18pages). |
Advisory Action for U.S. Appl. No. 14/698,746 dated May 15, 2018, (3pages). |
Response for U.S. Appl. No. 14/462,397 dated Feb. 20, 2018, (33pages). |
Response for U.S. Appl. No. 14/462,373 dated Feb. 28, 2018, (25pages). |
Office Action for U.S. Appl. No. 14/755,859 dated Mar. 21, 2018, (57pages). |
Response for U.S. Appl. No. 14/755,859 dated May 21, 2018, (8pages). |
Response for U.S. Appl. No. 14/448,886 dated Feb. 28, 2018, (31pages). |
Amendment for U.S. Appl. No. 14/448,922 dated Feb. 28, 2018, (27pages). |
Office Action for U.S. Appl. No. 14/448,922 dated May 16, 2018, (41pages). |
Office Action for U.S. Appl. No. 14/448,962 dated Apr. 13, 2018, (17pages). |
Office Action for U.S. Appl. No. 14/448,986 dated May 11, 2018, (15pages). |
Response for U.S. Appl. No. 14/448,986 dated May 8, 2018, (27pages). |
Response for U.S. Appl. No. 14/555,334 dated Apr. 4, 2018, (14pages). |
Advisory Action for U.S. Appl. No. 14/555,334 dated Apr. 17, 2018, (2pages). |
Response for U.S. Appl. No. 14/555,334 dated May 7, 2018, (41pages). |
Office Action for U.S. Appl. No. 14/673,646 dated Feb. 28, 2018, (19pages). |
Response for U.S. Appl. No. 14/673,646 dated Mar. 30, 2018, (22pages). |
Response for U.S. Appl. No. 14/701,087 dated Apr. 2, 2018, (41pages). |
Amendment After Allowance for U.S. Appl. No. 14/675,166, (5pages). |
Supplemental Notice of Allowability for U.S. Appl. No. 14/675,166, (3pages). |
Response for U.S. Appl. No. 14/555,296, (23pages). |
Response for U.S. Appl. No. 14/555,222, (8pages). |
Office Action for U.S. Appl. No. 14/700,981, (28pages). |
Office Action for U.S. Appl. No. 14/673,555, (43pages). |
H.R. Gregg; Decision Tables for Documentation and System Analysis; Oct. 3, 1967; Union Carbide Corporation, Nuclear Division, Computing Technology Center: pp. 5, 6, 18, 19, & 21 (Year: 1967). |
Solomon L. Pollack; Analysis of the Decision Rules in Decision Tables, May 1963; The Rand Corooration; pp. iii, iv, 1, 20, & 24 (Year: 1963). |
Amendment and Response dated Nov. 9, 2017 in U.S. Appl. No. 14/097,057, (31pgs.). |
Amendment and Response dated Oct. 26, 2017 in U.S. Appl. No. 14/553,347, (25pgs.). |
Amendment and Response dated Nov. 2, 2017 in U.S. Appl. No. 14/673,261, (30pgs.). |
Office Action dated Oct. 30, 2017 in U.S. Appl. No. 14/448,678, (39pgs.). |
Amendment and Response dated Oct. 30, 2017 in U.S. Appl. No. 14/555,553, (17pgs.). |
Notice of Allowance dated Nov. 3, 2017 in U.S. Appl. No. 14/529,736, (13pgs.). |
Interview Summary dated Sep. 28, 2017 in U.S. Appl. No. 14/529,736, (3pgs.). |
Office Action dated Sep. 14, 2017 in U.S. Appl. No. 14/530,159, (41pgs.). |
Amendment and Response dated Nov. 21, 2017 in U.S. Appl. No. 14/755,684, (23pgs.). |
Office Action dated Nov. 15, 2017 in U.S. Appl. No. 14/206,834, (100pgs.). |
Office Action dated Sep. 8, 2017 in U.S. Appl. No. 14/555,939, (92pgs.). |
Amendment and Response dated Sep. 28, 2017 in U.S. Appl. No. 14/207,121, (38pgs.). |
Office Action dated Sep. 14, 2017 in U.S. Appl. No. 14/557,335, (57pgs.). |
Amendment and Response dated Aug. 7, 2017 in U.S. Appl. No. 14/462,315, (10pgs.). |
Advisory Action dated Aug. 24, 2017 in U.S. Appl. No. 14/462,315, (3pgs.). |
Amendment and Response and Request for Continued Examination dated Sep. 6, 2017 in U.S. Appl. No. 14/462,315, (43pgs.). |
Amendment and Response dated Sep. 21, 2017 in U.S. Appl. No. 14/448,481, (44pgs.). |
Office Action dated Jun. 22, 2017 in U.S. Appl. No. 14/698,746, (50pgs.). |
Amendment and Response dated Sep. 22, 2017 in U.S. Appl. No. 14/698,746, (26pgs.). |
Office Action dated Oct. 13, 2017 in U.S. Appl. No. 141462,397, (72pgs.). |
Office Action dated Nov. 30, 2017 in U.S. Appl. No. 14/462,373, (72pgs.). |
Office Action dated Jun. 27, 2017 in U.S. Appl. No. 14/755,859, (174pgs.). |
Amendment and Response dated Nov. 27, 2017 in U.S. Appl. No. 14/755,859, (53pgs.). |
Amendment and Response dated Jun. 20, 2017 in U.S. Appl. No. 14/448,886, (14pgs.). |
Advisory Action dated Jul. 5, 2017 in U.S. Appl. No. 14/448,886, (4pgs.). |
Amendment and Response dated Aug. 21, 2017 in U.S. Appl. No. 14/448,886, (37pgs.). |
Office Action dated Nov. 28, 2017 in U.S. Appl. No. 14/448,886, (65pgs.). |
Amendment and Response and Request for Continued Examination dated Sep. 6, 2017 in U.S. Appl. No. 14/448,922, (36pgs.). |
Office Action dated Nov. 28, 2017 in U.S. Appl. No. 14/448,922, (65pgs.). |
Office Action dated Oct. 10, 2017 in U.S. Appl. No. 14/448,962, (27pgs.). |
Office Action dated Oct. 16, 2017 in U.S. Appl. No. 14/448,986, (30pgs.). |
OpenRules, Preparing a Tax Return Using OpenRules Dialog, Aug. 2011 (Year: 2011) (25pgs.). |
Amendment and Response and Request for Continued Examination dated Sep. 6, 2017 in U.S. Appl. No. 14/462,411, (24pgs.). |
Amendment and Response dated Nov. 7, 2017 in U.S. Appl. No. 14/555,334, (26pgs.). |
Advisory Action dated Nov. 22, 2017 in U.S. Appl. No. 14/555,334, (2pgs.). |
Office Action dated Oct. 11, 2017 in U.S. Appl. No. 14/701,030, (53pgs.). |
Office Action dated Aug. 25, 2017 in U.S. Appl. No. 14/673,646, (65pgs.). |
Office Action dated Jul. 10, 2017 in U.S. Appl. No. 14/555,222, (63pgs.). |
Amendment and Response dated Nov. 10, 2017 in U.S. Appl. No. 14/555,222, (25pgs.). |
Office Action dated Nov. 3, 2017 in U.S. Appl. No. 14/701,087, (103pgs.). |
Office Action dated Jun. 27, 2017 in U.S. Appl. No. 14/675,166, (46pgs.). |
Amendment and Response dated Oct. 27, 2017 in U.S. Appl. No. 14/675,166, (25pgs.). |
Response dated Jun. 23, 2017 in U.S. Appl. No. 14/555,296, (7pgs.). |
Office Action dated Oct. 20, 2017 in U.S. Appl. No. 14/555,296, (50pgs.). |
Office Action dated Aug. 18, 2017 in U.S. Appl. No. 14/555,543, (42pgs.). |
Interview Summary dated Oct. 25, 2017 in U.S. Appl. No. 14/555,543, (3pgs.). |
Office Action dated Sep. 25, 2017 in U.S. Appl. No. 14/700,981, (52pgs.). |
Office Action dated Aug. 25, 2017 in U.S. Appl. No. 14/673,555, (65pgs.). |
Office Action dated Sep. 28, 2017 in U.S. Appl. No. 14/701,149, (71pgs.). |
Vanderbilt University, “Free tax prep help available for Vanderbilt employees”, Feb. 6, 2014, Vanderbilt University, p. 1-3 [NPL-1]. |
Notice of Allowance and Fee(s) Due dated May 5, 2017 in U.S. Appl. No. 14/206,682, (30pages). |
PCT International Search Report for PCT/US2016/044094, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Apr. 24, 2017 (5pages). |
PCT Written Opinion of the International Search Authority for PCT/US2016/044094, Applicant: Intuit Inc., Form PCT/ISA/237, dated Apr. 24, 2017 (5pages). |
PCT International Search Report for PCT/US2016/067839, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Apr. 25, 2017 (5pages). |
PCT Written Opinion of the International Search Authority for PCT/US2016/067839, Applicant: Intuit Inc., Form PCT/ISA/237, dated Apr. 26, 2017 (12pages). |
Amendment dated May 3, 2017 in U.S. Appl. No. 14/462,411, filed Aug. 18, 2014, (5pages). |
Response dated May 15, 2017 in U.S. Appl. No. 14/448,962, filed Jul. 31, 2014, (30pages). |
Office Action dated May 15, 2017 in U.S. Appl. No. 14/462,345, filed Aug. 18, 2014, (57pages). |
Office Action dated May 15, 2017 in U.S. Appl. No. 14/555,902, filed Nov. 28, 2014, (8pages). |
PCT International Search Report for PCT/US2017/062777, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Feb. 21, 2018 (5pages). |
PCT Written Opinion of the International Search Authority for PCT/US2017/062777, Applicant: The Regents of the University of California, Form PCT/ISA/237, dated Feb. 21, 2018 (8pages). |
Office Action dated Feb. 22, 2018 in U.S. Appl. No. 14/673,261, filed Mar. 30, 2015, (46pages). |
Wikipedia, https://en.wikipedia.org/wiki/Data_structure, “Data Structures”, Jan. 12, 2012, entire page (Year:2012) (1page). |
Wikipedia, https://en.wikipedia.org/wiki/Tree_(data_structure), “Tree (data structure)”, May 15, 2005, entire page (Year:2005) (1page). |
Response to Rule 161 Communication dated Jan. 5, 2018 in European Patent Application No. 16843282.1, (16pages). |
Communication pursuant to Rules 161(2) and 162 EPC dated Jul. 26, 2017 in European Patent Application No. 16843282.1, (2pages). |
Response dated Aug. 10, 2017 in U.S. Appl. No. 14/448,678, (41pages). |
Response dated Jul. 5, 2017 in U.S. Appl. No. 14/555,902, (12pages). |
Response dated Aug. 7, 2017 in U.S. Appl. No. 14/462,315, (10pages). |
Request for Continued Examination and Amendment dated Sep. 6, 2017 in U.S. Appl. No. 14/462,411, (24pages). |
Office Action dated Nov. 17, 2016 in U.S. Appl. No. 14/448,922, filed Jul. 31, 2014, inventor: Gang Wang. |
Amendment dated Feb. 17, 2016 in U.S. Appl. No. 14/448,922, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Apr. 6, 2017 in U.S. Appl. No. 14/448,922, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Aug. 11, 2016 in U.S. Appl. No. 14/448,962, filed Jul. 31, 2014, inventor: Gang Wang. |
Amendment dated Nov. 11, 2016 in U.S. Appl. No. 14/448,962, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Jan. 13, 2017 in U.S. Appl. No. 14/448,962, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Aug. 23, 2016 in U.S. Appl. No. 14/448,986, filed Jul. 31, 2014, inventor: Gang Wang. |
Response dated Jan. 23, 2017 in U.S. Appl. No. 14/448,986, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Feb. 17, 2017 in U.S. Appl. No. 14/448,986, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Jan. 12, 2017 in U.S. Appl. No. 14/462,411, filed Aug. 18, 2014. |
Office Action dated Feb. 7, 2017 in U.S. Appl. No. 14/555,543, filed Nov. 26, 2014, inventor: Gang Wang. |
PCT International Search Report for PCT/US2016/039919, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Oct. 11, 2016. |
PCT Written Opinion of the International Search Authority for PCT/US2016/039919, Applicant: Intuit Inc., Form PCT/ISA/237, dated Oct. 11, 2016. |
PCT International Search Report for PCT/US2016/039917, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Oct. 11, 2016. |
PCT Written Opinion of the International Search Authority for PCT/US2016/039917, Applicant: Intuit Inc., Form PCT/ISA/237, dated Oct. 11, 2016. |
PCT International Search Report for PCT/US2016/039918, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Oct. 11, 2016. |
PCT Written Opinion of the International Search Authority for PCT/US2016/039918, Applicant: Intuit Inc., Form PCT/ISA/237, dated Oct. 11, 2016. |
PCT International Search Report for PCT/US2016/039913, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Oct. 21, 2016. |
PCT Written Opinion of the International Search Authority for PCT/US2016/039913, Applicant: Intuit Inc., Form PCT/ISA/237, dated Oct. 21, 2016. |
PCT International Search Report for PCT/US2016/039916, Applicant: Intuit Inc., Form PCT/ISA/210 and 220, dated Oct. 11, 2016. |
PCT Written Opinion of the International Search Authority for PCT/US2016/039916, Applicant: Intuit Inc., Form PCT/ISA/237, dated Oct. 11, 2016. |
http://en.wikipedia.org/wiki/Dependency_grammar#Semantic_dependencies, printed Mar. 11, 2014. |
http://www.webopedia.com/TERM/L/loose_coupling.html, printed Mar. 11, 2014. |
http://en.wikipedia.org/wiki/Loose_coupling, printed Mar. 11, 2014. |
www.turbotax.com, printed Mar. 11, 2014. |
https://turbotax.intuit.com/snaptax/mobile/, printed Mar. 11, 2014. |
http://www.jboss.org/drools/drools-expert.html, printed Mar. 11, 2014. |
http://en.wikipedia.org/wiki/Drools, printed Mar. 11, 2014. |
http://en.wikipedia.org/wiki/Declarative_programming, printed Mar. 11, 2014. |
http://www.wisegeek.com/what-is-declarative-programming.htm, printed Mar. 11, 2014. |
http://docs.jboss.org/drools/release/5.3.0.Final/drools-expert-docs/html/ch01.html, printed Mar. 11, 2014. |
http://quicken.intuit.com/support/help/tax-savings/simplify-tax-time/INF24047.html, updated Jul. 25, 2013, printed Jun. 24, 2014 (11 pages). |
http://quicken.intuit.com/support/help/income-and-expenses/how-to-assign-tax-form-line-items-to-a-category/GEN82142.html, updated Aug. 11, 2011, printed Jun. 24, 2014 (2 pages). |
http://quicken.intuit.com/support/help/reports--graphs-and-snapshots/track-the-earnings-taxes--deductions--or-deposits-from-paychecks/GEN82101.html, updated May 14, 2012, printed Jun. 24, 2014 (2 pages). |
NY State Dep of Taxation, NY State Personal Income Tax MeF Guide for Software Developers, 2012, NY State. |
Restriction Requirement dated May 22, 2015 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Response dated Jun. 30, 2015 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Office Action dated Oct. 2, 2015 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Response dated Feb. 29, 2016 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Final Office Action dated Apr. 8, 2016 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Pre-Appeal Brief dated Jun. 24, 2016 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Pre-Appeal Brief Conference Decision dated Aug. 15, 2016 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor Gang Wang. |
Amendment dated Sep. 13, 2016 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Office Action dated Nov. 4, 2016 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Amendment dated Feb. 6, 2017 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Final Rejection dated Mar. 9, 2017 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, inventor: Gang Wang. |
Office Action dated Dec. 23, 2016 in U.S. Appl. No. 14/462,345, filed Aug. 18, 2014, inventor: Gang Wang. |
Amendment dated Mar. 23, 2017 in U.S. Appl. No. 14/462,345, filed Aug. 18, 2014, inventor: Gang Wang. |
Office Action dated Mar. 10, 2017 in U.S. Appl. No. 14/448,678, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Jul. 8, 2015 in U.S. Appl. No. 14/206,682, filed Mar. 12, 2015, inventor: Gang Wang. |
Response dated Aug. 31, 2015 in U.S. Appl. No. 14/206,682, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Mar. 9, 2016 in U.S. Appl. No. 14/206,682, filed Mar. 12, 2015, inventor: Gang Wang. |
Amendment dated Jul. 11, 2016 in U.S. Appl. No. 14/206,682, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Sep. 16, 2016 in U.S. Appl. No. 14/206,682, filed Mar. 12, 2015, inventor: Gang Wang. |
Amendment dated Jan. 13, 2017 in U.S. Appl. No. 14/206,682, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Dec. 31, 2015 in U.S. Appl. No. 14/206,834, filed Mar. 12, 2015, inventor: Gang Wang. |
Amendment dated May 31, 2016 in U.S. Appl. No. 14/206,834, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Sep. 6, 2016 in U.S. Appl. No. 14/206,834, filed Mar. 12, 2015, inventor: Gang Wang. |
Amendment dated Jan. 6, 2017 in U.S. Appl. No. 14/206,834, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Apr. 30, 2015 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2015, inventor: Gang Wang. |
Response dated Apr. 30, 2015 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Jul. 30, 2015 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2015, inventor: Gang Wang. |
Response dated Nov. 30, 2015 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Apr. 29, 2016 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2015, inventor: Gang Wang. |
Amendment dated Aug. 29, 2016 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2015, inventor: Gang Wang. |
Office Action dated Dec. 14, 2016 in U.S. Appl. No. 14/462,315, filed Aug. 18, 2014, inventor: Gang Wang. |
Response dated Mar. 14, 2017 in U.S. Appl. No. 14/462,315, filed Aug. 18, 2014, inventor: Gang Wang. |
Office Action dated Mar. 21, 2017 in U.S. Appl. No. 14/448,481, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Nov. 29, 2016 in U.S. Appl. No. 14/448,886, filed Jul. 31, 2014, inventor: Gang Wang. |
Amendment dated Feb. 28, 2017 in U.S. Appl. No. 14/448,886, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Apr. 20, 2017 in U.S. Appl. No. 14/448,886, filed Jul. 31, 2014, inventor: Gang Wang. |
Office Action dated Dec. 28, 2017 in U.S. Appl. No. 14/097,057, filed Dec. 4, 2013, (10pages). |
Office Action dated Jan. 12, 2018 in U.S. Appl. No. 14/462,345, filed Aug. 18, 2014, (9pages). |
Office Action dated Jan. 30, 2018 in U.S. Appl. No. 14/553,347, filed Nov. 25, 2014, (40pages). |
PCT International Preliminary Report on Patentability (Chapter I of the Patent Cooperation Treaty) for PCT/US2016/039917, Applicant: Intuit, Inc., Form PCT/IB/326 and 373, dated Feb. 8, 2018 (13pages). |
Response dated Dec. 28, 2017 in U.S. Appl. No. 14/701,149, filed Apr. 30, 2015, (46pages). |
Interview Summary dated Jan. 19, 2018 in U.S. Appl. No. 14/673,555, filed Mar. 30, 2015, (3pages). |
Response dated Jan. 3, 2018 in U.S. Appl. No. 14/448,678, filed Jul. 31, 2014, (37pages). |
Advisory Action dated Feb. 5, 2018 in U.S. Appl. No. 14/448,678, filed Jul. 31, 2014, (7pages). |
Office Action dated Feb. 12, 2018 in U.S. Appl. No. 14/555,553, filed Nov. 26, 2014, (40pages). |
Notice of Allowability dated Dec. 22, 2017 in U.S. Appl. No. 14/529,736, filed Oct. 31, 2014, (13pages). |
Office Action dated Dec. 28, 2017 in U.S. Appl. No. 14/529,798, filed Oct. 31, 2014, (61pages). |
Response dated Jan. 16, 2018 in U.S. Appl. No. 14/530,159, filed Oct. 31, 2014, (13pages). |
Interview Summary dated Feb. 5, 2018 in U.S. Appl. No. 14/530,159, filed Oct. 31, 2014, (3pages). |
Office Action dated Jan. 12, 2018 in U.S. Appl. No. 14/755,684, filed Jun. 30, 2015, (31pages). |
PCT International Preliminary Report on Patentability (Chapter I of the Patent Cooperation Treaty) for PCT/US2016/039919, Applicant: Intuit Inc., Form PCT/IB/326 and 373, dated Jan. 11, 2018, (11pages). |
Response dated Feb. 15, 2018 in U.S. Appl. No. 14/206,834, filed Mar. 12, 2014, (36pages). |
Interview Summary dated Feb. 15, 2018 in U.S. Appl. No. 14/206,834, filed Mar. 12, 2014, (3pages). |
Response dated Jan. 5, 2018 in U.S. Appl. No. 14/555,902, filed Nov. 28, 2014, (14pages). |
Response dated Dec. 8, 2017 in U.S. Appl. No. 14/555,939, filed Nov. 28, 2014, (52pages). |
Office Action dated Jan. 18, 2018 in U.S. Appl. No. 14/207,121, filed Mar. 12, 2014, (22pages). |
Response dated Jan. 31, 2018 in U.S. Appl. No. 14/557,335, filed Dec. 1, 2014, (26pages). |
Office Action dated Feb. 9, 2018 in U.S. Appl. No. 14/462,315, filed Aug. 18, 2014, (38pages). |
Notice of Allowance and Fee(s) Due dated Jan. 25, 2018 in U.S. Appl. No. 14/448,481, filed Jul. 31, 2014, (62pages). |
Interview Summary dated Feb. 9, 2018 in U.S. Appl. No. 14/448,481, filed Jul. 31, 2014, (8pages). |
Response dated Dec. 22, 2017 in U.S. Appl. No. 14/698,746, filed Apr. 28, 2015, (15pages). |
Office Action dated Jan. 26, 2018 in U.S. Appl. No. 14/461,982, filed Aug. 18, 2014, (94pages). |
Interview Summary dated Dec. 15, 2017 in U.S. Appl. No. 14/755,859, filed Jun. 30, 2015, (4pages). |
PCT International Preliminary Report on Patentability (Chapter I of the Patent Cooperation Treaty) for PCT/US2016/039918, Applicant: Intuit Inc., Form PCT/IB/326 and 373, dated Jan. 11, 2018, (11pages). |
Response dated Jan. 10, 2018 in U.S. Appl. No. 14/448,962, filed Jul. 31, 2014, (27pages). |
Interview Summary dated Feb. 20, 2018 in U.S. Appl. No. 14/448,962, filed Jul. 31, 2014, (3pages). |
Response dated Feb. 16, 2018 in U.S. Appl. No. 14/448,986, filed Jul. 31, 2014, (16pages). |
Office Action dated Feb. 8, 2018 in U.S. Appl. No. 14/462,411, filed Aug. 18, 2014, (76pages). |
Office Action dated Feb. 5, 2018 in U.S. Appl. No. 14/555,334, filed Nov. 26, 2014, (52pages). |
Response dated Jan. 11, 2018 in U.S. Appl. No. 14/701,030, filed Apr. 30, 2015, (35pages). |
Response dated Dec. 22, 2017 in U.S. Appl. No. 14/673,646, filed Mar. 30, 2015, (22pages). |
Interview Summary dated Dec. 28, 2017 in U.S. Appl. No. 14/673,646, filed Mar. 30, 2015, (3pages). |
Response dated Feb. 13, 2018 in U.S. Appl. No. 14/462,397, filed Aug. 18, 2014, (33pages). |
Cronin, Julie-Anne et al., Distributing the Corporate Income Tax: Revised U.S. Treasury Methodology, May 2012, Department of Treasury, web, 2-31 (Year:2012) (34pages). |
Notice of Allowance and Fee(s) Due dated Feb. 20, 2018 in U.S. Appl. No. 14/675,166, filed Mar. 31, 2015, (52pages). |
Interview Summary dated Dec. 21, 2017 in U.S. Appl. No. 14/555,222, filed Nov. 26, 2014, (2pages). |
Office Action dated Feb. 5, 2018 in U.S. Appl. No. 14/555,222, filed Nov. 26, 2014, (4pages). |
Response dated Dec. 18, 2017 in U.S. Appl. No. 14/555,543, filed Nov. 26, 2014, (20pages). |
Advisory Action dated Jan. 17, 2018 in U.S. Appl. No. 14/555,543, filed Nov. 26, 2014, (3pages). |
Response dated Jan. 18, 2018 in U.S. Appl. No. 14/555,543, filed Nov. 26, 2014, (20pages). |
Office Action dated Feb. 14, 2018 in U.S. Appl. No. 14/555,543, filed Nov. 26, 2014, (18pages). |
Response dated Jan. 25, 2018 in U.S. Appl. No. 14/700,981, filed Apr. 30, 2015, (30pages). |
Response dated Dec. 26, 2017 in U.S. Appl. No. 14/673,555, filed Mar. 30, 2015, (22pages). |