The disclosed implementations relate generally to data visualization and more specifically to systems, methods, and user interfaces that enable users to interact with and explore datasets using a natural language interface.
Data visualization applications enable a user to understand a data set visually, including distribution, trends, outliers, and other factors that are important to making business decisions. Some data sets are very large or complex, and include many data fields. Various tools can be used to help understand and analyze the data, including dashboards that have multiple data visualizations. However, some functionality may be difficult to use or hard to find within a complex user interface. Most systems return only very basic interactive visualizations in response to queries, and others require expert modeling to create effective queries. Other systems require simple closed-ended questions, and then are only capable of returning a single text answer or a static visualization.
Natural language has garnered interest as a useful modality for creating and interacting with visualizations. Natural language interaction techniques offer the promise of easy, intuitive engagement with data even for non-experts by enabling users to express their analytical goals as natural language utterances. Natural language interaction is effective when it supports questions pertaining to a user's information needs. In this respect, however, many natural language interfaces fail to accurately determine a user's information needs.
To effectively support visual analytical workflows, it is critical to accurately infer the user's intent. However, existing natural language interfaces either do not infer intent, infer very limited aspects of intent, rely on explicitly named data attributes, values, and chart types, or restrict relevance of the automatically generated visualization responses.
Accordingly, there is a need for tools that infer user intent to produce more useful visualizations. There is also a need for tools that employ intent to allow users to effectively utilize functionality provided by data visualization applications. One solution to the problem is providing a natural language interface as part of a data visualization application (e.g., within the user interface for the data visualization application) for an interactive dialog that provides graphical results to natural language input. The natural language interface uses both context and intent to support analytical flow in conversations. The natural language interface models transitions in visual analytical conversations by characterizing users' goals.
In accordance with some implementations, a method executes at an electronic device with a display, one or more processors, and memory. For example, the electronic device can be a smart phone, a tablet, a notebook computer, or a desktop computer. The device receives user selection of a data source and a first natural language command, from the user, requesting information related to the data source. The device determines one or more first keywords from the first natural language command and determines, based on the one or more keywords, a first user intent to generate a data visualization. The device then generates a visual specification based on the first user intent. The visual specification specifies the data source (e.g., the user selected data source), a plurality of visual variables, and a plurality of data fields from the data source. Each of the visual variables is associated with a respective one or more of the data fields and each of the data fields is identified as either a dimension or a measure. The device then generates a data visualization based on the generated visual specification and displays the data visualization to the user.
In some implementations, the first user intent includes any of: determining an explicit user intent based on the one or more first keywords, determining a first context based on the plurality of data fields from the data source, and determining an implicit user intent based on the one or more first keywords and the first context.
In some implementations, when an explicit user intent and an implicit user intent are determined, the explicit user intent is prioritized over the implicit user intent.
In some implementations, the device displays a transcription of the first natural language command in an editable field in a user interface in response to receiving the first natural language command. The device may also receive user input to edit the transcription. In such cases, the one or more first keywords are determined based on the edited transcription of the first natural language command.
In some implementations, the device also receives user input to adjust one or more interactive filter controls and interactive legends in a user interface. In response to receiving the user input, the device modifies the plurality of visual variables based on the user input and displays a modified data visualization based on the modified plurality of visual variables.
In some implementations, the device also determines a data visualization type based on the first user intent and generates the data visualization in accordance with the determined data visualization type. In some implementations, the data visualization types include bar charts, line charts, scatterplots, pie charts, heatmaps, text tables, and maps.
In some implementations, the device also receives a second natural language command. In response to receiving the second natural language command, the device determines: (i) one or more second keywords from the second natural language command, (ii) a second user intent based on the one or more second keywords, and (iii) a second context based on the first natural language command and/or the data visualization. The device also modifies the data visualization based on the second user intent and the second context, and displays the modified data visualization.
In some implementations, the first natural language command includes verbal user input (e.g., captured by a microphone).
In some implementations, the first natural language command includes a user utterance.
In some implementations, the first natural language command includes user input of text into a natural language input field.
In accordance with some implementations, a method executes at an electronic device with a display, one or more processors, and memory. For example, the electronic device can be a smart phone, a tablet, a notebook computer, or a desktop computer. The device displays an initial data visualization according to a visual specification that specifies a data source, a plurality of visual variables, and a plurality of data fields from the data source. Each of the visual variables is associated with a respective one or more of the data fields and each of the data fields is identified as either a dimension or a measure. The device receives a user specified first natural language command requesting information related to the data source. The device extracts one or more first keywords from the first natural language command and determines a first user intent to modify the initial data visualization based on the one or more first keywords. The device then modifies the plurality of visual variables in the visual specification based on the first user intent and displays a modified data visualization to the user. The modified data visualization is based on the modified plurality of visual variables.
In some implementations, the first user intent includes any of: determining an explicit user intent based on the one or more first keywords, determining a first context based on the initial data visualization, and determining an implicit user intent based on the one or more first keywords and the first context.
In some implementations, when an explicit user intent and an implicit user intent are determined, the explicit user intent is prioritized over the implicit user intent.
In some implementations, the device also determines a data visualization type based on the first user intent and the data visualization is modified in accordance with the determined data visualization type.
In some implementations, the device displays a transcription of the first natural language command in an editable field in a user interface in response to receiving the first natural language command. The device may also receive a user input to edit the transcription. In such cases, the one or more first keywords are determined based on the edited transcription of the first natural language command.
In some implementations, the device also receives user input to adjust one or more interactive filter controls and interactive legends in a user interface. In response to receiving the user input, the device updates the modified plurality of visual variables based on the user input and displays a an updated data visualization based on the updated plurality of visual variables.
In some implementations, the device also receives a second natural language command that includes a request for information regarding the data source or a request to change the modified data visualization. In response to receiving the second natural language command, the device determines: (i) one or more second keywords from the second natural language command, (ii) a second user intent to change the modified data visualization based on the one or more second keywords, and (iii) a second context based the modified data visualization and/or the first natural language command. The device also updates the data visualization based on the second user intent and the second context, and displays the updated data visualization.
In some implementations, the first natural language command includes a verbal user input.
In some implementations, the first natural language command includes a user utterance.
In some implementations, the first natural language command includes a user input of text into a natural language input field.
Typically, an electronic device includes one or more processors, memory, a display, and one or more programs stored in the memory. The programs are configured for execution by the one or more processors and are configured to perform any of the methods described herein.
In some implementations, a non-transitory computer readable storage medium stores one or more programs configured for execution by a computing device having one or more processors, memory, and a display. The one or more programs are configured to perform any of the methods described herein.
Thus methods, systems, and graphical user interfaces are disclosed that allow users to efficiently generate and modify data displayed within a data visualization application by using natural language commands.
Both the foregoing general description and the following detailed description are exemplary and explanatory, and are intended to provide further explanation of the invention as claimed.
For a better understanding of the aforementioned systems, methods, and graphical user interfaces, as well as additional systems, methods, and graphical user interfaces that provide data visualization analytics, reference should be made to the Description of Implementations below, in conjunction with the following drawings in which like reference numerals refer to corresponding parts throughout the figures.
Reference will now be made to implementations, examples of which are illustrated in the accompanying drawings. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will be apparent to one of ordinary skill in the art that the present invention may be practiced without requiring these specific details.
Some implementations of an interactive data visualization application can provide data visualizations based on a natural language command input by a user. The natural language command can include any form of user input that can be understood and translated or transcribed. For instance, the natural language command may include verbal input, a user utterance, a text input, a symbolic input or even a user gesture. In most instances, the natural language command includes a request for information regarding the data source. For example, the natural language command may include a request to plot data included in the data source, or alternatively, may ask a question based on the data in the data source. For example, the data visualization application may receive a first natural language command to “create a scatter plot.” In some instances, in response to receiving the first natural language command, the first natural language command is displayed in the natural language input box 112 and a data visualization 120 is displayed in the data visualization region 114 of the user interface 110. In addition to generating a data visualization based on the data source and a natural language command received from a user, the data visualization application may also receive a natural language command to modify a current data visualization. For example, as shown in
In some implementations, the data visualization application is configured to provide responses to natural language commands so that there are no “dead ends.” By utilizing the methods described herein, the data visualization application maintains context from either existing data visualizations or from the natural language command in order to provide consistent modifications to the data visualization that are not unexpected for the user (e.g., the system maintains the context of the data prior to the natural language command). Additionally, the data visualization application may use techniques to extract or infer a user's intention based on the natural language command. By using methods that can handle underspecified or vague commands, as well as infer an implicit user intent from the natural language command, the data visualization application can adapt to each natural language command, remain interactive with the user, and allow the user to “undo” or “retry” when the data visualization application provides an unexpected data visualization. This avoids raising error message or not providing a response to an underspecified natural language command (e.g., “dead ends”).
The memory 206 includes high-speed random-access memory, such as DRAM, SRAM, DDR RAM, or other random-access solid-state memory devices; and may include non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid-state storage devices. In some implementations, the memory 206 includes one or more storage devices remotely located from the processors 202. The memory 206, or alternatively the non-volatile memory devices within the memory 206, includes a non-transitory computer-readable storage medium. In some implementations, the memory 206 or the computer-readable storage medium of the memory 206 stores the following programs, modules, and data structures, or a subset or superset thereof:
In some implementations, the data visualization application 230 includes a data visualization generation module 234, which takes user input (e.g., a visual specification 236), and generates a corresponding visual graphic. The data visualization application 230 then displays the generated visual graphic in the user interface 232. In some implementations, the data visualization application 230 executes as a standalone application (e.g., a desktop application). In some implementations, the data visualization application 230 executes within the web browser 226 or another application using web pages provided by a web server (e.g., a server-based application).
In some implementations, the information the user provides (e.g., user input) is stored as a visual specification 236. In some implementations, a visual specification 236 includes previous natural language commands received from a user or properties specified by the user through natural language commands.
In some implementations, the data visualization application 230 includes a language processing module 238 for processing (e.g., interpreting) commands provided by a user of the computing device. In some implementations, the commands are natural language commands (e.g., captured by the audio input device 220 or input via the touch surface 214 or the one or more input buttons such as buttons on a keyboard/mouse 216). In some implementations, the language processing module 238 includes sub-modules, such as an intent deduction module.
In some implementations, the memory 206 stores metrics and/or scores determined by the language processing module 238. In addition, the memory 206 may store thresholds and other criteria, which are compared against the metrics and/or scores determined by the language processing module 238. For example, the language processing module 238 may determine a relatedness metric (discussed in detail below) for an analytic word/phrase of a received command. Then, the language processing module 238 may compare the relatedness metric against a threshold stored in the memory 206.
Each of the above identified executable modules, applications, or sets of procedures may be stored in one or more of the previously mentioned memory devices, and corresponds to a set of instructions for performing a function described above. The above identified modules or programs (i.e., sets of instructions) need not be implemented as separate software programs, procedures, or modules, and thus various subsets of these modules may be combined or otherwise re-arranged in various implementations. In some implementations, the memory 206 stores a subset of the modules and data structures identified above. Furthermore, the memory 206 may store additional modules or data structures not described above.
Although
In some implementations, the memory 264 includes high-speed random-access memory, such as DRAM, SRAM, DDR RAM, or other random-access solid-state memory devices, and may include non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid-state storage devices. In some implementations, the memory 264 includes one or more storage devices remotely located from the CPUs 250. The memory 264, or alternatively the non-volatile memory devices within the memory 264, comprises a non-transitory computer readable storage medium.
In some implementations, the memory 264, or the computer readable storage medium of the memory 264, stores the following programs, modules, and data structures, or a subset thereof:
In some implementations, the data visualization web application 280 includes a data visualization generation module 234 and/or a language processing module 238, as described above for a client device 200. In some implementations, the data visualization web application 280 stores visual specifications 236, which are used to build data visualizations.
Although
Referring to
Note that the generated data visualization has binned the records according to age, such as defining children to be passengers whose ages were <18. The display of children/not children and survived/not survived is generally more useful than a data visualization that literally displayed only what the user asked (e.g., a single number based on filtering the data to just rows from the data source having Age <18 and Survived=True).
A user may provide a second natural language command 304 (“break this down by sex and age”). In response, the computing device 200 modifies the visual specification based on second keyword(s) determined from the second natural language command 304 and a user intent that is determined based on the second keyword(s). The computing device 200 then provides a second data visualization 314. The second data visualization 314 retains the attributes “Children Aboard?” and “Survived?” from the first data visualization 312, while adding data attributes “Sex” and “Age” in a way that preserves the previous structure of the bar chart in the first data visualization 312 (namely, a stacked bar chart that is further split into age bins).
Further, the user provides a third natural language command 306 (“what's the correlation between age, fare, and survival”), which has an implicit intent of asking for a “correlation.” The factors (“age,” “fare,” and “survival”) in the data source suggest a new visualization, such as a heat map, to depict the correlation between the data attributes “survived,” “Age,” and “Fare.” The computing device 200 responds by generating and displaying a third data visualization 314 that is a heat map, according to some implementations.
In some implementations, a natural language command may include an explicit intent request and/or an implicit intent. An explicit intent clearly states what a user expects to see and is prioritized over other inferences such as implicit intent. For example, if the first natural language command 342 includes “age and fare colored by survival status,” the user clearly indicates how the survival status attribute should be encoded. In this example, the an explicit intent is determined by the computing device 200 and is used in generating the visual specification (e.g., the visual specification will dictate that the survival status is encoded by color). Conversely, implicit intents do not directly specify details of a visual specification, but visualization best practices suggest taking implicit intent into consideration when generating the visual specification.
Further, the user may provide a second natural language command 344 (“add a trend line”) to modify the currently displayed data visualization 352. In response to receiving the second natural language command 344, the computing device 200 determines the user's intent to modify the existing data visualization by adding a trend line. The computing device 200 updates the visual specification based on the determined user intent and displays an updated data visualization 354 that retains the information and data visualization type of the initial data visualization 352 and adds trend lines that correspond to the displayed data. Note that two trend lines are created even though the user asked for “a trend line.” The two distinct trend lines are needed because a separate trend line is needed for each survival status.
Some additional examples of user intent include: elaborate (add new information to the visualization); adjust/pivot (adapt aspects of the visualization, such as apply/remove/modify a filter, or add/remove data fields); start new (create an altogether new visualization); retry (re-attempt a previous step that “failed”—either for technical reasons, such as a query timeout, or because the previous command failed to convey the desired visualization); and undo (return to the prior state).
Some examples of context include (i) currently selected data source, (ii) currently selected data fields, (iii) current data visualization state or type, (iv) currently applied filters, (v) currently displayed visual marks (including trend lines, highlighting, emphasize/deemphasize), and (vi) current color scheme.
The graphical user interface 110 also includes a data visualization region 414 for displaying a visual graphic (also referred to herein as a data visualization). In this example, the data visualization region displays a data visualization corresponding to data fields shown in the data information region 410. In some implementations, when no data source or data fields have been selected, the data visualization region 414 initially has no visual graphic (e.g., a data visualization is not displayed).
In some instances, the data visualization region 414 also includes a legend 416 and an interactive user affordance 418 for displaying and selecting a data visualization type. For example,
In some implementations, the graphical user interface 110 also includes a natural language processing region 420. The natural language processing region 420 includes a natural language input box 112 for receiving natural language commands. A user may interact with the input bar to provide commands. For example, the user may type a command in the natural language input box 112 to provide the command. In addition, the user may indirectly interact with the natural language input box 112 by speaking into a microphone (e.g., an audio input device 220) to provide commands. In some implementations, an initial data visualization may be generated based on a user's selection of a data source and one or more data fields. After the initial data visualization is generated and displayed (in the data visualization region 414), the user may use natural language commands (e.g., in the natural language processing region 420) to further explore the displayed data visualization. For example, the user may provide a command to create a relationship between two data elements. In response to receiving the command, an updated data visualization that shows a correlation between the two data elements is generated and displayed in the data visualization region 414.
In some implementations, the natural language input box 112 is an editable field. For example, the natural language command received from a user has been transcribed into the natural language input box 112 (“number of gold medals by year and gender”). In some cases, the transcription may have resulted in an incorrectly transcribed word, for example, the word “gold” may have been incorrectly transcribed as “gone” or the user may have included an incorrect spelling or a typographical error in entering the natural language command. In such cases, the text displayed in the natural language input box 112 is editable to either correct mistakes such as incorrect transcriptions or typographical errors as well as to change the natural language command. For instance, the user may provide a natural language command, “color code gender by green and blue” and after seeing that the two colors are too similar to one another, the user may change the text to “color code gender by orange and blue.” In response to this command, the visual specification and the data visualization are updated or modified to color code the genders by orange and blue.
In some implementations, as shown in
In some implementations, if a user deselects all options from any of the first or second set of interactive filter controls, the data visualization may show no data or the data visualization region 414 may be empty.
Some implementations include a filter attribute as an encoded visual variable in the visual specification that specifies the filter conditions. For example, by adjusting the interactive filter controls, a user updates the filter visual variable. Some implementations anticipate future user needs by adding more information than requested. As a simple example, when asked “how many people in first class survived?”, the system may respond with a histogram showing the number of people who survived in all class fares. The user may then deselect options corresponding to the 2nd and 3rd class fares and in response, the data visualization shown in
In accordance with some implementations, the computer receives (510) user selection of a data source. For example, referring to
The computer then determines (530) one or more first keywords from the first natural language command and based on the one or more first keywords, the computer determines (540) a first user intent to generate a data visualization. For example, referring to
The computer (550) generates a visual specification that specifies the data source, a plurality of visual variables, and a plurality of data fields from the data source. Each of the visual variables is associated with a respective one or more of the data fields and each of the data fields is identified as either a dimension or a measure. Each of the plurality of visual variables is specified based on the first user intent.
The computer then (560) generates a data visualization based on the generated visual specification and (570) displays the data visualization. Following the example in
In some implementations, a transcription of the first natural language command is displayed (522) in an editable field in a user interface and the computer receives (524) a first user input to edit the transcription. In such cases, the one or more keywords are determined (532) based in the edited transcription of the first natural language command. For example, as shown in
In some implementations, the computer determines one or more of: an explicit user intent (542) based on the one or more first keywords, an implicit user intent (544) based on the one or more keywords, and a data visualization type (546) based on the first user intent. When an explicit user intent and an implicit user intent are both determined, the explicit user intent is prioritized (548) over the implicit user intent. In some implementations, the data visualization is generated (562) in accordance with the determined data type.
In some implementations, the computer (580) receives a second user input to adjust one or more interactive filter controls and/or interactive legends in the user interface, modifies (582) the plurality of visual variables based on the second user input, and displays (584) a modified data visualization based on the modified plurality of visual variables. For example,
In some implementations, the computer also receives (590) a second natural language command and determines: one or more second keywords (591) from the second natural language command, a second user intent (592) based on the one or more keywords, and a second context (593) based on the first natural language command and/or the data visualization that is currently displayed. The computer then modifies (594) the data visualization based on the second user intent and the second context and displays (595) the modified data visualization. For example, in
In accordance with some implementations, the computer displays (610) an initial data visualization according to a visual specification that specifies the data source, a plurality of visual variables, and a plurality of data fields from the data source. Each of the visual variables is associated with a respective one or more of the data fields and each of the data fields is identified as either a dimension or a measure. For example,
The computer then determines (630) one or more first keywords from the first natural language command and based on the one or more first keywords, the computer determines (640) a first user intent to modify the initial data visualization. For example, referring to
The computer then modifies (650) the plurality of visual variables based on the first user intent and displays (660) a modified data visualization based on the modified plurality of visual variables. Following example 3B, the computer determines the user intent to split the data by the data field “survived?”. Thus, the computer modifies the visual variables in the visual specification to include two graphs in the data visualization, one which shows information regarding passengers of the Titanic who survived (“yes”) and a second graph that shows information regarding passengers of the Titanic who did not survive (“no”).
In some implementations, a transcription of the first natural language command is displayed (622) in an editable field in a user interface and the computer receives (624) a first user input to edit the transcription. In such cases, the one or more keywords are determined (632) based in the edited transcription of the first natural language command. An example is provided with respect to
In some instances, the computer determines one or more of: an explicit user intent (641) based on the one or more first keywords, a first context (642) based on the initial data visualization, an implicit user intent (643) based on the one or more keywords, and a data visualization type (644) based on the first user intent. When an explicit user intent and an implicit user intent are both determined, the explicit user intent is prioritized (645) over the implicit user intent. In some implementations, the data visualization is generated (562) in accordance with the determined data type.
In some instances, the computer receives (670) a second user input to adjust one or more interactive filter controls and/or interactive legends in user interface, updates (672) the modified plurality of visual variables based on the second user input, and displays (674) an updated data visualization based on the updated plurality of visual variables. For example,
In some instances, the computer receives (680) a second natural language command and determines: one or more second keywords (681) from the second natural language command, a second user intent (682) based on the one or more keywords, and a second context (683) based on the first natural language command and/or the modified data visualization. The computer then updates (684) the modified data visualization based on the second user intent and the second context and displays (685) an updated data visualization based on the updated plurality of visual variables.
The terminology used in the description of the invention herein is for the purpose of describing particular implementations only and is not intended to be limiting of the invention. As used in the description of the invention and the appended claims, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, steps, operations, elements, components, and/or groups thereof.
The foregoing description, for purpose of explanation, has been described with reference to specific implementations. However, the illustrative discussions above are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The implementations were chosen and described in order to best explain the principles of the invention and its practical applications, to thereby enable others skilled in the art to best utilize the invention and various implementations with various modifications as are suited to the particular use contemplated.
This application claims priority to U.S. Provisional Application Ser. No. 62/827,768, filed Apr. 1, 2019, entitled “Inferring Intent and Utilizing Context for Natural Language Expressions in a Data Visualization User Interface,” which is incorporated by reference herein in its entirety. This application is related to U.S. patent application Ser. No. 16/219,406, filed Dec. 13, 2018, entitled “Identifying Intent in Visual Analytical Conversations,” which is incorporated by reference herein in its entirety. This application is related to U.S. patent application Ser. No. 16/575,349, filed Sep. 18, 2019, entitled “Methods and Systems for Inferring Intent and Utilizing Context For Natural Language Expressions To Generate Data Visualizations in a Data Visualization Interface,” which is incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4800810 | Masumoto | Jan 1989 | A |
5036314 | Barillari et al. | Jul 1991 | A |
5060980 | Johnson et al. | Oct 1991 | A |
5144452 | Abuyama | Sep 1992 | A |
5169713 | Kumurdjian | Dec 1992 | A |
5265244 | Ghosh et al. | Nov 1993 | A |
5265246 | Li et al. | Nov 1993 | A |
5377348 | Lau et al. | Dec 1994 | A |
5383029 | Kojima | Jan 1995 | A |
5560007 | Thai | Sep 1996 | A |
5577241 | Spencer | Nov 1996 | A |
5581677 | Myers et al. | Dec 1996 | A |
5664172 | Antoshenkov | Sep 1997 | A |
5664182 | Nierenberg et al. | Sep 1997 | A |
5668987 | Schneider | Sep 1997 | A |
5794246 | Sankaran et al. | Aug 1998 | A |
5864856 | Young | Jan 1999 | A |
5893088 | Hendricks et al. | Apr 1999 | A |
5913205 | Jain et al. | Jun 1999 | A |
5933830 | Williams | Aug 1999 | A |
6031632 | Yoshihara et al. | Feb 2000 | A |
6032158 | Mukhopadhyay et al. | Feb 2000 | A |
6044374 | Nesamoney et al. | Mar 2000 | A |
6100901 | Mohda et al. | Aug 2000 | A |
6115744 | Robins et al. | Sep 2000 | A |
6154766 | Yost et al. | Nov 2000 | A |
6173310 | Yost et al. | Jan 2001 | B1 |
6188403 | Sacerdoti et al. | Feb 2001 | B1 |
6208990 | Suresh et al. | Mar 2001 | B1 |
6222540 | Sacerdoti et al. | Apr 2001 | B1 |
6247008 | Cambot et al. | Jun 2001 | B1 |
6253257 | Dundon | Jun 2001 | B1 |
6260050 | Yost et al. | Jul 2001 | B1 |
6269393 | Yost et al. | Jul 2001 | B1 |
6300957 | Rao et al. | Oct 2001 | B1 |
6301579 | Becker | Oct 2001 | B1 |
6317750 | Tortolani et al. | Nov 2001 | B1 |
6327628 | Anuff et al. | Dec 2001 | B1 |
6339775 | Zamanian et al. | Jan 2002 | B1 |
6377259 | Tenev et al. | Apr 2002 | B1 |
6397195 | Pinard et al. | May 2002 | B1 |
6400366 | Davies et al. | Jun 2002 | B1 |
6405195 | Ahlberg | Jun 2002 | B1 |
6405208 | Raghavan et al. | Jun 2002 | B1 |
6424933 | Agrawala et al. | Jul 2002 | B1 |
6490593 | Proctor | Dec 2002 | B2 |
6492989 | Wilkinson | Dec 2002 | B1 |
6522342 | Gagnon et al. | Feb 2003 | B1 |
6611825 | Billheimer et al. | Aug 2003 | B1 |
6643646 | Su et al. | Nov 2003 | B2 |
6707454 | Barg et al. | Mar 2004 | B1 |
6714897 | Whitney et al. | Mar 2004 | B2 |
6725230 | Ruth et al. | Apr 2004 | B2 |
6750864 | Anwar | Jun 2004 | B1 |
6768986 | Cras et al. | Jul 2004 | B2 |
6906717 | Couckuyt et al. | Jun 2005 | B2 |
7009609 | Miyadai | Mar 2006 | B2 |
7023453 | Wilkinson | Apr 2006 | B2 |
7089266 | Stolte et al. | Aug 2006 | B2 |
7117058 | Lin et al. | Oct 2006 | B2 |
7176924 | Wilkinson | Feb 2007 | B2 |
7290007 | Farber et al. | Oct 2007 | B2 |
7302383 | Valles | Nov 2007 | B2 |
7315305 | Crotty et al. | Jan 2008 | B2 |
7379601 | Yang et al. | May 2008 | B2 |
7426520 | Gorelik et al. | Sep 2008 | B2 |
7603267 | Wang et al. | Oct 2009 | B2 |
7716173 | Stolte et al. | May 2010 | B2 |
7882144 | Stolte et al. | Feb 2011 | B1 |
8082243 | Gorelik et al. | Dec 2011 | B2 |
8140586 | Stolte et al. | Mar 2012 | B2 |
8442999 | Gorelik et al. | May 2013 | B2 |
8473521 | Fot et al. | Jun 2013 | B2 |
8620937 | Jonas | Dec 2013 | B2 |
8713072 | Stotle et al. | Apr 2014 | B2 |
8751505 | Carmel et al. | Jun 2014 | B2 |
8874613 | Gorelik et al. | Oct 2014 | B2 |
8972457 | Stolte et al. | Mar 2015 | B2 |
9183235 | Stolte et al. | Nov 2015 | B2 |
9299173 | Rope et al. | Mar 2016 | B2 |
9336253 | Gorelik et al. | May 2016 | B2 |
9501585 | Gautam et al. | Nov 2016 | B1 |
9633091 | Stolte et al. | Apr 2017 | B2 |
9665662 | Gautam et al. | May 2017 | B1 |
9818211 | Gibb et al. | Nov 2017 | B1 |
9858292 | Setlur et al. | Jan 2018 | B1 |
9947314 | Cao et al. | Apr 2018 | B2 |
9983849 | Weingartner | May 2018 | B2 |
10042517 | Stolte et al. | Aug 2018 | B2 |
10042901 | Stolte et al. | Aug 2018 | B2 |
10331720 | Neels et al. | Jun 2019 | B2 |
10418032 | Mohajer et al. | Sep 2019 | B1 |
10515121 | Setlur et al. | Dec 2019 | B1 |
10546001 | Nguyen et al. | Jan 2020 | B1 |
10546003 | Gupta et al. | Jan 2020 | B2 |
10564622 | Dean et al. | Feb 2020 | B1 |
10817527 | Setlur et al. | Oct 2020 | B1 |
10956655 | Choe | Mar 2021 | B2 |
11080336 | van Dusen | Aug 2021 | B2 |
11114189 | Prosky et al. | Sep 2021 | B2 |
20010013036 | Judicibus | Aug 2001 | A1 |
20020002325 | Lliff | Jan 2002 | A1 |
20020059204 | Harris | May 2002 | A1 |
20020118192 | Couckuyt et al. | Aug 2002 | A1 |
20020123865 | Whitney et al. | Sep 2002 | A1 |
20020135610 | Ootani et al. | Sep 2002 | A1 |
20020154118 | McCarthy et al. | Oct 2002 | A1 |
20030200034 | Fellenberg et al. | Oct 2003 | A1 |
20040148170 | Acero et al. | Jul 2004 | A1 |
20040183800 | Peterson | Sep 2004 | A1 |
20040227759 | McKnight et al. | Nov 2004 | A1 |
20040243593 | Stolte et al. | Dec 2004 | A1 |
20050035966 | Pasquarette et al. | Feb 2005 | A1 |
20050035967 | Joffrain et al. | Feb 2005 | A1 |
20050060300 | Stolte et al. | Mar 2005 | A1 |
20050099423 | Brauss | May 2005 | A1 |
20060129913 | Vigesaa et al. | Jun 2006 | A1 |
20060136825 | Cory et al. | Jun 2006 | A1 |
20060206512 | Hanrahan et al. | Sep 2006 | A1 |
20070061344 | Dickerman et al. | Mar 2007 | A1 |
20070061611 | MacKinlay et al. | Mar 2007 | A1 |
20070126736 | Tolle | Jun 2007 | A1 |
20070129936 | Wang et al. | Jun 2007 | A1 |
20080016026 | Farber et al. | Jan 2008 | A1 |
20090313576 | Neumann et al. | Dec 2009 | A1 |
20110112837 | Kurki-Sounio et al. | May 2011 | A1 |
20110119047 | Ylonen | May 2011 | A1 |
20110184718 | Chen | Jul 2011 | A1 |
20120323948 | Li et al. | Dec 2012 | A1 |
20130249917 | Fanning et al. | Sep 2013 | A1 |
20140164362 | Syed et al. | Jun 2014 | A1 |
20140236579 | Kurz | Aug 2014 | A1 |
20160078354 | Petri et al. | Mar 2016 | A1 |
20160092090 | Stojanovic et al. | Mar 2016 | A1 |
20160171050 | Das | Jun 2016 | A1 |
20160179908 | Johnston et al. | Jun 2016 | A1 |
20170011023 | Ghannam et al. | Jan 2017 | A1 |
20170091277 | Zoch | Mar 2017 | A1 |
20170091902 | Bostik et al. | Mar 2017 | A1 |
20170118308 | Vigeant | Apr 2017 | A1 |
20170154089 | Sherman | Jun 2017 | A1 |
20170308571 | McCurley | Oct 2017 | A1 |
20180032576 | Romero | Feb 2018 | A1 |
20180039614 | Govindarajulu et al. | Feb 2018 | A1 |
20180144065 | Yellai | May 2018 | A1 |
20180158245 | Govindan | Jun 2018 | A1 |
20180203924 | Agrawal et al. | Jul 2018 | A1 |
20180210883 | Ang | Jul 2018 | A1 |
20180329987 | Tata et al. | Nov 2018 | A1 |
20190042634 | Stolte et al. | Feb 2019 | A1 |
20190102390 | Antunes et al. | Apr 2019 | A1 |
20190108171 | Stolte et al. | Apr 2019 | A1 |
20190115016 | Seok et al. | Apr 2019 | A1 |
20190120649 | Seok et al. | Apr 2019 | A1 |
20190121801 | Jethwa et al. | Apr 2019 | A1 |
20190138648 | Gupta et al. | May 2019 | A1 |
20190163807 | Jain et al. | May 2019 | A1 |
20190179607 | Thangarathnam et al. | Jun 2019 | A1 |
20190197605 | Sadler et al. | Jun 2019 | A1 |
20190236144 | Hou et al. | Aug 2019 | A1 |
20190272296 | Prakash et al. | Sep 2019 | A1 |
20190311717 | Kim et al. | Oct 2019 | A1 |
20190349321 | Cai et al. | Nov 2019 | A1 |
20190384815 | Patel et al. | Dec 2019 | A1 |
20200065385 | Dreher et al. | Feb 2020 | A1 |
20200089700 | Ericson et al. | Mar 2020 | A1 |
20200089760 | Ericson et al. | Mar 2020 | A1 |
20200090189 | Tutuk et al. | Mar 2020 | A1 |
20200104402 | Burnett et al. | Apr 2020 | A1 |
20200110803 | Djalali et al. | Apr 2020 | A1 |
20200134103 | Mankovskii | Apr 2020 | A1 |
20200327432 | Doebelin et al. | Oct 2020 | A1 |
Number | Date | Country |
---|---|---|
215657 | Jan 1994 | HU |
WO 2006060773 | Jun 2006 | WO |
Entry |
---|
Becker, Trellis Graphics Displays: A Multi-dimensional Data Visualization Tool for Data Mining, Aug. 1997, 13 pgs. |
Becker, Visualizing Decision Table Classifiers, 1998, 4 pgs. |
Beers, Office Action, U.S. Appl. No. 11/787,761, dated Jun. 12, 2008, 12 pgs. |
Beers, Office Action, U.S. Appl. No. 11/787,761, dated Dec. 17, 2008, 13 pgs. |
Bosch, Performance Analysis and Visualization of Parallel Systems Using SimOS and Rivet: A Case Study , Jan. 2000, 13 pgs. |
Bosch, Rivet: A Flexible Environment for Computer Systems Visualization, Feb. 2000, 9 pgs. |
Brunk, MineSet: An Integrated System for Data Mining, 1997, 4 pgs. |
Derthick, An Interactive Visual Query Environment for Exploring Data, 1997, 11 pgs. |
Freeze, Unlocking OLAP with Microsoft SQL Server and Excel 2000, 2000, 220 pgs. |
Fua, “Hierarchical Parallel Coordinates for Exploration of Large Datasets,” IEEE 1999, pp. 43-50 (Year: 1999). |
Eser Kandogan, “Star Coordinates: A Multi-dimensional Visualization Technique with Uniform Treatment of Dimensions,” www.citeseerx.st.psu.edu, pp. 1-4, 2000 (YearL 2000). |
Fua, Navigating Hierarchies with Structure-Based Brushes, 1999, 7 pgs. |
Gao, Tong, et al. “Datatone: Managing ambiguity in natural language interfaces for data visualization.” Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology. Nov. 2015, pp. 489-500. (Year: 2015). |
Goldstein, A Framework for Knowledge-Based Interactive Data Exploration, Dec. 1994, 30 pgs. |
Gray, Data Cube: A Relational Aggregation Operator Generalizing Group-By, 1997, 24 pgs. |
Hanrahan, Office Action, U.S. Appl. No. 11/005,652, dated Feb. 20, 2009, 11 pgs. |
Hanrahan, Office Action, U.S. Appl. No. 11/005,652, dated Jul. 24, 2008, 11 pgs. |
Hanrahan, Office Action, U.S. Appl. No. 11/005,652, dated Dec. 27, 2007, 11 pgs. |
Hanrahan, Specification, U.S. Appl. No. 11/005,652, dated Dec. 2, 2004, 104 pgs. |
Healey, On the Use of Perceptual Cues and Data Mining for Effective Visualization of Scientific Datasets, 1998, 8 pgs. |
Hearst, Office Action, U.S. Appl. No. 16/601,413, dated Nov. 3, 2020, 17 pgs. |
HU Search Report, HU P0700460, dated Oct. 9, 2007, 1 pg. |
John V. Carlis and Joseph A. Konstan, Interactive Visulaization of Serial Periodic Data, www. Courses.ischool.berkeley.edu, pp. 1-10, 1998 (Year: 1998). |
Joseph, Office Action, U.S. Appl. No. 13/734,694, dated Aug. 18, 2014, 46 pgs. |
Keim, VisDB: Datatbase Exploration Using Multidimensional Visualization, Aug. 1994, 27 pgs. |
Kohavi, Data Mining and Visualization, 2000, 8 pgs. |
Livney, M. et al., “DEVise: Integraed Querying and Visual Exploration of Large Datasets,” ACM, 1997, pp. 301-312, (Year: 1997). |
MacDonald, Creating Basic Charts, 2006, 46 pgs. |
MacKinlay, Automating the Design of Graphical Presentations of Relational Information, 1986, 34 pgs. |
MacKinlay, Office Action, U.S. Appl. No. 11/223,658, dated May 21, 2008, 20 pgs. |
MacKinlay, Office Action, U.S. Appl. No. 11/223,658, dated Feb. 23, 2009, 19 pgs. |
MacKinlay, Specification, U.S. Appl. No. 11/223,658, dated Sep. 9, 2005, 58 pgs. |
Matsushita, Mitsunori, Eisaku Maeda, and Tsuneaki Kato. “An interactive visualization method of numerical data based on natural language requirements.” International journal of human-computer studies 60.4, Apr. 2004, pp. 469-488. (Year: 2004). |
Perlin, An Alternative Approach to the Computer Interface, 1993, 11 pgs. |
Popescu, et al. “Towards a theory of natural language interfaces to databases.” Proceedings of the 8th international conference on Intelligent user interfaces. Jan. 2003, pp. 149-157. (Year: 2003). |
Rao, The Table Lens: Merging Graphical and Symbolic Representation in an Interactive Focus+Context Visualization for Tabular Information, Apr. 1994, 7 pgs. |
Roth, Interactive Graphic Design Using Automatic Presentation Knowledge, Apr. 24-28, 1994, 7 pgs. |
Roth, Visage: A User Interface Environment for Exploring Information, Oct. 28-29, 2006, 9 pgs. |
Screen Dumps for Microsoft Office Excel 2003 SP2, figures 1-36, 2003, pp. 1-19. |
Setlur, Preinterview First Office Action, U.S. Appl. No. 16/234,470, dated Sep. 24, 2020, 6 pgs. |
Setlur, First Action Interview Office Action, U.S. Appl. No. 16/234,470, dated Oct. 28, 2020, 4 pgs. |
Spenke, Focus: The Interactive Table for Product Comparison and Selection, Nov. 1996, 10 pgs. |
Stevens, On the Theory of Scales of Measurement, Jun. 7, 1946, 4 pgs. |
Stolte, Multiscale Visualization Using Data Cubes, 2002, 8 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 10/453,834, dated May 27, 2006, 9 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 11/488,407, dated Dec. 29, 1999, 8 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 13/019,227, dated Nov. 10, 2011, 8 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 13/425,300, dated Dec. 10, 2013, 10 pgs. |
Stolte, Office Action, U.S. Appl. No. 10/667,194, dated Jan. 7, 2008, 10 pgs. |
Stolte, Office Action, U.S. Appl. No. 10/667,194, dated Feb. 9, 2009, 11 pgs. |
Stolte, Office Action, U.S. Appl. No. 10/667,194, dated Aug. 14, 2007, 16 pgs. |
Stolte, Office Action, U.S. Appl. No. 10/667,194, dated Aug. 14, 2008, 10 pgs. |
Stolte, Office Action, U.S. Appl. No. 10/667,194, dated Jan. 18, 2007, 15 pgs. |
Stolte, Office Action, U.S. Appl. No. 10/667,194, dated Jun. 26, 2006, 13 pgs. |
Stolte, Office Action, U.S. Appl. No. 11/488,407, dated Apr. 3, 2009, 6 pgs. |
Stolte, Office Action, U.S. Appl. No. 13/019,227, dated Apr. 18, 2011, 9 pgs. |
Stolte, Office Action, U.S. Appl. No. 13/425,300, dated Mar. 15, 2013, 7 pgs. |
Stolte, Office Action, U.S. Appl. No. 14/937,836, dated Oct. 7, 2016, 10 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 14/937,836, dated Mar. 1, 2017, 8 pgs. |
Stolte, Office Action, U.S. Appl. No. 15/449,844, dated Jun. 29, 2017, 16 pgs. |
Stolte, Final Office Action, U.S. Appl. No. 15/449,844, dated Feb. 6, 2018, 8 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 15/449,844, dated May 18, 2018, 9 pgs. |
Stolte, Office Action, U.S. Appl. No. 15/582,478, dated Jul. 11, 2017, 16 pgs. |
Stolte, Final Office Action, U.S. Appl. No. 15/582,478, dated Mar. 8, 2018, 10 pgs. |
Stolte, Notice of Allowance U.S. Appl. No. 15/582,478, dated Jun. 26, 2017, 10 pgs. |
Stolte, Notice of Allowance U.S. Appl. No. 16/056,396, dated Apr. 16, 2019, 10 pgs. |
Stolte, Office Action, U.S. Appl. No. 16/056,819, dated Aug. 7, 2019, 12 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 16/056,819, dated Sep. 11, 2019, 8 pgs. |
Stolte Office Action, U.S. Appl. No. 16/220,240, dated Aug. 7, 2019, 11 pgs. |
Stolte, Notice of Allowance, U.S. Appl. No. 16/220,240, dated Sep. 11, 2019, 8 pgs. |
Stolte Notice of Allowance, U.S. Appl. No. 16/137,457, dated Sep. 6, 2019, 10 pgs. |
Stolte Notice of Allowance, U.S. Appl. No. 16/13 7,071, dated Sep. 11, 2019, 10 pgs. |
Stolte, Polaris: A System for Query, Analysis, and Visualization of Multidimensional Relational Databases, Jan. 2002, 14 pgs. |
Stolte, Query Analysis, and Visualization of Hierarchically Structured Data Using Polaris, Jul. 2002, 11 pgs. |
Stolte, Specification, U.S. Appl. No. 10/453,834, Jun. 2, 2003, 114 pgs. |
Stolte, Visualizing Application Behavior on Superscaler Processors, 1999, 9 pgs. |
Tableau Software, IPRP, PCT/US2005/043937, Jun. 5, 2007, 9 pgs. |
Tableau Software, IPRP, PCT/US2007/009810, Oct. 22, 2008, 7 pgs. |
Tableau Software, ISR/WO, PCT/US2005/043937, Apr. 18, 2007, 9 pgs. |
Tableau Software, ISR/WO, PCT/US2006/35300, Jul. 7, 2008, 6 pgs. |
Tableau Software, ISR/WO, PCT/US2007/009810, Jul. 7, 2008, 8 pgs. |
Tableau Software, Inc., International Search Report and Written Opinion, PCT/US2019/055169, dated Dec. 16, 2019, 12 pgs. |
The Board of Trustees . . . Stanford, IPRP, PCT/US04/18217, Oct. 19, 2006, 4 pgs. |
The Board of Trustees . . . Stanford, IPRP, PCT/US2004/30396, Jan 30, 2007, 3 pgs. |
The Board of Trustees . . . Stanford, ISR/WO, PCT/US04/18217, Feb. 7, 2006, 6 pgs. |
The Board of Trustees . . . Stanford, ISR/WO, PCT/US2004/30396, Aug. 24, 2006, 5 pgs. |
The Board of Trustees . . . Stanford, Supplementary ESR, EP 04754739.3, Dec. 17, 2007, 4 pgs. |
Thearling, Visualizing Data Mining Models, 2001, 14 pgs. |
Tory, First Action Preinterview Office Action, U.S. Appl. No. 16/219,406, dated Jul. 10, 2020, 7 pgs. |
Tory, Notice of Allowance U.S. Appl. No. 16/219,406, dated Sep. 9, 2020, 8 pgs. |
Tory, Office Action, U.S. Appl. No. 16/575,354, dated Nov. 3, 2020, 17 pgs. |
Tory, Office Action, U.S. Appl. No. 16/575,349, dated Oct. 13, 2020, 15 pgs. |
Ward, XmdvTool: Integrating Multiple Methods for Visualizing Multi-Variate Data, 9 pgs. |
Welling, Visualization of Large Multi-Dimensional Datasets, Aug. 11, 2000, 6 pgs. |
Wilkinson, nViZn: An Algebra-Based Visualization System, Mar. 21-23, 2001, 7 pgs. |
Wilkinson, Statistics and Computing—The Grammar of Graphics, 1999, 417 pgs. |
Hearst, Notice of Allowance, U.S. Appl. No. 16/601,413, dated Mar. 3, 2021, 10 pgs. |
Setlur, Final Office Action, U.S. Appl. No. 16/234,470, dated Jun. 2, 2021, 49 pgs. |
Tory, Office Action, U.S. Appl. No. 16/575,349, dated Feb. 3, 2021, 9 pgs. |
Setlur, Notice of Allowance, U.S. Appl. No. 16/234,470, dated Nov. 10, 2021, 14 pgs. |
Number | Date | Country | |
---|---|---|---|
62827768 | Apr 2019 | US |