A Table of Content (TOC) inside a document generally lists the parts of the document in the order they appear. The table of content might include a list of headers or titles of sections inside the document, and also may contain further levels inside each of the header referring to sub-sections. When textual content does not have any formatting information as part of its structure, it is a challenging task to determine which portions of its content is either a header, title, or should otherwise be included in a table of contents. Furthermore, text can appear in an unstructured manner in various scenarios, such as a result of an optical character reader (OCR) conversion, meeting notes, call center transcripts, and various documents often used inside an enterprise. In these unstructured documents, there is no indication of titles, headings, or section separators that identify the portions of the document that should be included in a table of content.
Traditional creation of a table of contents generally requires that the text in the document indicate which part of its document refers to headings and titles and should therefore be included in a table of contents. For example, table of contents generators in word processing software generate a table of contents based on the formatting information that is present inside the electronic document. While the document is being composed, the content is written according to whether it is a heading, a title, or a sub-section by choosing options present as part of the word processing software. The table of contents generator leverages this information and generates a table of contents for the document content automatically. Another example is the automatic generation of a table of contents from HTML files. HTML files include tags such as “<h1>”, “<h2>”, and the like that indicate if a content is a heading, a sub-heading or a title. Existing tools and frameworks leverage this HTML tagging information to generate a table of content based on the HTML tags found within the HTML document. A primary drawback of existing approaches, such as the examples discussed above, above is that such approaches rely upon existing indications of headers and titles in either a form of text format, or in the form of tags such as a mark-up tag in HTML. Given an unstructured text that is generated from a call transcript or a meeting note, such tools will not be able to generate the table of contents because the text upon which they work would lack such indicating information. In fact, the existing approached do not have the capability to semantically understand the document content and to determine the headings and titles that form a part of the table of contents.
An approach is provided for an information handling system that includes a processor and a memory to generate a table of contents pertaining to a document. The approach semantically analyzes the document to identify semantic relationships of proximate elements of the document. A number of candidate headings corresponding to a semantically related section of the document are identified and each of the candidate headings are scored. Based on the scores of each of the candidate headings, a section heading for the semantically related section of the document is selected. The selected heading is then included in the table of contents for the section of the document. The process of identifying candidate headings, scoring candidates, and selecting the section heading is repeated for other semantically related sections of the document.
The foregoing is a summary and thus contains, by necessity, simplifications, generalizations, and omissions of detail; consequently, those skilled in the art will appreciate that the summary is illustrative only and is not intended to be in any way limiting. Other aspects, inventive features, and advantages of the present invention, as defined solely by the claims, will become apparent in the non-limiting detailed description set forth below.
The present invention may be better understood, and its numerous objects, features, and advantages made apparent to those skilled in the art by referencing the accompanying drawings, wherein:
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer, server, or cluster of servers. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
QA system 100 may be configured to receive inputs from various sources. For example, QA system 100 may receive input from the network 102, a corpus of electronic documents 106 or other data, a content creator 108, content users, and other possible sources of input. In one embodiment, some or all of the inputs to QA system 100 may be routed through the network 102. The various computing devices 104 on the network 102 may include access points for content creators and content users. Some of the computing devices 104 may include devices for a database storing the corpus of data. The network 102 may include local network connections and remote connections in various embodiments, such that QA system 100 may operate in environments of any size, including local and global, e.g., the Internet. Additionally, knowledge manager 100 serves as a front-end system that can make available a variety of knowledge extracted from or represented in documents, network-accessible sources and/or structured data sources. In this manner, some processes populate the knowledge manager with the knowledge manager also including input interfaces to receive knowledge requests and respond accordingly.
In one embodiment, the content creator creates content in a document 106 for use as part of a corpus of data with QA system 100. The document 106 may include any file, text, article, or source of data for use in QA system 100. Content users may access QA system 100 via a network connection or an Internet connection to the network 102, and may input questions to QA system 100 that may be answered by the content in the corpus of data. In one embodiment, the process sends well-formed questions (e.g., natural language questions, etc.) to one or more components of the QA system. QA system 100 may interpret the question and provide a response to the content user containing one or more answers to the question. In some embodiments, QA system 100 may provide a response to users in a ranked list of answers.
In some illustrative embodiments, QA system 100 may be the IBM Watson™ QA system available from International Business Machines Corporation of Armonk, N.Y., which is augmented with the mechanisms of the illustrative embodiments described hereafter. The IBM Watson™ knowledge manager system may receive an input question which it then parses to extract the major features of the question, that in turn are then used to formulate queries that are applied to the corpus of data. Based on the application of the queries to the corpus of data, a set of hypotheses, or candidate answers to the input question, are generated by looking across the corpus of data for portions of the corpus of data that have some potential for containing a valuable response to the input question.
The IBM Watson™ QA system then performs deep analysis on the language of the input question and the language used in each of the portions of the corpus of data found during the application of the queries using a variety of reasoning algorithms. There may be hundreds or even thousands of reasoning algorithms applied, each of which performs different analysis, e.g., comparisons, and generates a score. For example, some reasoning algorithms may look at the matching of terms and synonyms within the language of the input question and the found portions of the corpus of data. Other reasoning algorithms may look at temporal or spatial features in the language, while others may evaluate the source of the portion of the corpus of data and evaluate its veracity.
The scores obtained from the various reasoning algorithms indicate the extent to which the potential response is inferred by the input question based on the specific area of focus of that reasoning algorithm. Each resulting score is then weighted against a statistical model. The statistical model captures how well the reasoning algorithm performed at establishing the inference between two similar passages for a particular domain during the training period of the IBM Watson™ QA system. The statistical model may then be used to summarize a level of confidence that the IBM Watson™ QA system has regarding the evidence that the potential response, i.e. candidate answer, is inferred by the question. This process may be repeated for each of the candidate answers until the IBM Watson™ QA system identifies candidate answers that surface as being significantly stronger than others and thus, generates a final answer, or ranked set of answers, for the input question. More information about the IBM Watson™ QA system may be obtained, for example, from the IBM Corporation website, IBM Redbooks, and the like. For example, information about the IBM Watson™ QA system can be found in Yuan et al., “Watson and Healthcare,” IBM developerWorks, 2011 and “The Era of Cognitive Systems: An Inside Look at IBM Watson and How it Works” by Rob High, IBM Redbooks, 2012.
Types of information handling systems that can utilize QA system 100 range from small handheld devices, such as handheld computer/mobile telephone 110 to large mainframe systems, such as mainframe computer 170. Examples of handheld computer 110 include personal digital assistants (PDAs), personal entertainment devices, such as MP3 players, portable televisions, and compact disc players. Other examples of information handling systems include pen, or tablet, computer 120, laptop, or notebook, computer 130, personal computer system 150, and server 160. As shown, the various information handling systems can be networked together using computer network 100. Types of computer network 102 that can be used to interconnect the various information handling systems include Local Area Networks (LANs), Wireless Local Area Networks (WLANs), the Internet, the Public Switched Telephone Network (PSTN), other wireless networks, and any other network topology that can be used to interconnect the information handling systems. Many of the information handling systems include nonvolatile data stores, such as hard drives and/or nonvolatile memory. Some of the information handling systems shown in
Northbridge 215 and Southbridge 235 connect to each other using bus 219. In one embodiment, the bus is a Direct Media Interface (DMI) bus that transfers data at high speeds in each direction between Northbridge 215 and Southbridge 235. In another embodiment, a Peripheral Component Interconnect (PCI) bus connects the Northbridge and the Southbridge. Southbridge 235, also known as the I/O Controller Hub (ICH) is a chip that generally implements capabilities that operate at slower speeds than the capabilities provided by the Northbridge. Southbridge 235 typically provides various busses used to connect various components. These busses include, for example, PCI and PCI Express busses, an ISA bus, a System Management Bus (SMBus or SMB), and/or a Low Pin Count (LPC) bus. The LPC bus often connects low-bandwidth devices, such as boot ROM 296 and “legacy” I/O devices (using a “super I/O” chip). The “legacy” I/O devices (298) can include, for example, serial and parallel ports, keyboard, mouse, and/or a floppy disk controller. The LPC bus also connects Southbridge 235 to Trusted Platform Module (TPM) 295. Other components often included in Southbridge 235 include a Direct Memory Access (DMA) controller, a Programmable Interrupt Controller (PIC), and a storage device controller, which connects Southbridge 235 to nonvolatile storage device 285, such as a hard disk drive, using bus 284.
ExpressCard 255 is a slot that connects hot-pluggable devices to the information handling system. ExpressCard 255 supports both PCI Express and USB connectivity as it connects to Southbridge 235 using both the Universal Serial Bus (USB) the PCI Express bus. Southbridge 235 includes USB Controller 240 that provides USB connectivity to devices that connect to the USB. These devices include webcam (camera) 250, infrared (IR) receiver 248, keyboard and trackpad 244, and Bluetooth device 246, which provides for wireless personal area networks (PANs). USB Controller 240 also provides USB connectivity to other miscellaneous USB connected devices 242, such as a mouse, removable nonvolatile storage device 245, modems, network cards, ISDN connectors, fax, printers, USB hubs, and many other types of USB connected devices. While removable nonvolatile storage device 245 is shown as a USB-connected device, removable nonvolatile storage device 245 could be connected using a different interface, such as a Firewire interface, etcetera.
Wireless Local Area Network (LAN) device 275 connects to Southbridge 235 via the PCI or PCI Express bus 272. LAN device 275 typically implements one of the IEEE 0.802.11 standards of over-the-air modulation techniques that all use the same protocol to wireless communicate between information handling system 200 and another computer system or device. Optical storage device 290 connects to Southbridge 235 using Serial ATA (SATA) bus 288. Serial ATA adapters and devices communicate over a high-speed serial link. The Serial ATA bus also connects Southbridge 235 to other forms of storage devices, such as hard disk drives. Audio circuitry 260, such as a sound card, connects to Southbridge 235 via bus 258. Audio circuitry 260 also provides functionality such as audio line-in and optical digital audio in port 262, optical digital output and headphone jack 264, internal speakers 266, and internal microphone 268. Ethernet controller 270 connects to Southbridge 235 using a bus, such as the PCI or PCI Express bus. Ethernet controller 270 connects information handling system 200 to a computer network, such as a Local Area Network (LAN), the Internet, and other public and private computer networks.
While
At predefined process 430, the process identifies the potential span and level/depth of the section headings that were identified in predefined process 410 and stored in memory area 420 (see
At predefined process 450, the process calculates heading scores for the potential headings and derives table of contents 320 (see
At step 520, the process gathers structural cues that pertain to the selected item (e.g., noun, n-gram, etc.). Structural cues can include cues such as whether the item is a bulleted or numbered item, whether there are gaps between lines or paragraphs, indentation of lines or paragraphs, and other symbols and structural cues. At step 530, the selected item is submitted to knowledge manager 104 that is, in one embodiment, trained in the domain regarding the item's semantic features. For example, if the input document is a medical transcript, then the domain might be a medical domain with the corpus including other medical transcripts and documents. At step 530, the process receives semantic data back from knowledge manager 104.
A decision is made by the process, based on the received semantic data, as to whether the selected item is a potential heading in the document (decision 550). If the selected item is a potential heading, then decision 550 branches to the “yes” branch whereupon, at step 560, the selected item is saved as a potential heading in memory area 420 along with the location of the selected item in the input document. At predefined process 570, the process identifies the potential span and level/depth of the potential section heading (see
A decision is made by the process as to whether there are more items in the input document to select and process (decision 580). If there are more items to select and process, then decision 580 branches to the “yes” branch which loops back to select and process the next item in the input document as described above. This looping continues until all of the items in the input document have been processed, at which point decision 595 branches to the “no” branch whereupon processing returns to the calling routine (see
A decision is made by the process as to whether there are more sentences in the document to process (decision 675). If there are more sentences in the document to process, then decision 675 branches to the “yes” branch which loops back to select and process the next sentence as described above. This looping continues until there are no more sentences to process, at which point decision 675 branches to the “no” branch.
A decision is made by the process as to whether there are more candidate headings to process from memory area 420 (decision 680). If there are more candidate headings to process, then decision 680 branches to the “yes” branch which loops back to select and process the next candidate heading as described above. This looping continues until there are no more candidate headings to process, at which point decision 680 branches to the “no” branch whereupon, at predefined process 690, the process identifies the potential level and depth of the candidate headings (see
A decision is made by the process as to whether the span of the selected candidate heading is the same as the span of the comparison candidate (decision 730). If the span of the selected candidate heading is the same as the span of the comparison candidate, then decision 730 branches to the “yes” branch whereupon, at step 740, the process merges the selected candidate heading and the comparison candidate as a new selected candidate heading and the process restarts the evaluation of this new selected candidate. On the other hand, if the span of the selected candidate heading is not the same as the span of the comparison candidate, then decision 730 branches to the “no” branch whereupon a decision is made by the process as to whether the span of the selected candidate heading is contained within the span of the comparison candidate (decision 750). If the span of the selected candidate heading is contained within the span of the comparison candidate, then decision 750 branches to the “yes” branch whereupon, at step 760, the process identifies the selected candidate heading as being sub-heading of the comparison candidate, with the selected candidate heading being at a lower level/depth than the comparison candidate. On the other hand, if the span of the selected candidate heading is not contained within the span of the comparison candidate, then decision 750 branches to the “no” branch bypassing step 760.
A decision is made by the process as to whether there are more comparison candidates to process and compare to the selected candidate heading as described above (decision 770). If there are more comparison candidates to process, then decision 770 branches to the “yes” branch which loops back to step 720 to select the next comparison candidate and process the next comparison candidate as described above. This looping continues until there are no more comparison candidates to process, at which point decision 770 branches to the “no” branch. A decision is made by the process as to whether there are more candidate headings to process (decision 780). If there are more candidate headings to process, then decision 780 branches to the “yes” branch which loops back to step 710 to select the next candidate heading and process it as described above. This looping continues until all of the candidate headings stored in memory area 420 have been processed, at which point decision 780 branches to the “no” branch and processing returns to the calling routine (see
A decision is made by the process as to whether there are more candidate headings to process (decision 850). If there are more candidate headings to process, then decision 850 branches to the “yes” branch which loops back to step 810 to select the next candidate heading from memory area 420 and process the candidate heading as described above with a decision ultimately being made as to whether to include the candidate heading as a potential heading that might be included in the table of contents. This looping continues until all candidate headings have been processed, at which point decision 850 branches to the “no” branch for further processing.
At step 860, the process initializes the current level to a base level (e.g., to zero, etc.). The current level is stored in memory area 870. At predefined process 875, the process visits each of the headings in the current level (see
At step 940, the process the first potential level heading from memory area 930 with the first selected heading being the heading with the highest score. At step 950, the span of the selected heading is compared with the spans of those headings that have already been identified for this level (if any) with such spans of other headings being retrieved from section headings memory area 460.
A decision is made by the process as to whether the span of the selected heading overlaps with the span of an already existing section heading at this level (decision 960). If the span of the selected heading does not overlap with the span of a heading already included in memory area 460, then decision 960 branches to the “no” branch whereupon, at step 970, the selected heading is included as a section heading along with the page number and span of the selected heading. The selected heading, page number, and span data are stored in memory area 460.
A decision is made by the process as to whether there are more potential headings in the current level (decision 980). If there are more potential headings in the current level, then decision 980 branches to the “yes” branch which loops back to select and process the next potential level heading from sorted memory area 930. This looping continues until all of the potential headings from the current level have been processed, at which point decision 980 branches of the “no” branch and processing returns to the calling routine (see
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
While particular embodiments of the present invention have been shown and described, it will be obvious to those skilled in the art that, based upon the teachings herein, that changes and modifications may be made without departing from this invention and its broader aspects. Therefore, the appended claims are to encompass within their scope all such changes and modifications as are within the true spirit and scope of this invention. Furthermore, it is to be understood that the invention is solely defined by the appended claims. It will be understood by those with skill in the art that if a specific number of an introduced claim element is intended, such intent will be explicitly recited in the claim, and in the absence of such recitation no such limitation is present. For non-limiting example, as an aid to understanding, the following appended claims contain usage of the introductory phrases “at least one” and “one or more” to introduce claim elements. However, the use of such phrases should not be construed to imply that the introduction of a claim element by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim element to inventions containing only one such element, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an”; the same holds true for the use in the claims of definite articles.
Number | Date | Country | |
---|---|---|---|
Parent | 14132173 | Dec 2013 | US |
Child | 15060789 | US |