The present invention relates to systems and methods for digitizing and mining handwritten or typed notes to derive insights and enable improved collaboration and real-time analysis and insights.
Global entities regularly use collaborative centers where attendees can explore data and discuss emerging technologies and their applications to complex business challenges. With a focus on facilitating collaboration and stimulating creativity, such collaborative centers provide a unique environment for entities and clients to work together to tackle client issues through data, analytics and AI (DA/AI), to ideate on challenges and opportunities related to strategy, operations, transformation, and use of DA/AI to drive business outcomes.
Generally, participants will write notes throughout a session to capture comments, questions, and insights. Participants may include client stakeholders and facilitators. At times, annotations pertain to artifacts presented or to live demonstrations. Other times, annotations are part of an interactive exercise. While the act of scribing is engaging and useful in the moment, it is difficult to propagate this content for subsequent use during and beyond the session. It is also difficult to synthesize discrete notes into broader, holistic themes as well as assign attribution to annotations.
It would be desirable, therefore, to have a system and method that could overcome the foregoing disadvantages of known systems.
According to an embodiment, the invention relates to a computer-implemented system that digitizes and mines handwritten notes to support real-time collaboration. The system comprises: an interface that communicates with a digital touch interface at a location during a collaboration session; and a computer processor that is connected to the interface and further programmed to perform the steps of: applying computer vision processing to parse an image displayed on the digital touch interface into a set of objects that comprise a set of handwritten notes; performing optical character recognition processing on the set of handwritten notes to identify a corresponding set of text strings; applying a natural language processing to correct errors wherein the errors comprise typographical errors and punctuation; applying the natural language processing to create a word cloud graphic based on the corresponding set of text strings wherein the word cloud graphic is based on a metric; applying the natural language processing to identify one or more keywords based on the set of text strings and making the one or more keywords available in real-time; applying the natural language processing to attribute authorship to each of the text strings based on a writing sample and making the authorship available in real-time; applying a recommender engine to identify one or more suggestions that support the set of text strings and making the one or more suggestions in real-time; and applying a classifier to identify relationships between the set of text strings and making the relationships available in real-time. In addition, an embodiment of the present invention may automatically summarize digital notes using Generative AI.
According to another embodiment, the invention relates to a computer-implemented method that digitizes and mines handwritten notes to support real-time collaboration. The method comprises the steps of: applying, via a computer processor, computer vision processing to parse an image displayed on a digital touch interface into a set of objects that comprise a set of handwritten notes, wherein the set of handwritten notes are captured during a collaboration session at a location; performing, via the computer processor, optical character recognition processing on the set of handwritten notes to identify a corresponding set of text strings; applying, via the computer processor, a natural language processing to correct errors wherein the errors comprise typographical errors and punctuation; applying, via the computer processor, the natural language processing to create a word cloud graphic based on the corresponding set of text strings wherein the word cloud graphic is based on a metric; applying, via the computer processor, the natural language processing to identify one or more keywords based on the set of text strings and making the one or more keywords available in real-time; applying, via the computer processor, the natural language processing to attribute authorship to each of the text strings based on a writing sample and making the authorship available in real-time; applying, via the computer processor, a recommender engine to identify one or more suggestions that support the set of text strings and making the one or more suggestions in real-time; and applying, via the computer processor, a classifier to identify relationships between the set of text strings and making the relationships available in real-time. In addition, an embodiment of the present invention may automatically summarize digital notes using Generative AI.
The invention also relates to computer-implemented system that digitizes and mines handwritten notes to support real-time collaboration, and to a computer-readable medium containing program instructions for executing a method for digitizing and mining handwritten notes to support real-time collaboration.
An embodiment of the present invention is directed to providing a solution that digitizes and then mines handwritten notes in real-time. The innovation may use an application program interface (API) to extract a graphical image containing annotations. The graphical image may be extracted from a canvas on a digital touch platform or other user interface. An embodiment of the present invention may leverage functionality described in U.S. Pat. No. 10,846,341, “System and Method for Analysis of Structured and Unstructured Data,” and continuation application U.S. patent application Ser. No. 17/100,019, the contents of which are incorporated by reference herein in their entirety.
An embodiment of the present invention may implement Computer Vision (CV) technology to parse an image into logical sub-regions, Optical Character Recognition (OCR) to convert handwritten notes to a string of text; and Natural Language Processing (NLP) to repair (e.g., spell check, etc.), to embellish (e.g., punctuation, capitalization, etc.), and to mine (e.g. keywords, topics, sentiment, etc.). The innovation may access the CV technology through external applications through an API.
For example, an embodiment of the present invention may use NLP to discern sentiment, for topic-modeling and to generate summarizations. Also, an embodiment of the present invention may be directed to submitting a query to a backend corpus of ingested content to support a “search” functionality in a question-answer paradigm.
An embodiment of the present invention may realize technical benefits and advantages including providing a captivating experience (e.g., Prioritized Roadmap, etc.); realizing potential to build benchmarks on session metadata (e.g., moments that matter, friction, common themes, etc.); ingesting content for downstream sessions; and generating synthesized, professional read-out for clients and other users and participants.
These and other advantages will be described more fully in the following detailed description.
In order to facilitate a fuller understanding of the present invention, reference is now made to the attached drawings. The drawings should not be construed as limiting the present invention, but are intended only to illustrate different aspects and embodiments of the invention.
Exemplary embodiments of the invention will now be described in order to illustrate various features of the invention. The embodiments described herein are not intended to be limiting as to the scope of the invention, but rather are intended to provide examples of the components, use, and operation of the invention.
An embodiment of the present invention is directed to automatically discerning creation of handwritten verbiage on a digital display; converting it to digital text objects that have been repaired (e.g., spelling, punctuation, etc.); and then inserting it as Scalable Vector Graphics onto the digital display. An embodiment of the present invention may execute a backend process that watches for insertion of a new note object. Once detected, the system may take action seamlessly or simply mark it for a subsequent action.
At step 110, a system may perform optical character recognition (OCR) on an image of text to convert the image into a text format. The image of text may represent a handwritten text on a small sized graphical note, such as a sticky note.
At step 112, natural language processing (NLP) may be applied to correct errors, such as typographical errors, punctuation, format, style, etc. NLP may be supported by systems as described in U.S. Pat. No. 10,846,341, and continuation application U.S. patent application Ser. No. 17/100,019, the contents of which are incorporated by reference herein in their entirety.
In addition, NLP may also support Named Entity Recognition features within the description of Natural Language Processing that may be applied to disambiguate acronyms, jargon as well as other terminology that may appear on various digital notes.
At step 114, NLP may be applied to create a word cloud or other graphic.
At step 116, the system may perform a search function to identify and highlight keywords including variations thereof.
At step 118, the system may provide the ability to move and/or adjust digital notes on an icon, a graphic or other visual interface, such as a roadmap.
At step 120, NLP may be applied to identify and highlight themes (e.g., financial, booster, blockers, etc.). For example, Themes may leverage a well-defined, curated ONTOLOGY or TAXONOMY.
At step 122, analytics may be applied to identify attributes, such as authorship.
At step 124, NLP may be applied to discern sentiment (e.g., positive, negative, neutral, etc.).
At step 126, a recommender engine may be applied to provide suggestions.
At step 128, a classifier may be applied to cluster related digital notes.
An embodiment of the present invention is directed to creating digital notes on a digital screen platform at a physical site. Another embodiment involves remote users who are working from home or other location, e.g., Location A 1414, Breakout Room 1416, Breakout Room 1418, Location B 1406, Location C 1408. In this example, remotes users may have software executing on a laptop, or access a software through a browser interface. These remote users may be active participants in an interactive session. According to an embodiment of the present invention, the software may generally represent a visual collaboration and dynamic multimedia presentation.
An embodiment of the present invention is directed to supporting multiple interfaces and canvases. For example, an embodiment of the present invention may concurrently support a plurality of collaboration and visualization platforms. The system may access data from across a collection of active canvases to execute a command. For example, the system may execute a command that inserts a job into an asynchronous queue where jobs may be processed on a first-in-first-out (FIFO) or other basis.
System 1410 may include Interface 1420 that supports various interactions on devices including mobile devices, smart devices, computers, laptops, tablets, etc. Interface 1420 may enable users and/or other entities to interact with Engine/Processor 1422. Interface 1420 may support various applications including browsers, mobile interfaces, dashboards, interactive interfaces, etc. For example, System 1410 may interact with Interface 1420 through an API. In this example, Interface 1420 may be provided through a separate service or provider.
Users may interact with an embodiment of the present invention using various devices including mobile devices and handhelds. Participants tend to value a session in direct proportion to the level of interaction and the quantity of “work” that is done toward analyzing a business problem. As such, many exercises seek to illicit individual responses to a prompt and then collaborate as a team to form a consensus. A challenge recognized by an embodiment of the present invention is that sometimes individuals are asked to engage in a task that is complex, time consuming, and based on content. Another challenge is identifying who provided a particular response and in what context. Yet another challenge involves compiling and analyzing responses accurately and expeditiously.
An embodiment of the present invention is directed to synchronizing inputs made on individual mobile devices or handhelds with an interface that is running at a location, e.g., in the front of the room. Using an API, an embodiment of the present invention may programmatically control the setup and/or extraction of content displayed on each individual handheld. An embodiment of the present invention may also apply external services and/or analytics to further process individual responses.
An embodiment of the present invention may then depict individual responses and/or the groups' collective work product onto a front display and presentation sequencing.
Using handhelds, participants may peruse content and provide detailed, individual responses. Each handheld may be assigned to a particular stakeholder so that the system may fully attribute “identity” and “context” as well as other metadata. Individual responses may be “revealed” on a front-monitor and/or compiled seamlessly into group work-product.
Engine/Processor 1422 may support various functions and processes via modules including Commands 1424, Database Interface 1426, Computer Vision 1428, OCR 1430, NLP 1432, Recommender Engine 1434, Classifier 1436, and Search Engine 1438. Other functions and features may be supported. Engine/Processor 1422 may interact with modules via an API or other interface. Other configurations may be supported.
Commands 1424 may represent a set of commands that facilitate user interaction and creation of content and other features. The following is an exemplary set of commands for illustration purposes. Other commands and variations may be implemented.
&SET GRID [number of rows] [number of columns] [color of digital notes] creates on-screen digital notes in the specified color, and in a grid with the requested number of rows and columns. With this command, a note grid may be named. For example, a user may click on a note to see the name. The creation of a well-defined and structured grid enables effective and algorithmic use of Computer Vision to detect edges.
&SCRIBE enables group OCR of an entire grid or group of digital notes. This command returns a pointer note which determines placement of OCR-ed results. A resulting note grid may be named. Generally, an embodiment of the present invention may include a return of SYSTEM MESSAGE digital notes. The SYSTEM MESSAGE note may then be manually moved by a facilitator to a location on the canvas, and the SVG graphic may be inserted as an offset. Meanwhile, the SYSTEM MESSAGE may be used for other purpose including: instructions; error checking, etc.
&KEYWORD [keyword] searches for keywords in OCR-ed digital notes. Resulting search hits may be highlighted. This may include a different color note and/or keyword text in all-caps. Other colors, fonts, graphics may be applied.
&WORD CLOUD is similar to &Scribe, but instead returns a word cloud graphic instead of a note grid. Other graphics and/or illustrations may be provided.
&DELETE [grid name] deletes an entire grid or group of digital notes, as referenced by a grid name or other identifier.
&DELETE ALL deletes all digital notes generated.
&TURN OFF sets the system to idle.
&TURN ON brings the system back from idle.
&SESSION START inserts a flag and timestamp into the database, indicating when a session starts. This supports data mining and other efforts.
&SESSION STOP inserts a flag and timestamp into the database, indicating when a session ends. This also supports data mining and other efforts.
&RELATED finds related digital notes for a specific note. This command may find and highlight a related set of digital notes (based on a clustering algorithm, for example). During sessions, this may provide a way to determine how an important “idea” appears throughout the digital notes that are created in a session.
&THEME enables a facilitator to contribute words to a particular taxonomy. For example, the word “Euro” might be added ad hoc. In this example, a record would then be added accordingly to a table that is LOCAL to the particular session, or a GLOBAL table that may apply to all subsequent sessions. This may be helpful when gathering insights and expanding databases, all in real time.
Other commands may include &ATTRIBUTE, &SUGGEST, &CLUSTER, &QUESTION, &SELECT and &SENTIMENT. For example, the &QUESTION command may invoke a search engine to peruse a library of artifacts (e.g., a database, document repository, content management system, and/or an interface to a Knowledge Graph) and provide answers to a user-specified question or query by using machine learning question-response algorithms.
Another command may include &GENERATE. In this example, a screenshot of specific digital notes may serve as arguments, where Generative AI is applied on the back end, and then one or more digital notes may be inserted as results onto a digital board, such as a video board.
For example, &GENERATE command may be applied to summarize a set of digital notes.
Another example of &GENERATE command may include a generation of use cases for Generative AI based on factors such as Industry only, Business Function only, or a combination of both “Industry AND Business Function” together.
Another example of &GENERATE may include creating digital notes that complement other digital notes that were written by another entity, such as clients, partners, vendors, users, etc.
An embodiment of the present invention may access a knowledge management (KM) system that stores and manages data that has been curated and refined for an organization's use. This data may be internal, managed by internal data managers, or external to an organization.
Another example of &GENERATE may include creating a contextual glossary on the fly that highlights digital notes based on keywords that are associated with a particular theme.
The commands may receive inputs and generate outputs in various forms including TEXT. Other variations may use JPEG files as inputs and generate graphic outputs.
As shown in
An embodiment of the present invention may be directed to mining metadata which may be performed intra- and inter-sessions. In addition to measuring adherence to an agenda, an embodiment of the present invention may take stock of delays, and potentially compress downstream exercises so that the clients can leave the session on time. Similarly, through attribution, an embodiment of the present invention may determine whether there is wide-spread participation/engagement in certain exercises. An embodiment of the present invention may determine whether ideas/questions that are socialized at the beginning of or during a session are ultimately answered by the end of a session. Other actions relating to ideation, problem-solving, etc. may be supported.
Computer Vision 1428 may be applied to parse an image into discrete objects (e.g., rectangular or square shaped objects, etc.). An embodiment of the present invention may discern edges of digital notes of varying shapes and sizes arranged in a grid or other format. For example, the system may implement functions for edge detection and finding contours to detect digital notes functions and also detect specific color schemes to positive OCR-ed text into a detected contour.
OCR 1430 may be applied to convert handwritten characters into digital text. Other character recognition technology and algorithms may be applied.
NLP 1432 may be applied to repair errors. This may include fixing typographical errors, correcting punctuations and addressing other grammatical issues. NLP may be supported by systems as described in U.S. Pat. No. 10,846,341, and continuation application U.S. patent application Ser. No. 17/100,019, the contents of which are incorporated by reference herein in their entirety. NLP 1432 may also be applied to locate keywords (e.g., explicit, thematic, etc.), then augment the string (e.g., uppercase, bold, italics, etc.) and change the note color (or add a graphic) to highlight and emphasize. In addition, NLP 1432 may be applied to compare ad hoc, handwritten digital notes to a set of pre-loaded samples, to attribute authorship and tag accordingly. Other features and analytics may be applied.
An embodiment of the present invention may insert a graphic, such as a Scalable Vector Graphic (SVG), onto a canvas with specific placement as an offset to a pointer.
According to an example, empty digital notes may be dropped altogether from the SVG graphics that were created. According to another example, the empty digital notes may be called forward into the graphics that are created.
Recommender Engine 1434 may provide suggestions that complement user-provided digital notes.
Classifier 1436 may group digital notes that have a relationship or other similarity.
Search Engine 1438 may access a database of artifacts and/or other source of information to respond to user-specified questions. This may include a database, document repository, content management system, and/or even an interface to a Knowledge Graph.
An embodiment of the present invention is directed to a collaboration feature. For example, collaboration sessions may be conducted at a location with interactive digital screens (e.g., collaboration and visualization platforms). This may also include sessions that involve paper notes or other physical annotations. An embodiment of the present invention may be extended to include functionality that captures digital notes via photos and/or mobile applications where the digital notes may be uploaded to a digital platform and commands may be executed.
An embodiment of the present invention may leverage Generative AI to create a Knowledge Framework and further use the Knowledge Framework to assist in prompt engineering of Generative AI. The Knowledge Framework may incorporate taxonomies, ontologies as well as curated artifacts from various sources. For example, an embodiment of the present invention may query the Knowledge Framework for information and context. The system may then publish anonymized client insights to this Knowledge Framework to expand its corpus. The details concerning the Knowledge Framework are provided in pending U.S. patent application Ser. No. 18/477,817, “Artificial Intelligence Enhanced Knowledge Framework,” filed Sep. 29, 2023, the contents of which are incorporated by reference.
System 1410 may store and manage data in various formats, including Databases 1440, 1442. An embodiment of the present invention may store information for later analysis and session improvement. This may involve excluding any data that is “personally identifiable” such as PII, PCI, PHI, etc. For example, if the digital note contains identifiable information, an embodiment of the present invention may REDACT identifiable information or EXCLUDE the note altogether.
Databases 1440, 1442 may represent a Postgres database deployed as a container and may store data in various forms, including tabular structure. An embodiment of the present invention may store NLP and OCR-related data in Postgres. In addition, as an optional feature, handwritten notes produced in sessions may be stored with consent of participants.
System 1410 may be communicatively coupled to Databases 1440, 1442. Databases 1440, 1442 may include any suitable data structure to maintain the information and allow access and retrieval of the information. Databases 1440, 1442 may be any suitable storage device or devices. The storage may be local, remote, or a combination thereof with respect to Databases 1440, 1442. Databases 1440, 1442 may have back-up capability built-in. Communications with Databases 1440, 1442 may be over a network, or communications may involve a direct connection between Databases 1440, 1442 and System 1410, as depicted in
Networks may be a wireless network, a wired network or any combination of wireless network and wired network. Although Network 1412 is depicted as one network for simplicity, it should be appreciated that according to one or more embodiments, Network 1412 may comprise a plurality of interconnected networks, such as, for example, a service provider network, the Internet, a cellular network, corporate networks, or even home networks, or any of the types of networks mentioned above. Data may be transmitted and received via Network 1412 utilizing a standard networking protocol or a standard telecommunications protocol.
While
The system 1400 of
An embodiment of the present invention may support various computer vision applications relating to edge and color detection, for example.
It will be appreciated by those persons skilled in the art that the various embodiments described herein are capable of broad utility and application. Accordingly, while the various embodiments are described herein in detail in relation to the exemplary embodiments, it is to be understood that this disclosure is illustrative and exemplary of the various embodiments and is made to provide an enabling disclosure. Accordingly, the disclosure is not intended to be construed to limit the embodiments or otherwise to exclude any other such embodiments, adaptations, variations, modifications and equivalent arrangements.
The foregoing descriptions provide examples of different configurations and features of embodiments of the invention. While certain nomenclature and types of applications/hardware are described, other names and application/hardware usage is possible and the nomenclature is provided by way of non-limiting examples only. Further, while particular embodiments are described, it should be appreciated that the features and functions of each embodiment may be combined in any combination as is within the capability of one skilled in the art. The figures provide additional exemplary details regarding the various embodiments.
Various exemplary methods are provided by way of example herein. The methods described can be executed or otherwise performed by one or a combination of various systems and modules.
The use of the term computer system in the present disclosure can relate to a single computer or multiple computers. In various embodiments, the multiple computers can be networked. The networking can be any type of network, including, but not limited to, wired and wireless networks, a local-area network, a wide-area network, and the Internet.
According to exemplary embodiments, the System software may be implemented as one or more computer program products, for example, one or more modules of computer program instructions encoded on a computer-readable medium for execution by, or to control the operation of, data processing apparatus. The implementations can include single or distributed processing of algorithms. The computer-readable medium can be a machine-readable storage device, a machine-readable storage substrate, a memory device, or a combination of one or more of them. The term “processor” encompasses all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include, in addition to hardware, software code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed for execution on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communications network.
A computer may encompass all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. It can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
The processes and logic flows described in this document can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
Computer-readable media suitable for storing computer program instructions and data can include all forms of nonvolatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
While the embodiments have been particularly shown and described within the framework for conducting analysis, it will be appreciated that variations and modifications may be affected by a person skilled in the art without departing from the scope of the various embodiments. Furthermore, one skilled in the art will recognize that such processes and systems do not need to be restricted to the specific embodiments described herein. Other embodiments, combinations of the present embodiments, and uses and advantages will be apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. The specification and examples should be considered exemplary.
The application claims priority to U.S. Provisional Application No. 63/419,390 (Attorney Docket No. 055089.0000097), filed Oct. 26, 2022, the contents of which are incorporated by reference herein in their entirety.
Number | Date | Country | |
---|---|---|---|
63419390 | Oct 2022 | US |