Many online and college level courses are taught to large number of student learners by single or a small group of teacher instructors. Classical testing and assessment techniques rely on quizzes, tests, and essays. Due to the large number of learners, quizzes and tests tend to be multiple choice, which requires considerable testing and selection of the questions via statistical evaluation in order for the multiple choice question to be a reliable and effective assessment tool.
The disclosure is better understood with reference to the following drawings. The elements of the drawings are not necessarily to scale relative to each other. Rather, emphasis has instead been placed upon clearly illustrating the claimed subject matter. Furthermore, like reference numerals designate corresponding similar parts through the several views.
Alternatively to multiple choice tests, essay tests potentially provide one free-form or open-ended method for an instructor to assess the deep level understanding of the learner. However, such essay tests require manual grading and when done with large number of students, they typically have to be graded (assessed) by several individuals and subjective differences among the graders lead to inconsistent scoring and thus is a low quality assessment technique. Accordingly, existing educational open-ended assessment techniques are not easily extended to large sets of people due to resource constraints and differences among graders. For example, in massively open online courses (MOOC) there may be thousands of students that are enrolled in a single class and manual open-ended assessment is simply not timely nor realistic. There is no present open-ended assessment method that provides for consistent and fair scoring that is cost effective.
Concept map testing is another worthy open-ended testing method for a learner to demonstrate their deep level understanding of targeted learning content. Relating new concepts to what is already known by the learner 104 helps to trigger meaningful deep learning. Learners 104 reveal their in-depth knowledge using concept maps by creating or choosing concepts and drawing edges and labeling the edges between concepts to show relationships. An instructor 102 then examines the learner concept maps, evaluates the quality of the concept map, and assigns a score. However, like other testing techniques, prior concept map grading is primarily done manually and presents a heavy workload to the instructor 102. Thus, it has been impractical with large numbers of learners 104. While multiple graders are one way to address the workload, the differing subjective judgment of the graders leads to unfair or inconsistent testing.
Presented within is a computer-implemented system and method for automating with software and computer processors the creation and assessment of concept or other content maps, also sometimes referred to as mind maps. By automating one or more of the various processes to create and assess concept maps, the workload of the instructor 102 may be significantly reduced. Additionally, learners can mentally focus their creativity and thought into generating concept maps rest assured that their concept maps are automatically and objectively assessed (graded or scored) consistent, fair, and relatively fast.
In addition, as a technique to increase one's learning retention rate, peer-review by the learners 104 and peer-feedback to the peer-reviewers is also available as learning and understanding are significantly increased by teaching others. Peer-review helps the learners 104 to increase their understanding or comprehension of the material by not only seeing how other learners 104 are understanding the material but also by having the learners 104 articulate their understanding of the conceptual relationships back to the peer-learner under review. Accordingly, in one example, each learner 104 plays two roles: learner and peer-reviewer. In this peer-review example, each learner 104 is evaluated by at least two factors: first, how well the learner has constructed concept relationships as scored by other peer-reviewers; and second, how well the learner's review comments help others as scored by other learners' feedback about the learner peer-review comments.
The computer-implemented systems and methods disclosed herein are able to automatically detect the relative importance of the various concepts and relationships and incorporate that information into the assessment. The importance is determined based on the relative importance of the concepts within the learning content 103, such as course material, books, lecture notes, references, articles, etc. Further, a “ground truth” or reference concept map 324 (
The assessment of a learners' test is done in comparison to difference and/or similarity matching with the reference concept map 324. Accordingly, the concept map matching may be done in at least one of multiple ways. In one example, the difference matching is done by matching the connection of edges between pairs of concepts in the learner's concept map 332 (
To help ease understanding, clarity, and conciseness, the following definitions will be used for discussion of additional details of the concept map assessment discussed herein unless context requires otherwise or the term is defined explicitly differently. The following definitions are not meant to be a complete definition of the respective terms but rather is intended to help the reader in understanding the following discussion.
Concepts—A concept is a labeling of an abstract idea, a plan, an instance, or some general notion generally formed by combination of characteristics or particulars of intuitive thought. Concepts may range from a collection of general to very detailed specific groupings or dissections of content. Concepts are generally labeled in a box, circle, nodes, cells, or other geometric form for visualization. Concepts may also include textual summaries beyond labeling and the textual summaries may include information on how particular labeled concepts related to other labeled concepts.
Edges—Edges are lines which are used to connect concepts and can be labeled text to show relationships between concepts. The edges may also be with directions noting hierarchy or they can be undirected.
Concept Maps—A concept map is a form of a structure diagram that illustrates conceptual knowledge and their relationships. The structure diagram may be hierarchical in some examples and not hierarchical in other examples. The concept map represents relationships among different concepts within a specific topic and in some instances may show relationships from general to specific concepts. A concept includes concept labels or summaries that are connected together by edges to other concept. One example concept map is a flow chart that is widely used in business to gain an insight on the overview idea being constructed. Other example concept maps are used to brain-storm new ideas and developments. For instance,
Learner—A student, apprentice, observer, or other individual that is working on acquiring knowledge in new or familiar topical areas.
Instructor—A teacher, professor, educator, mentor, or other individual who possesses or communicates knowledge that learners wish to absorb and understand.
Assessment—A test, quiz, essay, competition, exercise, examination, or other evaluation used to assess a learner's understanding of taught material.
Learning content—The educational material put forward by an instructor or other individual that contains the targeted learning material for the learners to review, absorb, and understand to acquire in-depth knowledge of the subject matter in order to understand the targeted concepts and the relations between them. It may include course material, books, lecture notes, references, articles, etc. from a global document set as just some examples.
Summary—A description of the relationship between differing concepts and their relations. The summary may be in addition to the labeled concepts and edges or in place thereof. Summaries may be textual blocks which are amenable to summarization techniques to determine textual similarities.
The instructor 102 may auto-create a reference concept map within a concept map generation module 107 using a content segmentation module 110 to automatically parse and segment the learning content 103 into various content blocks. The content blocks may then be further examined with a concept extraction module 112 to mine and group concepts and edges within the content blocks. A concept calibration module 114 allows an instructor 102 to modify, such as by editing, adding, or deleting, etc. the auto extracted concepts and edges based on his/her preferences or intimate knowledge of the subject matter. The instructor 102 then develops lessons 116 and teaches the concepts 118 to the learners 104 and creates assessments 120 using the concept material.
The learner 104 in turn learns concepts 119 and demonstrates that learning by taking the assessment from instructor 102 which requires the learner 104 using a concept map construction module 121 to construct a learner concept map. The completed learner concept map is given to the instructor 102 to so that it may be evaluated and assessed in the auto-assessment module 122. The auto-assessment module 120 may perform the grading using one or more techniques, such as using differences 126, similarity 128, and/or peer-review 130.
In some examples, the learner concept map may also be presented to the peer concept map assessment module 123 for peer-review by other learners for ratings and/or comments. The reviews by other learners is then presented to learner 104 in the review reviewers module 125 to allow for feedback ratings to individual other learners on how helpful their peer-reviews on the learner's concept map were to the learner 104. In some examples, the peer reviews are also communicated to the auto-assessment module 122 to allow for augmentation or incorporation into the learner's performance assessment and grading.
Accordingly, the instructor 102 may review the assessed test result along with the peer-review comments and feedback on the peer-reviews to arrive at a final assessment for the learner. In some examples, there may be a final assessment dashboard module 132 that the instructor 102 can use to view the overall assessments of the learners 104 in the class and to provide new suggested learning content material to each learner 104 based on the final assessment and how the learner's concept map compared to the instructor's reference concept map. The learner 104 may use the review assessment/performance module 127 to review the final assessment and any new suggested learning content material provided by the instructor 102 from the suggested learning content suggested in module 132.
For instance,
In other examples, the modules may be implemented in software, firmware, or logic as one or more modules and sub-modules alone or in combination as best suits the particular implementation. Further the modules may be implemented and made available in standalone applications or apps and also as application programming interfaces (APIs) for web or cloud services. While a particular example module organization is shown for understanding, those of skill in the art will recognize that the software may be organized in any particular order or combinations that implements the described functions and still meet the intended scope of the claims.
The CRM may include a storage area for holding programs and/or data and may also be implemented in various levels of hierarchy, such as various levels of cache, dynamic random access memory (DRAM), virtual memory, file systems of non-volatile memory, and physical semiconductor, nanotechnology materials, and magnetic/optical media or combinations thereof. In some examples, all the memory may be non-volatile memory or partially non-volatile such as with battery backed up memory. The non-volatile memory may include magnetic, optical, flash, EEPROM, phase-change memory, resistive RAM memory, and/or combinations.
The processor 302 may be one or more central processing unit (CPU) cores, hyper threads, or one or more separate CPU units in one or more physical machines. For instance, the CPU may be a multi-core Intel™ or AMD™ processor or it may consist of one or more server implementations, either physical or virtual, operating separately or in one or more datacenters, including the use of cloud computing services.
The processor 302 is also communicatively coupled via channel 318 to a network interface 316 or other communication channel to allow interaction with the instructor 102 via an instructor client computer 322 and one or more learners 104 and their respective learner client computer 330. The client computers may be desktop, workstations, laptops, notebooks, tablets, smart-phones, personal data assistants, or other computing devices. In addition, the client computes may be virtual computer instances. The instructor client computer 322, the learner client computer 330 and the network 316 are coupled via a cloud network 320, which may be an intranet, Internet, local area network, virtual private network, or combinations thereof and implemented physically with wires, photonic cables, or wireless technologies, such as radio frequency and optical, infra-red, UV, microwave, etc. In some examples, the cloud network 320 may be implemented at least partially virtually in software or firmware operating on the processor 302 or various components of the cloud network 320. In some examples, the learning content 103 may be stored on one or more databases 330 and coupled to the cloud network 320. In other examples, the learning content 103 may be stored in memory 304 or other processor coupled memory or storage for direct access by the processor 302 without use of the network interface 316.
In one example, a memory 304 is coupled to the processor 302 and the memory 304 includes instructions 306 that when read by the processor 302 cause the processor 302 to perform at least one of the operations to: a) provide a set of key concepts 326 and a fixed set of edges 204 to allow learners 104 to create a learner concept map 332 and automatically compare the learner concept map 332 to a reference concept map 324; b) generate 308 the set of candidate concepts 206 from content blocks in a set of learning content 103 in a database 330 and to allow learners 104 to connect pairs of candidate concepts 206 with edges 204 defining the pairs of candidate concept 206 inter-relationships; c) provide a peer-review assessment system 312 to the learners 104 to review and provide feedback 334 to a set of other learner concept maps and to allow the learner 104 to review feedback from other learners for the learner concept map and provide 326 feedback on feedback ; and d) provide a similarity service 314 using summarization techniques to detect similarities and/or differences between the learner concept map 332, the reference concept map 324 and the set of learning content 103 in the database 330.
Providing the set of candidate concepts 206 and a fixed set of edges 204 may be implemented by having the learning content 103 segmented into a set of content blocks using content segmentation 110. The content blocks are extracted by the processor 302 using the various table of contents in the learning content 103 and/or topic distribution within the learning content 103 using semantic content segmentations. Then, for each content block, the underlying key concepts are extracted by the processor 302 in concept extraction module 112. The key concepts may be extracted from the concept blocks using one or more of a) word frequency analysis, b) using a document corpus, c) word co-occurrence relationships, d) using lexical chains, and e) key phrase extraction using a Bayes-classifier.
Key concept extraction may not provide exactly the key concepts depending on the learning content 103 and the key concept extraction techniques used. Accordingly, concept calibration module 114 can be used to allow the instructor 102 to modify the extracted key concepts to allow specific concepts/terms the instructor 102 wants to emphasize based on their teaching experiences. In concept calibration, the processor 102 provides a user interface, such as a graphical user interface (GUI) to the instructor's client computer 322 to allow for adding, deleting, and editing the key concepts and/or the fixed set of edges.
For instance,
As noted, there are two type of edges in this example concept map. One type of the edge has a hierarchical property, such as one concept consists of another concepts (e.g. “animal” consists of “mammal”). The other type of edge has a peer-to-peer (bi-directional) property, such as one concept is strongly related to another (e.g. “cloud” is related to “rain”). In other examples, the instructor 102 may use more than two types of edges or multiple edges of varying types. Each edge has two factors to define the type of an edge; 1) whether it is directed or undirected, and 2) the label of the edge which defines its semantic meaning.
In one example, after an instructor 102 creates a reference (or “ground truth”) concept map as in
There are several ways of auto-assessing the learner concept maps 322 using a processor 302. A comparison may be made by having a processor 302 check the similarity between the learner concept map 322 and the reference concept map 324. A software graphical user interface (GUI) can be executed by the processor 302 to allow a peer-review of the learner concept map 322 may be done by other learners 104 in the class and review of the reviewers used to help provide a check on the peer review as well as provide for additional assessment of the learner 104 based on the feedback on his/her feedback to other learners 104.
Another approach is to have a processor 302 compare the learner concept map 322 with the reference concept map and look for differences of the edges between pairs of concepts in each concept map. Further differences for the processor 302 to check is the difference of paths between the reference concept map 324 and the learner concept map 322. Each candidate concept is given a different weight, which is determined by the processor 302 by the importance of the concept in the learning content 103 and the centrality of the candidate concept in the reference concept map 324. Each type of edge is also given a weight, which is given a default value by the processor and the instructor 102 can change the values of each of the edge weights. More detail follows.
One method of having the processor 302 automatically comparing the learner concept map 332 to a reference concept map 324 involves checking the difference in the relationship of each pair of concepts in the learner's concept map and the corresponding pair of the concepts in the reference concept map. The relationship of each pair of concepts is divided into two parts by the software. One part is the difference of the edges between the same pair of concepts in the two concept maps: Δf (eij, e′ij), where eij is the reference edge between concept I and concept j and e′ij is the edge between concept I and concept j in the learner's concept map. The other part is the difference of the paths between the same pair of concepts in the reference and learner concept maps: Δg(Pij, P′ij) where Pij is the set of paths between concept I and concept j in the reference concept map, and P′ij is the set of paths between concept I and concept j in the learner's concept map. The assessment (grade or score) of the learner's concept map is implemented in software instructions 306 on the processor 302 by the following formula:
where K is the total number of candidate concepts, θ1 and θ2 are two parameters, Smax is the maximum score that one can get.
To further calculate Δf (eij, e′ij) and Δg(Pij,P′ij), weights are assigned to each type of edges. The weight of each type of edge has a default value given by the processor 302, with a directed edge having a weight w1 and an undirected edge having a weight w2. In one example, it is assumed that w1>w2 (e.g., w1=2*w2). In other examples, the instructor may change the weight value for each type of edge. A weight is then calculated for each concept. With those weights, Δ(eij, e′ij) and Δg(Pij, P′ij) are calculated and the details of which are described in
For the processor 302 to calculate the weight of a concept, the importance of the concept in the learning content 103 is considered based on its frequency in the learning content 103 and the Inverse Document Frequency (IDF) in all global books, the global document set 105, which are widely used in the information retrieval area and known to those of skill in the art. Concepts or terms used to describe the concepts that appear in many documents are considered common and relatively less important. An IDF value of a concept may be measured using instructions 306 on the processor 302 with the following equation:
where n is the total number of documents in the global document set; nc
To be adaptive to the concept relationship, the centrality value of the concept in the instructor reference concept map 324 network is considered by processor 302 for the weight of the concept. For a metric to calculate the centrality value of a concept in the reference concept map 324 network, a modified Page Rank technique may be implemented in software instructions 306 that takes into consideration the weight of the edges 204. The following equation may be used by processor 302 to determine the modified Page Rank values for the nodes in the reference concept map 324 with weighted edges 204. It is a recursive equation:
where d is a parameter with a typical value of 0.85. Note that d can be optimized for a given type of document set. PR(ci) is the modified PageRank value of node ci, PR(cj) is the modified PageRank value of node cj, w(cj,ck) is the weight of the edge 204 from node cj to node ck, w(cj,ci) is the weight of the edge 204 from node cj to node ci, N is the total number of nodes in the reference concept map 324.
The weight of concept ci is determined by processor 302 as:
R(ci)=Ø1*TF(ci)*IDF(ci)+Ø2*PR(ci)
where Ø1 and Ø2 are two parameters and R(ci) is the output weight for concept ci.
Using both the weights of the edge and the weights of the concepts, the following formula may be used by processor 302 and software instructions 306 to determine the difference of the edges between the same pair of concepts in the learner's concept map 322 and the reference concept map 324:
where R(ci) is the weight of concept ci and R(cj) is the output weight for concept cj, wij is the edge weight between concepts ci and cj in the reference concept map 324.
where on the right side of the above formula, the first term is from the path of length 1, and the second term is from the paths of length 2, and the third term is from the paths of length 3, . . . . The variable “∝” is a predefined decay parameter with a small value much less than 1, which makes sure that the longer the path, the less of a contribution it will give to the relevancy score. The contribution value of each path with length L to the score is calculated by multiplying the weight of each edge in the path and multiply the decay factor ∝ for (L-1) times. Accordingly, the path difference is determined by the processor 302 and software instructions 306 using the following formula:
66 g(Pij, P′ij)=Pij−P′ij
where Pij is the path score from concept ci to concept cj in the reference concept map. P′ij is the path score from concept ci to concept cj in the learner concept map.
While providing a score based on a learner's understanding of the concept map is useful, the computer-implemented system and method herein provide an opportunity to not only assess the learner's understanding of the learning content 103 but also to improve the understanding by having the learners 104 participate in a peer-review process. This peer-review process helps to engage the learner 104 and helps minimize the time instructors 102 put into helping/tutoring learners 104 in large classes by placing the learners 104 in dual roles of “learner” and “reviewer”. That is, in order to help make the scoring of these constrained or open-ended concept maps scalable for a large set of people (a large on-ground class or an online learning scenario such as a MOOC), the peer concept assessment module 123 is implemented by the processor 302 and instructions 306 to provide a GUI interface that allows learners to review concept maps from other learners, assign review scores and/or comments and then provide feedback to the reviewers of their learner concept maps, which is also used to help calculate an overall score for an individual learner 104. In total, every learner 104 may be evaluated by a sub-set of their class peers for the learner's understanding of the concepts (average or median review scores (or ratings) from the peer-reviewers), and each learner 104 is evaluated for their ability to provide meaningful and constructive feedback to others. That is, every learner's performance may be assessed in both roles, learner and reviewer, and assigned a final assessment result by processor 302.
Another computer-implemented approach to automated assessment that allows for less constraints on the concept and edge labels used is similarity detection of textual blocks such as learner summarization of the concepts conveyed in the context map. In the similarity module 128 of
In the similarity module 128, a set of summaries related to the topic under assessment for the course is selected from the learning contents by the instructor 102. For instance, the summary can be the reference concept map 324 which can be auto generated by processor 302 as described along with textual summaries or the textual summaries created/modified by the instructor 102 or combinations thereof. The instructor 102 may suggest to processor 302 via a GUI interface reserve material within the learning content 103 to be provided to the learners 104 as complimentary topics. The learner's review the topic learning content and enter their own textual summaries in learner concept maps from candidate concepts 206 as chosen by the computer-implemented system or the instructor 102.
At assessment time, the similarity module 128 uses similarity detection using summarization techniques to detect the textual similarity and/or differences between the summary or learner concept map 322, the instructor 102 reference concept map 324, and the topical material in the learning content 103. The textual summarization techniques may use software-based statistical language models and/or singular value decomposition (SVD) and Karhunen-Loeve (KL) expansion divergence methods to check the learner 104 understanding. Other such textual summarization techniques may include: glossaries, computational linguistics, natural language processing, domain ontology, subject indexing, taxonomy, terminology, text mining, and text simplification to just name a few. In addition, various text similarity detection APIs exist and are known and available to those of skill in the art.
The similarity module may also include a reserve material suggestion module that uses the assessment scores as weightings for the atomic components (words, etc.) of each topic previously defined. These weighting can then be used to direct the final weightings of one or more software-based summarization (or key word extraction) engine modules. The extraction summary obtained for the output may then be compared (subtracted from) the summary obtained from the reference material 103. The lower an assessment for a learner 104 in a specific topic, the larger the weight to be used to choose summaries to provide additional content for that particular topic, and vice versa.
Topical content in the learning content 103 that is not in the summary (or learner content map) and not indicated as successfully understood by the learner 104 is then indicated by processor 302 for the learner 104 and may be sent to the instructor 102 prior to the learner 104 for review, approval, changes, etc. using a software-based GUI interface as needed in the final assessment dashboard 132 in
Additionally, a computer-implemented assign weights and concepts method 1416 may include wherein different weights are assigned to each concept in the second set of concepts based on a determination of the relative importance of the respective concept to the other concepts in the second set of concepts. The assign weights to concepts method 1416 includes wherein the different weights assigned to each concept by processor 302 in the second set of concepts is based on at least an inverse document frequency in references used in the field of the concepts and wherein the different weights assigned by processor 302 to each concept in the second set of concepts is based on at least a page-rank algorithm that considers the weight of the edges in the second set of edges.
The computer-implemented method 1400 may include an assign weights to edges method 1418 wherein the instructor determines the second set of edges and different weights are assigned by the instructor to each edge in the second set of edges or wherein each edge type in the first and second set of edges is defined by whether the respective edge is directed or undirected and a label that defines its semantic meaning. Other methods to method 1400 include wherein the edge difference determined by processor 302 is a product of the weight of a first concept and the weight of a second concept and the edge weight between the first concept and the second concept in the second concept map if the edge types are different or wherein the path difference is determined by processor 302 based on a weight of each edge between two concepts and a decay factor based on path length in the first concept map.
While the claimed subject matter has been particularly shown and described with reference to the foregoing examples, those skilled in the art will understand that many variations may be made therein without departing from the intended scope of subject matter in the following claims. This description should be understood to include all novel and non-obvious combinations of elements described herein, and claims may be presented in this or a later application to any novel and non-obvious combination of these elements. The foregoing examples are illustrative, and no single feature or element is essential to all possible combinations that may be claimed in this or a later application. Where the claims recite “a” or “a first” element of the equivalent thereof, such claims should be understood to include incorporation of one or more such elements, neither requiring nor excluding two or more such elements.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2015/055168 | 10/12/2015 | WO | 00 |