The disclosure relates generally to a digital learning platform, and more particularly to an assessment engine that uses a neural network architecture to determine a knowledge state of a student from a set of possible knowledge states of a knowledge structure.
In an educational environment, it is desirable to track a student's progress relating to a corresponding academic field. Educational products are employed to administer course materials associated with an academic field to a number of students and model and assess a student's knowledge state. A knowledge state represents a set of problem types or items within an academic field of study that a student “knows” or has established a threshold level of proficiency. Certain systems use a knowledge space theory (KST) approach to mathematically model and assess a student's knowledge state from among a set of feasible or possible knowledge states (i.e., a collection of feasible knowledge states referred to as a “knowledge structure”). For example, the knowledge structure can be used to define a prerequisite relation between topics or items, where if every feasible knowledge state containing item b also contains item a, then item a is a prerequisite to item b.
Increasing amounts of educational content (e.g., number of problem types) are administered and managed by the respective course product to assess and determine a student's knowledge state associated with a subject or academic field. The increased amount of content results in the expansion of the number of feasible knowledge states of the knowledge structure associated with each subject.
Certain knowledge space theory-based approaches perform a probabilistic search of the knowledge structure to identify a student's knowledge state. However, these approaches fail to account for the large amount of data available in online learning systems in assessing a knowledge state of a student. Furthermore, these probabilistic approaches fail to output a probability after each question presented to a student that is consistent with and adapted to the set of feasible states, which leads to inconsistencies in actual assessments. In such cases, these systems generate an assessment of a student's performance that is compatible with multiple different feasible states, which results in measurable uncertainty and inaccuracy in the state returned by the assessment.
The disclosure will be understood more fully from the detailed description given below and from the accompanying drawings of various embodiments of the disclosure. The drawings, however, should not be taken to limit the disclosure to the specific embodiments, but are for explanation and understanding only.
Aspects of the present disclosure are directed to a system (herein referred to as a “knowledge state assessment system”) to execute one or more methods or processes to perform functions and operations to determine a knowledge state of a student from a set of feasible knowledge states in accordance with an adaptive neural network-based assessment. According to embodiments, the knowledge state is a representation of a proficiency level of a student with respect to a subject area or body of knowledge.
According to embodiments, the knowledge state assessment system generates a knowledge structure including a set of feasible or possible knowledge states associated with a subject area or body of knowledge (e.g., an academic field, a topic, a subject, etc.). The knowledge state assessment system uses the established knowledge structure to discover or determine a current knowledge state associated with a student, where the student's current knowledge state is selected or determined from among the feasible knowledge states of the knowledge structure.
According to embodiments, the knowledge state assessment system 100 determines a current knowledge state of the student using a methodology including a neural network-based probabilistic determination (e.g., to account for the student's careless errors and guesses) and a neural network-based adaptive determination to adaptively select an optimized “next” or subsequent item (e.g., a question) to present to the student in view of the one or more prior or previous answers submitted by the student in response to the one or more previously presented items (e.g., the one or more previously asked questions). Advantageously, the neural network-based probabilistic determination enables the knowledge state assessment system 100 to more accurately identify a final knowledge state of a student from among the feasible knowledge states of the knowledge structure.
According to embodiments, the knowledge state assessment system 100 executes an assessment of a knowledge state of a student associated with a student system 150. In an embodiment, at the beginning of the assessment, the knowledge state assessment system 100 reads a set of initialization parameters from the knowledge database 110, selects an item (e.g., a question) to present to the student and serves the selected item to the student system (e.g., via a web-based program or application). According to embodiments, the one or more knowledge databases 110 can be communicatively coupled to the knowledge state assessment engine 120 via one or more networks. In an embodiment, each time the student submits an answer (i.e., a response) corresponding to a selected item, the knowledge state assessment system 100 identifies a current state of the assessment from the database 110, processes the answer, updates the database 110 with the new state of the assessment, and selects a new item (e.g., a new question) to serve to the student system 150. The assessment process can be executed iteratively until a last or final item is reached (i.e., a last item in a set of items or in a particular time period). After the student system submits an answer to the last item of the assessment, the knowledge state assessment system 100 determines a final knowledge state associated with the student. In an embodiment, the final knowledge state can be provided to one or more systems (e.g., a student system 150, a teacher system, a parent system, etc.) via a knowledge state report 160.
According to embodiments, the knowledge state assessment system 100 includes a knowledge state assessment engine 120 configured to manage and execute a neural network model 105 (e.g., a recurrent neural network (RNN) model employed in the iterative processing of items and answers to assess a current knowledge state of the student). In an embodiment, the neural network model 105 updates probabilities of all items in view of and accordance with the student's answer to the previously selected item. In an embodiment, in view of the student's response or answer, the knowledge state assessment engine 120 executes the neural network model 105 to generate an updated probability for each item which represents an estimated probability that the student has demonstrated knowledge or “knows” the respective item.
The current knowledge state of the student is updated (e.g., a knowledge state that would be returned by the knowledge state assessment system 100 if the iterative assessment process were to terminate or stop) and a new item is selected. According to embodiments, the neural network model 105 updates the item probabilities directly to generate probability estimates that are consistent with a set of possible or feasible knowledge states, as described in greater detail below. Advantageously, the knowledge state assessment system 100 employs the trainable neural network model 105 to adaptively generate the probability estimates in alignment with the possible or feasible knowledge states.
As illustrated, the knowledge state assessment engine 120 executes the neural network model 105 to iteratively generate selected items (e.g., questions) to provide to the student system 150 and receive corresponding answers or responses to each of the selected items. According to embodiments, the knowledge state assessment engine 120 processes the student's answers and generates the probability estimates associated with the set of feasible knowledge states.
In an embodiment, the knowledge state assessment system 100 may be communicatively coupled to the one or more student systems 150 via any suitable network, interface or communication protocol, such as, for example, application programming interfaces (APIs), a web browser, JavaScript, etc.
According to embodiments, the knowledge state assessment system 100 includes one or more memory devices 140 to store instructions executable by one or more processing devices 130 to perform the instructions to execute the operations, features, and functionality described in detail herein.
The term “computing system”, “computer” or “computer platform” is intended to include any data processing device, such as a desktop computer, a laptop computer, a tablet computer, a mainframe computer, a server, a handheld device, a digital signal processor (DSP), an embedded processor, or any other device able to process data. The computer/computer platform is configured to include one or more microprocessors communicatively connected to one or more non-transitory computer-readable media and one or more networks. The term “communicatively connected” is intended to include any type of connection, whether wired or wireless, in which data may be communicated. The term “communicatively connected”, “communicatively coupled”, or “coupled” is intended to include, but not limited to, a connection between devices and/or programs within a single computer or between devices and/or separate computers over a network. The term “network” is intended to include, but not limited to, OTA (over-the-air transmission, ATSC, DVB-T), packet-switched networks (TCP/IP, e.g., the Internet), satellite (microwave, MPEG transport stream or IP), direct broadcast satellite, analog cable transmission systems (RF), and digital video transmission systems (ATSC, HD-SDI, HDMI, DVI, VGA), etc.
According to embodiments, the neural network model 105 updates the probabilities Pi(xj) that the student demonstrates knowledge or “knows” item xj, for each of the items in the course product. The probabilities Pi(xj) are updated by first updating the feature vectors Xi shown in
In an embodiment, the data used for the development and generation of both the knowledge structures (e.g., as stored in the knowledge database 110) and the neural network model(s) 105 is collected, sourced, aggregated, retrieved, obtained, etc. from a large number (e.g., tens of millions) of student assessments taken over a large period of time (e.g., a period of many years). In an embodiment, during each assessment, an item unrelated to the adaptive assessment is randomly and uniformly selected from the items in the course product. The student's answer to this item is not used by the adaptive assessment, but is recorded in the database. This item and its answer may be used as a label (e.g., target variable) of the assessment in the training of the neural network model.
According to embodiments, with reference to
According to embodiments, the neural network model 105 has multiple variations dependent on whether the assessment is the first assessment taken by the student (i.e., an initial assessment) or a subsequent assessment, as described in greater detail herein. In an embodiment, the knowledge state assessment engine 120 performs operations to handle an initial assessment of a student, when no prior specific information relating to the student is available. As shown in
In an embodiment, following the initial assessment, the neural network model 105 performs operations to maintain and update the knowledge state of the student iteratively over time. Consequently, the neural network model 105 leverages information specific to the student (e.g., a pre-assessment knowledge state associated with the student). As shown in
As shown in
For example, the neural network model 105 can generate a student's knowledge state having a number of items (e.g., 103 items) out of a possible larger set of items (e.g., 556 items) associated with a subject area (e.g., Pre-calculus). In this example, this knowledge state may have 50 items pertaining to Algebra and Geometry Review, 25 in Equations and Inequalities, and 28 in Graphs and Functions.
According to embodiments, the neural network model 105 of the knowledge state assessment system 100 may be constrained based on one or more prerequisite properties. For example, item arith033 (e.g., an item associated with the “greatest common factor of 2 numbers”) is designated as a prerequisite to alge738 (e.g., an item associated with “factoring out a monomial from a polynomial: Univariate”) because the identified feasible states that contains alge738 also contains arith033. Formally, (alge738 in state K) implies (arith033 is in state K), and accordingly can be expressed as: arith033<=alge738.
In an embodiment, the knowledge state assessment engine 120 uses the neural network model 105 to determine the knowledge state of a student by constraining the probability estimates of the items to be consistent with the prerequisite relationship determined by the knowledge structure, as described in greater detail herein.
In an embodiment, for a given knowledge structure, the prerequisite relationships between the items are the partial order ≤such that, for any two items q and r and any feasible knowledge state K in the knowledge structure, as represented by the following expression:
(q≤r)⇔ (wherein r is in K implies q is in K). Expression A
According to embodiments, the knowledge state assessment system 100 defines the prerequisite relationship defined by “r is in K implies q is in K” as a partial order on the items. Accordingly, the partial order indicates a reflexive relationship (e.g., “r is in K implies r is in K”), anti-symmetric relationship (e.g., if “r is in K implies q is in K” and “q is in K implies r is in K”, then “r=q”), and a transitive relationship (e.g., if “r is in K implies q is in K” and “q is in K implies p is in K”, then “r is in K implies p is in K”). The partial order relationship can be written as to denote the corresponding prerequisite relationship.
In an embodiment, the neural network model 105 generates probability estimates that are consistent with the prerequisite relationship if, for any two items x and y and their probability estimates P(x) and P(y), the following expression is satisfied:
x<y⇒P(x)≥P(y); Expression B
where x is a prerequisite to y (and x is distinct to y). For example, if “item y is in K” implies “item x is in K”, then any probability estimate attributed to item x by the knowledge state assessment engine 120 is determined to be greater or equal to the probability estimate attributed to item y.
In an embodiment, the prerequisite relation induces a partition of the set Q of the items in the course products into “levels”, where Q represents a set of all terms (e.g., x_1, x_2, . . . , x_n), as shown in
In an embodiment, as shown in
In an embodiment, for level 1, the neural network model 105 applies the logistic function S to each β1, x to get the probability P(x) for each item x in level 1, in accordance with the following expression:
The above-identified function is applied between the “Linear” and “Probability” layers shown in
In an embodiment, for level(s) n and n>1, for each item x in level n, the neural network model 105 computes a maximum probability of all the items q for which x is a prerequisite, in accordance with the following expression:
In an embodiment, the neural network model 105 converts Mx back to the logistic scale and uses the converted Mx to compute the probability P(x) of item x, in accordance with the following expression:
According to embodiments, the knowledge state assessment system 100 determines the knowledge state of a student based on probability estimates of the items that are “positively correlated” with the student's answer. Accordingly, a correct answer to an item should make it more likely that the item is in the student's knowledge state, while an incorrect answer should make it less likely. To enforce this property, the neural network model 105 is trained using a process that adjusts the probability estimates by penalizing the model whenever the estimates are not positively correlated with the student's response.
In an embodiment, let an
where α is a tunable hyperparameter. In an embodiment, the penalty can be added to the loss function during the training of the neural network model 105.
In an embodiment, as shown in
At operation 710, the processing logic determines, by a neural network model executed by the processing logic, an initial knowledge state of a student relating to a subject. At operation 720, the processing logic generates a vector representation of a set of items associated with the subject. At operation 730, the processing logic executes an assessment comprising at least a portion of the set of items relating to the subject. At operation 740, the processing logic provides a first item of the set of items to the student. At operation 750, the processing logic receives, from the student, a first response to the first item.
At operation 760, the processing logic generates an updated vector representation based on the first response to the first item. At operation 770, the processing logic generates, based on the updated vector representation and the initial knowledge state, a first set of probabilities associated with an updated knowledge state of the student corresponding to the set of items relating to the subject.
The example computer system 800 may comprise a processing device 802 (also referred to as a processor or CPU), a main memory 804 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM), etc.), a static memory 806 (e.g., flash memory, static random access memory (SRAM), etc.), and a secondary memory (e.g., a data storage device 816), which may communicate with each other via a bus 830.
Processing device 802 represents one or more general-purpose processing devices such as a microprocessor, central processing unit, or the like. More particularly, the processing device may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computer (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets. Processing device 802 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. Processing device 802 is configured to execute a proficiency profile and recommendation system for performing the operations and steps discussed herein. For example, the processing device 802 may be configured to execute instructions implementing the processes and methods described herein, for supporting the knowledge state assessment system 100, in accordance with one or more aspects of the disclosure.
Example computer system 800 may further comprise a network interface device 822 that may be communicatively coupled to a network 825. Example computer system 800 may further comprise a video display 810 (e.g., a liquid crystal display (LCD), a touch screen, or a cathode ray tube (CRT)), an alphanumeric input device 812 (e.g., a keyboard), a cursor control device 814 (e.g., a mouse), and an acoustic signal generation device 820 (e.g., a speaker).
Data storage device 816 may include a computer-readable storage medium (or more specifically a non-transitory computer-readable storage medium) 824 on which is stored one or more sets of executable instructions 826 of the knowledge state assessment system 100. In accordance with one or more aspects of the disclosure, executable instructions 826 may comprise executable instructions encoding various functions of the knowledge state assessment system 100 in accordance with one or more aspects of the disclosure.
Executable instructions 826 may also reside, completely or at least partially, within main memory 804 and/or within processing device 802 during execution thereof by example computer system 800, main memory 804 and processing device 802 also constituting computer-readable storage media. Executable instructions 826 may further be transmitted or received over a network via network interface device 822.
While computer-readable storage medium 824 is shown as a single medium, the term “computer-readable storage medium” should be taken to include a single medium or multiple media. The term “computer-readable storage medium” shall also be taken to include any medium that is capable of storing or encoding a set of instructions for execution by the machine that cause the machine to perform any one or more of the methods described herein. The term “computer-readable storage medium” shall accordingly be taken to include, but not be limited to, solid-state memories, and optical and magnetic media.
Some portions of the detailed descriptions above are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise, as apparent from the following discussion, it is appreciated that throughout the description, discussions utilizing terms such as “identifying,” “determining,” “analyzing,” “selecting,” “receiving,” “presenting,” “generating,” “deriving,” “providing” or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
Examples of the disclosure also relate to an apparatus for performing the methods described herein. This apparatus may be specially constructed for the required purposes, or it may be a general-purpose computer system selectively programmed by a computer program stored in the computer system. Such a computer program may be stored in a computer readable storage medium, such as, but not limited to, any type of disk including optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic disk storage media, optical storage media, flash memory devices, other type of machine-accessible storage media, or any type of media suitable for storing electronic instructions, each coupled to a computer system bus.
The methods and displays presented herein are not inherently related to any particular computer or other apparatus. Various general-purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct a more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear as set forth in the description below. In addition, the scope of the disclosure is not limited to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the disclosure.
It is to be understood that the above description is intended to be illustrative, and not restrictive. Many other embodiment examples will be apparent to those of skill in the art upon reading and understanding the above description. Although the disclosure describes specific examples, it will be recognized that the systems and methods of the disclosure are not limited to the examples described herein, but may be practiced with modifications within the scope of the appended claims. Accordingly, the specification and drawings are to be regarded in an illustrative sense rather than a restrictive sense. The scope of the disclosure should, therefore, be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled.
This application claims the benefit of U.S. Provisional Application No. 63/432,116, filed Dec. 13, 2022, the entirety of which is incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
63432116 | Dec 2022 | US |